would love your take when you do!
07.12.2024 15:56 β π 5 π 0 π¬ 0 π 0@finbarr.bsky.social
building the future research at midjourney, deepmind. slinging ai hot takes π₯at artfintel.com
would love your take when you do!
07.12.2024 15:56 β π 5 π 0 π¬ 0 π 0π₯Έ
07.12.2024 15:53 β π 5 π 0 π¬ 0 π 0This is one of my all time favorite papers:
openreview.net/forum?id=ByJ...
It shows that, under fair experimental evaluation, lstms do just as well as a bunch of βimprovementsβ
Iβll find it one sec
07.12.2024 15:48 β π 1 π 0 π¬ 1 π 0Very likely.
06.12.2024 00:57 β π 0 π 0 π¬ 0 π 0π
06.12.2024 00:56 β π 0 π 0 π¬ 0 π 0Fun fact: I recently encountered (well, saw on the news) the only other person named finbarr in Canada Iβve ever seen.
The only issue is, he was an arsonist who set a ton of fires in Edmonton.
Really fun conversation with @natolambert.bsky.social!
05.12.2024 20:43 β π 5 π 0 π¬ 0 π 0This is mckernan! What I thought was a nice neighborhood π
03.12.2024 18:19 β π 1 π 0 π¬ 1 π 0Apparently there *is* another finbar(r) in Alberta.
03.12.2024 06:19 β π 5 π 0 π¬ 0 π 0New homeowner fear unlocked; someone hit and ran my neighborβs garage
03.12.2024 06:18 β π 3 π 0 π¬ 1 π 0I thought that was ai gen at first!
02.12.2024 12:32 β π 3 π 0 π¬ 1 π 0thereβs a type of βnot tryingβ which means not executing at the level of competence of a $XX billion corporation
this is the complaint about eg Google products. Theyβre good! better than most startups! But not βtrillion dollar corporation famed for engineering expertiseβ good.
would also accept Austria
30.11.2024 04:57 β π 4 π 0 π¬ 2 π 0I watched too many ski movies and now am trying to convince my wife we should move to Alaska
30.11.2024 04:56 β π 5 π 0 π¬ 2 π 0building my own mlp implementation from scratch in numpy, including backprop, remains one of the most educational exercises Iβve done
30.11.2024 04:18 β π 19 π 0 π¬ 2 π 0Welcome!
30.11.2024 04:18 β π 2 π 0 π¬ 0 π 0π
26.11.2024 23:38 β π 1 π 0 π¬ 0 π 0Ha, itβs been on my list of todos for a while! Iβm glad someone got to it.
26.11.2024 23:38 β π 1 π 0 π¬ 0 π 0Love this. Very clean implementations of various inference optimizations.
26.11.2024 23:20 β π 6 π 0 π¬ 1 π 0Agreed! Folk knowledge is worth publishing!
26.11.2024 22:53 β π 10 π 0 π¬ 1 π 0I mailed this out like a month ago and just never did the promo π
26.11.2024 16:52 β π 2 π 0 π¬ 1 π 0Force of habit!
26.11.2024 16:30 β π 5 π 0 π¬ 0 π 0Ahh youβre right!
26.11.2024 16:30 β π 1 π 0 π¬ 0 π 0again, link is: www.artfintel.com/p/papers-ive...
26.11.2024 16:19 β π 4 π 0 π¬ 0 π 0seems like we're seeing convergence in VLM design. most recent models (Pixtral, PaliGemma, etc) are moving away from complex fusion techniques toward simpler approaches
as usual, the bitter lesson holds: better to learn structure than impose it
incompleteideas.net/IncIdeas/Bit...
open source VLMs use relatively little compute compared to what you might expect:
LLaVa: 768 A100 hours
DeepSeek-VL: 61,440 A100 hours
PaliGemma: ~12k A100 hours
(for reference, Stable Diffusion used 150k A100 hours)
what i found interesting: VLMs are way simpler than they first appear. current SOTA is basically:
1. ViT encoder (init from SigLIP/CLIP)
2. pretrained LLM base
3. concat image features with text
4. finetune