bilal's Avatar

bilal

@bilal2vec.bsky.social

twitter.com/bilaltwovec

102 Followers  |  106 Following  |  19 Posts  |  Joined: 10.11.2024  |  1.8616

Latest posts by bilal2vec.bsky.social on Bluesky

it's Space Quebec, because they speak a language with the same phonology as French which is incomprehensible to Francophones

06.05.2025 22:13 β€” πŸ‘ 221    πŸ” 26    πŸ’¬ 15    πŸ“Œ 1
Preview
Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction We present Visual AutoRegressive modeling (VAR), a new generation paradigm that redefines the autoregressive learning on images as coarse-to-fine "next-scale prediction" or "next-resolution prediction...

there’s been some interesting work lately on multiscale autoregressive image modeling arxiv.org/abs/2404.029...

29.03.2025 02:55 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

city of stairs and the tainted cup both very good πŸ‘

28.03.2025 12:00 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

once again coming crawling back to AdamW after every paper published after 2015 has failed me again

27.03.2025 04:24 β€” πŸ‘ 68    πŸ” 4    πŸ’¬ 6    πŸ“Œ 0
28.03.2025 11:56 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

i think they’re just starting to realize what they unleashed w the tweet

25.03.2025 12:43 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

from the other app xd

25.03.2025 12:35 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Love to log on to the Horrors app to catch up on today's Horrors

15.03.2025 18:27 β€” πŸ‘ 20    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

we’re doing ai exorcisms in 2025 huh

04.03.2025 19:29 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

like one of the big things about 404 media, Brian Merchant, Paris Marx, Ed Zitron, etc., is that they neither know nor care how the subject of their criticism actually works

23.01.2025 20:18 β€” πŸ‘ 179    πŸ” 15    πŸ’¬ 12    πŸ“Œ 10

I wish academic ML was a bit more skeptical of papers and less skeptical of industry. I get that it sucks to not have visibility on details, but it doesn’t invalidate the results. On the flip side, there are too many papers whose message are parroted despite sketchy experiments.

11.01.2025 21:47 β€” πŸ‘ 9    πŸ” 2    πŸ’¬ 3    πŸ“Œ 0

we’re going shopping

09.01.2025 04:38 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

an LLM that uses streetview to pre-drive down the route and assemble comments like "at the big red barn, turn left" "when you get to the sorta squiggly road, take the exit" like a farmer would

09.01.2025 00:44 β€” πŸ‘ 6    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

if you squint hard enough everything in ml is either a special case of the KL div or newton’s method

04.01.2025 16:27 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

a lot of machine learning research is about discovering which parts of mathematics are actually L2 regularization and which parts of mathematics are actually Adam

04.01.2025 06:59 β€” πŸ‘ 76    πŸ” 5    πŸ’¬ 6    πŸ“Œ 1
20.12.2024 09:17 β€” πŸ‘ 148    πŸ” 8    πŸ’¬ 1    πŸ“Œ 0

justine tunney the libc mutex micro optimizations person??

19.12.2024 05:21 β€” πŸ‘ 23    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

This guy needs to read Manufacturing Consent! You’re not supposed to do this yourself you gotta hire editors who already agree with you, this is amateur hour shit…

12.12.2024 22:00 β€” πŸ‘ 6    πŸ” 2    πŸ’¬ 3    πŸ“Œ 0
08.12.2024 02:19 β€” πŸ‘ 643    πŸ” 82    πŸ’¬ 3    πŸ“Œ 0

pov: post training researchers learning what pretraining researchers do while waiting for the model to train

06.12.2024 22:21 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

accidentally typed rm -fr and i’m using that now

03.12.2024 03:16 β€” πŸ‘ 613    πŸ” 43    πŸ’¬ 46    πŸ“Œ 13

congrats!!

03.12.2024 19:16 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

thanks for cleaning it up

01.12.2024 01:57 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

ai generated slavoj zizek voice on slop video of some bizarre rural chinese cooking

30.11.2024 20:45 β€” πŸ‘ 58    πŸ” 2    πŸ’¬ 1    πŸ“Œ 1

incredible new forms of postings emerging

01.12.2024 01:55 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

interesting is there anywhere i can read more about this

28.11.2024 20:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

its all approximating numbers w other numbers all the way down. everything else is an implementation detail! πŸ˜›

28.11.2024 18:43 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

if your values do matter replace them w values similar to them aka ones (parameter sharing / shared kv cache / factorizing a large matrix into two small ones / lora / adafactor)

28.11.2024 18:42 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

i love how every efficiency advance in machine learning is approximating [expensive operation] by either ones (just pass it straight through) or zeros (doesn’t matter, just don’t compute it/sparsity)

28.11.2024 18:38 β€” πŸ‘ 20    πŸ” 1    πŸ’¬ 2    πŸ“Œ 0

@bilal2vec is following 20 prominent accounts