Still no bookmarks?
05.02.2025 07:04 β π 0 π 0 π¬ 0 π 0Still no bookmarks?
05.02.2025 07:04 β π 0 π 0 π¬ 0 π 0Hi, the Microsoft Translator research team is looking for an intern for the summer. If you a PhD student in Machine Translation, Natural Language Processing, or related, check it out: aka.ms/mtintern
28.01.2025 17:55 β π 5 π 4 π¬ 0 π 0
Just 10 days after o1's public debut, weβre thrilled to unveil the open-source version of the technique behind its success: scaling test-time compute
By giving models more "time to think," Llama 1B outperforms Llama 8B in mathβbeating a model 8x its size. The full recipe is open-source!
Rant: Apparently every vector-based sentence alignment tool insists on having an unusable file-based API.
16.12.2024 21:49 β π 0 π 0 π¬ 0 π 0It's more confusing than that. It does exist and seems to mean crucifix which had me even more confused. Suddenly very high stakes π
16.12.2024 18:26 β π 0 π 0 π¬ 1 π 0Rood?
16.12.2024 18:19 β π 0 π 0 π¬ 1 π 0Missing bookmarks are a much bigger deal for me. But I think it's funny that they didn't go for one of the most requested features. Seemed like an easy win.
16.12.2024 18:11 β π 0 π 0 π¬ 1 π 0
Wrote up some notes on Microsoft's new Phi-4 LLM. They trained it on a LOT of synthetic data, and the details of how and why they did that are really interesting.
https://simonwillison.net/2024/Dec/15/phi-4-technical-report/
Chart of time vs: - number of cameras (exponentially increasing ), - giant squid footage (exponentially increasing ), - bigfoot footage (small and not increasing), and - good quality UFO footage (small and not increasing)
It's messier, but I think this one slaps the point home a bit stronger by adding the giant squid footage. I think unique weather, like lighting sprites, would make the point just as well.
14.12.2024 20:25 β π 91 π 3 π¬ 2 π 0
the anthropomorphizing in this LLM scheming paper is through the roof and the interpretations are wild, but still a cute set of experiments and a fun skim, showing some interesting behaviors.
arxiv.org/abs/2412.04984
π Introducing the Byte Latent Transformer (BLT) β A LLM architecture that scales better than Llama 3 using patches instead of tokens π€―
Paper π dl.fbaipublicfiles.com/blt/BLT__Pat...
Code π οΈ github.com/facebookrese...
So... no edit button, huh?
13.12.2024 23:33 β π 2 π 0 π¬ 2 π 0Oh no no. VSCode is a an actual recommendation. My actual favorite piece of software that I didn't write.
13.12.2024 23:25 β π 2 π 0 π¬ 1 π 0Whispers in Microsoft: VSCode
13.12.2024 23:11 β π 2 π 0 π¬ 1 π 0This place needs bookmarks.
13.12.2024 18:28 β π 0 π 0 π¬ 0 π 0Leave links. I am old.
12.12.2024 20:47 β π 1 π 0 π¬ 1 π 0And they laughed at us when we pursued PhDs. Who's laughing now?
12.12.2024 19:50 β π 3 π 0 π¬ 0 π 0Ah, are these size-limited? And you guys continue with running numbering?
12.12.2024 19:46 β π 0 π 0 π¬ 2 π 0Oh. There you are. Where's that starter pack?
12.12.2024 19:40 β π 1 π 0 π¬ 1 π 0You convinced me π
12.12.2024 17:46 β π 1 π 0 π¬ 2 π 0Hi!
11.12.2024 08:13 β π 10 π 1 π¬ 1 π 0