Sayan

Sayan

@shockroborty.bsky.social

ML/LLM research. Prev @BrownUniversity

54 Followers 374 Following 17 Posts Joined Oct 2024
9 months ago

lmao instant follow

1 0 0 0
11 months ago

You can totally see how this mf bankrupted a casino.

60,350 10,322 799 354
1 year ago
Preview
moonshotai (Moonshot AI) Org profile for Moonshot AI on Hugging Face, the AI community building the future.

Moonshot + Muon

A new 16B model

The Muon optimizer is 2x more data efficient than AdamE, but only for matrix parameters

note: this is a big deal

huggingface.co/moonshotai

19 3 1 0
1 year ago

super valuable stuff

1 0 0 0
1 year ago

👀🙏

1 0 0 0
1 year ago
Post image

In case it interests anyone, I managed to set up a demo of GRPO RL training in Colab. It’s an adaptation of Will Brown instant classic for math reasoning. Replace llama 1B with qwen 0.5b and inference with vllm. Full training in about 2 hours.

colab.research.google.com/drive/1bfhs1...

38 5 1 1
1 year ago

yeah, academia moved over here but the engineers and researchers at frontier labs are still at twitter sadly

3 0 1 0
1 year ago
Post image

I don’t understand this eval. why compare their deep research model with gemini thinking, when gemini deep research exists

2 0 0 0
1 year ago

at this point, gpt3 and claude sonnet/haiku could easily be open sourced

1 0 0 0
1 year ago

little disappointed seeing reactions of researchers from frontier labs on deepseek. science is not a zero sum game. we should really applaud the open weights, reproducibility, MIT license and detailed report which we hardly see in this decade. gracefulness besides the bias would’ve been nice

3 0 0 0
1 year ago

The inference speed is amazing!

11 1 0 0
1 year ago

Just saw ScaleAI's front page ad on "America must win the AI war".

I'm afraid in the AI war only Palantir wins.

0 0 0 0
1 year ago

this is how I prepare for interviews, especially around research. Super helpful

1 0 0 0
1 year ago

internal search is very interesting, i hope the implementation is easy to read through

0 0 0 0
1 year ago

How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this:

Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢

🧵⬇️

854 138 36 24
1 year ago

The most realistic reason to be pro open source AI is to reduce concentration of power.

190 36 6 5
1 year ago

Most elaborate game of chinese whisper

0 0 0 0
1 year ago

Good thread explaining general public reaction on data scraping

0 0 0 0
1 year ago

I believe o1 will be replicated soon. First by meta and then a truly open source release with datasets and training recipe by @ai2.bsky.social team

0 0 0 0
1 year ago

Outside tech, I see a lot of AI fear and hatred. Usually the argument is on AI taking jobs and creative tasks. I don't remember seeing this kind of general consensus of hatred and fear about a new technology before

2 0 1 0
1 year ago
The OLMo 2 models sit at the Pareto frontier of training FLOPs vs model average performance.

Meet OLMo 2, the best fully open language model to date, including a family of 7B and 13B models trained up to 5T tokens. OLMo 2 outperforms other fully open models and competes with open-weight models like Llama 3.1 8B — As always, we released our data, code, recipes and more 🎁

151 36 5 12
1 year ago

So cool!! Someone needs to create a feed for papers and models!

0 0 0 0
1 year ago

i keep forgetting to include this cause i always assume people do this by default. Any time there is an exponent or a norm, you should be working in the highest practical precision

25 1 0 0
1 year ago

Posting a call for help: does anyone know of a good way to simultaneously treat both POTS and Ménière’s disease? Please contact me if you’re either a clinician with experience doing this or a patient who has found a good solution. Context in thread

128 71 15 5
1 year ago

There is something similar. Check if your discord channel is covered by swyx's ai news newsletter. If not, you can pay for customization.

0 0 0 0
1 year ago

📢 Ultimate test of #NLP bluesky:

I need emergency reviewers for NAACL submissions on encoders (one multilingual, one for sentence embeddings). Help a desperate editor abandoned by the ACs! Author response starts tomorrow, so that's a true emergency.

If you're my hero, lmk your openreview profile.

30 26 1 1