@apoorvkh.com.bsky.social
cs phd student at brown https://apoorvkh.com
Will be at ACL this week! #ACL2025 #ACL2025NLP
Presenting Tian Yunβs paper on abstract reasoners at CoNLL on Thursday.
Iβve been investigating how LLMs internally compose functions lately. Happy to chat about that (among other things) and hang out in Vienna!
Curious how many papers were assigned to reviewers on average! Review quality seems better than average from my small sample size. Wondering if that correlates with a lower reviewer load? E.g. I only received 2 papers to review.
29.05.2025 18:49 β π 0 π 0 π¬ 0 π 0Scores of R1, Flash-thinking, Claude 4.7, QwQ, o1-pro, o3-mini on USAMO 2025. Scores less than 5% of max score.
Tests on USAMO immediately after problems were posted yield surprisingly bad model performance. Suggests there's much more training on test than expected.
arxiv.org/abs/2503.219...
Just read that AIβs energy consumption in data centers is nothing to be worried about because most of the hyperscale datacenters running AI are "powered by renewable energy or low-carbon nuclear power."
Let's debunk that, shall we?
If you're in the northeastern US and you're submitting a paper to COLM on March 27, you should absolutely be sending its abstract to New England NLP on March 28.
19.03.2025 19:59 β π 8 π 3 π¬ 0 π 0+ No system pre-reqs, multi-stage PyTorch workflows in one script, CLI integrations, catching system failures as exceptions, SLURM support, better logging, and so much more!
Additional fine-tuning examples in our docs with:
@pytorch.org, Deepspeed, @lightningai.bsky.social, HF Accelerate
A cool side-effect: fine-tune any LLM (from
@huggingface
transformers) on any text dataset *with multiple nodes* in just *one command*.
torchrun.xyz/examples/tra...
It's a replacement for CLI tools, like "torchrun".
Most basic usage: specify some (SSH-enabled) machines you want to parallelize your code on. Then launch a function onto that configuration.
All from inside your Python script!
We made a library (torchrunx) to make multi-GPU / multi-node PyTorch easier, more robust, and more modular! π§΅
github.com/apoorvkh/tor...
Docs: torchrun.xyz
`(uv) pip install torchrunx` today!
(w/ the very talented, Peter Curtin, Brown CS '25)
Paper: A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers
β¨How does the depth of a transformer affect its reasoning capabilities? New preprint by myself and @Ashish_S_AI shows that a little depth goes a long way to increase transformersβ expressive power
We take this as encouraging for further research on looped transformers!π§΅
(1/9) Excited to share my recent work on "Alignment reduces LM's conceptual diversity" with @tomerullman.bsky.social and @jennhu.bsky.social, to appear at #NAACL2025! π
We want models that match our values...but could this hurt their diversity of thought?
Preprint: arxiv.org/abs/2411.04427
I started a blog! First post is everything I know about setting up (fast, reproducible, error-proof) Python project environments using the latest tools. These methods have saved me a lot of grief. Also a short guide to CUDA in appendix :)
blog.apoorvkh.com/posts/projec...
I think typing my code and using a linter (ruff) + static type checker (pyright) saves me a lot of grief.
25.01.2025 18:49 β π 1 π 0 π¬ 0 π 0Can GANs compete in 2025? In 'The GAN is dead; long live the GAN! A Modern GAN Baseline', we show that a minimalist GAN w/o any tricks can match the performance of EDM with half the size and one-step generation - github.com/brownvc/r3gan - work of Nick Huang, @skylion.bsky.social, Volodymyr Kuleshov
10.01.2025 19:08 β π 69 π 14 π¬ 3 π 1A couple sources for academic talks that I really like!
Cohere For AI (www.youtube.com/playlist?lis...)
Simons Institute (www.youtube.com/@SimonsInsti...)
Let he who hath not \usepackage[subtle]{savetrees}
18.12.2024 01:27 β π 13 π 1 π¬ 1 π 0Slides from the tutorial are now posted here!
neurips.cc/media/neurip...
βThey said it could not be doneβ. Weβre releasing Pleias 1.0, the first suite of models trained on open data (either permissibly licensed or uncopyrighted): Pleias-3b, Pleias-1b and Pleias-350m, all based on the two trillion tokens set from Common Corpus.
05.12.2024 16:39 β π 251 π 85 π¬ 12 π 19I am an ex-Paperpile user and am liking Zotero lately! Free storage from the university helps.
27.11.2024 05:15 β π 0 π 0 π¬ 0 π 0Lots of folks talking about scaling LLM inference over this last year
Internally, Iβve been developing and using a library that makes this extremely easy, and I decided to open-source it
Meet the decoding library: github.com/benlipkin/de...
1/7
βTurnβ a decoder into an encoder with LLM2Vec (github.com/McGill-NLP/l...). Seen at COLM 2024 :)
If you want the naive, training-free / model-agnostic approach: their related work section says it is most common to using the final tokenβs last hidden state.
Okay genius idea to improve quality of #nlp #arr reviews. Literally give gold stars to the best reviewers, visible on open review next to your anonymously ID during review process.
Hereβs why it would work, and why would you should RT this fab idea:
Thanks and great! Hope you are likewise doing well!
21.11.2024 21:29 β π 1 π 0 π¬ 1 π 0Would be great to join, thanks!
21.11.2024 21:15 β π 1 π 0 π¬ 1 π 0Excited to release Tulu 3! We worked hard to try and make the best open post-training recipe we could, and the results are good!
I was lucky enough to work on almost every stage of the pipeline in one way or another. Some comments + highlights β¬οΈ
You can find the βauthorsβ cutβ at: arxiv.org/abs/2410.23261
21.11.2024 16:23 β π 1 π 0 π¬ 1 π 0Nature wrote a nice article about our work!
www.nature.com/articles/d41...