Nikhil Garg's Avatar

Nikhil Garg

@nkgarg.bsky.social

I study algorithms/learning/data applied to democracy/markets/society. Asst. professor at Cornell Tech. https://gargnikhil.com/. Helping building personalized Bluesky research feed: https://bsky.app/profile/paper-feed.bsky.social/feed/preprintdigest

2,785 Followers  |  1,277 Following  |  357 Posts  |  Joined: 23.11.2023
Posts Following

Posts by Nikhil Garg (@nkgarg.bsky.social)

A picture of Joe Halpern smiling in green shirt in front of a blue background.

A picture of Joe Halpern smiling in green shirt in front of a blue background.

Today arXiv remembers our colleague Joe Halpern, who was instrumental in founding arXiv's CS section.

Joe's passions ranged far & wide and we're lucky that arXiv was one of them. Joe, thank you for giving so much to arXiv - you are missed.

blog.arxiv.org/2026/02/27/remembering-joe-halpern

27.02.2026 18:38 β€” πŸ‘ 41    πŸ” 11    πŸ’¬ 2    πŸ“Œ 2
Cornell University, Computer Science Job #AJO31698, Lecturer/Senior Lecturer, Computer Science, Cornell University, New York, New York, US

Care about preparing people to contribute responsibly to building the next generation of AI and technology?

Full-time (or at least 50%) lecturer position at Cornell Tech just posted, teaching computer science or related topics.

academicjobsonline.org/ajo/jobs/31698

17.02.2026 22:37 β€” πŸ‘ 12    πŸ” 10    πŸ’¬ 0    πŸ“Œ 2
Screenshot from the paper with a figure showing 15 scatterplots in a grid., evaluating LLM-as-judge on HELM. In each plot, one model is used as the judge. Each dot is another model; the y-axis is the
accuracy inflation (compared to ground truth) of using the given model as the judge, and the x-axis is the model’s true accuracy. A
vertical red line corresponds to the true accuracy of the judge. Each judge tends to inflate the accuracy of models that are less accurate
than itself, especially models from the same provider or family

Screenshot from the paper with a figure showing 15 scatterplots in a grid., evaluating LLM-as-judge on HELM. In each plot, one model is used as the judge. Each dot is another model; the y-axis is the accuracy inflation (compared to ground truth) of using the given model as the judge, and the x-axis is the model’s true accuracy. A vertical red line corresponds to the true accuracy of the judge. Each judge tends to inflate the accuracy of models that are less accurate than itself, especially models from the same provider or family

yeah! In this paper in ICML25, we found both directions of this -- in LLM as judge, using a bigger/more accurate model inflates accuracies bc of correlated errors, and using a worse model deflates them for the reason in the above paper

arxiv.org/abs/2506.07962

23.02.2026 12:36 β€” πŸ‘ 10    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

Wonderful, congrats!

17.02.2026 18:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Toy Models of Superposition

Se Kenny's thread, and e.g., this blog post: transformer-circuits.pub/2022/toy_mod... by Anthropic. Kenny formalizes some of the claims made therein

17.02.2026 18:44 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Really glad to share this work by @kennypeng.bsky.social, toward better theoretical foundation for the linear representation hypothesis.

17.02.2026 18:42 β€” πŸ‘ 13    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

This call is still open. I am looking to recruit, as well as many other faculty at Cornell. We review folders as they come, and will send offers until all positions are filled.

Please share with your network πŸ™

17.02.2026 04:50 β€” πŸ‘ 10    πŸ” 8    πŸ’¬ 0    πŸ“Œ 0

Thanks for sharing! Beautiful

08.02.2026 20:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This is an intriguing paper with numerous examples of using an LLM to generate formal proofs and ideas, and a nice cookbook for how to do it. Refreshingly, the paper stays away from hype and as far as possible states very clearly how the LLMs are used and the key nature of human-AI interaction.

08.02.2026 06:14 β€” πŸ‘ 14    πŸ” 6    πŸ’¬ 0    πŸ“Œ 0

What does that mean?

08.02.2026 20:05 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Preview
First Proof To assess the ability of current AI systems to correctly answer research-level mathematics questions, we share a set of ten math questions which have arisen naturally in the research process of the au...

Ten math problems with proofs known the authors. Proofs are encrypted until Feb 13. For all problems, authors claim both AI-based literature searches and zero-shot attempts at proofs failed. If you want to take a crack, you have until next Friday (2/13)!

08.02.2026 15:59 β€” πŸ‘ 26    πŸ” 5    πŸ’¬ 0    πŸ“Œ 3
Video thumbnail

Our paper β€œInferring fine-grained migration patterns across the United States” is now out in @natcomms.nature.com! We released a new, highly granular migration dataset. 1/9

05.02.2026 17:30 β€” πŸ‘ 70    πŸ” 27    πŸ’¬ 2    πŸ“Œ 5

yeah, I'm not sure it's "useful yet" but perhaps "I should start learning to be on the cutting edge for when it is imminently useful"

04.02.2026 22:37 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Until seeing this paper, my view was, "Lean is promising but doesn't have the libraries for the type of math I typically do." And it seems like folks have made substantial progress in fixing that issue

04.02.2026 19:21 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

I have been waiting for something like this to onboard onto Lean (alongside an LLM agent), and it seems like that moment is here

04.02.2026 19:20 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🚨New WP "@Grok is this true?"
We analyze 1.6M factcheck requests on X (grok & Perplexity)
πŸ“ŒUsage is polarized, Grok users more likely to be Reps
πŸ“ŒBUT Rep posts rated as false more oftenβ€”even by Grok
πŸ“ŒBot agreement with factchecks is OK but not great; APIs match fact-checkers
osf.io/preprints/ps...

03.02.2026 21:55 β€” πŸ‘ 118    πŸ” 48    πŸ’¬ 2    πŸ“Œ 3

ha I had seen snow before, but before I moved to NYC the coldest/densest place I'd really lived was SF

31.01.2026 13:53 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Wow! I admit I thought that bodies of water freezing over was a movies thing...

31.01.2026 13:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
A search for factors for algorithm understanding results in multiple terms displayed as documents, including available, compact, and aligned. These are shown to be necessary and sufficient. Other, similar terms are shown in the background faded, like intuitive, rule-based, grounded, modular, linear, decomposable, accurate, symbolic, causal, and personalized.

A search for factors for algorithm understanding results in multiple terms displayed as documents, including available, compact, and aligned. These are shown to be necessary and sufficient. Other, similar terms are shown in the background faded, like intuitive, rule-based, grounded, modular, linear, decomposable, accurate, symbolic, causal, and personalized.

Is the only way we can create algorithms that people understand to make them trivially simple? We argue, no.

People can predict the behavior of algorithms that are arbitrarily complex, if and only if they are available, compact and aligned.

arxiv.org/abs/2601.18966

29.01.2026 18:49 β€” πŸ‘ 38    πŸ” 10    πŸ’¬ 2    πŸ“Œ 3
Preview
Fairness in PCA-Based Recommenders

πŸŽ™οΈ I had a great time joining the Data Skeptic podcast to talk about my work on recommender systems

If you're interested in embeddings, aligning group preferences, or music recommendations, check out the episode below πŸ‘‡

open.spotify.com/episode/6IsP...

28.01.2026 16:22 β€” πŸ‘ 14    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0
Bluesky Network Analyzer Find accounts that you don't follow (yet) but are followed by lots of accounts that you do follow.

There is a third-party app for this, you can find it here: bsky-follow-finder.theo.io

My impression with bluesky is that a lot of things exist but there is just no way of finding them.

@nkgarg.bsky.social

23.01.2026 16:14 β€” πŸ‘ 17    πŸ” 4    πŸ’¬ 3    πŸ“Œ 3

ah amazing thanks for sharing!

23.01.2026 16:23 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I've been playing with various side projects with codex. And even though I do not have time for next few weeks, this post is very close to sniping me into finally building this...

23.01.2026 15:48 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yeah, I don't understand the need to ask certain questions during a talk

22.01.2026 23:04 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This is indeed more or less exactly what happened...

22.01.2026 22:00 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Check out our new paper at #AAAI 2026! I’ll be presenting in Singapore at Saturday’s poster session (12–2pm). This is joint work with @shuvoms.bsky.social, @bergerlab.bsky.social, @emmapierson.bsky.social, and @nkgarg.bsky.social. 1/9

20.01.2026 16:08 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

Diversity is definitely top of mind for us!

17.01.2026 17:06 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Congrats team!

17.01.2026 03:11 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Thank you!

16.01.2026 04:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yep we filtered those out!

bsky.app/profile/sjgr...

16.01.2026 03:54 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0