I really enjoyed my MLST chat with Tim @neuripsconf.bsky.social about the research we've been doing on reasoning, robustness and human feedback. If you have an hour to spare and are interested in AI robustness, it may be worth a listen π§
Check it out at youtu.be/DL7qwmWWk88?...
19.03.2025 15:11 β
π 8
π 3
π¬ 0
π 0
"Rather than being animals that *think*, we are *animals* that think"; the last sentence of Tom Griffiths's characterisation of human intelligence through limited time, compute, and communication hits different today than it did 4 years ago.
22.12.2024 11:04 β
π 16
π 1
π¬ 1
π 0
leave parrots alone!!
15.12.2024 18:46 β
π 2
π 0
π¬ 0
π 0
Sometimes o1's thinking time almost feels like a slight. o1 is like "oh I thought about this uninvolved question of yours for 7 seconds and here is my 20 page essay on it"
15.12.2024 17:38 β
π 18
π 2
π¬ 1
π 0
The broader spectrum of in-context learning
The ability of language models to learn a task from a few examples in context has generated substantial interest. Here, we provide a perspective that situates this type of supervised few-shot learning...
What counts as in-context learning (ICL)? Typically, you might think of it as learning a task from a few examples. However, weβve just written a perspective (arxiv.org/abs/2412.03782) suggesting interpreting a much broader spectrum of behaviors as ICL! Quick summary thread: 1/7
10.12.2024 18:17 β
π 123
π 32
π¬ 2
π 1
Big congratulations to Dr. @jumelet.bsky.social for obtaining his PhD today and crafting a beautiful thesis full of original and insightful work!! π arxiv.org/pdf/2411.16433?
10.12.2024 15:07 β
π 15
π 3
π¬ 0
π 0
I'll be at NeurIPS tues-sun, send me a message if you'd like to chat!
08.12.2024 16:51 β
π 7
π 0
π¬ 0
π 0
Cool that those experiments changed your mind :) The referenced appendix was important to convince myself of what we eventually concluded (that the correlations indicate procedural knowledge). And thank you for the praise!! What kind of ideas did you get?
01.12.2024 16:20 β
π 1
π 0
π¬ 0
π 0
If that's how you define retrieval, then they are doing retrieval under your definition. The heavy lifting is of course done by the word "synthesize", how do they do that? That's what we're characterising in the paper
01.12.2024 14:51 β
π 2
π 0
π¬ 1
π 0
This is an incredible paper that I've longed to do for a long time. However the engineering challenges were far too daunting, so my collaborators and I settled for indirect evidence for this hypothesis instead (or did other things).
30.11.2024 17:10 β
π 53
π 2
π¬ 2
π 0
It should be much less computationally expensive to do for fine tuning data
01.12.2024 09:10 β
π 2
π 0
π¬ 0
π 0
Just want to add to Stella's responses that the reason we went with procedural knowledge very much came from the correlation results; documents influence each query with the same underlying task similarly, even though the task is applied to different numbers for different queries.
01.12.2024 09:07 β
π 2
π 0
π¬ 1
π 0
Definitely! Next time will be Christmas so I presume that's not ideal, but I can reach out when I know the next time I will be there?
01.12.2024 09:02 β
π 0
π 0
π¬ 0
π 0
What did you think
01.12.2024 09:01 β
π 0
π 0
π¬ 0
π 0
Yeah. I do think as you become more senior you become better at determining from the intro whether a paper is likely to be good or bad. The point is just that we should still actively keep an open mind when reading the rest of the paper
28.11.2024 12:31 β
π 1
π 0
π¬ 0
π 0
I learn so much from reviewing, itβs the papers I review that I keep coming back to for my own ideas and citations. They broaden and deepen my view on the field. Letβs give it the time it deserves.
27.11.2024 17:25 β
π 21
π 2
π¬ 1
π 0
Itβs actually pretty cool if you as a reviewer get to make papers better by suggesting improvements. This cycle, Iβve given an 8 where all other reviewers gave a rejecting rating. Now, the scores are 8, 5, 8, 6, 8. Pretty exciting, if you ask me.
27.11.2024 17:25 β
π 11
π 0
π¬ 1
π 1
You donβt have to add these to the review (unless itβs TMLR). But hold yourself accountable when you are rejecting it. What could the authors do to lift your scores? If the answer is nothing, be sure to have a good reason for this. If there is something, tell the authors.
27.11.2024 17:25 β
π 15
π 1
π¬ 1
π 0
Thereβs an easy way to hold yourself accountable (thanks TMLR guidelines βοΈ): "make a list of proposed adjustments to the submission, specifying for each whether they are critical to securing your recommendation for acceptance or would simply strengthen the work in your view."
27.11.2024 17:25 β
π 19
π 1
π¬ 2
π 0
The art of rebuttal is to learn how to stick firmly to the points you believe are important, while at the same time allowing yourself to be wrong. Admitting when you might be misunderstanding (after all, the authors probably spent about ~1000x more time thinking about it).
27.11.2024 17:25 β
π 16
π 2
π¬ 1
π 0
The hardest part is to keep an open mind all the way down π’. The rebuttal phase is the kicker. If you donβt spend enough time in this phase, just donβt sign up to be a reviewer, because itβs incredibly demoralising to people who work months to years on a submission.
27.11.2024 17:25 β
π 17
π 1
π¬ 2
π 1
Iβve heard people say they know whether they will accept or reject a paper after reading the abstract/intro. Thatβs great, but what is even greater is when you realise that is *just presentation*, and the soundness and contribution are *not* going to be determined by that part.
27.11.2024 17:25 β
π 23
π 2
π¬ 2
π 1
Reviewing requires constant questioning of the motive behind your responses, every step of the way. Which btw, according to chatty, will help you become a better scientist yourself.
27.11.2024 17:25 β
π 16
π 0
π¬ 1
π 0
The art is to lift up the best bits of the paper together with the authors, not to call missing baseline and be done with it.
27.11.2024 17:25 β
π 24
π 1
π¬ 1
π 0
Personally, reviewing for NeurIPS a couple years back changed me as a reviewer. For one paper I rejected, I kept citing it throughout the year to people for a finding it had. This made me realise it was a good paper, it just had some easy targets for rejection.
27.11.2024 17:25 β
π 67
π 8
π¬ 2
π 1
To paraphrase Dennett (rip π), the goal of reviewing is to determine truth, not to conquer your opponent.
Too many reviewers seem to not have internalised this. In my opinion, this is the hardest lesson a reviewer has to learn, and I want to share some thoughts.
27.11.2024 17:25 β
π 47
π 9
π¬ 3
π 1
Do you know what rating youβll give after reading the intro? Are your confidence scores 4 or higher? Do you not respond in rebuttal phases? Are you worried how it will look if your rating is the only 8 among 3βs? This thread is for you.
27.11.2024 17:25 β
π 77
π 20
π¬ 4
π 3
didnt mean to ask chatgpt how to do it, but rather ask chatgpt for pointers to papers on the topic ;)
26.11.2024 11:33 β
π 0
π 0
π¬ 1
π 0