Excited to have the Big Picture workshop back for another iteration this year at @aclmeeting.bsky.social
Submit your big picture ideas, consolidation work, phd thesis distillation, etc. by March 5th!
www.bigpictureworkshop.com
w/ Allyson Ettinger, @norakassner.bsky.social, @sebruder.bsky.social
03.02.2026 14:44 β π 9 π 3 π¬ 0 π 0
But I guess this is where the "position" of our paper gets in. At least it wasn't rejected from arXiv π
31.01.2026 20:10 β π 0 π 0 π¬ 0 π 0
a) but not a reasonable one. there are many such papers, many good ones, and it's not like arXiv is this sanctuary of high-quality papers from the first place. I don't see why should we enforce this quasi appearance of quality suddenly.
b) I don't think this is true anymore.
31.01.2026 20:06 β π 0 π 0 π¬ 1 π 0
correct! I combined them together here for brevity (some may say laziness. but in fact arXiv treats them the same, and so do we (at least we try, in our evaluations and instructions).
31.01.2026 19:39 β π 2 π 0 π¬ 1 π 0
Please read the paper. Percentages are far from the only metric to look at.
31.01.2026 19:38 β π 0 π 0 π¬ 1 π 0
Job application process in 2026:
Applicant: I have written 3 papers
Interviewer: where?
Applicant: arXiv
Interviewer: hired!
29.01.2026 18:07 β π 4 π 0 π¬ 0 π 1
No, we unfortunately couldn't get our hands on such data. But from our understanding such data (in the worse case scenario) shouldn't affect the trends we found in our results.
29.01.2026 16:33 β π 2 π 0 π¬ 1 π 0
CS ArXiv recently banned βreview and positionβ papers, but what are those? Do they include more generated content? Who is most affected by this change? @yanai.bsky.social and I dug into the data to find out!
Nearly 50% of Computers & Society papers might be censored, vs 3% of Computer Vision βΌοΈ
29.01.2026 14:14 β π 42 π 19 π¬ 2 π 0
and lastly, a huge thanks to @pangram.com who provided us with credits to detect AI slop in scientific papers!
29.01.2026 14:32 β π 3 π 0 π¬ 0 π 0
In addition, when considering different subfields, we find striking differences in how we classify 'position' papers, leading to huge differences in how this policy affects different CS subfields.
29.01.2026 14:00 β π 6 π 0 π¬ 1 π 1
However! when considering the number of papers from each type, we find there are significantly more non-review papers that are LLM-generated in all categories (cs and physics in the first tweet, math and stats here).
These results question the main motivation of this new policy.
29.01.2026 14:00 β π 5 π 0 π¬ 1 π 0
We then calculate the percentage of LLM-generated papers in each category, and find that indeed, 'position' papers have higher percentages of LLM-generated papers, across the past three years.
29.01.2026 14:00 β π 4 π 0 π¬ 1 π 1
We establish two classifiers that predict based on papers' abstracts* whether they are 'position' vs non-position papers, and whether they are LLM-generated, using two existing classifiers.
We also experiment with classifying the full papers, and reach similar conclusions.
29.01.2026 14:00 β π 4 π 0 π¬ 2 π 0
In this new paper (w/ @mariaa.bsky.social ) we study these questions, and try to provide some answers.
Overall, we find evidence that this decision does not make much sense, and that it should be reconsidered!
29.01.2026 14:00 β π 5 π 0 π¬ 1 π 0
However, they provided no empirical evidence to support this decision. How many LLM-generated papers are being posted on arXiv? How many position papers are posted on arXiv? Do these ratios justify this abrupt solution? What are 'review' papers anyway?
29.01.2026 14:00 β π 4 π 0 π¬ 2 π 0
π¨ New Study π¨
@arxiv.bsky.social has recently decided to prohibit any 'position' paper from being submitted to its CS servers.
Why? Because of the "AI slop", and allegedly higher ratios of LLM-generated content in review papers, compared to non-review papers.
29.01.2026 14:00 β π 29 π 9 π¬ 2 π 2
Microsoft word wizard
16.01.2026 19:03 β π 1 π 0 π¬ 0 π 0
"at least it's not Linkedin!" - if that's not shitposting idk what is
16.01.2026 16:23 β π 3 π 0 π¬ 1 π 0
We'll see about that...
05.01.2026 19:58 β π 3 π 0 π¬ 0 π 0
I think I'm in a secret competition with my co-author about who uses more footnotes in the paper.
05.01.2026 19:12 β π 4 π 0 π¬ 1 π 0
Thanks! I've actually read it, and I feel like it falls mostly under the "for people who don't know how LLMs work" category, for which it can be super useful! But I'm looking for something else.
01.12.2025 13:59 β π 0 π 0 π¬ 0 π 0
Any good tutorials/guides on how to code with LLMs?
What I found so far targeted folks who don't know what they are, mostly setting expectations and saying what LLMs can or cannot do.
I'm beyond that. I'm looking for best practices. Kind of design patterns for vibe coding.
01.12.2025 13:25 β π 3 π 0 π¬ 4 π 0
Estimating the Causal Effect of Early ArXiving on Paper Acceptance
What is the effect of releasing a preprint of a paper before it is submitted for peer review? No randomized controlled trial has been conducted, so we turn to observational data to answer this questio...
A data point in favor is from an observational data study we performed on the effect of early arxiving (= potentially breaking anonymity) on paper acceptance. We measured very small effects, and perhaps more importantly, no difference in the effect on different groups.
arxiv.org/abs/2306.13891
30.11.2025 22:29 β π 1 π 0 π¬ 0 π 0
Interested in interpretability, data attribution, evaluation, and similar topics?
Interested in doing a postdoc with me?
Apply to the prestigious Azrieli program!
Link below π
DMs are open (email is good too!)
13.11.2025 13:59 β π 5 π 1 π¬ 1 π 0
You had one job!
25.10.2025 10:25 β π 7 π 0 π¬ 0 π 1
would you like to have a pop tart?
17.10.2025 17:27 β π 1 π 0 π¬ 1 π 0
cool, let's have a chat then!
13.10.2025 19:45 β π 1 π 0 π¬ 0 π 0
Research Scientist at Meta β’ ex Cohere, Google DeepMind β’ https://www.ruder.io/
Philosopher working on AI alignment, governance and adaptation
Lab: https://mintresearch.org
Self: https://sethlazar.org
Newsletter: https://philosophyofcomputing.substack.com
Prof at EPFL
AI β’ Climbing
That guy who makes visual essays about software at https://samwho.dev.
Developer Educator @ ngrok.com. Want to pair on something ngrok related? Let's do it! https://cal.com/samwho/workhours
He/him.
We're arXiv, the open access repository sharing discoveries and breaking down barriers to cutting edge science for over 30 years.
Postdoc @ TakeLab, UniZG | previously: Technion; TU Darmstadt | PhD @ TakeLab, UniZG
Faithful explainability, controllability & safety of LLMs.
π On the academic job market π
https://mttk.github.io/
William H. Neukom Professor, Stanford Law School. Partner, Lex Lumina LLP. I teach and write in IP, antitrust, internet, and video game law
Independent AI researcher, creator of datasette.io and llm.datasette.io, building open source tools for data journalism, writing about a lot of stuff at https://simonwillison.net/
Postdoc @ Princeton AI Lab
Natural and Artificial Minds
Prev: PhD @ Brown, MIT FutureTech
Website: https://annatsv.github.io/
Assistant Prof @sbucompsc @stonybrooku.
Researcher β @SFResearch
Ph.D. β @ColumbiaCompSci
Human Centered AI / Future of Work / AI & Creativity
Phd student @ University of Mannheim | Social NLP | she/her
π CS Prof at UCLA
π§ Researching reasoning and learning in artificial intelligence: tactable deep generative models, probabilistic circuits, probabilistic programming, neurosymbolic AI
https://web.cs.ucla.edu/~guyvdb/
Sr. Principal Research Manager at Microsoft Research, NYC // Machine Learning, Responsible AI, Transparency, Intelligibility, Human-AI Interaction // WiML Co-founder // Former NeurIPS & current FAccT Program Co-chair // Brooklyn, NY // http://jennwv.com
Assistant Professor the Polaris Lab @ Princeton (https://www.polarislab.org/); Researching: RL, Strategic Decision-Making+Exploration; AI+Law
EMNLP 2025 Program Chair. NLP researcher, working on Responsible AI for multimodal LLMs, automated fact checking. When not looking after two wee ones, I dabble in woodworking, 3D printing, and most recently single-engine flying.
Associate Professor @ UW
Computational Social Science
PhD student at Cambridge University. Causality & language models. Passionate musician, professional debugger.
pietrolesci.github.io