Thrilled to announce the 1st Workshop on Computational Developmental Linguistics (CDL) at ACL 2026 π A new venue at the intersection of development linguistics Γ modern NLP, spearheaded by @fredashi.bsky.social @marstin.bsky.social, and and outstanding team of colleagues!
A thread π§΅
20.01.2026 11:26 β π 22 π 9 π¬ 2 π 0
new year boi
02.01.2026 23:51 β π 2 π 0 π¬ 0 π 0
NEPA
NEPA: Next-Embedding Predictive Autoregression
sihanxu.me/nepa/
Key ideas:
- One self-supervised signal: cosine-style next-embedding prediction
- Autoregression runs directly on the model's native embeddings
- No pixel decoder (& loss), no contrastive pairs, no task-specific heads, no random masks
20.12.2025 00:41 β π 0 π 0 π¬ 0 π 0
Test-Time Training Done Better: From Plastic Adaptation to Elastic Memory
Elastic Test-Time Training (ETTT) that prevents catastrophic forgetting at inference time and overfitting during pretraining LaCT.
In my new blog, βTest-Time Training Done Better: From Plastic Adaptation to Elastic Memory Consolidation,β I introduce a long-context modeling architecture that learns to adapt and memorize at test time by updating a subset of the modelβs weights during inference.
mars-tin.github.io/blogs/posts/...
11.12.2025 16:21 β π 0 π 0 π¬ 0 π 0
Gosh, Iβm getting way too emotional writing my thesis acknowledgements...
24.11.2025 06:59 β π 1 π 0 π¬ 0 π 0
Will be at #NeurIPS2025 (San Diego) Dec 1-9, then in the Bay Area until the 14th. Hmu if you wanna grab coffee and talk about totally random stuff.
Thread with a few things Iβm excited about.
P.S. 4 NeurIPS papers all started pre-May 2024 and took ~1 year of polishing...so proud of the team!
23.11.2025 18:54 β π 0 π 0 π¬ 0 π 0
Trying to decide what to do on the first day of #NeurIPS2025?
Check out my, @marstin.bsky.social and @xiangyue96.bsky.social's tutorial, "The Science of Benchmarking: What's Measured, What's Missing, What's Next" on December 2 from 1:30 to 4:00pm.
benchmarking.science
What will we cover?
1/3
18.11.2025 03:48 β π 21 π 4 π¬ 2 π 1
An Open-Notebook Exploration of Emergent Grounding in LMs
How We Did This Curiosity-Driven Research? An Open-Notebook Exploration of Emergent Grounding in LMs
@fredashi.bsky.social and I wrote a blog for our new mechinterp paper (arxiv.org/abs/2510.13796), including many unpublished and even negative results that we found meaningful to share.
An Open-Notebook Exploration of Emergent Grounding in LMs mars-tin.github.io/blogs/posts/...
21.10.2025 01:40 β π 7 π 0 π¬ 0 π 1
Regrettably canβt attend #COLM2025 due to deadlines, but
Jane and Joyce will be presenting our work. :)
Jane is an exceptional undergraduate researcher and a great collaborator! Go meet her at COLM if youβre curious about her work on mechanistic interpretability, multimodality, & pragmatics!
05.10.2025 22:43 β π 2 π 0 π¬ 0 π 0
π ACL ARR is looking for a Co-CTO to join me lead our amazing tech team and drive the future of our workflow. If youβre interested or know someone who might be, letβs connect!
RTs & recommendations appreciated.
29.09.2025 21:30 β π 4 π 3 π¬ 1 π 0
Unfortunately, Iβll be missing #ACL2025NLP this year β but here are a few things Iβm excited about! π
28.07.2025 06:19 β π 1 π 0 π¬ 0 π 0
Congratulations!!
22.07.2025 05:35 β π 1 π 0 π¬ 1 π 0
with @fredashi.bsky.social / Jiayuan Mao / @djiafei.bsky.social / @manlingli.bsky.social / David Hsu / Parisa Kordjamshidi
14.07.2025 20:16 β π 1 π 0 π¬ 0 π 0
π£ Excited to announce SpaVLE: #NeurIPS2025 Workshop on Space in Vision, Language, and Embodied AI!
Join us in San Diego to push the frontiers of spatial understanding and reasoning across CV, NLP, and robotics!
π space-in-vision-language-embodied-ai.github.io
14.07.2025 20:13 β π 2 π 0 π¬ 1 π 0
#CoreCognition #LLM #multimodal #GrowAI We spent 3 years to curate 1503 classic experiments spanning 12 core concepts in human cognitive development and evaluated on 230 MLLMs with 11 different prompts for 5 times to get over 3.8 millions inference data points.
A thread (1/n) - #ICML2025 β
30.06.2025 06:07 β π 13 π 9 π¬ 1 π 0
New Paper Alert βΌοΈ Current VLMs completely fail human gaze understanding π and scaling does NO help βΌοΈ
However, humans, since an extremely age π§, are extremely sensitive to other people's gaze π π
No mentors, no labs, only pre-doc students, 111 VLMs, and we did it π
11.06.2025 23:21 β π 6 π 5 π¬ 1 π 1
SimWorld
SimWorld: A World Simulator for Scaling Photorealistic Multi-Agent Interactions
& @tianminshu.bsky.social (+ @marstin.bsky.social, @zhitinghu.bsky.social, βͺ@lianhui.bsky.social & more) will present βSimWorld: A World Simulator for Scaling Photorealistic Multi-Agent Interactions,β an @unrealengine.bsky.social-based sim that generates unlimited/diverse urban environments: (13/14)
10.06.2025 19:45 β π 1 π 1 π¬ 1 π 0
At Albuquerque Now :)
30.04.2025 05:49 β π 1 π 0 π¬ 0 π 0
See you at #NAACL2025! I will talk about grounded lexicon acquisition and scaling mechanistically grounded vision language models. Happy to chat if you are around :)
29.04.2025 23:28 β π 1 π 0 π¬ 0 π 0
VLMs Are Not Pragmatically Competent in Referring Expression Generation
VLMs fail to refer like humans. Our study reveals widespread pragmatic issues in GPT-4o, LLaVA, and others, showing how their expressions often violate Gricean maxims.
We introduce RefOI, a new dataset of 1.5k objects, each with 3 written and 2 spoken human-produced referring expressions. We also release RefOI-TLHF, a large dataset of token-level human feedback for 10.6k referring expressions.
πhttps://vlm-reg.github.io/
πhttps://arxiv.org/abs/2504.16060
23.04.2025 17:55 β π 1 π 0 π¬ 0 π 0
Vision-Language Models are not yet pragmatically optimal.
We identify 3 key failures of pragmatic competence in referring expression generation with VLMs: (1) cannot uniquely refer to the referent, (2) include excessive or irrelevant information, and (3) misalign with human pragmatic preferences.
23.04.2025 17:55 β π 4 π 3 π¬ 1 π 1
πΉ ICLR BiAlign Workshop:
Weβre hosting the Bidirectional Human-AI Alignment Workshop (BiAlign).
π Apr 28, (Garnet 216β214)
Website: bialign-workshop.github.io
Iβll join remotely β huge thanks to @huashen.bsky.social for leading this!
19.04.2025 01:53 β π 4 π 0 π¬ 1 π 0
πΉ ICLR Oral Paper:
Do Vision-Language Models Represent Space and How?
π Oral: Apr 25, 3:42β3:54 a.m. (Session 4C)
π Poster: Thu, Apr 24, 10 p.m.β12:30 a.m. (Hall 3 + 2B, #212)
Website: spatial-comfort.github.io
Big thanks to @fredashi.bsky.social for presenting on site!
19.04.2025 01:53 β π 0 π 0 π¬ 1 π 0
I wonβt be attending #ICLR2025 in person since #NAACL2025 follows right after, but here are a few things Iβm excited about (all time in EDT) β¬οΈ
19.04.2025 01:53 β π 0 π 0 π¬ 1 π 0
π Out of these, 72 papers were accepted, including 5 tiny papers. 10 papers were selected for oral presentations: 2 at CHI and 8 at ICLR. Award winners will be announced during the workshop!
15.04.2025 20:55 β π 0 π 0 π¬ 1 π 0
π¬ We received over 100 submissions, each reviewed by 2β4 expert reviewers, with ethical assessments included when appropriate. Our program committee features leading researchers in NLP, RL, HCI, ML, and AI/ML Ethics, carefully selected based on scholarly merit and expertise.
15.04.2025 20:55 β π 0 π 0 π¬ 1 π 0
π Special thanks to Tammy Masterson, Technical Partnerships Lead at the AI Security Institute, who will be joining us as a panelist.
15.04.2025 20:55 β π 0 π 0 π¬ 1 π 0
π We are grateful to our gold sponsors, Prolific and Layer 6 AI of TD Bank Group, for their generous support in funding paper awards and travel grants.
15.04.2025 20:55 β π 0 π 0 π¬ 1 π 0
Multimodal Communication and Learning in Social Interactions (CoCoDev team). Associate Professor of Computer/Cognitive Science at Aix-Marseille University.
afourtassi.github.io
Assistant Professor at UCSD. Artificial Intelligence. https://zhiting.ucsd.edu
NLP/ML Assistant Professor @UCSD
He teaches information science at Cornell. http://mimno.infosci.cornell.edu
Assi. Prof @UofTCompSci. Postdoc @MPI_IS w/ @bschoelkopf. Research on (1) @CausalNLP and (2) NLP4SocialGood @NLP4SG. Mentor & mentee @ACLMentorship.
Associate professor at CMU, studying natural language processing and machine learning. Co-founder All Hands AI
Google Chief Scientist, Gemini Lead. Opinions stated here are my own, not those of Google. Gemini, TensorFlow, MapReduce, Bigtable, Spanner, ML things, ...
Postdoc @LTIatCMU. PhD from Ohio State @osunlp. Author of MMMU, MAmmoTH. Training & evaluating foundation models. Previously @MSFTResearch. Opinions are my own.
Postdoc at UW and Doctor of NLP/Vision+Language from UCSB
Evals, metrics, multilinguality, multiculturality, multimodality, and (dabbling in) reasoning
100% Product of public schools
https://saxon.me/
Assistant Professor at UNC Chapel Hill, previously a postdoc at Meta AI, PhD from UPenn, a basketball enthusiast π.
π https://www.gedasbertasius.com/
π https://scholar.google.com/citations?hl=en&user=8FWkjw8AAAAJ
Waiting on a robot body. All opinions are universal and held by both employers and family. ML/NLP professor.
nsaphra.net
Robotics PhD student @uwcse|Graduate Student Researcher @allen_ai |Ex-@NVIDIA |@ASTARsg scholars|BEng from @ntueee. Research in robot learning and embodied AI
www.duanjiafei.com
https://growing-ai-like-a-child.github.io/
Philosopher, Scientist, Engineer
https://hokindeng.github.io/
Data Science junior @Umich
Β· Starting research in vision-language models and pragmatic generation Β· Exploring how AI communicates like humans
homepage: jingding-ai.github.io
Assistant Professor at UIUC. Interested in Trustworthy ML, Data Attribution, GenAI Copyright.
https://jiaqima.github.io/
foundations of cogsci @UMich. prev. @UCL @Edin. interested in consciousness, the self, emotion, language, philosophy & theories of AI