Martin Ziqiao Ma's Avatar

Martin Ziqiao Ma

@marstin.bsky.social

https://mars-tin.github.io member of technical staff @ __ member of less technical stuff @ aclmentorship phd @ umich; bs @ sjtu prev. @ mit-ibm-watson, adobe, amazon Herborium Lover, Fortune Teller, PokΓ©mon Trainer, Szechuan Cuisine Chef.

183 Followers  |  67 Following  |  46 Posts  |  Joined: 13.11.2024  |  1.7953

Latest posts by marstin.bsky.social on Bluesky

Post image

Thrilled to announce the 1st Workshop on Computational Developmental Linguistics (CDL) at ACL 2026 πŸŽ‰ A new venue at the intersection of development linguistics Γ— modern NLP, spearheaded by @fredashi.bsky.social @marstin.bsky.social, and and outstanding team of colleagues!

A thread 🧡

20.01.2026 11:26 β€” πŸ‘ 22    πŸ” 9    πŸ’¬ 2    πŸ“Œ 0
Post image

new year boi

02.01.2026 23:51 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
NEPA

NEPA: Next-Embedding Predictive Autoregression
sihanxu.me/nepa/

Key ideas:
- One self-supervised signal: cosine-style next-embedding prediction
- Autoregression runs directly on the model's native embeddings
- No pixel decoder (& loss), no contrastive pairs, no task-specific heads, no random masks

20.12.2025 00:41 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Test-Time Training Done Better: From Plastic Adaptation to Elastic Memory Elastic Test-Time Training (ETTT) that prevents catastrophic forgetting at inference time and overfitting during pretraining LaCT.

In my new blog, β€œTest-Time Training Done Better: From Plastic Adaptation to Elastic Memory Consolidation,” I introduce a long-context modeling architecture that learns to adapt and memorize at test time by updating a subset of the model’s weights during inference.

mars-tin.github.io/blogs/posts/...

11.12.2025 16:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Gosh, I’m getting way too emotional writing my thesis acknowledgements...

24.11.2025 06:59 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Will be at #NeurIPS2025 (San Diego) Dec 1-9, then in the Bay Area until the 14th. Hmu if you wanna grab coffee and talk about totally random stuff.

Thread with a few things I’m excited about.
P.S. 4 NeurIPS papers all started pre-May 2024 and took ~1 year of polishing...so proud of the team!

23.11.2025 18:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Trying to decide what to do on the first day of #NeurIPS2025?

Check out my, @marstin.bsky.social and @xiangyue96.bsky.social's tutorial, "The Science of Benchmarking: What's Measured, What's Missing, What's Next" on December 2 from 1:30 to 4:00pm.

benchmarking.science

What will we cover?

1/3

18.11.2025 03:48 β€” πŸ‘ 21    πŸ” 4    πŸ’¬ 2    πŸ“Œ 1
Preview
An Open-Notebook Exploration of Emergent Grounding in LMs How We Did This Curiosity-Driven Research? An Open-Notebook Exploration of Emergent Grounding in LMs

@fredashi.bsky.social and I wrote a blog for our new mechinterp paper (arxiv.org/abs/2510.13796), including many unpublished and even negative results that we found meaningful to share.

An Open-Notebook Exploration of Emergent Grounding in LMs mars-tin.github.io/blogs/posts/...

21.10.2025 01:40 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1

Regrettably can’t attend #COLM2025 due to deadlines, but
Jane and Joyce will be presenting our work. :)

Jane is an exceptional undergraduate researcher and a great collaborator! Go meet her at COLM if you’re curious about her work on mechanistic interpretability, multimodality, & pragmatics!

05.10.2025 22:43 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

πŸš€ ACL ARR is looking for a Co-CTO to join me lead our amazing tech team and drive the future of our workflow. If you’re interested or know someone who might be, let’s connect!

RTs & recommendations appreciated.

29.09.2025 21:30 β€” πŸ‘ 4    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Post image

Unfortunately, I’ll be missing #ACL2025NLP this year β€” but here are a few things I’m excited about! πŸ‘‡

28.07.2025 06:19 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Congratulations!!

22.07.2025 05:35 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

with @fredashi.bsky.social / Jiayuan Mao / @djiafei.bsky.social / @manlingli.bsky.social / David Hsu / Parisa Kordjamshidi

14.07.2025 20:16 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸ“£ Excited to announce SpaVLE: #NeurIPS2025 Workshop on Space in Vision, Language, and Embodied AI!

Join us in San Diego to push the frontiers of spatial understanding and reasoning across CV, NLP, and robotics!

πŸ‘‰ space-in-vision-language-embodied-ai.github.io

14.07.2025 20:13 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

#CoreCognition #LLM #multimodal #GrowAI We spent 3 years to curate 1503 classic experiments spanning 12 core concepts in human cognitive development and evaluated on 230 MLLMs with 11 different prompts for 5 times to get over 3.8 millions inference data points.

A thread (1/n) - #ICML2025 βœ…

30.06.2025 06:07 β€” πŸ‘ 13    πŸ” 9    πŸ’¬ 1    πŸ“Œ 0
Post image

New Paper Alert ‼️ Current VLMs completely fail human gaze understanding πŸ™€ and scaling does NO help ‼️

However, humans, since an extremely age πŸ§’, are extremely sensitive to other people's gaze πŸ™„ πŸ‘€

No mentors, no labs, only pre-doc students, 111 VLMs, and we did it 😎

11.06.2025 23:21 β€” πŸ‘ 6    πŸ” 5    πŸ’¬ 1    πŸ“Œ 1
SimWorld SimWorld: A World Simulator for Scaling Photorealistic Multi-Agent Interactions

& @tianminshu.bsky.social (+ @marstin.bsky.social, @zhitinghu.bsky.social, β€ͺ@lianhui.bsky.social & more) will present β€œSimWorld: A World Simulator for Scaling Photorealistic Multi-Agent Interactions,” an @unrealengine.bsky.social-based sim that generates unlimited/diverse urban environments: (13/14)

10.06.2025 19:45 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

At Albuquerque Now :)

30.04.2025 05:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

See you at #NAACL2025! I will talk about grounded lexicon acquisition and scaling mechanistically grounded vision language models. Happy to chat if you are around :)

29.04.2025 23:28 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
VLMs Are Not Pragmatically Competent in Referring Expression Generation VLMs fail to refer like humans. Our study reveals widespread pragmatic issues in GPT-4o, LLaVA, and others, showing how their expressions often violate Gricean maxims.

We introduce RefOI, a new dataset of 1.5k objects, each with 3 written and 2 spoken human-produced referring expressions. We also release RefOI-TLHF, a large dataset of token-level human feedback for 10.6k referring expressions.

πŸ‘€https://vlm-reg.github.io/
πŸ“„https://arxiv.org/abs/2504.16060

23.04.2025 17:55 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Vision-Language Models are not yet pragmatically optimal.

We identify 3 key failures of pragmatic competence in referring expression generation with VLMs: (1) cannot uniquely refer to the referent, (2) include excessive or irrelevant information, and (3) misalign with human pragmatic preferences.

23.04.2025 17:55 β€” πŸ‘ 4    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1
Preview
Do Vision-Language Models Have Internal World Models? Towards an... Internal world models (WMs) enable agents to understand the world's state and predict transitions, serving as the basis for advanced deliberative reasoning. Recent large Vision-Language Models...

πŸ”Ή Workshop Paper at World Models:
Do Vision-Language Models Have Internal World Models?
πŸ—“ Apr 27, 9 p.m. (Peridot 201&206)

Paper: openreview.net/forum?id=tpP...

Excited for this collaboration with MaitrixOrg, details coming soon :)

19.04.2025 01:53 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

πŸ”Ή ICLR BiAlign Workshop:
We’re hosting the Bidirectional Human-AI Alignment Workshop (BiAlign).
πŸ—“ Apr 28, (Garnet 216–214)

Website: bialign-workshop.github.io

I’ll join remotely β€” huge thanks to @huashen.bsky.social for leading this!

19.04.2025 01:53 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ”Ή ICLR Oral Paper:
Do Vision-Language Models Represent Space and How?

πŸ—“ Oral: Apr 25, 3:42–3:54 a.m. (Session 4C)
πŸ—“ Poster: Thu, Apr 24, 10 p.m.–12:30 a.m. (Hall 3 + 2B, #212)

Website: spatial-comfort.github.io

Big thanks to @fredashi.bsky.social for presenting on site!

19.04.2025 01:53 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

I won’t be attending #ICLR2025 in person since #NAACL2025 follows right after, but here are a few things I’m excited about (all time in EDT) ⬇️

19.04.2025 01:53 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
BiAlign: ICLR'25 Workshop on Bidirectional Human-AI Alignment The official website for the ICLR BiAlign: Workshop on Bidirectional Human-AI Alignment

πŸ“„ View the full list of accepted papers: bialign-workshop.github.io#/papers

We look forward to seeing you there!

15.04.2025 20:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

πŸŽ‰ Out of these, 72 papers were accepted, including 5 tiny papers. 10 papers were selected for oral presentations: 2 at CHI and 8 at ICLR. Award winners will be announced during the workshop!

15.04.2025 20:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ“¬ We received over 100 submissions, each reviewed by 2–4 expert reviewers, with ethical assessments included when appropriate. Our program committee features leading researchers in NLP, RL, HCI, ML, and AI/ML Ethics, carefully selected based on scholarly merit and expertise.

15.04.2025 20:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ™ Special thanks to Tammy Masterson, Technical Partnerships Lead at the AI Security Institute, who will be joining us as a panelist.

15.04.2025 20:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ™ We are grateful to our gold sponsors, Prolific and Layer 6 AI of TD Bank Group, for their generous support in funding paper awards and travel grants.

15.04.2025 20:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@marstin is following 20 prominent accounts