Shadab Choudhury's Avatar

Shadab Choudhury

@namer.bsky.social

(He/Him). Previously at LARA Lab @UMBC, Mohsin Lab @BRACU | Accessibility, Explainability and Multimodal DL. My opinions are mine. I'm on the PhD application cycle for Fall '26! www.shadabchy.com

404 Followers  |  455 Following  |  290 Posts  |  Joined: 24.08.2023  |  1.8713

Latest posts by namer.bsky.social on Bluesky


That's not remotely the issue here though. LLMs can already generate stories/code/designs. World models aren't the same thing as persistent memory.

It can't generate *good* stories for reasons of verifiability an un-RL-ability, as said above, and world models won't change that at all.

10.02.2026 11:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

The neatest thing is just how much it looks like mold growing on the fruits. A nicely picked example.

06.02.2026 22:25 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
A twitter post:

Kyunghyun Cho
@kchonyc
i was made aware of miscitations thanks to the GPTZero team (cc 
@alexcdot
). ji won and i quickly checked them ourselves and have posted what happened on openreview: https://openreview.net/forum?id=IiEtQPGVyV&noteId=W66rrM5XPk. we have already notified NeurIPS'25 PC's about this issue. 

i truly thank the GPTZero team for bringing this to our attention as well as raising the awareness of this serious issue (https://gptzero.me/news/neurips/), and at the same time i sincerely apologize to all for our error.

There is an image inside the post:

We identify the causes of miscitations in this work and provide fixes for them. These miscitations were identified and reported by the GPTZero team. The full report can be found at

Nazar Shmatko, Alex Adam and Paul Esau. GPTZero finds 100 new hallucinations in NeurIPS 2025 accepted papers. January 2026. https://gptzero.me/news/neurips/

We (Park & Cho) used a large-scale language model (LLM; specifically ChatGPT) to generate the citations after giving it author-year in-text citations, titles, or their paraphrases. Most, if not all, of these miscitations, either hallucinated, typoed or misattributed, are incorrect or non-existent bibtex entries fetched by the LLM. We categorize and analyze these citations below.

We are submitting an updated version to arXiv (https://arxiv.org/abs/2510.21310) and have already uploaded the updated version at https://tinyurl.com/ymb99d7s. We will shortly reach out to the program chairs of NeurIPS'25 as well.

We sincerely apologize to the whole community, especially the authors affected, for this grave mistake and thank the GPTZero team for bringing this issue to our attention.

(1) Prior work directly relevant to the paper
Below are miscitations for papers. Methods in these papers were implemented, or considered for implementation, in our paper as baseline uncertainty quantification methods, baseline sampling methods, or as the base LMs.

A twitter post: Kyunghyun Cho @kchonyc i was made aware of miscitations thanks to the GPTZero team (cc @alexcdot ). ji won and i quickly checked them ourselves and have posted what happened on openreview: https://openreview.net/forum?id=IiEtQPGVyV&noteId=W66rrM5XPk. we have already notified NeurIPS'25 PC's about this issue. i truly thank the GPTZero team for bringing this to our attention as well as raising the awareness of this serious issue (https://gptzero.me/news/neurips/), and at the same time i sincerely apologize to all for our error. There is an image inside the post: We identify the causes of miscitations in this work and provide fixes for them. These miscitations were identified and reported by the GPTZero team. The full report can be found at Nazar Shmatko, Alex Adam and Paul Esau. GPTZero finds 100 new hallucinations in NeurIPS 2025 accepted papers. January 2026. https://gptzero.me/news/neurips/ We (Park & Cho) used a large-scale language model (LLM; specifically ChatGPT) to generate the citations after giving it author-year in-text citations, titles, or their paraphrases. Most, if not all, of these miscitations, either hallucinated, typoed or misattributed, are incorrect or non-existent bibtex entries fetched by the LLM. We categorize and analyze these citations below. We are submitting an updated version to arXiv (https://arxiv.org/abs/2510.21310) and have already uploaded the updated version at https://tinyurl.com/ymb99d7s. We will shortly reach out to the program chairs of NeurIPS'25 as well. We sincerely apologize to the whole community, especially the authors affected, for this grave mistake and thank the GPTZero team for bringing this issue to our attention. (1) Prior work directly relevant to the paper Below are miscitations for papers. Methods in these papers were implemented, or considered for implementation, in our paper as baseline uncertainty quantification methods, baseline sampling methods, or as the base LMs.

Like, if you don't know how to I'm happy to show you. It's not hard or something and it just speeds up your workflow.

Copying and pasting each of your citations into ChatGPT is just silly.

(img source: the other site)

22.01.2026 10:58 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Why on earth would you even do this in the first place?

Pasting the details into ChatGPT, then asking it to generate the citation is a way bigger hassle than simply having Scholar + Zotero (or another bib manager) extensions set up correctly to grab metadata and generate citations.

22.01.2026 10:53 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Lowkey, I think this is intentional, and it seems like the kind of thing I would do if I was making Claude more attractive to students or other people who don't want their code to be easily recognized as LLM-generated.

08.01.2026 21:26 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Poster titled "Launching HCCS: Inauguration and research dialogue with the HCI Pioneers of Bangladesh". On the bottom left, there are 3 portraits, named Dr Syed Ishtiaque Ahmed, Dr. Hasan Shahid Ferdous and Dr. Sharifa Sultana

Poster titled "Launching HCCS: Inauguration and research dialogue with the HCI Pioneers of Bangladesh". On the bottom left, there are 3 portraits, named Dr Syed Ishtiaque Ahmed, Dr. Hasan Shahid Ferdous and Dr. Sharifa Sultana

A stage with 4 people sitting on chairs in the middle. A poster in the background shows the previous image's poster, as well as a QR code to the panel discussion question submission.

A stage with 4 people sitting on chairs in the middle. A poster in the background shows the previous image's poster, as well as a QR code to the panel discussion question submission.

It was great listening to Dr's Ishtiaque, Ferdous, and Sultana talk about their experiences!

There's a *lot* of HCI work to be done in the scope of Bangladesh, and IMO not enough folks working on them, so HCCS should be a great addition. Looking forward to the work soon to come out of the lab.

04.01.2026 13:29 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

lowkey I appreciate folks aren't posting linkedinisms like "In 2025 I achieved X, Y, Z" this time around.

It's fine to celebrate your wins, but I imagine for most people, 2025 was...

A Year.

...and I think that's all that needs to be said.

02.01.2026 23:56 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Nahhh my high school friends would've also found that name funny as fuck 10 years ago

01.01.2026 09:21 โ€” ๐Ÿ‘ 7    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Seen on LinkedIn.

How does someone raise $5 million and then write job ads like this?

smh...

25.12.2025 14:41 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I don't think "money" is the simple answer, since every frontier lab is a black hole of money rn, and gene editing could've also been ludicrously profitable.

Was it the political climate? The everyday accessibility of GenAI? Lower levels of scruples in the community (not to accuse anyone directly)?

21.12.2025 14:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

In the mid 2010s biology folks figured out how to do human gene editing. The community took one look, realized the consequences would be so dire for humanity, and put a hard stop to it. People like Jiankui He were excoriated for illegally editing embryos.

Why wasn't this the case with GenAI?

21.12.2025 14:26 โ€” ๐Ÿ‘ 3    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

And it's not ML, it's "GenAI" that invokes certain concerns.

People don't mind when ML's used to detect cancer or study whale speech. Those models aren't trained on human inputs and used to take human jobs. GenAI specifically, however, is trained on human inputs and used to take human jobs.

17.12.2025 14:35 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Nah. That article's 8mo old and says they're "not there yet". Vincke's recent comment, however, explicitly mentions flesh out PowerPoint presentations, develop concept art" and these tasks are explicitly creative jobs lost.

17.12.2025 14:31 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

by perchance is it specifically like these 5 companies?

16.12.2025 22:45 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I hope the Peer Reviewer Recognition Policy actually puts the <$30~ per paper reviewed (based on 3 reviewers per $100 paper minus waived papers) to good use.

04.12.2025 19:25 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Primary Paper Initiative: IJCAI-ECAI 2026 is launching the Primary Paper Initiative in response to the international AI research communityโ€™s call to address challenges and to revitalize the peer review process, while strengthening the reviewers and authors in the process. Under the IJCAI-ECAI 2026 Primary Paper Initiative, every submission is subject to a fee of USD 100. That paper submission fee is waived for primary papers, i.e., papers for which none of the authors appear as an author on any other submission to IJCAI-ECAI 2026. The initiative applies to the main track, Survey Track, and all special tracks, excluding the Journal Track, the Sister Conferences Track, Early Career Highlights, Competitions, Demos, and the Doctoral Consortium. All proceeds generated from the Primary Paper Initiative will be exclusively directed toward the support of the reviewing community of IJCAI-ECAI 2026. To recognize the reviewersโ€™ contributions, the initiative introduces Peer Reviewer Recognition Policy with clearly defined standards (which will be published on the conference web site). The initiative aims to enhance review quality, strengthen accountability, and uphold the scientific excellence of the conference. Details and the FAQ will be published on the IJCAI-ECAI 2026 website.

Primary Paper Initiative: IJCAI-ECAI 2026 is launching the Primary Paper Initiative in response to the international AI research communityโ€™s call to address challenges and to revitalize the peer review process, while strengthening the reviewers and authors in the process. Under the IJCAI-ECAI 2026 Primary Paper Initiative, every submission is subject to a fee of USD 100. That paper submission fee is waived for primary papers, i.e., papers for which none of the authors appear as an author on any other submission to IJCAI-ECAI 2026. The initiative applies to the main track, Survey Track, and all special tracks, excluding the Journal Track, the Sister Conferences Track, Early Career Highlights, Competitions, Demos, and the Doctoral Consortium. All proceeds generated from the Primary Paper Initiative will be exclusively directed toward the support of the reviewing community of IJCAI-ECAI 2026. To recognize the reviewersโ€™ contributions, the initiative introduces Peer Reviewer Recognition Policy with clearly defined standards (which will be published on the conference web site). The initiative aims to enhance review quality, strengthen accountability, and uphold the scientific excellence of the conference. Details and the FAQ will be published on the IJCAI-ECAI 2026 website.

How was no one talking about this?* IJCAI-ECAI 2026 @ijcai.org levying a $100 fee per submission unless every author on the paper is only on that one submitted paper.

* rhetorical question. I assume the ICLR drama drowned it

04.12.2025 19:22 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Basically, if your Altmetric score is higher than your Accesses, you just got ratioed.

29.11.2025 13:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Case in point.

Good grief.

27.11.2025 22:37 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Took me about 5 minutes to dig out the identity of the 40 questions reviewer โ€” after someone posted it on the other site; I dunno how to search Xiaohongshu directly.

Honestly, I don't think western academics are going to feel a fraction of the shitstorm that the Chinese ML community's probably in.

27.11.2025 19:29 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1

This is the site I unfortunately have to be professional on ๐Ÿ˜ญ

20.11.2025 18:17 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

many such cases

Simpsons-tier prescience

20.11.2025 07:53 โ€” ๐Ÿ‘ 0    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I understand the sentiment behind this, but I'm just extremely bearish on rankings that do fancy mathematical tricks or use black-box algorithms because the more of that you do, the more you can bias it towards a specific outcome.

The closer the metric is to the raw data instead, the better.

16.11.2025 13:14 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Two issues: first, this is not transparent. There's NO way to tell *which* papers were counted, nor how 'most important papers to this paper' is computed. The papers contributing to the ranking should be listed.

Second, CSRankings is CC BY-NC-ND 4.0 so I'm pretty sure this is copyright infringement

16.11.2025 13:09 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

I can't believe I spent the evening skimming through every Visual Reasoning paper at ICLR instead of finishing my SoP.

15.11.2025 14:58 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

It may be in bad taste to call out a reviewer like this, but I don't believe anyone who gives weaknesses like this, as if it isn't empirical common sense for anyone working with MLLMs that larger models give better outcomes when inference speed isn't relevant, is acting in good faith.

15.11.2025 14:48 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

There's a reviewer at ICLR who apparently always writes *exactly* 40 weaknesses and comments no matter what paper he's reviewing.

Exhibit A: openreview.net/forum?id=8qk...
Exhibit B: openreview.net/forum?id=GlX...
Exhibit C: openreview.net/forum?id=kDh...

15.11.2025 14:42 โ€” ๐Ÿ‘ 9    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 2
Preview
Introduction to Latent Variable Energy-Based Models: A Path Towards Autonomous Machine Intelligence Current automated systems have crucial limitations that need to be addressed before artificial intelligence can reach human-like levels and bring new technological revolutions. Among others, our socie...

Tbh, it's better to go through Yann LeCun's lectures on JEPA first before trying to comprehend the mathematical foundations of LeJEPA, since this is a specific optimization on top of the original JEPA theory.

The lecture notes were easier to work out initially.

14.11.2025 13:59 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
SIMA 2: A Gemini-Powered AI Agent for 3D Virtual Worlds Introducing SIMA 2, the next milestone in our research creating general and helpful AI agents. By integrating the advanced capabilities of our Gemini models, SIMA is evolving from an instruction-follโ€ฆ

Google DeepMind also dropped something extremely similar (SIMA 2), trained on different games rather than Genshin, but they didn't release their technical report yet so I'm valuing this higher

13.11.2025 20:07 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Weitao Tan, @WeihaoTan64 on Twitter, posts: ๐Ÿš€Introducing Lumine, a generalist AI agent trained within Genshin Impact that can perceive, reason, and act in real time, completing hours-long missions and following diverse instructions within complex 3D open-world environments.๐ŸŽฎ

Website: http://lumine-ai.org

Weitao Tan, @WeihaoTan64 on Twitter, posts: ๐Ÿš€Introducing Lumine, a generalist AI agent trained within Genshin Impact that can perceive, reason, and act in real time, completing hours-long missions and following diverse instructions within complex 3D open-world environments.๐ŸŽฎ Website: http://lumine-ai.org

Saw an insanely cool paper (Weihao Tan @ NTU (sg) plus Bytedance and PekingU collabs) on the other site: a VLM agent pretrained on human gameplay videos + action primitives can fully play video games.

Very impressive work, esp. the inference process.

www.lumine-ai.org

arxiv.org/abs/2511.08892

13.11.2025 20:03 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 2

*Urgently* looking for emergency reviewers for the ARR October Interpretability track ๐Ÿ™๐Ÿ™

ReSkies much appreciated

11.11.2025 10:29 โ€” ๐Ÿ‘ 1    ๐Ÿ” 9    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@namer is following 20 prominent accounts