Deqing Fu's Avatar

Deqing Fu

@deqing.bsky.social

CS PhD Student @USC. deqingfu.github.io

176 Followers  |  469 Following  |  14 Posts  |  Joined: 18.11.2024  |  1.4496

Latest posts by deqing.bsky.social on Bluesky


I gave a talk earlier today as Stanford NLP seminar. Here are the slides if you are interested: deqingfu.github.io/_docs/202505...

22.05.2025 22:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

At @naaclmeeting.bsky.social this week! I’ll be presenting our work on LLM domain induction with @thomason.bsky.social on Thu (5/1) at 4pm in Hall 3, Section I.

Would love to connect and chat about LLM planning, reasoning, AI4Science, multimodal stuff, or anything else. Feel free to DM!

30.04.2025 18:38 β€” πŸ‘ 4    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1

It seems I haven't posted any research related posts on this platform. Starting to do it now.
bsky.app/profile/deqi...

08.02.2025 05:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I would like to thank my intern mentor Lawrence Chen from Meta, and all other peers Tong Xiao, Rui Wang, Guan Pang, and Pengchuan Zhang. Big thanks to my lab mate @billzhu.bsky.social for valuable discussions and my advisor @robinjia.bsky.social for thoughtful inputs.

08.02.2025 05:29 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Finally, token-level annotations given by TLDR model could speedup human annotators to fix image captions that are slightly off. In fact, it can speed up human annotation by 3 times!

08.02.2025 05:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Next, there is something interesting. After finishing training the TLDR model, one can simply remove the reward model head and re-attach the original language model head, to, obviously, become a new vision-language model. It's shown that these new models become better.

08.02.2025 05:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

TLDR has rich usefulness. First, it can serve as a hallucination rate evaluation metric. As shown in the table, GPT-4o is still the best vision language model in the token level while open-weight models such as Llama-3.2-90B is catching up in the sentence and response level.

08.02.2025 05:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

TLDR is trained on synthetic hard negatives generated via a perturbation-based method. The architecture is very simple. Instead of applying the reward model head to the last token, as many RMs are doing, TLDR applies the reward model head to every token.

08.02.2025 05:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Excited to share that my intern work at Meta GenAI is accepted to @iclr-conf.bsky.social #ICLR2025

Introducing TLDR: Token-Level Detective Reward Model For Large Vision Language Models.

TLDR provides fine-grained annotations to
each text token.

πŸ”—arXiv: arxiv.org/abs/2410.04734

08.02.2025 05:29 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 1    πŸ“Œ 1

I think it may come from pretraining data and how numbers are presented by humans. We are still investigating how/why these features emerge from LLMs and will keep you updated with any new findings!

06.02.2025 18:22 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Pre-trained Large Language Models Use Fourier Features to Compute Addition Pre-trained large language models (LLMs) exhibit impressive mathematical reasoning capabilities, yet how they compute basic arithmetic, such as addition, remains unclear. This paper shows that pre-tra...

we have a very much similar results in NeurIPS 2024: arxiv.org/abs/2406.03445

06.02.2025 05:32 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I'll be at #NeurIPS2024! My group has papers analyzing how LLMs use Fourier Features for arithmetic and how TFs learn higher-order optimization for ICL (led by @deqing.bsky.social), plus workshop papers on backdoor detection and LLMs + PDDL (led by @billzhu.bsky.social)

09.12.2024 22:21 β€” πŸ‘ 23    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1

Can add add me please? Thanks!

23.11.2024 23:48 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Thanks for making this pack. Can you add me please? Thank you!

23.11.2024 23:48 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

πŸ™Œ

19.11.2024 23:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

USC NLP folks are on Bluesky!
Follow my amazing colleagues here

go.bsky.app/KUwSZ6W

12.11.2024 17:44 β€” πŸ‘ 17    πŸ” 5    πŸ’¬ 3    πŸ“Œ 2

Happy to join a new social media platform. I work on theory/science behind modern LLMs, and how to make them more robust and explainable.

19.11.2024 08:23 β€” πŸ‘ 12    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@deqing is following 19 prominent accounts