's Avatar

@liang-weixin.bsky.social

12 Followers  |  9 Following  |  22 Posts  |  Joined: 21.11.2024  |  1.8603

Latest posts by liang-weixin.bsky.social on Bluesky

๐ŸŽ‰ Excited to share: "๐Œ๐ข๐ฑ๐ญ๐ฎ๐ซ๐ž-๐จ๐Ÿ-๐“๐ซ๐š๐ง๐ฌ๐Ÿ๐จ๐ซ๐ฆ๐ž๐ซ๐ฌ (๐Œ๐จ๐“)" has been officially accepted to TMLR (March 2025) and the code is now open-sourced!

๐Ÿ“Œ GitHub repo: github.com/facebookrese...
๐Ÿ“„ Paper: arxiv.org/abs/2411.04996

09.05.2025 05:35 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

๐Ÿšจ Our new study reveals widespread LLM adoption across society:

๐Ÿ“ŠBy late 2024, LLMs assist in writing:
- 18% of financial consumer complaints
- 24% of corporate press releases
- Up to 15% of job postings (esp. in small/young firms)
- 14% of UN press releases

arxiv.org/abs/2502.09747

01.03.2025 22:28 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Joint work with Junhong Shen, Genghan Zhang @zhang677.bsky.social, Ning Dong, Luke Zettlemoyer, Lili Yu

#LLM #MultiModal #pretraining

06.02.2025 05:31 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

โœ… Mamba w/ Chameleon setting (discrete tokens): Dense-level image/text at 42.5% and 65.4% FLOPs

โœ… Mamba w/ Three-modality (image, text, speech): Dense-level speech at 24.8% FLOPs

Takeaway:

Modality-aware sparsity isnโ€™t just for Transformersโ€”it thrives in SSMs like Mamba too! ๐Ÿ

06.02.2025 05:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image Post image Post image

๐Ÿš€ Want 2x faster pretraining for your multi-modal LLM?

๐Ÿงต Following up on Mixture-of-Transformers (MoT), we're excited to share Mixture-of-Mamba (MoM)!

โœ… Mamba w/ Transfusion setting (image + text): Dense-level performance with just 34.76% of the FLOPs

Full paper: ๐Ÿ“š arxiv.org/abs/2501.16295

06.02.2025 05:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 1    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

w/ Genghan Zhang(zhang677.github.io), Olivia Hsu (weiya711.github.io), and Prof. Kunle Olukotun (arsenalfc.stanford.edu/kunle/)

06.02.2025 04:42 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Genghan Zhang A simple, whitespace theme for academics. Based on [*folio](https://github.com/bogoli/-folio) design.

2/ Our system learns through curriculum-based experience, programming complex ML operators like attention and MoE from scratch.

3/ Bonus: We're releasing a clean benchmark with STeP - a new programming language never seen in training data. A true test of reasoning ability!๐ŸŽฏ

06.02.2025 04:42 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image Post image Post image

๐Ÿ“ข Can LLMs program themselves to run faster? ๐Ÿƒโฑ๏ธ

LLM self-taught to code for next-gen AI hardware!
arxiv.org/abs/2502.02534

1/ Programming AI accelerators is a major bottleneck in ML. Our self-improving LLM agent learns to write optimized code for new hardware, achieving 3.9x better results.

06.02.2025 04:41 โ€” ๐Ÿ‘ 0    ๐Ÿ” 1    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image

Honored that #Nature has highlighted our work again in their latest piece examining #ChatGPT's transformative impact on scientific research and academia over the past two years. h/t @natureportfolio.bsky.social

www.nature.com/articles/d41...

04.12.2024 17:15 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

*Our survey was opt-in and could have selection bias.

**Reviewers should still engage w/ papers independently w/o relying on LLM.

(4/n)

22.11.2024 20:56 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Findings (cont.):

๐Ÿ‘Ž GPT4 can struggle with in-depth critique of study methods; sometimes more generic.

Takeaway: high-quality human feedback still necessary; #LLM could help authors improve early drafts before official peer review**.

Collaborators: Yuhui, Hancheng, and many others ๐Ÿ‘

(3/n)

22.11.2024 20:55 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Findings:
๐Ÿ‘ Most authors found GPT4 generated feedback helpful*
๐Ÿ‘ >50% of points raised by GPT4 also raised by >=1 human reviewer.
๐Ÿ‘ Overlap between GPT4 and human feedback similar to overlap between 2 human reviewers.

(2/n)

22.11.2024 20:52 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Can large language models provide useful feedback on research papers? A large-scale empirical analysis Expert feedback lays the foundation of rigorous research. However, the rapid growth of scholarly production and intricate knowledge specialization challenge the conventional scientific feedback mechan...

How well can #GPT4 provide scientific feedback on research projects? We study this in: arxiv.org/abs/2310.01783

We created a pipeline using GPT4 to read 1000s papers (from #Nature, #ICLR, etc.) and generate feedback (eg suggestions for improvement). Then we compare with human expert reviews.

(1/n)

22.11.2024 20:51 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0
Post image

Our new study in #COLM2024 estimates that ~17% of recent CS arXiv papers used #LLMs substantially in its writing.

Around 8% for bioRxiv papers.

Paper: arxiv.org/abs/2404.01268 ๐Ÿงต

22.11.2024 20:49 โ€” ๐Ÿ‘ 0    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Excited that our paper quantifying #LLMs usage in paper reviews is selected as an #ICML2024 oral (top 1.5% of submissions)! ๐Ÿš€

Main results๐Ÿ‘‡
proceedings.mlr.press/v235/liang24...

Media Coverage: The New York Times
nyti.ms/3vwQhdi

22.11.2024 20:48 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
ChatGPT is transforming peer review โ€” how can we use it responsibly? At major computer-science publication venues, up to 17% of the peer reviews are now written by artificial intelligence. We need guidelines before things get out of hand.

Excited to share that our recent work on LLM in peer review and responsible LLM use is featured in #Nature!

Many thanks to my collaborators for their insights and dedication to advancing fair and ethical AI practices in scientific publishing. #AI #PeerReview

www.nature.com/articles/d41...

22.11.2024 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Takeaway:
Modality-aware sparsity in MoT offers a scalable path to efficient, multi-modal AI with reduced pretraining costs.

Work of a great team with Lili Yu, Liang Luo, Srini Iyer, Ning Dong, Chunting Zhou, Gargi Ghosh, Mike Lewis, Scott Wen-tau Yih, Luke Zettlemoyer, Victoria Lin.

(6/n)

22.11.2024 01:40 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

โœ… System profiling shows MoT achieves dense-level image quality in 47% and text quality in 75.6% of the wall-clock time**

**Measured on AWS p4de.24xlarge instances with NVIDIA A100 GPUs.

(5/n)

22.11.2024 01:39 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

โœ… Transfusion setting (text autoregressive + image diffusion): MoT matches dense model quality using one-third of the FLOPs.

(4/n)

22.11.2024 01:38 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

โœ… Chameleon setting (text + image generation): Our 7B MoT matches dense baseline quality using just 55.8% of the FLOPs.
Extended to speech as a third modality, MoT achieves dense-level speech quality with only 37.2% of the FLOPs.

(3/n)

22.11.2024 01:38 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

At Meta, we introduce Mixture-of-Transformers (MoT), a sparse architecture with modality-aware sparsity for every non-embedding transformer parameter (e.g., feed-forward networks, attention matrices, and layer normalization).

MoT achieves dense-level performance with up to 66% fewer FLOPs!

(2/n)

22.11.2024 01:37 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image Post image

How can we reduce pretraining costs for multi-modal models without sacrificing quality? We study this Q in our new work: arxiv.org/abs/2411.04996

โœ… MoT achieves dense-level 7B performance with up to 66% fewer FLOPs!

22.11.2024 01:36 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 5    ๐Ÿ“Œ 1

@liang-weixin is following 8 prominent accounts