Surprising new results from Owain Evans and Anthropic: Training on the outputs of a model can change the model's behavior, even when those outputs seem unrelated. Training only on completions of 3-digit numbers was able to transmit a love of owls. alignment.anthropic.com/2025/sublimi...
22.07.2025 17:14 β π 31 π 5 π¬ 5 π 2
MorphScore got an update! MorphScore now covers 70 languages πππ We have a new-preprint out and we will be presenting our paper at the Tokenization Workshop @tokshop.bsky.social at ICML next week! @marisahudspeth.bsky.social @brenocon.bsky.social
10.07.2025 16:09 β π 11 π 4 π¬ 1 π 1
Our work finding universal concepts in vision models is accepted at #ICML2025!!!
My first major conference paper with my wonderful collaborators and friends @matthewkowal.bsky.social @thomasfel.bsky.social
@Julian_Forsyth
@csprofkgd.bsky.social
Working with y'all is the best π₯Ή
Preprint β¬οΈ!!
01.05.2025 22:57 β π 15 π 4 π¬ 0 π 1
Contribute to the first global archive of soniferous freshwater life, The Freshwater Sounds Archive, and receive recognition as a co-author in a resulting data paper!
Pre-print now available. New deadline: 31st Dec, 2025.
See link π4 more fishsounds.net/freshwater.js
09.06.2025 18:21 β π 39 π 14 π¬ 4 π 2
π Interested in Neuro-Symbolic Learning and attending #ICRA2025? π§ π€
Do not miss Leon Keller presenting βNeuro-Symbolic Imitation Learning: Discovering Symbolic Abstractions for Skill Learningβ.
Joint work of Honda Research Institute EU and @jan-peters.bsky.social (@ias-tudarmstadt.bsky.social).
19.05.2025 08:51 β π 11 π 2 π¬ 1 π 0
Prasoon Bajpai, Tanmoy Chakraborty
Multilingual Test-Time Scaling via Initial Thought Transfer
https://arxiv.org/abs/2505.15508
23.05.2025 02:49 β π 2 π 1 π¬ 0 π 0
In-Context Learning Boosts Speech Recognition via Human-like Adaptation to Speakers and Language Varieties
ArXiv link for In-Context Learning Boosts Speech Recognition via Human-like Adaptation to Speakers and Language Varieties
A study shows in-context learning in spoken language models can mimic human adaptability, reducing word error rates by nearly 20% with just a few utterances, especially aiding low-resource language varieties and enhancing recognition across diverse speakers. https://arxiv.org/abs/2505.14887
23.05.2025 03:10 β π 1 π 1 π¬ 0 π 0
"Interdimensional Cable", shorts made with Veo 3 ai. By CodeSamurai on Reddit
22.05.2025 02:51 β π 152 π 26 π¬ 11 π 30
Bingda Tang, Boyang Zheng, Xichen Pan, Sayak Paul, Saining Xie
Exploring the Deep Fusion of Large Language Models and Diffusion Transformers for Text-to-Image Synthesis
https://arxiv.org/abs/2505.10046
16.05.2025 07:35 β π 1 π 1 π¬ 0 π 0
Learning Nonlinear Dynamics in Physical Modelling Synthesis using Neural Ordinary Differential Equations
Victor Zheleznov, Stefan Bilbao, Alec Wright, Simon King
A neural ODE model combined modal decomposition with a neural network to model nonlinear string vibrations, generating synthetic data and sound examples.
16.05.2025 11:05 β π 2 π 1 π¬ 0 π 0
Omni-R1: Do You Really Need Audio to Fine-Tune Your Audio LLM?
ArXiv link for Omni-R1: Do You Really Need Audio to Fine-Tune Your Audio LLM?
Research unveils Omni-R1, a fine-tuning method for audio LLMs that boosts audio performance via text training, achieving MMAU results. Findings reveal how enhanced text reasoning affects audio capacities, suggesting new model optimization directions. https://arxiv.org/abs/2505.09439
15.05.2025 11:10 β π 1 π 1 π¬ 0 π 0
Yeah we finally have a model report with an actual data section. Thanks Qwen 3! github.com/QwenLM/Qwen3...
13.05.2025 18:51 β π 53 π 11 π¬ 1 π 0
FLAM: Frame-Wise Language-Audio Modeling
ArXiv link for FLAM: Frame-Wise Language-Audio Modeling
FLAM, a novel audio-language model, enables frame-wise localization of sound events in an open-vocabulary format. With large-scale synthetic data and advanced training methods, FLAM enhances audio understanding and retrieval, aiding multimedia indexing and access. https://arxiv.org/abs/2505.05335
10.05.2025 01:40 β π 2 π 1 π¬ 0 π 0
#ICML2025
Is standard RLHF optimal in view of test-time scaling? Unsurprisingly no.
We show a simple change to standard RLHF framework that involves π«ππ°ππ«π πππ₯π’ππ«πππ’π¨π§ and π«ππ°ππ«π ππ«ππ§π¬ππ¨π«π¦πππ’π¨π§ (suited to test-time procedure) is optimal!
09.05.2025 00:20 β π 17 π 6 π¬ 1 π 0
Is Best-of-N really the best we can do for language model inference?
New paper (appearing at ICML) led by the amazing Audrey Huang (ahahaudrey.bsky.social) with Adam Block, Qinghua Liu, Nan Jiang, and Akshay Krishnamurthy (akshaykr.bsky.social).
1/11
03.05.2025 17:40 β π 22 π 5 π¬ 1 π 1
Congratulations to the #AABI2025 Workshop Track Outstanding Paper Award recipients!
29.04.2025 20:54 β π 21 π 8 π¬ 0 π 1
Why not?
Reinforcement Learning for Reasoning in Large Language Models with One Training Example
Applying RLVR to the base model Qwen2.5-Math-1.5B, they identify a single example that elevates model performance on MATH500 from 36.0% to 73.6%,
30.04.2025 02:55 β π 20 π 2 π¬ 2 π 2
An incomplete list of Chinese AI:
- DeepSeek: www.deepseek.com. You can also access AI models via API.
- Moonshot AI's Kimi: www.kimi.ai
- Alibaba's Qwen: chat.qwen.ai. You can also access AI models via API.
- ByteDance's Doubaob (only in Chinese): www.doubao.com/chat/
27.04.2025 17:30 β π 22 π 7 π¬ 1 π 0
I really liked this approach by @matthieuterris.bsky.social et al.They propose learning a unique lightweight model for multiple inverse problems by conditioning it with the forward operator A. Thanks to self-supervised fine-tuning, it can tackle unseen inverse pb.
π° https://arxiv.org/abs/2503.08915
26.04.2025 16:02 β π 7 π 1 π¬ 0 π 0
Excited to be presenting our spotlight ICLR paper Simplifying Deep Temporal Difference Learning today! Join us in Hall 3 + Hall 2B Poster #123 from 3pm :)
25.04.2025 22:56 β π 7 π 1 π¬ 0 π 0
Balinese text-to-speech dataset as digital cultural heritage https://pubmed.ncbi.nlm.nih.gov/40275973/
26.04.2025 03:04 β π 1 π 1 π¬ 0 π 0
Kimi.ai releases Kimi-Audio! Our new open-source audio foundation model advances capabilities in audio understanding, generation, and conversation.
Paper: github.com/MoonshotAI/K...
Repo: github.com/MoonshotAI/K...
Model: huggingface.co/moonshotai/K...
25.04.2025 16:54 β π 13 π 2 π¬ 1 π 0
Our #ICLR2025 poster "Discrete Codebook World Models for Continuous Control" (Aidan Scannell, Mohammadreza Nakhaeinezhadfard, Kalle KujanpÀÀ, Yi Zhao, Kevin Luck, Arno Solin, Joni Pajarinen)
ποΈ Hall 3 + Hall 2B #415, Thu 24 Apr 10 a.m. +08 β 12:30 p.m. +08
π Preprint: arxiv.org/abs/2503.00653
21.04.2025 15:38 β π 10 π 3 π¬ 2 π 0
Andrew Kiruluta
Wavelet-based Variational Autoencoders for High-Resolution Image Generation
https://arxiv.org/abs/2504.13214
21.04.2025 07:14 β π 1 π 1 π¬ 0 π 0
7/ Large Language Models to Diffusion Finetuning
Paper: openreview.net/forum?id=Wu5...
Workshop: workshop-llm-reasoning-planning.github.io
New finetuning method empowering pre-trained LLMs with some of the key properties of diffusion models and the ability to scale test-time compute.
21.04.2025 09:56 β π 4 π 3 π¬ 1 π 0
10/ Sakana AI Co-Founder and CEO, David Ha, will be giving a talk at the #ICLR2025 World Models Workshop, at a panel to discuss the Current Development and Future Challenges of World Models.
Workshop Website: sites.google.com/view/worldmo...
21.04.2025 09:59 β π 12 π 3 π¬ 0 π 1
Duy A. Nguyen, Quan Huu Do, Khoa D. Doan, Minh N. Do: Are you SURE? Enhancing Multimodal Pretraining with Missing Modalities through Uncertainty Estimation https://arxiv.org/abs/2504.13465 https://arxiv.org/pdf/2504.13465 https://arxiv.org/html/2504.13465
21.04.2025 06:02 β π 1 π 1 π¬ 1 π 0
Yixuan Even Xu, Yash Savani, Fei Fang, Zico Kolter: Not All Rollouts are Useful: Down-Sampling Rollouts in LLM Reinforcement Learning https://arxiv.org/abs/2504.13818 https://arxiv.org/pdf/2504.13818 https://arxiv.org/html/2504.13818
21.04.2025 06:07 β π 2 π 3 π¬ 1 π 1
Deep learning, computational chemistry, generative modeling, AI for Science. Principal Research Manager at Microsoft Research AI for Science.
Autonomous agents | Robotics | LLMs
Chief Models Officer @ Stealth Startup; Inria & MVA - Ex: Llama @AIatMeta & Gemini and BYOL @GoogleDeepMind
Principal Researcher in AI/ML/RL Theory @ Microsoft Research NE/NYC. Previously @ MIT, Cornell. http://dylanfoster.net
RL Theory Lecture Notes: https://arxiv.org/abs/2312.16730
Assistant Professor & Faculty Fellow, NYU.
AI Fellow, Georgetown University.
Probabilistic methods for robust and transparent ML & AI Governance.
Prev: Oxford, Yale, UC Berkeley.
https://timrudner.com
Unofficial bot by @vele.bsky.social w/ http://github.com/so-okada/bXiv https://arxiv.org/list/cs.LG/new
List https://bsky.app/profile/vele.bsky.social/lists/3lim7ccweqo2j
ModList https://bsky.app/profile/vele.bsky.social/lists/3lim3qnexsw2g
language is irreducibly contextual and multimodal.
bizarre hybrid AI researcher / fullstack dev. currently working on https://talktomehuman.com/ & consulting (uname = domain)
previously:
- buncha travel
- phd @ uw (nlp)
- eng @ google (kubernetes)
NLP + applied math @BYU
https://vin.how
Postdoc with @summerfieldlab.bsky.social at University of Oxford studying learning in humans and machines
Interested in collective intelligence, metascience, philosophy of science
You can find me on Mastodon at https://neuromatch.social/@jess
Retired software engineer. AI enthusiast. Deadhead. I implemented the regex operator (=~) in Bash.
Cat parent, bad guitar player, AI researcher
Opinions are mine (and sometimes my cat's)
Audio Tech | XR | Engineer @ Dolby
Co-founder and CEO, Mistral AI
Professor of Marketing at NYU Stern School of Business, serial entrepreneur, and host of the Prof G and Pivot Podcasts.
Doctoral Student at IIT Hyderabad, India
An eccentric dreamer in search of truth and happiness for all.
http://www.jlcstudios.com
Explainability of deep neural nets and causality https://tfjgeorge.github.io/
AI Research @Hugging Face π€
Contributing to the Chinese ML community.
PhD fellow in XAI, IR & NLP
βοΈ Mila - Quebec AI Institute | University of Copenhagen π°
#NLProc #ML #XAI
Recreational sufferer