's Avatar

@mgaido91.bsky.social

29 Followers  |  74 Following  |  3 Posts  |  Joined: 07.01.2025  |  1.4479

Latest posts by mgaido91.bsky.social on Bluesky


Qualtrics Survey | Qualtrics Experience Management The most powerful, simple and trusted way to gather experience data. Start your journey to experience management and try a free account today.

๐Ÿ” Stiamo studiando come l'AI viene usata in Italia e per farlo abbiamo costruito un sondaggio!

๐Ÿ‘‰ bit.ly/sondaggio_ai...

(รจ anonimo, richiede ~10 minuti, e se partecipi o lo fai girare ci aiuti un sacco๐Ÿ™)

Ci interessa anche raggiungere persone che non si occupano e non sono esperte di AI!

03.06.2025 10:24 โ€” ๐Ÿ‘ 16    ๐Ÿ” 18    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
FAMA - a FBK-MT Collection The First Large-Scale Open-Science Speech Foundation Model for English and Italian

๐Ÿš€ New tech report out! Meet FAMA, our open-science speech foundation model family for both ASR and ST in ๐Ÿ‡ฌ๐Ÿ‡ง English and ๐Ÿ‡ฎ๐Ÿ‡น Italian.

The models are live and ready to try on @hf.co:
๐Ÿ”— huggingface.co/collections/...

๐Ÿ“„ Preprint: arxiv.org/abs/2505.22759

#ASR #ST #OpenScience #MultilingualAI

30.05.2025 15:35 โ€” ๐Ÿ‘ 7    ๐Ÿ” 3    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Reserved topic scholarships | Doctoral Program - Information Engineering and Computer Science

๐Ÿ“ข Come and join our group!
We offer a fully funded 3-year PhD position:

๐Ÿ“” Automatic translation with large multimodal models: iecs.unitn.it/education/ad...

๐Ÿ“Full details for application: iecs.unitn.it/education/ad...

๐Ÿ“… Deadline May 12, 2025

#NLProc #FBK

22.04.2025 10:14 โ€” ๐Ÿ‘ 9    ๐Ÿ” 9    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
https://arxiv.org/abs/2501.04561

Interesting to see multimodal LLM built by combining modality encoders and LLM with adapters, as in the SFM+LLM paradigm, independently for each modality. This modularity may ease the creation of more MLMs from collaborations of single-modality experts. arxiv.org/abs/2501.04561

16.04.2025 13:20 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1

๐Ÿ“ข The evaluation period of the Instruction Following task at
@iwslt.bsky.social just started!

๐Ÿ–ฅ๏ธ Consider submitting your speech-to-text system!

The outputs can be easily uploaded on the SPEECHM platform developed in the Meetween project (www.meetween.eu)!
โžก๏ธ iwslt2025.speechm.cloud.cyfronet.pl

01.04.2025 12:39 โ€” ๐Ÿ‘ 9    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

While we look forward to a sunny Geneva, why wait to join the conversation?

Weโ€™ve created a starter pack for our #GITT2025 friends!
๐Ÿ•ต๏ธ Follow researchers working on gender bias in MT
๐Ÿ’ฌ Stay up to date and dive into the discussion!

All info at sites.google.com/tilburgunive...

28.02.2025 09:22 โ€” ๐Ÿ‘ 21    ๐Ÿ” 16    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Preview
AlignFormer: Modality Matching Can Achieve Better Zero-shot Instruction-Following Speech-LLM Integrating speech into LLM (speech-LLM) has gaining increased attention recently. The mainstream solution is to connect a well-trained speech encoder and LLM with a neural adapter. However, the lengt...

very interesting to see more and more methods to close the length mismatch between speech and text sequences (aka length adapter -- see arxiv.org/abs/2402.12025) for SFM+LLM models! This one merging CTC and Q-former sounds very cool to me:
arxiv.org/abs/2412.01145

14.02.2025 10:34 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Simultaneous track Home of the IWSLT conference and SIGSLT.

Next up: simultaneous speech translation!

๐ŸŽฏ Goal: to explore ways to translate speech into another language like simultaneous interpreting.

๐Ÿ”— Link: iwslt.org/2025/simulta...

30.01.2025 19:31 โ€” ๐Ÿ‘ 6    ๐Ÿ” 4    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Instruction-following Speech Processing track Home of the IWSLT conference and SIGSLT.

First up, a new task for 2025:
*Instruction-following for speech processing!*

Explore instruction-following for speech โ‡จ
Integrate speech foundation models with LLMs across tasks such as speech translation, recognition, summarization, and QA.

๐Ÿ”—: iwslt.org/2025/instruc...

28.01.2025 18:13 โ€” ๐Ÿ‘ 8    ๐Ÿ” 6    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Today's task: model compression!!
๐Ÿ†• New at IWSLT! But no less exciting ๐Ÿ”ฅ

๐ŸŽฏ Goal: Compress a large, general-purpose multimodal model, making speech translation more efficient โšก๏ธ, deployable ๐Ÿ“ฒ, and sustainable โ™ป๏ธ, while preserving translation quality โญ๏ธ
#AI #SpeechTech #ModelCompression #LLMcompression

29.01.2025 16:48 โ€” ๐Ÿ‘ 8    ๐Ÿ” 5    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Prepending or Cross-Attention for Speech-to-Text? An Empirical Comparison Following the remarkable success of Large Language Models (LLMs) in NLP tasks, there is increasing interest in extending their capabilities to speech -- the most common form in communication. To integ...

I'm happy to share that our paper "Prepending or Cross-Attention for Speech-to-Text? An Empirical Comparison" has been accepted at @naaclmeeting.bsky.social 2025! #NAACL2025

@mgaido91.bsky.social ๐Ÿ‘

๐Ÿ“ƒ Preprint: arxiv.org/abs/2501.02370
โฐ Code will be released soon

#NLProc #Speech

23.01.2025 08:44 โ€” ๐Ÿ‘ 10    ๐Ÿ” 3    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
a polar bear cub is laying in a pile of branches . ALT: a polar bear cub is laying in a pile of branches .

Hello world! ๐Ÿ‘‹ We're coming out of hibernation to bring you this happy news:
1) We're organising the 3rd edition of GITT at #MTSummit! Working on #gender & #translation #technology? We'll see you there!
2) We're moving away from Twitter, so share the news and help us find old and new GITT friends!

22.01.2025 12:17 โ€” ๐Ÿ‘ 26    ๐Ÿ” 15    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Instruction-following Speech Processing track Home of the IWSLT conference and SIGSLT.

Our #iwslt 2025 task on instruction-following speech models is out! Submission by April 15th. Check it out at: iwslt.org/2025/instruc...

09.01.2025 09:43 โ€” ๐Ÿ‘ 4    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1

@mgaido91 is following 20 prominent accounts