6/6: We hope our work will inspire further research on the intrinsic difficulty of translating and generating different languages in the age of LLMs, particularly through experimentation with alternative decoding strategies.
For now, I'm looking forward to presenting our work in Rabat, Morocco π²π¦
08.02.2026 16:56 β π 1 π 0 π¬ 0 π 0
5/6: In the context of
searching for the modelβs highest-probability translation, we found that languages with more complex morphology and flexible word order benefit more from wider beam size.
In other words, the standard practice of left-to-right beam search may be suboptimal for these languages.
08.02.2026 16:56 β π 1 π 0 π¬ 1 π 0
Spearman correlations between continuous language properties and NLLB-200 chrF++ translation quality scores at beam size k = 5. Source language is English. Sample sizes (i.e. number of target languages) for each property are indicated next to their respective bars. Correlations significant at p < 0.05 are marked with *, at p < 0.01 with **, at p < 0.001 with ***.
4/6: Through correlation and regression experiments, we found that language properties like typological distance, type/token ratio, and head-finality drive translation quality of both NMT models, even after controlling for more trivial factors such as language resourcedness and script similarity.
08.02.2026 16:56 β π 1 π 0 π¬ 1 π 0
Tower+ 9B chrF++ scores vs. NLLB-200 3.3B chrF++ scores at beam size k = 7. Each point denotes a language pair and is colored by source language, while βΌ denotes target languages officially supported
by Tower+. The blue and orange shaded regions indicate language pairs for which either NLLB-200 or Tower+ scores are higher, respectively. Sample size is n = 7 Γ 52 = 364.
3/6: We analyze 2 NMT models, NLLB-200 and Tower+.
Although current SOTA has shifted to prompting decoder-only LLMs such as Tower+, we find that NLLB achieves higher chrF++ scores on all languages outside Tower's coverage, reaffirming the relevance of encoder-decoders for low-resourced languages.
08.02.2026 16:56 β π 1 π 0 π¬ 1 π 0
2/6: First, we compile a broad set of fine-grained typological and morphosyntactic features for 212 languages in the FLORES+ MT benchmark. We release this set publicly: github.com/v-hirak/expl...
08.02.2026 16:56 β π 1 π 0 π¬ 1 π 0
Henry Cavill is a creep though
17.06.2025 10:54 β π 0 π 0 π¬ 0 π 0
They aren't canonizing anything, this show is gonna be as canon as the millions of other people's playthroughs. It's just their take on the story
17.06.2025 10:53 β π 0 π 0 π¬ 1 π 0
Thank you from a Ukrainian, Kala, sincerely π I love your Mass Effect content
28.02.2025 22:34 β π 4 π 0 π¬ 0 π 0
wondering how humans and computers learn and use language πΆπ§ π£οΈπ₯οΈπ¬
the work is mysterious and important, see bbunzeck.github.io
phd at @clausebielefeld.bsky.social
NLP assistant prof at KU Leuven, PI @lagom-nlp.bsky.social. I like syntax more than most people. Also multilingual NLP, interpretability, mountains and beer. (She/her)
PhD student at the University of Cape Town, working on text generation for low-resource, morphologically complex languages.
https://francois-meyer.github.io/
Cape Town, South Africa
PhD student at Aalborg University, mostly working on NLP and linguistic typology
Associate Professor at GroNLP ( @gronlp.bsky.social⬠) #NLP | Multilingualism | Interpretability | Language Learning in Humans vs NeuralNets | Mum^2
Head of the InClow research group: https://inclow-lm.github.io/
Postdoc @rug.nl with Arianna Bisazza.
Interested in NLP, interpretability, syntax, language acquisition and typology.