Thrilled to be heading to Suzhou with a big team of GroNLP'ers ๐ฎ
Interested in Interpretable, Cognitively inspired, Low-resource LMs? Don't miss our posters & talks #EMNLP2025!
@frap98.bsky.social
2nd year PhD Student at @gronlp.bsky.social ๐ฎ - University of Groningen Language Acquisition - NLP
Thrilled to be heading to Suzhou with a big team of GroNLP'ers ๐ฎ
Interested in Interpretable, Cognitively inspired, Low-resource LMs? Don't miss our posters & talks #EMNLP2025!
...ii) a direct quality reward from a teacher model, and iii) a reward based on the log probabilities of a teacher model (and its dialogue continuations). While these rewards did not improve our models performance, two different DPO approaches did!
28.10.2025 12:55 โ ๐ 1 ๐ 1 ๐ฌ 1 ๐ 0Dialogue Is Not Enough to Make a Communicative BabyLM (But Neither Is Developmentally Inspired Reinforcement Learning) Francesca Padovani1โ Bastian Bunzeck2โ Manar Ali2 Omar Momen2 Arianna Bisazza1 Hendrik Buschmeier2 Sina Zarrieร2 1Center for Language and Cognition (CLCG), University of Groningen 2CRC 1646 โ Linguistic Creativity in Communication, Bielefeld University f.padovani@rug.nl bastian.bunzeck@uni-bielefeld.de
As part of this year's BabyLM challenge, we (researchers from @gronlp.bsky.social and @clausebielefeld.bsky.social diverged from established pretraining paradigm by training only on dialogue data from CHILDES.
28.10.2025 12:53 โ ๐ 16 ๐ 3 ๐ฌ 1 ๐ 0Donโt hesitate to reach out with any questions or doubts :)
Every contribution is more than welcome! ๐๐
These are living datasets that researchers around the world can enrich as new resources become available
On the website of the project babylm.github.io/babybabellm/
you can find information about the dataset creation pipeline, training, and evaluation.
This was made possible thanks to the dedication of each language-specific expert and the coordination of senior researchers, @jumelet.bsky.social and @lchoshen.bsky.social among others.
14.10.2025 17:01 โ ๐ 4 ๐ 0 ๐ฌ 1 ๐ 0๐๐จ ๐ฒ๐จ๐ฎ ๐ซ๐๐๐ฅ๐ฅ๐ฒ ๐ฐ๐๐ง๐ญ ๐ญ๐จ ๐ฌ๐๐ ๐ฐ๐ก๐๐ญ ๐ฆ๐ฎ๐ฅ๐ญ๐ข๐ฅ๐ข๐ง๐ ๐ฎ๐๐ฅ ๐๐๐๐จ๐ซ๐ญ ๐ฅ๐จ๐จ๐ค๐ฌ ๐ฅ๐ข๐ค๐? ๐จ๐ณ๐ฎ๐ฉ๐ธ๐ช
Hereโs the proof! ๐๐๐๐ฒ๐๐๐๐๐ฅ๐๐ is the first Multilingual Benchmark of Developmentally Plausible Training Data available for 45 languages to the NLP community ๐
arxiv.org/abs/2510.10159
This was made possible thanks to the dedication of each language-specific expert and the coordination of senior researchers, in particular @jumelet.bsky.social ! ๐คฉ
14.10.2025 16:53 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Computational Psycholinguistics Meeting 2025
cpl2025.sites.uu.nl
When: December 18โ19, 2025
Where: Utrecht, the Netherlands
Abstract submission deadline: June 15, 2025
Organizers: Jakub Dotlaฤil, Lena Jรคger, Bruno Nicenboim, Ece Takmaz
Jane Goodall (1934-2025) ๐ค
01.10.2025 20:02 โ ๐ 469 ๐ 96 ๐ฌ 3 ๐ 5My very first book review is out now ๐
Muchas gracias to @stefanhartmann.bsky.social for inviting me, looking forward to our next project(s) ๐
๐๐ก๐ ๐๐๐๐๐๐ ๐ฉ๐ซ๐จ๐ฃ๐๐๐ญ (๐๐๐ ๐๐จ๐, ๐๐ ๐๐ง๐ ๐๐ฎ๐๐ซ๐๐๐ซ๐จ๐) ๐ข๐ฌ ๐ฅ๐จ๐จ๐ค๐ข๐ง๐ ๐๐จ๐ซ ๐ ๐ง๐๐ฐ ๐ญ๐๐๐ฆ ๐ฆ๐๐ฆ๐๐๐ซ! โฃ๐
โฃ
We have an opening for a ๐๐จ๐ฌ๐ญ๐๐จ๐ ๐ฉ๐จ๐ฌ๐ข๐ญ๐ข๐จ๐ง in ๐๐ซ๐๐๐ญ๐ข๐ฏ๐ข๐ญ๐ฒ, ๐๐ซ๐๐ง๐ฌ๐ฅ๐๐ญ๐ข๐จ๐ง ๐๐ง๐ ๐๐๐๐ก๐ง๐จ๐ฅ๐จ๐ ๐ฒ for a starting duration of 12 months extendible to 30 months.โฃ
For further details -> www.rug.nl/about-ug/wor...
๐ง๐ต๐ฒ ๐๐๐ฟ๐ผ๐ฝ๐ฒ๐ฎ๐ป ๐ฅ๐ฒ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต๐ฒ๐ฟ๐โ ๐ก๐ถ๐ด๐ต๐ ๐๐ฎ๐ ๐ฎ๐ป๐๐ถ๐ฐ๐ถ๐ฝ๐ฎ๐๐ฒ๐ฑ ๐ฏ๐ ๐ฎ๐ป ๐ฒ๐พ๐๐ฎ๐น๐น๐ ๐ฒ๐
๐ฐ๐ถ๐๐ถ๐ป๐ด ๐ฎ๐ณ๐๐ฒ๐ฟ๐ป๐ผ๐ผ๐ป! ๐
Here are some photos of our GroNLP station at the Forum, where we welcomed high school classes and potential future researchers. ๐งโ๐ฌ๐ค
Quizzes, puzzles, demos, card games, and of course, lots of chocolate! ๐ซ
Thanks to the entire CIRM team and the @ilcb.bsky.social for organizing this truly special week !
12.09.2025 09:52 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0I can only ๐ต๐ถ๐ด๐ต๐น๐ ๐ฟ๐ฒ๐ฐ๐ผ๐บ๐บ๐ฒ๐ป๐ฑ ๐ฝ๐ฎ๐ฟ๐๐ถ๐ฐ๐ถ๐ฝ๐ฎ๐๐ถ๐ป๐ด in this training week in its future editions! ๐ณ
A special thanks to my amazing colleagues from Amsterdam, @mdhk.net and @cpouw.bsky.social , and to Felicia Stich from Gรถttingen :))
I had the chance to meet fellow researchers full of brilliant ideas and a genuine willingness to share, attend lectures, and listen to invited speakers who were not only knowledgeable but eager to engage in active intellectual exchange.
12.09.2025 09:52 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0Set in a breathtaking location amid the Mediterranean scrub of Parc des Calanques, the program was intense and incredibly enriching.
12.09.2025 09:52 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0The ๐๐๐๐ ๐ฆ๐๐บ๐บ๐ฒ๐ฟ ๐ฆ๐ฐ๐ต๐ผ๐ผ๐น in Marseille went beyond all my expectations! ๐ฏ
A week has already flown by since I had one of the most formative experiences of my PhD so far. ๐ฉโ๐จ
We recharged our batteries with a wonderful weekend together in the province of Drenthe! ๐ณ๐
As per tradition, we kick off the academic year with plenty of sharing, smiles, and sports ๐
We wish everyone a wonderful start of the academic year <3 may it be inspiring and exciting ๐
Rowena Garcia is presenting research at #ISB15 showing that #crosslinguistic #priming occurs across #typologically unrelated languages & without overlap of constituent order. The degree of overlap modulated the magnitude of the effect.
#Tagalog #Indonesian #English
@isbilingualism.bsky.social
#ACL2025 has just finished and it was great to meet with old friends and connect with new colleagus. In case you missed our papers/presentations, this was us ๐ฎ๐
@facultyofartsug.bsky.social
Fabulous blog about screentime from Gonzales, Golinkoff & Hirsh-Pasek here 1/ childandfamilyblog.com/children-and...
02.07.2025 14:58 โ ๐ 5 ๐ 2 ๐ฌ 1 ๐ 0Children are incredible language learning machines. But how do they do it? Our latest paper, just published in TICS, synthesizes decades of evidence to propose four components that must be built into any theory of how children learn language. 1/
www.cell.com/trends/cogni... @mpi-nl.bsky.social
Excited to welcome a new member to the family of language-specific BLiMPs ๐ this time putting Turkish in the spotlight!
Thanks to the meticulous, linguistically grounded work of the brilliant @ezgibasar.bsky.social. We hope will help push forward research on typologically diverse languages!
CDL may still hold unique value when used in models that better reflect the dynamics of real-world language learning ๐ค ๐คฒ
Let's keep investigating!
Trained on static text without interaction or feedback, they miss the communicative and social context central to how children learn. Rather than dismissing CDL, we argue for rethinking its role within more interactive and cognitively grounded learning frameworks ๐
30.05.2025 07:39 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0While our findings suggest that Child-Directed Language (CDL) does not consistently enhance syntax learning in current models, they also highlight a key limitation: our models learn in ways fundamentally different from humans. ๐ถ๐คตโโ๏ธ
30.05.2025 07:39 โ ๐ 2 ๐ 0 ๐ฌ 2 ๐ 0We found that models trained on CDL consistently underperform compared to those trained on Wikipedia, regardless of language, model architecture, or syntactic evaluation task, even when controlling for lexical frequency effects ๐
30.05.2025 07:39 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0This work is developed in close collaboration with my amazing supervisors @arianna-bis.bsky.social @yevgenm.bsky.social and colleague @jumelet.bsky.social to whom Iโm deeply grateful!
30.05.2025 07:39 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0โChild-Directed Language Does Not Consistently Boost Syntax Learning in Language Modelsโ
Iโm happy to share that the preprint of my first PhD project is now online!
๐ Paper: arxiv.org/abs/2505.23689