Jenny Kunz's Avatar

Jenny Kunz

@jeku.bsky.social

Postdoc at Linköping University🇸🇪. Doing NLP, particularly explainability, language adaptation, modular LLMs. I‘m also into🌋🏕️🚴.

560 Followers  |  300 Following  |  12 Posts  |  Joined: 15.12.2023  |  1.8361

Latest posts by jeku.bsky.social on Bluesky

Video thumbnail

We are opening an investigation into Grok because we believe that X may have breached the DSA.

We have seen antisemitic content, non-consensual deepfakes of women, and child sexual abuse material.

In Europe, no company will make money by violating our fundamental rights.

link.europa.eu/Fh8h84

26.01.2026 13:24 — 👍 2421    🔁 846    💬 110    📌 125
Preview
KI-Inhalte auf Social Media: Stopp den Slop Extraflache KI-Inhalte, genannt Slop, fluten das Internet. Wie kommen wir da jemals wieder raus?

ich werfe mich hiermit in den deutungskampf rund um definition, bedeutung und konsequenzen von "slop" (free link!)
www.zeit.de/2026/04/ki-i...

21.01.2026 16:15 — 👍 39    🔁 5    💬 4    📌 1

Kevin Glocker, K\"atriin Kukk, Romina Oji, Marcel Bollmann, Marco Kuhlmann, Jenny Kunz
Grow Up and Merge: Scaling Strategies for Efficient Language Adaptation
https://arxiv.org/abs/2512.10772

12.12.2025 05:31 — 👍 1    🔁 1    💬 0    📌 0
Post image

Introducing Global PIQA, a new multilingual benchmark for 100+ languages. This benchmark is the outcome of this year’s MRL shared task, in collaboration with 300+ researchers from 65 countries. This dataset evaluates physical commonsense reasoning in culturally relevant contexts.

29.10.2025 15:50 — 👍 22    🔁 10    💬 1    📌 5
Screenshot of the Viabundus website.

Screenshot of the Viabundus website.

A neat tool I just came across: Viabundus, a digital road map of northern Europe 1350-1650, that lets you calculate contemporary travel routes/times. In 1500, going Amiens → Köln by horse took almost 7 days and 13 toll payments.

#medievalsky

www.landesgeschichte.uni-goettingen.de/handelsstras...

24.10.2025 22:58 — 👍 990    🔁 379    💬 27    📌 48
Post image

📢 Announcing the First Workshop on Multilingual and Multicultural Evaluation (MME) at #EACL2026 🇲🇦

MME focuses on resources, metrics & methodologies for evaluating multilingual systems! multilingual-multicultural-evaluation.github.io

📅 Workshop Mar 24–29, 2026
🗓️ Submit by Dec 19, 2025

20.10.2025 10:37 — 👍 34    🔁 15    💬 1    📌 0

“no deanonymized preprint may be posted in the month prior to submission.” Is that a mistake?

02.09.2025 18:00 — 👍 2    🔁 0    💬 1    📌 0
AI Narrative Test

how well do *you* understand how AI reasoning works? test yourself here:
do-you-understand-ai.com

27.08.2025 21:31 — 👍 9    🔁 1    💬 4    📌 2

Sjukt att kyrkflytten marknadsförs som slow tv när det bokstavligen är världens just nu snabbaste kyrka

19.08.2025 06:46 — 👍 544    🔁 143    💬 17    📌 3

We have over 200 volunteers now for 90+ languages! We are hoping to expand the diversity of our language coverage and are still looking for participants who speak these languages. Check out how to get involved below, and please help us spread the word!

18.08.2025 15:52 — 👍 3    🔁 3    💬 1    📌 0

An abbreviation (ABB) in a journal article (JA) or Grant Application (GA) is rarely worth the words it saves. Every ABB requires cognitive resources (CR) and at my age by the time I'm halfway through a JA or GA I no longer have the CR to remember what your ABB stood for.

15.08.2025 09:39 — 👍 402    🔁 123    💬 12    📌 22
Preview
Have LLMs Finally Mastered Geolocation? - bellingcat We tasked LLMs from OpenAI, Google, Anthropic, Mistral and xAI to geolocate our unpublished holiday snaps. Here's how they did.

We all know AI models can now create realistic images and videos but how do they fare at identifying where a real image was taken? Bellingcat researchers have put Large Language Models to the test: www.bellingcat.com/resources/ho...

06.06.2025 08:09 — 👍 335    🔁 98    💬 5    📌 13

Pope Leo XIV has a degree in mathematics from Villanova university. To get where he is he has had to demonstrate a sound knowledge of original sin, but he will be the first pope to completely grasp original cos and original tan.

09.05.2025 06:55 — 👍 317    🔁 101    💬 13    📌 6
Neuralatex: A machine learning library written in pure LATEX Neuralatex: A machine learning library written in pure LATEX

Are you tired of context-switching between coding models in @pytorch.org and paper writing on @overleaf.com?

Well, I’ve got the fix for you, Neuralatex! An ML library written in pure Latex!

neuralatex.com

To appear in Sigbovik (subject to rigorous review process)

01.04.2025 11:23 — 👍 84    🔁 18    💬 2    📌 12
Executive Summary
A pro-Russia content aggregation network, Pravda, appears to be set up to flood large-language models with pro-Kremlin content, The American Sunlight Project has found. 

Over the past several months, ASP researchers have investigated 108 new domains and subdomains belonging to the Pravda network, a previously-established ecosystem of largely identical, automated web pages that previously targeted many countries in Europe as well as Africa and Asia with pro-Russia narratives about the war in Ukraine. ASP’s research, in combination with that of other organizations, brings the total number of associated domains and subdomains to 182. The network’s older targets largely consisted of states belonging to or aligned with the West. 

Notably, this latest expansion includes many countries in Africa, the Asia-Pacific, the Middle East, and North America. It also includes entities other than countries as targets, specifically non-sovereign nations, international organizations, audiences for specific languages, and prominent heads of state. The top objective of the network appears to be duplicating as much pro-Russia content as widely as possible. With one click, a single article could be autotranslated and autoshared with dozens of other sites that appear to target hundreds of millions of people worldwide. 

ASP researchers also believe the network may have been custom-built to flood large language models (LLMs) with pro-Russia content. The network is unfriendly to human users; sites within the network boast no search function, poor formatting, and unreliable scrolling, among other usability issues. This final finding poses foundational implications for the intersection of disinformation and artificial intelligence (AI), which threaten to turbocharge highly automated, global information operations in the future.

Executive Summary A pro-Russia content aggregation network, Pravda, appears to be set up to flood large-language models with pro-Kremlin content, The American Sunlight Project has found. Over the past several months, ASP researchers have investigated 108 new domains and subdomains belonging to the Pravda network, a previously-established ecosystem of largely identical, automated web pages that previously targeted many countries in Europe as well as Africa and Asia with pro-Russia narratives about the war in Ukraine. ASP’s research, in combination with that of other organizations, brings the total number of associated domains and subdomains to 182. The network’s older targets largely consisted of states belonging to or aligned with the West. Notably, this latest expansion includes many countries in Africa, the Asia-Pacific, the Middle East, and North America. It also includes entities other than countries as targets, specifically non-sovereign nations, international organizations, audiences for specific languages, and prominent heads of state. The top objective of the network appears to be duplicating as much pro-Russia content as widely as possible. With one click, a single article could be autotranslated and autoshared with dozens of other sites that appear to target hundreds of millions of people worldwide. ASP researchers also believe the network may have been custom-built to flood large language models (LLMs) with pro-Russia content. The network is unfriendly to human users; sites within the network boast no search function, poor formatting, and unreliable scrolling, among other usability issues. This final finding poses foundational implications for the intersection of disinformation and artificial intelligence (AI), which threaten to turbocharge highly automated, global information operations in the future.

A pro-Russia content aggregation network is churning out at least 3 MILLION pieces of propaganda per year, all on sites that are virtually unusable by humans.

So what's the goal? We explore the idea that it might be to flood LLMs with pro-Russia content:
static1.squarespace.com/static/6612c... 1/

27.02.2025 13:49 — 👍 1127    🔁 557    💬 22    📌 83
Post image

Sug på den, era jävla smålänningar

28.02.2025 15:17 — 👍 105    🔁 17    💬 13    📌 7
Gravestone with a picture of a dog, on Wiktionary page for “här ligger en hund begraven”

Gravestone with a picture of a dog, on Wiktionary page for “här ligger en hund begraven”

I appreciate how someone on Wiktionary found a great illustration of this Swedish idiom (“a dog lies buried here”) in 2009, long before AI slop

12.02.2025 13:45 — 👍 2    🔁 0    💬 0    📌 0
Postdoc in Natural Language Processing We are looking for a postdoc within the TrustLLM project. Full-time position for two years.

📢 We’re looking to hire a postdoc within the TrustLLM project!

Full-time position, two years, no teaching obligation. Research areas include language adaptation, modularisation, tokenization, and evaluation for multilingual LLMs.

Apply by 2025-02-05!

▶️ liu-nlp.ai/postdoc-trus...

15.01.2025 14:54 — 👍 9    🔁 6    💬 0    📌 2

This is a limited study, only considering one language and only evaluating summarization performance, but so far these findings seem to hold even in follow-up experiments in other setups. There’ll be more on this coming up next year!

18.12.2024 09:25 — 👍 0    🔁 0    💬 0    📌 0

Even with the quite limited data I used, having more trainable parameters consistently lead to better scores.

18.12.2024 09:21 — 👍 0    🔁 0    💬 1    📌 0

The best setup was LoRA in the feed-forward module, followed by bottleneck adapters. LoRA in the attention module was less stable & performed worse, especially considering the number of parameters added. Prefix tuning and (IA)^3 didn’t really work in comparison.

18.12.2024 09:19 — 👍 0    🔁 0    💬 1    📌 0
Preview
Train More Parameters But Mind Their Placement: Insights into Language Adaptation with PEFT Smaller LLMs still face significant challenges even in medium-resourced languages, particularly when it comes to language-specific knowledge -- a problem not easily resolved with machine-translated da...

New short paper, to appear at NoDaLiDa 2025! I’ve written up findings from ablation studies on language adaptation of small LLMs with limited data/compute using different PEFT methods and setups: arxiv.org/abs/2412.12674

18.12.2024 09:15 — 👍 21    🔁 2    💬 1    📌 0

Hello world!!!!

We are now on BlueSky as well 😁.

#NLP #NLProc #nodalida #baltichlt

11.12.2024 14:07 — 👍 11    🔁 4    💬 0    📌 2

Great points! The example prompt on the first page is truly painful💀

12.12.2024 19:59 — 👍 2    🔁 0    💬 0    📌 0

I had a great couple of days at SLTC in Linköping. The Swedish NLP community is doing well. Thanks @jeku.bsky.social and others for organizing!

29.11.2024 20:56 — 👍 6    🔁 2    💬 0    📌 0

1) I want his job
2) I need to create a dataset that contains this sentence

24.11.2024 21:12 — 👍 3    🔁 0    💬 0    📌 0

I’d like to be on there :)

22.11.2024 20:03 — 👍 1    🔁 0    💬 0    📌 0

bsky.app/profile/did:...

20.11.2024 20:20 — 👍 1    🔁 0    💬 0    📌 0

I’d like to be added! :)

20.11.2024 20:19 — 👍 0    🔁 0    💬 0    📌 0

Is there no starter pack for multilingual NLP yet?🧐

19.11.2024 14:08 — 👍 1    🔁 0    💬 1    📌 0

@jeku is following 20 prominent accounts