's Avatar

@nhminhle.bsky.social

11 Followers  |  1 Following  |  7 Posts  |  Joined: 30.05.2025  |  1.5261

Latest posts by nhminhle.bsky.social on Bluesky

Preview
OWL: Probing Cross-Lingual Recall of Memorized Texts via World Literature Large language models (LLMs) are known to memorize and recall English text from their pretraining data. However, the extent to which this ability generalizes to non-English languages or transfers...

Check out our dataset and findings more in details๐Ÿ‘‡

๐ŸŒ github.com/emirkaan5/OWL/
๐Ÿ“œ arxiv.org/abs/2505.22945

Work done at @UMassNLP by @alishasrivas.bsky.socialโ€ฌโ€ฌ, @emirdotexe.bsky.socialโ€ฌ, @nhminhle.bsky.socialโ€ฌ, @chautmpham.bsky.social, @markar.bsky.social, and @miyyer.bsky.social.

30.05.2025 15:37 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

๐Ÿงฎ Does quantization (GPTQ) impact cross-lingual knowledge transfer?

LLaMA-3.1-70B: 4-bit > 8-bit โฉ accuracy drops MORE at 8-bit (up to 25%)
LLaMA-3.1-8B: 8-bit > 4-bit โฉ accuracy drops MORE at 4-bit (up to 8%)

๐Ÿง  Bigger models arenโ€™t always more robust to quantization.

30.05.2025 15:37 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

๐Ÿ”Š LLMs can transfer knowledge across modalities (Text โ†’ Audio).

On GPT-4o-Audio vs Text:
๐Ÿ“– Direct Probing: 75.5% (vs. 92.3%)
๐Ÿ‘ค Name Cloze: 15.9% (vs. 38.6%)
โœ๏ธ Prefix Probing: 27.2% (vs. 22.6%)

Qwen-Omni shows similar trends but lower accuracy.

30.05.2025 15:37 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

What if we perturb the text?

๐Ÿงฉ shuffled text
๐ŸŽญ masked character names
๐Ÿ™…๐Ÿปโ€โ™€๏ธ passages w/o characters

๐ŸšจReduce accuracy with the degree varying across languages BUT models can still identify the books better than newly published books (0.1%) ๐Ÿšจ

30.05.2025 15:37 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

LLMs can identify book titles and authors across languages - even those not seen during pre-training:

63.8% accuracy on English passages
47.2% on official translations (Spanish, Turkish, Vietnamese) ๐Ÿ‡ช๐Ÿ‡ธ ๐Ÿ‡น๐Ÿ‡ท ๐Ÿ‡ป๐Ÿ‡ณ
36.5% on completely unseen languages like Sesotho & Maithili ๐ŸŒ

30.05.2025 15:37 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

OWL has aligned excerpts from 20 EN novels, with translations in ES ๐Ÿ‡ช๐Ÿ‡ธ, TR ๐Ÿ‡น๐Ÿ‡ท, VI ๐Ÿ‡ป๐Ÿ‡ณ + 6 new low-resource languages ๐ŸŒ & EN audio ๐Ÿ”Š

We probe LLMs to:
1๏ธโƒฃ identify book/author (direct probing)
2๏ธโƒฃ predict masked names (name cloze)
3๏ธโƒฃ generate continuation (prefix probing)

30.05.2025 15:37 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

LLMs memorize novels ๐Ÿ“š in English. But what about existing translations? Or translations into new languages?

Our ๐Ÿฆ‰OWL dataset (31K/10 languages) shows GPT4o recognizes books:
92% English
83% official translations
69% unseen translations
75% as audio (EN)

30.05.2025 15:37 โ€” ๐Ÿ‘ 7    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 3

@nhminhle is following 1 prominent accounts