Weโre having a (human) language acquisition meetup at #ACL2025. RSVP on Whova for updates!
28.07.2025 15:08 โ ๐ 2 ๐ 1 ๐ฌ 0 ๐ 0@antonisa.bsky.social
Assistant Prof at GMU. NLP, CompLing, ML, and other things language+humans
Weโre having a (human) language acquisition meetup at #ACL2025. RSVP on Whova for updates!
28.07.2025 15:08 โ ๐ 2 ๐ 1 ๐ฌ 0 ๐ 0Proud to work with John Pavlopoulos and @antonisa.bsky.social on this publication!
Check out the data and code here: github.com/andhmak/rule...
4/4
I'm attending #ACL in Vienna this week. ๐ฆ๐น We're running a BoF on Language Technologies for Crisis Response and Preparedness, co-hosted w/Will Lewis
๐ Wed. 30th, 11am. Room 1.33.
You can join us virtually too. DM me if youโre interested โจ
@wildlewis.bsky.social @antonisa.bsky.social
Looking forward to this year's edition! With great speakers: Ryan McDonald Yulan He @vn-ml.bsky.social @antonisa.bsky.social Raquel Fernandez @annarogers.bsky.social Preslav Nakov @mohitbansal.bsky.social @eunsol.bsky.social Marie-Catherine de Marnefffe !
06.06.2025 09:10 โ ๐ 6 ๐ 3 ๐ฌ 0 ๐ 0๐ข It's official! Save the Date!
The #AthNLP Summer School is coming!
๐
4-10 September 2025
๐ Athens, Greece
@athnlp.bsky.social a top NLP summer school, offers a week of lectures, workshops, and networking.
๐ athnlp.github.io/2025/index.h...
#AthNLP2025 #NLP #AI #SummerSchool
๐ข I am looking for a postdoc for the next academic year!
(Due to the funding source, US persons preferred)
Interested in multimodal LLMs and their application to education domains (as well as multilinguality, cross-lingual, and low-resource learning)?
If yes, send me a message here or an email!
Usually on an ipad these days...
I think I just hate having to write notes in the middle of a line in the 1-col papers, so it's probably about how close the space is to the text as opposed to how abundant it is
I find the 2-col format easier for reviewing/note-taking/suggesting edits, because the info is spread out vertically and I have more margin space for notes closer to the actual text.
But for just reading, agreed, we should just produce dynamic pubs that people can customize to their preferences.
A recently discharged Division 252 officer describes the arbitrary nature of this boundary: "For the division, the kill zone extends as far as a sniper can see." But the issue goes beyond geography. "We're killing civilians there who are then counted as terrorists," he says. "The IDF spokesperson's announcements about casualty numbers have turned this into a competition between units. If Division 99 kills 150 [people], the next unit aims for 200." These accounts of indiscriminate killing and the routine classification of civilian casualties as terrorists emerged repeatedly in Haaretz's conversations with recent Gaza veterans.
Fucking hell... Seriously, there's no going back from this. If it were not from Haaretz, nobody would believe it. Worse part? Nobody fucking cares.
archive.ph/NVG4p#select...
Another example of a point that supports this argument is the observation that more than 50% of the "facts" that are available in Wikipedia/Wikidata, they are only available or retrievable in a _single_ language. The observation is hidden somewhere in this paper
aclanthology.org/2020.emnlp-m...
Hm, I think I have a couple of examples.
The first is the linked PNAS paper, titled "Language extinction triggers the loss of unique medicinal knowledge" www.pnas.org/doi/epdf/10....
Excited to announce the launch of our ML-SUPERB 2.0 challenge @interspeech.bsky.social 2025! Join us in pushing the boundaries of multilingual ASR and LID! ๐
๐ป multilingual.superbbenchmark.org
In the above examples, some people had a problem, and a computer scientist steps in to help produce a solution, and they write a paper about it so that if anyone else has a similar problem in the future, there's a guide to solving it. How's that not enough of a contribution?
05.12.2024 04:09 โ ๐ 16 ๐ 0 ๐ฌ 2 ๐ 0next thing you know you realize that you need and you starr building a simplification dataset for the contact language (let's pick Rioplatense Spanish for this example), or building NER tools that can handle the specific regional orthographic variations of Cypriot Greek.
05.12.2024 04:09 โ ๐ 14 ๐ 0 ๐ฌ 1 ๐ 0Or they might result from the specific needs of a scientific (or not) team. Hypothetical example: a sociologist teams up with a meteorologist and a computer scientist to figure out how to best convey changing climate threats to an indigenous community, and ...
05.12.2024 04:09 โ ๐ 14 ๐ 0 ๐ฌ 1 ๐ 0Or the leaders of a different community might actually want an LLM to ensure their language has the same perceived prestige and tool access as a more dominant language that might be threatening theirs.
05.12.2024 04:09 โ ๐ 16 ๐ 0 ๐ฌ 1 ๐ 0A lot of the "narrow"-focus datasets on otherwise underserved languages might be the result of the specific needs of the community: a community might not need an LLM, but they might need a morphosyntactic analyser that they can deploy in a classroom to teach their language.
05.12.2024 04:09 โ ๐ 16 ๐ 1 ๐ฌ 1 ๐ 0
Sure, If your space of scientific questions is only "how can I train a model to do X?", then the slight variation of "how can I train a model to do X in language Y?" is not too interesting in and of itself (although there might be arguments just for that, see above)
The extent to which we understand them, in our current setting, is measured by the datasets that you complain about.
Yes, we might _believe_ that model X will be able to perform task Y in some language Z and context W, but we don't _know_, not until we actually try (and often find "...Not quite").
Human beings have come up with 7000+ ways to communicate. And each of these modes encodes unique sociocultural, historical, (and potentially more types of *al) information. So being able to understand (or have machines understand) them ensures that we don't lose part of our collective knowledge.
05.12.2024 04:09 โ ๐ 28 ๐ 4 ๐ฌ 2 ๐ 1Late to the party, but after reading most subthreads, I'll bite.
I think your whole premise suggests a very narrow view of what is science or scientific contributions.
These are good tips!
03.12.2024 18:22 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0interspeech2025.org challenge MultiLingual Speech processing Universal PERformance Benchmark Organizers: Antonios Anastasopoulos, Martijn Bartelds, William Chen, Dan Jurafsky, Hung-yi Lee, Karen Livescu, Chutong Meng, Jiatong Shi, Hsiu-Hsuan Wang, Shih-Heng Wang, Shinji Watanabe
๐๐ค ML-SUPERB 2.0 Challenge at #Interspeech2025: Push the boundaries of cross-lingual speech processing!
๐ 154 languages & 200+ accents/dialects
๐ Live leaderboard & online evaluation! Join now: multilingual.superbbenchmark.org
We had a great discussion with @robertlemos.bsky.social from Dark Reading about our new paper "Hacking Back the AI-Hacker: Prompt Injection as a Defense Against LLM-driven Cyberattacks"(arxiv.org/abs/2410.20911). Mantis turns the hardness of dealing with prompt injections into an opportunity!
21.11.2024 13:22 โ ๐ 7 ๐ 4 ๐ฌ 1 ๐ 0The multi-number phonebook retrieval task entails retrieving several phone numbers from a phonebook at once, given names. Hawk trained using Birdie strongly outperforms Hawk trained using Next Token Prediction on the multi-number phonebook retrieval task. Hawk trained using Next Token Prediction performs just above random guessing when retrieving 1 and 4 phone numbers, and falls to random performance when retrieving more than 4 phone numbers. In contrast, Hawk trained using Birdie gets 100% accuracy when retrieving 1 phone number. That 100% score slowly decays to about 80% accuracy when retrieving up to 32 phone numbers simultaneously. Two Transformers are included, one trained using Birdie, and the other trained using Next Token Prediction. They both always achieve about 100% accuracy, even when retrieving 32 phone numbers.
Meet Birdie ๐ค!
Our EMNLP 2024 paper boosts SSMs like Mamba and Hawk on long-range, context-heavy tasks, closing the gap with Transformers.
Proud to work with @jimmysmith1919.bsky.social, @antonisa.bsky.social, & Amarda Shehu.
๐ Paper: arxiv.org/abs/2411.01030
๐ป Code: github.com/samblouir/bi...
Are there any resources for helping introverted students navigate the craziness that is conference attendance, so that they make the most of it?
I mean, I can tell my students to "go talk to people", but I'm looking for something more comprehensive.
I (extrovert advisor) am struggling.
This ๐
the other place has been not-fun for a while -- I'll try to post here as well!