Manuel's Avatar

Manuel

@chasinggradients.bsky.social

NLP - mostly representation learning #NLP #NLProc

66 Followers  |  317 Following  |  11 Posts  |  Joined: 29.10.2023  |  2.0906

Latest posts by chasinggradients.bsky.social on Bluesky

Preview
Discontinuation of the MS Word Template ARR abandons the MS Word Template for conference submissions. The submissions based on the Word template will be desk-rejected starting from March 2026.

๐Ÿ“ Discontinuation of the MS Word Template
ARR will now fully adopt the LaTeX template to streamline formatting and reduce review workload.
Starting March 2026, submissions using the MS Word template will be desk-rejected.
Check details here: aclrollingreview.org/discontinuat...
#ARR #NLProc

30.10.2025 21:40 โ€” ๐Ÿ‘ 9    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

The MTEB team has just released MTEB v2, an upgrade to their evaluation suite for embedding models!

Their blogpost covers all changes, including easier evaluation, multimodal support, rerankers, new interfaces, documentation, dataset statistics, a migration guide, etc.

๐Ÿงต

20.10.2025 14:36 โ€” ๐Ÿ‘ 7    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
PTEB: Towards Robust Text Embedding Evaluation via Stochastic Paraphrasing at Evaluation Time with LLMs Current evaluations of sentence embedding models typically rely on static test beds such as the Massive Text Embedding Benchmark (MTEB). While invaluable, repeated tuning on a fixed suite can inflate ...

PTEB: Towards Robust Text Embedding Evaluation via Stochastic Paraphrasing at Evaluation Time with LLMs

Introduces a dynamic evaluation protocol that generates meaning-preserving paraphrases at evaluation time to assess embedding model robustness.

๐Ÿ“ arxiv.org/abs/2510.06730

09.10.2025 02:32 โ€” ๐Ÿ‘ 1    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
AI bots wrote and reviewed all papers at this conference Event will assess how reviews by models compare with those written by humans.

๐Ÿงช A new computer science conference, Agents4Science, will feature papers written and peer-reviewed entirely by AI agents. The event serves as a sandbox to evaluate the quality of machine-generated research and its review process.
#MLSky

15.10.2025 15:33 โ€” ๐Ÿ‘ 4    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
ACL 2026 Visa Invitation Letter Request Please use this form to request a visa invitation letter. The letter will be sent to the email address provided below.

โœˆ๏ธ Visa Letter Requests for ACL 2026
If you intend to commit your paper to ACL 2026 and require an invitation letter for visa purposes, please fill out the visa request form as soon as possible.
(docs.google.com/forms/d/e/1F...)

#ARR #ACL #NLProc

14.10.2025 15:27 โ€” ๐Ÿ‘ 8    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Illustration of a woman wearing a graduation cap and a lab coat, holding a magnifying glass and examining a document. A mechanical parrot with gears and circuits is perched on her shoulder.

Illustration of a woman wearing a graduation cap and a lab coat, holding a magnifying glass and examining a document. A mechanical parrot with gears and circuits is perched on her shoulder.

๐Ÿš€ ๐—Ÿ๐—ฎ๐˜๐—ฒ๐˜€๐˜ ๐—ฃ๐—ฒ๐—ฒ๐—ฟ ๐—ฅ๐—ฒ๐˜ƒ๐—ถ๐—ฒ๐˜„ ๐——๐—ฎ๐˜๐—ฎ๐˜€๐—ฒ๐˜ ๐—ฅ๐—ฒ๐—น๐—ฒ๐—ฎ๐˜€๐—ฒ ๐—ณ๐—ฟ๐—ผ๐—บ ๐—”๐—ฅ๐—ฅ ๐Ÿฎ๐Ÿฌ๐Ÿฎ๐Ÿฑ!
tudatalib.ulb.tu-darmstadt.de/handle/tudat...

๐Ÿ“Š ๐—ก๐—ฒ๐˜„๐—น๐˜† ๐—ฎ๐—ฑ๐—ฑ๐—ฒ๐—ฑ ๐—”๐—–๐—Ÿ ๐Ÿฎ๐Ÿฌ๐Ÿฎ๐Ÿฑ ๐—ฑ๐—ฎ๐˜๐—ฎ:
โœ… ๐Ÿฎ๐—ธ papers
โœ… ๐Ÿฎ๐—ธ reviews
โœ… ๐Ÿด๐Ÿฐ๐Ÿต meta-reviews
โœ… ๐Ÿญ.๐Ÿฑ๐—ธ papers with rebuttals

(1/๐Ÿงต)

08.10.2025 06:57 โ€” ๐Ÿ‘ 3    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Google just released a 270M parameter Gemma model. As a tiny model lover I'm excited. Models in this size class are usually barely coherent, I'll give it a try today to see how this does. developers.googleblog.com/en/introduci...

14.08.2025 16:38 โ€” ๐Ÿ‘ 48    ๐Ÿ” 2    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1
Post image

I just released Sentence Transformers v4.1; featuring ONNX and OpenVINO backends for rerankers offering 2-3x speedups and improved hard negatives mining which helps prepare stronger training datasets.

Details in ๐Ÿงต

15.04.2025 13:54 โ€” ๐Ÿ‘ 11    ๐Ÿ” 4    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
a cat holding a sign that says help ALT: a cat holding a sign that says help

๐Ÿ—ฃ๏ธCall for emergency reviewers

I am serving as an AC for #ICML2025, seeking emergency reviewers for two submissions

Are you an expert of Knowledge Distillation or AI4Science?

If so, send me DM with your Google Scholar profile and OpenReview profile

Thank you!

20.03.2025 05:25 โ€” ๐Ÿ‘ 2    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Post image

We've just released MMTEB, our multilingual upgrade to the MTEB Embedding Benchmark!

It's a huge collaboration between 56 universities, labs, and organizations, resulting in a massive benchmark of 1000+ languages, 500+ tasks, and a dozen+ domains.

Details in ๐Ÿงต

21.02.2025 15:06 โ€” ๐Ÿ‘ 23    ๐Ÿ” 4    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Preview
Can Cross Encoders Produce Useful Sentence Embeddings? Cross encoders (CEs) are trained with sentence pairs to detect relatedness. As CEs require sentence pairs at inference, the prevailing view is that they can only be used as re-rankers in information r...

Can Cross Encoders Produce Useful Sentence Embeddings?

IBM discovered that early cross encoders layers can produce effective sentence embeddings, enabling 5.15x faster inference while maintaining comparable accuracy to full dual encoders.

๐Ÿ“ arxiv.org/abs/2502.03552

07.02.2025 03:34 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Same issue for me

30.01.2025 19:53 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Distiling DeepSeek reasoning to ModernBERT classifiers How can we use the reasoning ability of DeepSeek to generate synthetic labels for fine tuning a ModernBERT model?

Why choose between strong #LLM reasoning and efficient models?

Use DeepSeek to generate high-quality training data, then distil that knowledge into ModernBERT for fast, efficient classification.

New blog post: danielvanstrien.xyz/posts/2025/d...

29.01.2025 10:07 โ€” ๐Ÿ‘ 58    ๐Ÿ” 11    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 4
Post image

People often claim they know when ChatGPT wrote something, but are they as accurate as they think?

Turns out that while general population is unreliable, those who frequently use ChatGPT for writing tasks can spot even "humanized" AI-generated text with near-perfect accuracy ๐ŸŽฏ

28.01.2025 14:55 โ€” ๐Ÿ‘ 188    ๐Ÿ” 66    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 19
Preview
A Test So Hard No AI System Can Pass It โ€” Yet (Gift Article) The creators of a new test called โ€œHumanityโ€™s Last Examโ€ argue we may soon lose the ability to create tests hard enough for A.I. models.

I wrote about a new AI evaluation called "Humanity's Last Exam," a collection of 3,000 questions submitted by leading academics to try to stump leading AI models, which mostly find today's college-level tests too easy.

www.nytimes.com/2025/01/23/t...

23.01.2025 16:41 โ€” ๐Ÿ‘ 208    ๐Ÿ” 46    ๐Ÿ’ฌ 17    ๐Ÿ“Œ 15
Post image

I just released Sentence Transformers v3.4.0, featuring a memory leak fix (memory not being cleared upon model & trainer deletion), compatibility between the powerful Cached... losses and the Matryoshka loss modifier, and a bunch of fixes & small features.

Details in ๐Ÿงต

23.01.2025 16:44 โ€” ๐Ÿ‘ 12    ๐Ÿ” 4    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
10th Workshop on Representation Learning for NLP - Call for Papers The 10th Workshop on Representation Learning for NLP (RepL4NLP 2025), co-located with NAACL 2025 in Albuquerque, New Mexico, invites papers of a theoretical or experimental nature describing recent ad...

Disappointed with #ICLR or #NAACL reviews? Consider submitting your work at #Repl4NLP, whether it's full papers, extended abstracts, or cross-submissions. ๐Ÿ”ฅ

Details on submissions ๐Ÿ‘‰ sites.google.com/view/repl4nl...

โฐ Deadline January 30

23.01.2025 16:30 โ€” ๐Ÿ‘ 2    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Preview
Introducing Phi-4: Microsoftโ€™s Newest Small Language Model Specializing in Complex Reasoning | Microsoft Community Hub Today we are introducing Phi-4, our 14B parameter state-of-the-art small language model (SLM) that excels at complex reasoning in areas such as math, in...

Microsoftโ€™s latest small language model - phi-4 - is open source and now available on Hugging Face techcommunity.microsoft.com/blog/aiplatf...

09.01.2025 15:40 โ€” ๐Ÿ‘ 10    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
LLMs are Also Effective Embedding Models: An In-depth Overview Large language models (LLMs) have revolutionized natural language processing by achieving state-of-the-art performance across various tasks. Recently, their effectiveness as embedding models has gaine...

LLMs are Also Effective Embedding Models: An In-depth Overview

Provides a comprehensive analysis on adopting LLMs as embedding models, examining both zero-shot prompting and tuning strategies to derive competitive text embeddings vs traditional models.

๐Ÿ“ arxiv.org/abs/2412.12591

18.12.2024 06:40 โ€” ๐Ÿ‘ 1    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

You and users of the "but humans"-argument assume different goals for AI.

The argument assumes that the goal is to develop human-level AI. (Or it's used to counter statements claiming AI systems are less intelligent than humans.) It's not a direct argument for their usefulness.

01.01.2025 10:23 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Not sure about this idea (and also the objective of maximizing impact), but I really like the "plain language summary" I've seen in some medical papers.

25.12.2024 10:05 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
How Hallucinatory A.I. Helps Science Dream Up Big Breakthroughs (Gift Article) Hallucinations, a bane of popular A.I. programs, turn out to be a boon for venturesome scientists eager to push back the frontiers of human knowledge.

The upside of A.I. hallucination
gift link www.nytimes.com/2024/12/23/s...

24.12.2024 15:05 โ€” ๐Ÿ‘ 103    ๐Ÿ” 30    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 4
Preview
What happened to BERT & T5? On Transformer Encoders, PrefixLM and Denoising Objectives โ€” Yi Tay A Blogpost series about Model Architectures Part 1: What happened to BERT and T5? Thoughts on Transformer Encoders, PrefixLM and Denoising objectives

Good blog post on good old encoder-style models.

Glad to see ModernBERT recently brought something new to the field. So don't count BERT as GOFAI yet.

www.yitay.net/blog/model-a...

24.12.2024 12:16 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image Post image

Announcement #1: our call for papers is up! ๐ŸŽ‰
colmweb.org/cfp.html
And excited to announce the COLM 2025 program chairs @yoavartzi.com @eunsol.bsky.social @ranjaykrishna.bsky.social and @adtraghunathan.bsky.social

17.12.2024 15:48 โ€” ๐Ÿ‘ 66    ๐Ÿ” 24    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Preview
Computer Science Conference Deadlines Map Interactive world map of Computer Science, AI, and ML conference deadlines

Unsure where to submit your next research paper to now that aideadlin.es is not updated anymore? And letโ€™s be honest, is the location not as important as the conference itself?

๐Ÿ—บ๏ธ Check out my latest side-project: deadlines.pieter.ai

23.12.2024 14:39 โ€” ๐Ÿ‘ 13    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

๐Ÿงช New pre-print explores generative AIโ€™s in medicine, highlighting applications for clinicians, patients, researchers, and educators. It also addresses challenges like privacy, transparency, and equity.
Additional details from the author linked below.
๐Ÿฉบ๐Ÿ–ฅ๏ธ
Direct link: arxiv.org/abs/2412.10337

22.12.2024 15:03 โ€” ๐Ÿ‘ 20    ๐Ÿ” 4    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Instead of listing my publications, as the year draws to an end, I want to shine the spotlight on the commonplace assumption that productivity must always increase. Good research is disruptive and thinking time is central to high quality scholarship and necessary for disruptive research.

20.12.2024 11:18 โ€” ๐Ÿ‘ 1154    ๐Ÿ” 375    ๐Ÿ’ฌ 21    ๐Ÿ“Œ 57

IMO, there's a great discussion over there (in my timeline, not the for yoy tab) with interesting insights from the OpenAI team.

Bluesky isn't there (yet).

22.12.2024 13:24 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Re your question at the end:

22.12.2024 13:18 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

I'll get straight to the point.

We trained 2 new models. Like BERT, but modern. ModernBERT.

Not some hypey GenAI thing, but a proper workhorse model, for retrieval, classification, etc. Real practical stuff.

It's much faster, more accurate, longer context, and more useful. ๐Ÿงต

19.12.2024 16:45 โ€” ๐Ÿ‘ 620    ๐Ÿ” 147    ๐Ÿ’ฌ 19    ๐Ÿ“Œ 34

@chasinggradients is following 20 prominent accounts