Yannis Kalantidis's Avatar

Yannis Kalantidis

@skamalas.bsky.social

he/him; Researcher at NAVER LABS Europe. Greek, resident a Barcelona. https://www.skamalas.com/

844 Followers  |  450 Following  |  24 Posts  |  Joined: 18.11.2024  |  2.1313

Latest posts by skamalas.bsky.social on Bluesky

Preview
Annie and Lakeesha struggle in school. AI teacher assistants treated them very differently. A Common Sense Media study found that prominent teacher assistants that use AI generated recommendations that appeared to be rooted in racial stereotypes based on students’ names. About a third of tea...

"Asked to generate intervention plans for struggling students, AI teacher assistants recommended more-punitive measures for hypothetical students with Black-coded names and more supportive approaches for students the platforms perceived as white" www.chalkbeat.org/2025/08/06/a...

06.08.2025 19:25 β€” πŸ‘ 547    πŸ” 320    πŸ’¬ 9    πŸ“Œ 69

most people want a quick and simple answer to why AI systems encode/exacerbate societal and historical bias/injustice and due to the reductive but common thinking of "bias in, bias out," the obvious culprit often is training data but this is not entirely true

1/

24.11.2024 16:26 β€” πŸ‘ 611    πŸ” 226    πŸ’¬ 26    πŸ“Œ 44
Post image

The biggest update in 3D reconstruction world - VGGT (new weights) has been released for commercial usage as well.
Kudos to @jianyuanwang.bsky.social to make this happen!
github.com/facebookrese...

04.08.2025 08:26 β€” πŸ‘ 12    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0
AI and Fraternity, Abeba Birhane, AI Accountability Lab  

I envision a future where human dignity, justice, peace, kindness, care, respect, accountability, and rights and freedoms serve as the north stars that guide AI development and use. Realising these ideals can’t happen without intentional tireless work, dialogues, and confrontations of ugly realities – even if they are uncomfortable to deal with. This starts with deciphering hype from reality. Pervasive narratives portray AI as a magical, fully autonomous entity approaching a God-like omnipotence and omniscience. In reality, audits of AI systems reveal a consistent failure to deliver on grandiose promises and suffer from all kinds of shortcomings, issues often swept under the rug. AI in general, and GenAI in particular, encodes and exacerbates historical stereotypes, entrenches harmful societal norms, and amplifies injustice. A robust body of  evidence demonstrates that β€” from hiring, welfare allocation, medical care allocation to anything in between β€” deployment of AI is widening inequity, disproportionately impacting people at the margins of society and concentrating power and influence in the hands of few. Major actorsβ€”including Google, Microsoft, Amazon, Meta, and OpenAIβ€”have willingly aligned with authoritarian regimes and proactively abandoned their pledges to fact-check, prevent misinformation, respect diversity and equity, refrain from using AI for weapons development, while retaliating against critique. The aforementioned vision can’t and won’t happen without confrontation of these uncomfortable facts. This is precisely why we need active resistance and refusal of unreliable and harmful AI systems; clearly laid out regulation and enforcement; and shepherding of the AI industry towards transparency and accountability of responsible bodies. "Machine agency" must be in service of human agency and empowerment, a coexistence that isn't a continuation of modern tech corporations’ inequality-widening,

AI and Fraternity, Abeba Birhane, AI Accountability Lab I envision a future where human dignity, justice, peace, kindness, care, respect, accountability, and rights and freedoms serve as the north stars that guide AI development and use. Realising these ideals can’t happen without intentional tireless work, dialogues, and confrontations of ugly realities – even if they are uncomfortable to deal with. This starts with deciphering hype from reality. Pervasive narratives portray AI as a magical, fully autonomous entity approaching a God-like omnipotence and omniscience. In reality, audits of AI systems reveal a consistent failure to deliver on grandiose promises and suffer from all kinds of shortcomings, issues often swept under the rug. AI in general, and GenAI in particular, encodes and exacerbates historical stereotypes, entrenches harmful societal norms, and amplifies injustice. A robust body of evidence demonstrates that β€” from hiring, welfare allocation, medical care allocation to anything in between β€” deployment of AI is widening inequity, disproportionately impacting people at the margins of society and concentrating power and influence in the hands of few. Major actorsβ€”including Google, Microsoft, Amazon, Meta, and OpenAIβ€”have willingly aligned with authoritarian regimes and proactively abandoned their pledges to fact-check, prevent misinformation, respect diversity and equity, refrain from using AI for weapons development, while retaliating against critique. The aforementioned vision can’t and won’t happen without confrontation of these uncomfortable facts. This is precisely why we need active resistance and refusal of unreliable and harmful AI systems; clearly laid out regulation and enforcement; and shepherding of the AI industry towards transparency and accountability of responsible bodies. "Machine agency" must be in service of human agency and empowerment, a coexistence that isn't a continuation of modern tech corporations’ inequality-widening,

so I am one of the 12 people (including the β€œgod-fathers of AI”) that will be at the Vatican this September for a two full-day working group on the Future of AI

here is my Vatican approved short provocation on 'AI and Fraternity' for the working group

04.08.2025 11:31 β€” πŸ‘ 532    πŸ” 157    πŸ’¬ 31    πŸ“Œ 16

Deadlines for ICLR 2026 have been communicated on the official site:
Abstract: Sep 19 '25 (Anywhere on Earth)
Paper: Sep 24 '25 (Anywhere on Earth)
iclr.cc/Conferences/...

ICLR 2026 will happen in Brazil! ❀️

(Don't fall for the predatory conferences with the same name)

25.06.2025 09:06 β€” πŸ‘ 23    πŸ” 8    πŸ’¬ 0    πŸ“Œ 0
Post image

New research from MIT found that those who used ChatGPT can’t remember any of the content of their essays.

Key takeaway: the product doesn’t suffer, but the process does. And when it comes to essays, the process *is* how they learn.

arxiv.org/pdf/2506.088...

18.06.2025 07:32 β€” πŸ‘ 3743    πŸ” 1542    πŸ’¬ 204    πŸ“Œ 382
Post image

When were #CVPR2025 papers available on arXiv? πŸ‘‡

17.06.2025 11:52 β€” πŸ‘ 39    πŸ” 9    πŸ’¬ 2    πŸ“Œ 0
Post image Post image Post image

#CVPR2025 attendance per country, comparing 2025 and 2024, sorted by diff. Numbers were extracted from these two slides presented by the respective program chairs.

(Totals on the slide seem to include virtual registrations, but not the per country numbers)

Diff between western and eastern Europe?

14.06.2025 21:49 β€” πŸ‘ 19    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

I have created a starter pack with researchers from Naver Labs Europe @naverlabseurope.bsky.social: we are in Grenoble, France, and we do research in AI for robotics, computer vision, NLP, machine learning, HRI.
go.bsky.app/JdTFu4Q

14.06.2025 19:05 β€” πŸ‘ 33    πŸ” 6    πŸ’¬ 1    πŸ“Œ 1
Post image

Our work on "Reasoning in visual navigation..." presented as a "Highlight" by Boris Chidlovskii and Francesco Giuliari at #cvpr2025!

Interactive site, play around with dynamical models:
europe.naverlabs.com/research/pub...

Thanks @weinzaepfelp.bsky.social for the photo.

@steevenj7.bsky.social

14.06.2025 17:33 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

During today's #CVPR2025 workshops, I will present:
- What matters in ImageNav: architecture, pre-training, sim settings, pose (poster & highlight at the Embodied AI workshop)
- CondiMen: Conditional Multi-person Human Mesh Recovery (Poster at the Rhobin workshop and at the 3D Humans workshop)

12.06.2025 12:18 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0

Paper: arxiv.org/abs/2503.19777
Code: github.com/vladan-stojn...
Demo: huggingface.co/spaces/stojn...

Work with @skamalas.bsky.social, Jiri Matas, and @gtolias.bsky.social.

13.06.2025 12:02 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Wanna the outstanding performance of MASt3R while using a ViT-B or ViT-S encoder instead of its ViT-L one? Don't miss how we build DUNE, a single encoder for diverse 2D & 3D tasks, at this afternoon #CVPR2025 poster session (poster #376).

paper: arxiv.org/abs/2503.14405
code: github.com/naver/dune

15.06.2025 12:02 β€” πŸ‘ 18    πŸ” 6    πŸ’¬ 0    πŸ“Œ 0

Are you at @cvprconference.bsky.social? Come by our poster!
πŸ“… Sat 14/6, 10:30-12:30
πŸ“ Poster #395, ExHall D

13.06.2025 05:09 β€” πŸ‘ 17    πŸ” 9    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸ’œ Huge thanks to our amazing organizers for making #WiCV
@cvprconference.bsky.social 2025 possible! πŸ™Œ

After 6+ months of near-daily work by 9 dedicated organizers, it’s exciting to see everything come together.
So proud of this amazing team! πŸ’ͺ✨

12.06.2025 20:55 β€” πŸ‘ 15    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

wait a minute... is this the Acro... :P

11.06.2025 13:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

You woke up early in the morning jet-lagged and having a hard time deciding for a workshop today @cvprconference.bsky.social ?

Here's a reliable choice for you: our workshop on πŸ›Ÿ Uncertainty Quantification for Computer Vision!

πŸ—“οΈ Day: Wed, Jun 11
πŸ“Room: 102 B
#CVPR2025 #UNCV2025

11.06.2025 11:33 β€” πŸ‘ 9    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Post image

Have you heard about HD-EPIC?
Attending #CVPR2025
Multiple opportunities to know about the most highly-detailed video dataset with a digital twin, long-term object tracks, VQA,…
hd-epic.github.io

1. Find any of the 10 authors attending @cvprconference.bsky.social
– identified by this badge.

🧡

10.06.2025 21:48 β€” πŸ‘ 5    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

Check out our new paper in #CVPR2025! We test the limits of multi-teacher distillation to get one of the most versatile encoders yet. Mixing DINOv2 semantics πŸ¦•, MASt3R multi-view reconstruction πŸ“ΈπŸ“Έ...πŸ“Έ and Human Mesh Recovery πŸ•ΊπŸ’ƒ...πŸŒοΈβ€β™‚οΈ
Check our project page: europe.naverlabs.com/research/pub...

09.06.2025 14:05 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

10/ πŸ’¬ Both papers tackle generalization head-on β€”
One by building a universal encoder from heterogeneous knowledge, the other by refining language-driven predictions without training.

Don't forget to check them out at #CVPR2025 in Nashville! 🎸

09.06.2025 11:07 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
GitHub - vladan-stojnic/LPOSS: Code for LPOSS: Label Propagation Over Patches and Pixels for Open-vocabulary Semantic Segmentation (CVPR2025) Code for LPOSS: Label Propagation Over Patches and Pixels for Open-vocabulary Semantic Segmentation (CVPR2025) - vladan-stojnic/LPOSS

9/ πŸ“ˆ Results:
β†’ Substantial mIoU boost on 8 datasets
β†’ LPOSS+ especially shines for boundary precision
β†’ Outperforms many VLM-based methods without training

Work w/ Vladan @stojnicv.xyz and Giorgos @gtolias.bsky.social from CTU in Prague
Code: github.com/vladan-stojn...
Demo: hf.co/spaces/stojn...

09.06.2025 11:07 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

8/ πŸ” LPOSS stages:
πŸ”Έ LPOSS: Refines patch-level predictions
πŸ”Ή LPOSS+: Refines on pixel-level using upsampled LPOSS output
βœ… Uses DINO features + spatial cues for affinity
βœ… Totally training-free!

09.06.2025 11:07 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

7/ 🚫 VLMs are great at training-free segmentation, but:

* Predictions are patch-based (low-res)
* Predictions are performed independently at each patch (noisy)

LPOSS improves this with test-time label propagation, without fine-tuning.

09.06.2025 11:07 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

6/ πŸ“„ Paper 2:
"LPOSS: Label Propagation Over Patches and Pixels for Open-Vocabulary Semantic Segmentation"

Can graph-based label propagation refine weak, patch-level predictions from VLMs like CLIP? We say yes β€” introducing LPOSS and LPOSS+.

09.06.2025 11:07 β€” πŸ‘ 7    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Preview
DUNE: Distilling a Universal Encoder from heterogenous 2D and 3D teachers CVPR 2025 publication

5/ Project page: europe.naverlabs.com/dune

Work with Bulent Sariyildiz @mbsariyildiz.bsky.social Diane Larlus @dlarlus.bsky.social Philippe Weinzaepfel @weinzaepfelp.bsky.social Pau de Jorge @pdejorge.bsky.social and Thomas Lucas at NAVER LABS Europe @naverlabseurope.bsky.social

09.06.2025 11:07 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
GitHub - naver/dune: Code repository for "DUNE: Distilling a Universal Encoder from Heterogeneous 2D and 3D Teachers" Code repository for "DUNE: Distilling a Universal Encoder from Heterogeneous 2D and 3D Teachers" - naver/dune

4/ πŸ’‘ Results:
βœ… Strong performance on diverse tasks
βœ… #1 on Niantic's Map-free Relocalization leaderboard, inproving over MASt3R with a smaller encoder (ViT-Base vs. ViT-Large)

Code and models: github.com/naver/dune

09.06.2025 11:07 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

3/ πŸ” DUNE is trained from heterogeneous teachers using heterogeneous data:
🟧 DINOv2 (2D foundation model - Oquab et al @ Meta)
πŸŸ₯ MAST3R (3D foundation model - @vincentleroy.bsky.social et al @ NAVER LABS Europe - )
🟨 Multi-HMR (3D human perception - @fbaradel.bsky.social et al @ NAVER LABS Europe)

09.06.2025 11:07 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

2/ 🧠 Why DUNE?
β†’ 2D & 3D vision models (e.g., segmentation, mono- or binocular depth, human mesh, 3D reconstruction) are siloed.
β†’ Can we distill their knowledge into a single universal encoder?
Yes β€” using multi-teacher distillation with transformer projectors, teacher dropping, and more.

09.06.2025 11:07 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

1/ πŸ“„ Paper 1:
"DUNE: Distilling a UNiversal Encoder from Heterogeneous 2D and 3D Teachers"
We propose DUNE: a ViT-based encoder distilled from multiple specialized 2D & 3D foundation models to unify visual tasks across 2D, 3D and human understanding.

09.06.2025 11:07 β€” πŸ‘ 18    πŸ” 5    πŸ’¬ 2    πŸ“Œ 2

🧡 Two new papers at #CVPR2025 on generalization in visual representations β€” covering both universal encoders for 2D and 3D tasks, and open-vocabulary semantic segmentation

Let's dive in! πŸ‘‡

09.06.2025 11:07 β€” πŸ‘ 23    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

@skamalas is following 20 prominent accounts