Super interesting to see pure SSL outperforms text alignement on a super competitive but text-aligned suited task π€―
18.08.2025 15:44 β π 2 π 0 π¬ 0 π 0@gastruc.bsky.social
2nd Year PhD Student from Imagine-ENPC/IGN/CNES Working on Self-supervised Cross-modal Geospatial Learning. Personal WebPage: https://gastruc.github.io/
Super interesting to see pure SSL outperforms text alignement on a super competitive but text-aligned suited task π€―
18.08.2025 15:44 β π 2 π 0 π¬ 0 π 0π°οΈ At #CVPR2025 presenting "AnySat: An Earth Observation Model for Any Resolutions, Scales, and Modalities" - Saturday afternoon, Poster 355!
If you're here and want to discuss geolocation or geospatial foundation models, let's connect!
π’ FLAIR-HUB dataset
A new large-scale, multimodal dataset for land cover and crop type mapping
π€ Dataset: huggingface.co/datasets/IGN...
π Preprint: arxiv.org/abs/2506.07080
π€ Pretrained models: huggingface.co/collections/...
π» Code: github.com/IGNF/FLAIR-HUB
π Project : arxiv.org/abs/2506.07080
I will be presenting our work on the detection of archaeological looting with satellite image time series at CVPR 2025 EarthVision workshop tomorrow!
Honored and grateful that this paper received the best student paper award!
π’ New preprint!
βWhen majority rules, minority loses: bias amplification of gradient descentβ
We often blame biased data but training also amplifies biases. Our paper explores how ML algorithms favor stereotypes at the expense of minority groups.
β‘οΈ arxiv.org/abs/2505.13122
(1/3)
We've added new experiments demonstrating robust generalization capabilities! Notably, AnySat shows strong performance on HLS Burn Scars - a sensor never seen during pretraining! π₯π°οΈ
Check it out:
π Paper: arxiv.org/abs/2412.14123
π Project: gastruc.github.io/anysat
Looking forward to #CVPR2025! We will present the following papers:
30.04.2025 13:04 β π 28 π 7 π¬ 1 π 1Introducing HySCDG #CVPR2025, a generative pipeline for creating a large hybrid semantic change detection for Earth Observation using Stable Diffusion and ControlNet ! πΊοΈπ©οΈ
π Paper: arxiv.org/abs/2503.15683
π»We've released the code for our #CVPR2025 paper MAtCha!
π΅MAtCha reconstructs sharp, accurate and scalable meshes of both foreground AND background from just a few unposed images (eg 3 to 10 images)...
...While also working with dense-view datasets (hundreds of images)!
π₯π₯π₯ CV Folks, I have some news! We're organizing a 1-day meeting in center Paris on June 6th before CVPR called CVPR@Paris (similar as NeurIPS@Paris) π₯πΎπ₯π·
Registration is open (it's free) with priority given to authors of accepted papers: cvprinparis.github.io/CVPR2025InPa...
Big π§΅π with details!
Starter pack including some of the lab members: go.bsky.app/QK8j87w
14.03.2025 10:34 β π 24 π 11 π¬ 0 π 1π§© Excited to share our paper "RUBIK: A Structured Benchmark for Image Matching across Geometric Challenges" (arxiv.org/abs/2502.19955) accepted to #CVPR2025! We created a benchmark that systematically evaluates image matching methods across well-defined geometric difficulty levels. π
28.02.2025 15:23 β π 19 π 7 π¬ 2 π 0Weights for CAD are finally available. It's one of the smallest diffusion models on the market, achieving performance close to SD and Pixart, featuring a Perceiver-like architecture.
We leverage our coherence aware training to improve the textual understanding
π Check it out:
π Paper: arxiv.org/abs/2412.14123
π Project: gastruc.github.io/anysat
π€ HuggingFace: huggingface.co/g-astruc/Any...
π GitHub: github.com/gastruc/AnySat
π Even better: AnySat supports linear probing for semantic segmentation!
That means you can fine-tune just a few thousand parameters and achieve SOTA results on challenging tasksβall with minimal effort.
AnySat achieves SOTA performance on 6 tasks across 10 datasets:
π± Land cover mapping
πΎ Crop type segmentation
π³ Tree species classification
π Flood detection
π Change detection
We trained AnySat on 5 multimodal datasets simultaneously:
π‘ 11 distinct sensors
π Resolutions: 0.2mβ500m
π Revisit: single date to weekly
ποΈ Scales: 0.3β150 hectares
The pretrained model can adapt to truly diverse data, and probably yours too!
πThanks to our modified JEPA training scheme and scale-adaptive spatial encoders, AnySat trains on datasets with diverse scales, resolutions, and modalities!
π§ 75% of its parameters are shared across all inputs, enabling unmatched flexibility.
π€ What if embedding multimodal EO data was as easy as using a ResNet on images?
Introducing AnySat: one model for any resolution (0.2mβ250m), scale (0.3β2600 hectares), and modalities (choose from 11 sensors & time series)!
Try it with just a few lines of code:
Guillaume Astruc, Nicolas Gonthier, Clement Mallet, Loic Landrieu
AnySat: An Earth Observation Model for Any Resolutions, Scales, and Modalities
https://arxiv.org/abs/2412.14123
β οΈReconstructing sharp 3D meshes from a few unposed images is a hard and ambiguous problem.
βοΈWith MAtCha, we leverage a pretrained depth model to recover sharp meshes from sparse views including both foreground and background, within mins!π§΅
πWebpage: anttwo.github.io/matcha/
π Guessing where an image was taken is a hard, and often ambiguous problem. Introducing diffusion-based geolocationβwe predict global locations by refining random guesses into trajectories across the Earth's surface!
πΊοΈ Paper, code, and demo: nicolas-dufour.github.io/plonk
Hi, I am a PhD student from @imagineenpc.bsky.social. Could you also add us both please?
25.11.2024 15:55 β π 4 π 0 π¬ 1 π 0