Anthony Fuller's Avatar

Anthony Fuller

@anthonyfuller.bsky.social

PhD Student at Carleton University (Ottawa, Canada) https://antofuller.github.io/

149 Followers  |  1,465 Following  |  5 Posts  |  Joined: 17.11.2024
Posts Following

Posts by Anthony Fuller (@anthonyfuller.bsky.social)

Thanks for comparing with LookHere. Looking forward to reading this πŸ™‚

04.02.2026 14:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Galileo, created by Mila researchers Gabriel Tseng and David Rolnick, uses AI to uncover trends across decades of satellite and human activity dataβ€”revealing early signals about our planet’s health and helping us act in time.
mila.quebec/en/article/d...

24.10.2025 14:33 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

LookWhere has been accepted to NeurIPS 2025!

LookWhere accelerates inference and fine-tuning by approximating full, deep representations with adaptive computation of predictions learned from distillation.

Paper: arxiv.org/abs/2505.18051

Code and weights: github.com/antofuller/l...

29.09.2025 13:54 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Our remote sensing foundation model Galileo has been accepted to ICML 2025!

Galileo outperforms state-of-the-art across different input data modalities and shapes, and using it requires only minimal data and compute.

More at:
arxiv.org/abs/2502.09356

08.06.2025 20:31 β€” πŸ‘ 24    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Elbows up, Canada.

22.03.2025 22:11 β€” πŸ‘ 52328    πŸ” 13353    πŸ’¬ 1482    πŸ“Œ 2283
Preview
LookHere: Vision Transformers with Directed Attention Generalize and Extrapolate High-resolution images offer more information about scenes that can improve model accuracy. However, the dominant model architecture in computer vision, the vision transformer (ViT), cannot effectivel...

Lookhere position encoding could help with this: arxiv.org/abs/2405.13985

27.02.2025 17:04 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Awesome, thanks for the explanation!

04.02.2025 16:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
DeBERTa: Decoding-enhanced BERT with Disentangled Attention Recent progress in pre-trained neural language models has significantly improved the performance of many natural language processing (NLP) tasks. In this paper we propose a new model architecture DeBE...

Cool work? I think the position encoding method looks similar to DeBERTa’s disentangled attention: arxiv.org/abs/2006.03654

04.02.2025 14:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Writing a good scientific paper

For those who missed this post on the-network-that-is-not-to-be-named, I made public my "secrets" for writing a good CVPR paper (or any scientific paper). I've compiled these tips of many years. It's long but hopefully it helps people write better papers. perceiving-systems.blog/en/post/writ...

20.11.2024 10:18 β€” πŸ‘ 260    πŸ” 64    πŸ’¬ 4    πŸ“Œ 8