Foundation Models in Astronomy: Why They Matter

Feb 10, 2025·
Matthieu Le Lain
Matthieu Le Lain
· 1 min read
blog

Foundation models — large neural networks pre-trained on massive datasets — are starting to transform astronomical data analysis. Models like DINOv2 or CLIP, originally trained on natural images, can be fine-tuned for astronomical tasks with surprisingly good results. At EAS 2024, I presented early comparisons of these models on galaxy morphological classification, and my GRETSI 2025 paper digs into how to best specialize DINOv2 for astronomy.

The key challenge is the domain gap: astronomical images (multi-band, high dynamic range, specific noise) look nothing like everyday photos. Choosing the right fine-tuning strategy turns out to matter a lot — and is the core focus of my current work.

On the multimodal side, I am involved in the UniverseTBD collaboration, where I contribute to AstroLLaVA — a vision-language model for astronomy presented at ICLR 2025.

Matthieu Le Lain
Authors
AI for astronomy & astrophysics
PhD student at IRISA, Université Bretagne Sud (expected 2026), and lecturer in computer science, working on foundation models for astronomy and astrophysics, with a broader interest in deep learning applied to scientific data. Also contributing to the UniverseTBD collaboration on vision–language models for astronomy.