DOE PAGES title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: AstroCLIP: a cross-modal foundation model for galaxies

Journal Article · · Monthly Notices of the Royal Astronomical Society

ABSTRACT We present AstroCLIP, a single, versatile model that can embed both galaxy images and spectra into a shared, physically meaningful latent space. These embeddings can then be used – without any model fine-tuning – for a variety of downstream tasks including (1) accurate in-modality and cross-modality semantic similarity search, (2) photometric redshift estimation, (3) galaxy property estimation from both images and spectra, and (4) morphology classification. Our approach to implementing AstroCLIP consists of two parts. First, we embed galaxy images and spectra separately by pre-training separate transformer-based image and spectrum encoders in self-supervised settings. We then align the encoders using a contrastive loss. We apply our method to spectra from the Dark Energy Spectroscopic Instrument and images from its corresponding Legacy Imaging Survey. Overall, we find remarkable performance on all downstream tasks, even relative to supervised baselines. For example, for a task like photometric redshift prediction, we find similar performance to a specifically trained ResNet18, and for additional tasks like physical property estimation (stellar mass, age, metallicity, and specific-star-formation rate), we beat this supervised baseline by 19 per cent in terms of R2. We also compare our results with a state-of-the-art self-supervised single-modal model for galaxy images, and find that our approach outperforms this benchmark by roughly a factor of two on photometric redshift estimation and physical property prediction in terms of R2, while remaining roughly in-line in terms of morphology classification. Ultimately, our approach represents the first cross-modal self-supervised model for galaxies, and the first self-supervised transformer-based architectures for galaxy images and spectra.

Sponsoring Organization:
USDOE
Grant/Contract Number:
AC02-05CH11231
OSTI ID:
2377235
Journal Information:
Monthly Notices of the Royal Astronomical Society, Journal Name: Monthly Notices of the Royal Astronomical Society Journal Issue: 4 Vol. 531; ISSN 0035-8711
Publisher:
Oxford University PressCopyright Statement
Country of Publication:
United Kingdom
Language:
English

References (36)

The Dawes Review 10: The impact of deep learning for the analysis of galaxy surveys journal January 2023
Photometric redshifts from SDSS images using a convolutional neural network journal December 2018
Identifying strong lenses with unsupervised machine learning using convolutional autoencoder journal April 2020
Detecting outliers in astronomical images with deep generative networks journal June 2020
Galaxy Zoo DECaLS: Detailed visual morphology measurements from volunteers and deep learning for 314 000 galaxies journal September 2021
Model compression
  • Buciluǎ, Cristian; Caruana, Rich; Niculescu-Mizil, Alexandru
  • Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '06 https://doi.org/10.1145/1150402.1150464
conference January 2006
Overview of the DESI Legacy Imaging Surveys journal April 2019
Dimensionality Reduction of SDSS Spectra with Variational Autoencoders journal June 2020
Mapping the Diversity of Galaxy Spectra with Deep Unsupervised Machine Learning journal January 2022
The DESI Bright Galaxy Survey: Final Target Selection, Design, and Validation journal May 2023
Autoencoding Galaxy Spectra. I. Architecture journal July 2023
Autoencoding Galaxy Spectra. II. Redshift Invariance and Outlier Detection journal July 2023
LSST: From Science Drivers to Reference Design and Anticipated Data Products journal March 2019
The DESI PRObabilistic Value-added Bright Galaxy Survey (PROVABGS) Mock Challenge journal March 2023
Self-supervised Representation Learning for Astronomical Images journal April 2021
Outlier Detection in the DESI Bright Galaxy Survey journal October 2023
Neural Machine Translation by Jointly Learning to Align and Translate preprint January 2014
Density estimation using Real NVP preprint January 2016
Spreading vectors for similarity search preprint January 2018
Representation Learning with Contrastive Predictive Coding preprint January 2018
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding preprint January 2018
Understanding the Limitations of Variational Mutual Information Estimators preprint January 2019
A Simple Framework for Contrastive Learning of Visual Representations preprint January 2020
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale preprint January 2020
Estimating Galactic Distances From Images Using Self-supervised Representation Learning preprint January 2021
SEED: Self-supervised Distillation For Visual Representation preprint January 2021
On the Opportunities and Risks of Foundation Models preprint January 2021
Self-supervised similarity search for large scientific datasets preprint January 2021
Masked Autoencoders Are Scalable Vision Learners preprint January 2021
iBOT: Image BERT Pre-Training with Online Tokenizer preprint January 2021
ClimaX: A foundation model for weather and climate preprint January 2023
EVA-CLIP: Improved Training Techniques for CLIP at Scale preprint January 2023
DINOv2: Learning Robust Visual Features without Supervision preprint January 2023
A Cookbook of Self-Supervised Learning preprint January 2023
The Early Data Release of the Dark Energy Spectroscopic Instrument text January 2023
Multiple Physics Pretraining for Physical Surrogate Models preprint January 2023

Similar Records

Related Subjects