A Generalist Model Including Evolved Star Mass and Age
Determining precise stellar ages and masses for evolved giants is crucial for Galactic archaeology but challenged by spectral degeneracies. Gaia’s low-resolution XP spectra offer a unique opportunity to infer these parameters on a massive scale using data-driven methods. We extend a transformer-based astronomical foundation model to evolved stars, establishing a unified framework to simultaneously predict atmospheric parameters ($T_{\mathrm{eff}}$, $\log g$, $[\mathrm{M}/\mathrm{H}]$) and evolutionary labels (mass, age) with physical consistency. Treating spectra as token sequences, we integrated mass and age into the model’s vocabulary. The model is trained on Gaia XP spectra cross-matched with the APOGEE DR17 DistMass catalog. Our generative approach enables flexible input handling, including spectral inpainting and parameter-to-spectrum generation. On an independent test set, the model achieves a prediction scatter of $σ\approx 0.114 , M_{\odot}$ for mass and $σ\approx 1.334$ Gyr for age. Beyond numerical accuracy, it successfully reproduces the giant branch’s mass-luminosity relation and autonomously disentangles interstellar extinction from intrinsic temperature variations without explicit physical priors. It also robustly recovers missing spectral data and estimates reliable uncertainties. Validating that foundation models can internalize stellar physics from data, this physically-aware, probabilistic framework offers a powerful tool for unraveling Milky Way history using large-scale spectroscopic surveys.
💡 Research Summary
This paper presents a novel, data‑driven framework for estimating stellar masses and ages of evolved giant stars on a Galactic scale, leveraging the low‑resolution BP/RP (XP) spectra from Gaia and high‑precision labels from the APOGEE DR17 DistMass catalog. Recognizing the limitations of traditional discriminative machine‑learning models—fixed‑size inputs, poor handling of missing data, and difficulty capturing subtle evolutionary signatures—the authors extend a transformer‑based “foundation model” originally introduced by Leung & Bovy (2024).
The training set comprises 319,585 stars after quality cuts (mass 0.5–4 M⊙, age 0–14 Gyr, metallicity –2.5 to +0.7 dex). For each star, a 120‑dimensional vector is assembled: 110 normalized Gaia XP coefficients, broadband colors, APOGEE atmospheric parameters (T_eff, log g,
Comments & Academic Discussion
Loading comments...
Leave a Comment