異なる望遠鏡データを処理する恒星データAIモデルを開発(Chinese Researchers Develop AI Model to Process Stellar Data from Different Telescopes)

2026-02-26 中国科学院(CAS)

中国科学院国家天文台や中国科学院大学などの研究チームは、異なる望遠鏡由来の恒星スペクトルを統合解析できるAIモデル「SpecCLIP」を開発した。LAMOSTと欧州のGaia衛星は分解能や波長域が異なり、従来は直接比較が困難だった。研究チームは大規模言語モデルに着想を得た対照学習を導入し、異種データ間の内在的関係を自律的に学習する枠組みを構築。恒星大気パラメータや元素組成の同時推定、スペクトル類似検索、特異天体の同定を可能にした。銀河考古学や系外惑星探査における大規模データ解析の効率化に貢献する。

<関連情報>

SpecCLIP: 星の分光測定のアラインメントと変換 SpecCLIP: Aligning and Translating Spectroscopic Measurements for Stars

Xiaosheng Zhao, Yang Huang, Guirong Xue, Xiao Kong, Jifeng Liu, Xiaoyu Tang, Timothy C. Beers, Yuan-Sen Ting, and A-Li Luo
The Astrophysical Journal  Published: 2026 February 11
DOI:10.3847/1538-4357/ae2c7e

異なる望遠鏡データを処理する恒星データAIモデルを開発(Chinese Researchers Develop AI Model to Process Stellar Data from Different Telescopes)

Abstract

In recent years, large language models (LLMs) have transformed natural language understanding through vast data sets and large-scale parameterization. Inspired by this success, we present SpecCLIP, a foundation model framework that extends LLM-inspired methodologies to stellar spectral analysis. Stellar spectra, akin to structured language, encode rich physical and chemical information about stars. By training foundation models on large-scale spectral data sets, our goal is to learn robust and informative embeddings that support diverse downstream applications. As a proof of concept, SpecCLIP involves pretraining on two spectral types—LAMOST low-resolution and Gaia XP—followed by contrastive alignment using the Contrastive Language–Image Pretraining (CLIP) framework, adapted to associate spectra from different instruments. This alignment is complemented by auxiliary decoders that preserve spectrum-specific information and enable translation (prediction) between spectral types, the former being achieved by maximizing mutual information between embeddings and input spectra. The result is a cross-spectrum framework that enables intrinsic calibration and flexible applications across instruments. We demonstrate that fine-tuning these models on moderate-sized labeled data sets improves adaptability to tasks such as stellar-parameter estimation and chemical-abundance determination. SpecCLIP also enhances the accuracy and precision of parameter estimates benchmarked against external survey data. In addition, its similarity search and cross-spectrum prediction capabilities offer potential for anomaly detection. Our results suggest that contrastively trained foundation models enriched with spectrum-aware decoders can advance precision stellar spectroscopy. Our code SpecCLIP is publicly available on GitHub ✎.

1603情報システム・データ工学
ad
ad
Follow
ad
タイトルとURLをコピーしました