论文标题
J-Plus:支持向量回归以测量恒星参数
J-PLUS: Support Vector Regression to Measure Stellar Parameters
论文作者
论文摘要
语境。恒星参数是恒星研究中最重要的特征之一,这些特征基于传统方法中的大气模型。但是,时间成本和亮度限制了光谱观测的效率。 J-Plus是一个观察活动,旨在获得12个频段中的光度法。由于其特征,J-Plus数据已成为恒星研究的宝贵资源。机器学习提供了有效分析大型数据集的强大工具,例如J-Plus的数据集,并使我们能够将研究域扩展到恒星参数。目标。这项研究的主要目的是在J-Plus观测活动的第一个数据发布中构建SVR算法来估计星星的恒星参数。方法。参数回归的训练数据以J-Plus的12波带光度为特征,并与基于频谱的目录进行了交叉识别。这些目录来自Lamost,Apogee和Segue。然后,我们将其标记为恒星有效温度,表面重力和金属性。持有百分之十的样品进行盲测。我们开发了一种新方法,一种多模型方法,以便完全考虑大小和恒星参数的不确定性。该方法利用200多个模型应用不确定性分析。结果。我们提出了一个目录为2,493,424颗恒星,有效温度回归中均方根误差为160K,表面重力回归中为0.35,金属性回归中的均方根误差为0.25。我们还讨论了这种多模型方法的优势,并将其与其他机器学习方法进行比较。
Context. Stellar parameters are among the most important characteristics in studies of stars, which are based on atmosphere models in traditional methods. However, time cost and brightness limits restrain the efficiency of spectral observations. The J-PLUS is an observational campaign that aims to obtain photometry in 12 bands. Owing to its characteristics, J-PLUS data have become a valuable resource for studies of stars. Machine learning provides powerful tools to efficiently analyse large data sets, such as the one from J-PLUS, and enable us to expand the research domain to stellar parameters. Aims. The main goal of this study is to construct a SVR algorithm to estimate stellar parameters of the stars in the first data release of the J-PLUS observational campaign. Methods. The training data for the parameters regressions is featured with 12-waveband photometry from J-PLUS, and is cross-identified with spectrum-based catalogs. These catalogs are from the LAMOST, the APOGEE, and the SEGUE. We then label them with the stellar effective temperature, the surface gravity and the metallicity. Ten percent of the sample is held out to apply a blind test. We develop a new method, a multi-model approach in order to fully take into account the uncertainties of both the magnitudes and stellar parameters. The method utilizes more than two hundred models to apply the uncertainty analysis. Results. We present a catalog of 2,493,424 stars with the Root Mean Square Error of 160K in the effective temperature regression, 0.35 in the surface gravity regression and 0.25 in the metallicity regression. We also discuss the advantages of this multi-model approach and compare it to other machine-learning methods.