Abstract
Taking inspiration from natural language embeddings, we present ASTROMER, a transformer-based model to create representations of light curves. ASTROMER was pre-trained in a self-supervised manner, requiring no human-labeled data. We used millions of R-band light sequences to adjust the ASTROMER weights. The learned representation can be easily adapted to other surveys by re-training ASTROMER on new sources. The power of ASTROMER consists in using the representation to extract light curve embeddings that can enhance the training of other models, such as classifiers or regressors. As an example, we used ASTROMER embeddings to train two neural-based classifiers that use labeled variable stars from MACHO, OGLE-III, and ATLAS. In all experiments, ASTROMER-based classifiers outperformed a baseline recurrent neural network trained on light curves directly when limited labeled data were available. Furthermore, using ASTROMER embeddings decreases the computational resources needed while achieving state-of-the-art results. Finally, we provide a Python library that includes all the functionalities employed in this work.
Subject
Space and Planetary Science,Astronomy and Astrophysics
Reference42 articles.
1. Abadi M., Agarwal A., Barham P., et al. 2015, TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, software available from tensorflow.org
2. Calibration of the MACHO Photometry Database
3. The MACHO Project: Microlensing Results from 5.7 Years of Large Magellanic Cloud Observations
4. Alcock C., Allsman R., Alves D., et al. 2003, VizieR Online Data Catalog: II/247
5. Allam Jr T., & McEwen J. D. 2021, ArXiv e-prints [arXiv:2105.06178]
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献