Back to Search
Start Over
Photometry of high-redshift blended galaxies using deep learning.
- Source :
- Monthly Notices of the Royal Astronomical Society; 01/11/2020, Vol. 491 Issue 2, p2481-2495, 15p
- Publication Year :
- 2020
-
Abstract
- The new generation of deep photometric surveys requires unprecedentedly precise shape and photometry measurements of billions of galaxies to achieve their main science goals. At such depths, one major limiting factor is the blending of galaxies due to line-of-sight projection, with an expected fraction of blended galaxies of up to 50 per cent. This proof-of-concept work explores for the first time the use of deep neural networks to estimate the photometry of blended pairs of galaxies in space-based monochrome images similar to the ones that will be delivered by the Euclid space telescope under simplified idealized conditions. Using a clean sample of isolated galaxies from the CANDELS survey, we artificially blend them and train two different network models to recover the photometry of the two galaxies. We show that our approach can recover the original photometry of the galaxies before being blended with |$\sim 7{{\ \rm per\ cent}}$| mean absolute percentage error on flux estimations without any human intervention and without any assumption on the galaxy shape. This represents an improvement of at least a factor of 4 compared to the classical  SExtractor approach. We also show that, forcing the network to simultaneously estimate fractional segmentation maps results in a slightly improved photometry. All data products and codes have been made public to ease the comparison with other approaches on a common data set. See https://github.com/aboucaud/coindeblend. [ABSTRACT FROM AUTHOR]
- Subjects :
- DEEP learning
PHOTOMETRY
GALAXIES
SPACE telescopes
SHAPE measurement
Subjects
Details
- Language :
- English
- ISSN :
- 00358711
- Volume :
- 491
- Issue :
- 2
- Database :
- Complementary Index
- Journal :
- Monthly Notices of the Royal Astronomical Society
- Publication Type :
- Academic Journal
- Accession number :
- 140823257
- Full Text :
- https://doi.org/10.1093/mnras/stz3056