1. Training a neural network to learn other dimensionality reduction removes data size restrictions in bioinformatics and provides a new route to exploring data representations
- Author
-
Ian S. Gilmore, Piotr Grabowski, Mariia Yuneva, Arafath Kaja Najumudeen, Bin Yan, Panina Y, Teresa Murta, Andrew D. Campbell, Chelsea J. Nikula, Simon T. Barry, Alex Dexter, Owen J. Sansom, Kenneth N. Robinson, Richard J. A. Goodwin, Josephine Bunch, Spencer A. Thomas, John G. Swales, Adam J. Taylor, Zoltan Takats, Efstathios A Elia, Gregory Hamm, and Rory T. Steven
- Subjects
Artificial neural network ,Computer science ,business.industry ,Dimensionality reduction ,Deep learning ,Feature extraction ,Hyperspectral imaging ,Sample (statistics) ,Machine learning ,computer.software_genre ,Embedding ,Artificial intelligence ,business ,computer ,Sentence - Abstract
High dimensionality omics and hyperspectral imaging datasets present difficult challenges for feature extraction and data mining due to huge numbers of features that cannot be simultaneously examined. The sample numbers and variables of these methods are constantly growing as new technologies are developed, and computational analysis needs to evolve to keep up with growing demand. Current state of the art algorithms can handle some routine datasets but struggle when datasets grow above a certain size. We present a training deep learning via neural networks on non-linear dimensionality reduction, in particular t-distributed stochastic neighbour embedding (t-SNE), to overcome prior limitations of these methods.One Sentence SummaryAnalysis of prohibitively large datasets by combining deep learning via neural networks with non-linear dimensionality reduction.
- Published
- 2020