1. Research on neural processes with multiple latent variables
- Author
-
Xiao‐Han Yu, Shao‐Chen Mao, Lei Wang, Shi‐Jie Lu, and Kun Yu
- Subjects
encoder–decoder ,multiple latent variables ,neural process ,regression ,Photography ,TR1-1050 ,Computer software ,QA76.75-76.765 - Abstract
Abstract Neural Process (NP) fully combines the advantages of neural network and Gaussian Process (GP) to provide an efficient method for solving regression problems. Nonetheless, limited by the dimensionality of the latent variable, NP has difficulty fitting the observed data completely and predicting the targets perfectly. To remedy these drawbacks, the authors propose a concise and effective improvement of the latent path of NP, which the authors term Multi‐Latent Variables Neural Process (MLNP). MLNP samples multiple latent variables and integrates the representations corresponding to the latent variables in the decoder with adaptive weights. MLNP inherits the desirable property of linear computation scales of NP and learns the approximate distribution over objective functions from contexts more flexibly and accurately. By applying MLNP to 1‐D regression, real‐world image completion, which can be seen as a 2‐D regression task, the authors demonstrate its significant improvement in the accuracy of prediction and contexts fitting capability compared with NP. Through ablation experiments, the authors also verify that the number of latent variables has a great impact on the prediction accuracy and fitting capability of MLNP. Moreover, the authors also analyze the roles played by different latent variables in reconstructing images.
- Published
- 2023
- Full Text
- View/download PDF