1. Acorn
- Author
-
Gordon Wetzstein, Marco Monteiro, Julien N. P. Martel, David B. Lindell, Eric R. Chan, and Connor Z. Lin
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,010103 numerical & computational mathematics ,02 engineering and technology ,01 natural sciences ,Machine Learning (cs.LG) ,Rendering (computer graphics) ,Octree ,Computer Science - Graphics ,0202 electrical engineering, electronic engineering, information engineering ,Quadtree ,Computer vision ,Polygon mesh ,0101 mathematics ,Block (data storage) ,Network architecture ,business.industry ,020207 software engineering ,Computer Graphics and Computer-Aided Design ,Graphics (cs.GR) ,Artificial intelligence ,Geometric modeling ,business ,Encoder - Abstract
Neural representations have emerged as a new paradigm for applications in rendering, imaging, geometric modeling, and simulation. Compared to traditional representations such as meshes, point clouds, or volumes they can be flexibly incorporated into differentiable learning-based pipelines. While recent improvements to neural representations now make it possible to represent signals with fine details at moderate resolutions (e.g., for images and 3D shapes), adequately representing large-scale or complex scenes has proven a challenge. Current neural representations fail to accurately represent images at resolutions greater than a megapixel or 3D scenes with more than a few hundred thousand polygons. Here, we introduce a new hybrid implicit-explicit network architecture and training strategy that adaptively allocates resources during training and inference based on the local complexity of a signal of interest. Our approach uses a multiscale block-coordinate decomposition, similar to a quadtree or octree, that is optimized during training. The network architecture operates in two stages: using the bulk of the network parameters, a coordinate encoder generates a feature grid in a single forward pass. Then, hundreds or thousands of samples within each block can be efficiently evaluated using a lightweight feature decoder. With this hybrid implicit-explicit network architecture, we demonstrate the first experiments that fit gigapixel images to nearly 40 dB peak signal-to-noise ratio. Notably this represents an increase in scale of over 1000x compared to the resolution of previously demonstrated image-fitting experiments. Moreover, our approach is able to represent 3D shapes significantly faster and better than previous techniques; it reduces training times from days to hours or minutes and memory requirements by over an order of magnitude., Comment: J. N. P. Martel and D. B. Lindell equally contributed to this work
- Published
- 2021
- Full Text
- View/download PDF