Papers
Topics
Authors
Recent
Search
2000 character limit reached

Compact Neural Graphics Primitives with Learned Hash Probing

Published 28 Dec 2023 in cs.CV and cs.GR | (2312.17241v1)

Abstract: Neural graphics primitives are faster and achieve higher quality when their neural networks are augmented by spatial data structures that hold trainable features arranged in a grid. However, existing feature grids either come with a large memory footprint (dense or factorized grids, trees, and hash tables) or slow performance (index learning and vector quantization). In this paper, we show that a hash table with learned probes has neither disadvantage, resulting in a favorable combination of size and speed. Inference is faster than unprobed hash tables at equal quality while training is only 1.2-2.6x slower, significantly outperforming prior index learning approaches. We arrive at this formulation by casting all feature grids into a common framework: they each correspond to a lookup function that indexes into a table of feature vectors. In this framework, the lookup functions of existing data structures can be combined by simple arithmetic combinations of their indices, resulting in Pareto optimal compression and speed.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (64)
  1. Discrete cosine transform. IEEE transactions on Computers 100, 1 (1974), 90–93.
  2. Nonlinear transform coding. IEEE Journal of Selected Topics in Signal Processing 15, 2 (2020), 339–353.
  3. Variational image compression with a scale hyperprior. In International Conference on Learning Representations. https://openreview.net/forum?id=rkcQFMZRb
  4. State-of-the-art in compressed GPU-based direct volume rendering. In Computer Graphics Forum, Vol. 33. Wiley Online Library, 77–100.
  5. Rendering from compressed textures. In Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques. 373–378.
  6. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432 (2013).
  7. 3D Scene Compression through Entropy Penalized Neural Representation Functions. In 2021 Picture Coding Symposium (PCS). 1–5. https://doi.org/10.1109/PCS50896.2021.9477505
  8. Deep local shapes: Learning local SDF priors for detailed 3D reconstruction. In ECCV. Springer, 608–625.
  9. Efficient Geometry-aware 3D Generative Adversarial Networks. In CVPR.
  10. TensoRF: Tensorial Radiance Fields. In European Conference on Computer Vision (ECCV).
  11. Ricardo L. De Queiroz and Philip A. Chou. 2016. Compression of 3D point clouds using a region-adaptive hierarchical transform. IEEE Transactions on Image Processing 25, 8 (2016), 3947–3956.
  12. COIN: COmpression with Implicit Neural representations. ICLR 2021 Neural Compression Workshop Spotlight, arXiv preprint arXiv:2103.03123 (2021).
  13. COIN++: Neural compression across modalities. Transactions on Machine Learning Research 2022, 11 (2022).
  14. K-Planes: Explicit Radiance Fields in Space, Time, and Appearance. In CVPR.
  15. Plenoxels: Radiance Fields without Neural Networks. In CVPR.
  16. On Quantizing Implicit Neural Representations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 341–350.
  17. Vivek K. Goyal. 2001. Theoretical foundations of transform coding. IEEE Signal Processing Magazine 18, 5 (2001), 9–21.
  18. Robert M. Gray. 1984. Vector quantization. IEEE ASSP Magazine 1, 2 (1984), 4–29.
  19. Robert M. Gray and David L. Neuhoff. 1998. Quantization. IEEE Transactions on Information Theory 44, 6 (1998), 2325–2383.
  20. Alfred Haar. 1909. Zur Theorie der orthogonalen Funktionensysteme. Georg-August-Universität, Göttingen.
  21. David A. Huffman. 1952. A method for the construction of minimum-redundancy codes. Proceedings of the IRE 40, 9 (1952), 1098–1101.
  22. LVAC: Learned Volumetric Attribute Compression for Point Clouds using Coordinate Based Networks. Frontiers in Signal Processing 2 (2022). https://doi.org/10.3389/frsip.2022.1008812
  23. NeuralVDB: High-resolution Sparse Volume Representation using Hierarchical Neural Networks. (2022). https://doi.org/10.48550/arXiv.2208.04448
  24. Donald Knuth. 1963. Notes on “Open” Addressing. https://web.archive.org/web/20160303225949/http://algo.inria.fr/AofA/Research/11-97.html.
  25. On Controllable Sparse Alternatives to Softmax. In Advances in Neural Information Processing Systems, S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Eds.), Vol. 31. Curran Associates, Inc. https://proceedings.neurips.cc/paper_files/paper/2018/file/6a4d5952d4c018a1c1af9fa590a10dda-Paper.pdf
  26. DropMax: Adaptive Variational Softmax. In Advances in Neural Information Processing Systems, S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Eds.), Vol. 31. Curran Associates, Inc. https://proceedings.neurips.cc/paper_files/paper/2018/file/389bc7bb1e1c2a5e7e147703232a88f6-Paper.pdf
  27. Compressing Volumetric Radiance Fields to 1 MB. (June 2023), 4222–4231.
  28. Magic3D: High-Resolution Text-to-3D Content Creation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 300–309.
  29. Bacon: Band-limited Coordinate Networks for Multiscale Scene Representation. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 16231–16241. https://doi.org/10.1109/CVPR52688.2022.01577
  30. Neural Sparse Voxel Fields. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 15651–15663. https://proceedings.neurips.cc/paper_files/paper/2020/file/b4b758962f17808746e9bb832a6fa4b8-Paper.pdf
  31. Compressive Neural Representations of Volumetric Scalar Fields. Computer Graphics Forum 40, 3 (2021), 135–146. https://doi.org/10.1111/cgf.14295
  32. ACORN: Adaptive Coordinate Networks for Neural Representation. ACM Trans. Graph. (SIGGRAPH) (2021).
  33. André F. T. Martins and Ramón F. Astudillo. 2016. From Softmax to Sparsemax: A Sparse Model of Attention and Multi-Label Classification. In Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48 (New York, NY, USA) (ICML’16). JMLR.org, 1614–1623.
  34. NeRF: Representing scenes as neural radiance fields for view synthesis. In European conference on computer vision. Springer, 405–421.
  35. Instant Neural Graphics Primitives with a Multiresolution Hash Encoding. ACM Trans. Graph. 41, 4, Article 102 (July 2022), 15 pages. https://doi.org/10.1145/3528223.3530127
  36. Real-time Neural Radiance Caching for Path Tracing. ACM Trans. Graph. 40, 4, Article 36 (Aug. 2021), 16 pages. https://doi.org/10.1145/3450626.3459812
  37. Ken Museth. 2021. NanoVDB: A GPU-friendly and portable VDB data structure for real-time rendering and simulation. In ACM SIGGRAPH 2021 Talks. 1–2.
  38. OpenVDB. In ACM SIGGRAPH 2019 Courses. 1–56.
  39. Gradient Estimation with Stochastic Softmax Tricks. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 5691–5704. https://proceedings.neurips.cc/paper_files/paper/2020/file/3df80af53dce8435cf9ad6c3e7a403fd-Paper.pdf
  40. Convolutional Occupancy Networks. In European Conference on Computer Vision (ECCV).
  41. Sparse Sequence-to-Sequence Models. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 1504–1519. https://doi.org/10.18653/v1/P19-1146
  42. DreamFusion: Text-to-3D using 2D Diffusion. In The Eleventh International Conference on Learning Representations. https://openreview.net/forum?id=FjNys5c7VyY
  43. Nathan Reed. 2012. Understanding BCn Texture Compression Formats. https://www.reedbeta.com/blog/understanding-bcn-texture-compression-formats/. Online; accessed 24 January 2023.
  44. Masked Wavelet Representation for Compact Neural Radiance Fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 20680–20690.
  45. Radu Alexandru Rosu and Sven Behnke. 2023. PermutoSDF: Fast Multi-View Reconstruction with Implicit Surfaces using Permutohedral Lattices. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
  46. MINER: Multiscale Implicit Neural Representation. In Computer Vision – ECCV 2022, Shai Avidan, Gabriel Brostow, Moustapha Cissé, Giovanni Maria Farinella, and Tal Hassner (Eds.). Springer Nature Switzerland, Cham, 318–333.
  47. Vector regression functions for texture compression. ACM Transactions on Graphics (TOG) 35, 1 (2015), 1–10.
  48. Jacob Ström and Tomas Akenine-Möller. 2005. iPACKMAN: High-quality, low-complexity texture compression for mobile phones. In Proceedings of the ACM SIGGRAPH/EUROGRAPHICS Conference on Graphics Hardware. 63–70.
  49. Implicit neural representations for image compression. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXVI. Springer, 74–91.
  50. Variable bitrate neural fields. In ACM SIGGRAPH 2022 Conference Proceedings. 1–9.
  51. Neural geometric level of detail: Real-time rendering with implicit 3D shapes. In CVPR. 11358–11367.
  52. Kaolin Wisp: A PyTorch library and engine for neural fields research.
  53. Real-time compression and streaming of 4D performances. ACM Transactions on Graphics (TOG) 37, 6 (2018), 1–11.
  54. Deep implicit volume compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1293–1303.
  55. Optimized Spatial Hashing for Collision Detection of Deformable Objects. In Proceedings of VMV’03, Munich, Germany. 47–54.
  56. Lossy Image Compression with Compressive Autoencoders. In International Conference on Learning Representations. https://openreview.net/forum?id=rJiNwv9gg
  57. Random-Access Neural Compression of Material Textures. In Proceedings of SIGGRAPH.
  58. Neural discrete representation learning. Advances in neural information processing systems 30 (2017).
  59. Gregory K. Wallace. 1992. The JPEG still picture compression standard. IEEE transactions on consumer electronics 38, 1 (1992), xviii–xxxiv.
  60. Lossy Point Cloud Geometry Compression via End-to-End Learning. IEEE Transactions on Circuits and Systems for Video Technology 31, 12 (2021), 4909–4923. https://doi.org/10.1109/TCSVT.2021.3051377
  61. Li-Yi Wei and Marc Levoy. 2000. Fast texture synthesis using tree-structured vector quantization. In Proceedings of the 27th annual conference on Computer graphics and interactive techniques. 479–488.
  62. Neural Fields in Visual Computing and Beyond. Computer Graphics Forum 41, 2 (2022), 641–676. https://doi.org/10.1111/cgf.14505
  63. PlenOctrees for Real-time Rendering of Neural Radiance Fields. In ICCV.
  64. Loss Functions for Image Restoration With Neural Networks. IEEE Transactions on Computational Imaging PP (12 2016), 1–1. https://doi.org/10.1109/TCI.2016.2644865
Citations (15)

Summary

  • The paper demonstrates that Compact NGP significantly improves storage efficiency and inference speed by blending hash table lookups with learned probing.
  • The method achieves a competitive quality-size trade-off on image datasets, markedly reducing memory footprints and computational costs.
  • The technique paves the way for advanced multimedia applications, including texture compression in gaming and real-time rendering.

Introduction

The paper introduces an innovative approach to representing and compressing multimedia content across different formats, including images and volume data used in immersive experiences. It seeks to address the ever-growing demand for higher fidelity content, a challenge that current multimedia formats struggle to meet efficiently.

Current Compression Techniques

Lossy compression is a common strategy for reducing multimedia file sizes, generally involving steps such as transform coding, quantization, and entropy coding. However, the methods commonly used are often tailored for specific multimedia types and may not be well-suited for the latest high-dimensional formats such as volumetric video, requiring a more versatile approach.

Neural Graphic Primitives (NGP)

NGPs have emerged as a flexible solution capable of representing various data types and rendering novel views in applications ranging from image generation to light caching. These models typically use trained feature grids composed of latent embeddings, decoded by a neural network. While promising, these feature grids pose challenges due to either their large memory footprint or slow performance.

Compact NGP Methodology

The core contribution of the paper is the Compact Neural Graphics Primitives (Compact NGP), which improves storage efficiency and accelerates data access by blending traditional hash table speed with the compactness of learned probing techniques. This is achieved by managing lookup functions that index into a compact feature codebook, a technique that combines spatial hashing with learning indices. Reducing the feature codebook size and increasing cache utilization, the Compact NGP demonstrates improved performance during both training and inference as well as notable compression rates.

Performance and Results

The methodology proposed shows impactful results when tested against established benchmarks. On image datasets, the Compact NGP exhibits a competitive quality-size trade-off, especially noticeable in high-quality targets where existing methods struggle. Additionally, the storage cost is managed more efficiently as the representation becomes increasingly dominated by integer parameters rather than floating-point ones.

Future Work and Applications

Looking forward, the technique outlined in the paper has potential applications beyond what has been demonstrated. These include, but are not limited to, video game texture compression, real-time renderers handling texture compression, and live-streaming of high-dimensional multimedia formats. The authors suggest that future research into data-adaptive float quantization or integer entropy minimization could yield even more effective compression strategies.

Conclusion

The paper encapsulates the authors' efforts in proposing a versatile compression scheme via Compact NGP that not only addresses the storage and computational challenges of modern multimedia representation but is also capable of adapting to a wide range of multimedia types and quality requirements. This new approach is poised to play a significant role in the development and distribution of next-generation digital content.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 3 likes about this paper.