Papers
Topics
Authors
Recent
Search
2000 character limit reached

NeuSG Framework: 3D Surface Reconstruction

Updated 6 February 2026
  • NeuSG Framework is a 3D surface reconstruction method that combines neural implicit SDFs with explicit 3D Gaussian splatting to recover highly detailed surfaces from multi-view images.
  • The joint optimization leverages photometric, scale, and normal alignment losses to mutually refine both the SDF network and Gaussian parameters for enhanced accuracy.
  • This approach overcomes the limitations of over-smoothed depth maps and sparse point clouds, delivering state-of-the-art performance with improved computational efficiency.

The NeuSG framework is a 3D surface reconstruction methodology that integrates neural implicit @@@@1@@@@ (SDF)-based surfaces with explicit 3D Gaussian Splatting, establishing a mutual optimization pipeline that enables the recovery of highly detailed surfaces from calibrated multi-view images. The framework exploits the dense geometric priors provided by Gaussian Splatting to regularize the implicit neural surface, while simultaneously leveraging the predicted surface normals of the implicit model to refine the explicit Gaussians. This approach enables NeuSG to overcome the limitations of over-smoothed depth maps and sparse point clouds in previous multi-view reconstruction methods, achieving state-of-the-art (SOTA) performance with improved computational efficiency (Chen et al., 2023).

1. NeuSG System Overview

NeuSG is designed to recover a complete, high-fidelity 3D surface from multi-view RGB images, combining the geometrical strength of neural implicit SDFs (NeuS) with the structural detail of 3D Gaussian Splatting. The core pipeline consists of three primary phases:

  1. 3D Gaussian Splatting: Standard differentiable splatting is performed on the multi-view images, projecting them into a dense set of 3D Gaussians. Each Gaussian is parameterized by center pip_i, covariance Σi\Sigma_i, spherical-harmonic color coefficients HiH_i, and opacity αi\alpha_i.
  2. Implicit SDF Network Construction: A neural implicit SDF f(x;θ)f(\mathbf{x};\theta), parameterized by a multi-layer perceptron (MLP), is erected to define the surface as the zero-level set, along with a radiance/color predictor c(x,d;θn)c(\mathbf{x}, \mathbf{d};\theta_n).
  3. Joint Optimization: Parameters of both the SDF network and the Gaussians are refined together using a shared photometric loss and two new regularizers: a scale regularizer which flattens Gaussians onto the surface and a normal alignment term to align their principal axes with local surface normals.

By alternating or interleaving optimization steps for SDF parameters (θ\theta) and Gaussian parameters (pi,si,ri,Hi,αip_i, s_i, r_i, H_i, \alpha_i), NeuSG establishes a bi-directional refinement: the splatted Gaussians feed dense structural guidance into the SDF network, while the SDF-derived normals improve Gaussian localization.

2. Mathematical Formulation

2.1 Neural Implicit Surface (NeuS)

A neural implicit surface is modeled as an SDF f:R3Rf: \mathbb{R}^3 \to \mathbb{R}, with the surface defined by {xf(x)=0}\{ \mathbf{x}\mid f(\mathbf{x})=0 \}. The function f(x;θ)f(\mathbf{x};\theta) is represented by an MLP, potentially using hash encoding for efficiency. Volume rendering for view synthesis samples points xi=o+tidx_i = o + t_i d along each camera ray, with opacity computed as:

Φs(u)=11+exp(u/s),αi=max(Φs(f(xi))Φs(f(xi+1))Φs(f(xi)),0)\Phi_s(u) = \frac{1}{1 + \exp(-u/s)}, \quad \alpha_i = \max\left( \frac{\Phi_s(f(x_i)) - \Phi_s(f(x_{i+1}))}{\Phi_s(f(x_i))}, 0 \right)

The synthesized color is:

C^(o,d)=i=1Nwic(xi,d),wi=Tiαi,Ti=j=1i1(1αj)\hat{C}(o,d) = \sum_{i=1}^N w_i c(x_i, d), \quad w_i = T_i \alpha_i, \quad T_i = \prod_{j=1}^{i-1} (1 - \alpha_j)

2.2 3D Gaussian Splatting

Each Gaussian is specified as:

  • Center piR3p_i \in \mathbb{R}^3
  • Covariance Σi=RiSiSiTRiT\Sigma_i = R_i S_i S_i^T R_i^T, where scaling matrix Si=diag(si1,si2,si3)S_i = \operatorname{diag}(s_{i1}, s_{i2}, s_{i3}), rotation RiR_i (from quaternion rir_i)
  • Spherical harmonics color HiH_i, opacity αi\alpha_i

Differentiable splatting provides rendered color/opacity for each view, with rendering loss analogous to the SDF photometric loss.

2.3 Loss Function Components

Loss Term Mathematical Formulation Purpose
Photometric loss LRGB=raysCC^1L_{RGB} = \sum_{rays} \| C - \hat{C}\|_1 Color/radiance fidelity to ground-truth images
Eikonal regularizer Leik=1Ni=1N(f(xi)21)2L_{eik} = \frac{1}{N}\sum_{i=1}^N ( \| \nabla f(x_i) \|_2 - 1 )^2 Gradient normalization of SDF
Point cloud prior Lpt=if(pi)1L_{pt} = \sum_i | f(p_i) |_1 Draws Gaussians’ centers to the true surface
Scale regularizer Lscale=imin(si1,si2,si3)1L_{scale} = \sum_i | \min(s_{i1},s_{i2},s_{i3}) |_1 Flattens Gaussian ellipsoids on the surface
Normal alignment Lalign=i1nw,if(pi)1L_{align} = \sum_i | 1 - | n_{w,i}^\top \nabla f(p_i) | |_1 Aligns thin axis of Gaussian with surface normal

The total optimization objective is:

LNeuSG=LRGB+λ1Leik+λ2Lpt+λ3Lscale+λ4LalignL_{NeuSG} = L_{RGB} + \lambda_1 L_{eik} + \lambda_2 L_{pt} + \lambda_3 L_{scale} + \lambda_4 L_{align}

with recommended hyperparameters λ1=0.1\lambda_1=0.1, λ2=1.0\lambda_2=1.0, λ3=100\lambda_3=100, λ4=1.0\lambda_4=1.0 (Chen et al., 2023).

3. 3D Gaussian Splatting Guidance and Mutual Refinement

NeuSG’s innovation is in the iterative mutual refinement strategy:

  • Gaussian Generation: Initial Gaussians are fit via standard differentiable 3D Splatting (Kerbl et al., 2023), producing a dense point cloud whose centers do not strictly coincide with the object’s surface.
  • Flattening and Alignment: The scale regularizer (LscaleL_{scale}) pushes each ellipsoid to have a negligible axis, coercing its center to align with the SDF zero-crossing. Its principal (thin) axis, extracted as nc,i=OneHot(argmin(si1,si2,si3))n_{c,i} = \mathrm{OneHot}(\arg\min (s_{i1}, s_{i2}, s_{i3})), is rotated to world space (nw,i=Rinc,in_{w,i} = R_i n_{c,i}) and aligned to the SDF-derived normal via LalignL_{align}.
  • Parameter Updates: Gaussian parameters (scale, rotation) are updated by back-propagating LscaleL_{scale} and LalignL_{align}. Centers are moved through the photometric loss and LptL_{pt}. The color/shading coefficients remain under LRGBL_{RGB} supervision.

This dual strategy ensures that the Gaussian set remains surface-faithful and that their implied normals denoise and enhance surface geometry, feeding back to the SDF for subsequent training rounds.

4. Training Protocols and Computational Details

  • Preprocessing: Given NN calibrated RGB images with camera poses, the pipeline first generates the initial Gaussian set and randomly initializes the SDF+color MLPs (hash encoding optional).
  • Optimization Schedule: Uses Adam optimizer (lr=103\text{lr}=10^{-3} for both MLP and Gaussian parameters). Training comprises 500k SDF steps with, every 100k, a 30k-step Gaussian-centric mini-optimization focused on Lscale+LalignL_{scale}+L_{align} regularization. Each iteration samples 1024 rays.
  • Hardware: Employs a single NVIDIA RTX 4090 GPU, requiring approximately 16 hours for end-to-end optimization (significantly more efficient than prior SOTA approaches).
  • This regimen allows SDF and Gaussian parameters to be updated under the shared, multi-component LNeuSGL_{NeuSG} objective, realizing the mutual guidance principle.

5. Experimental Validation and Ablation Analysis

Quantitative Benchmarks

NeuSG was evaluated on the Tanks & Temples benchmark, providing F1-scores:

Method Mean F1 Score Training Time Hash Size
NeuS (baseline) 0.38 - -
Geo-Neus / MonoSDF / RegSDF 0.35–0.39 - -
NeuralAngelo 0.50 128 hours 2222^{22}
NeuSG 0.49 16 hours 2192^{19}

NeuSG matches or exceeds SOTA while using 8×\sim 8\times fewer hash-encoding parameters and 8×\sim 8\times less training time (Chen et al., 2023).

Qualitative Assessments

NeuSG reconstructions exhibit substantially fewer geometric holes (e.g., in roofs and freestanding objects) and recover fine structural detail (panel seams, railings) that prior approaches routinely miss.

Ablation Results

Ablative experiments demonstrate:

  • Raw Gaussian point cloud guidance alone can lower performance due to noise (F1=0.59F1=0.59).
  • Introducing only the scale regularizer yields F1=0.69F1=0.69.
  • Full scale and normal alignment brings mean F1F1 to $0.73$, with regularizers converging rapidly (Lscale108L_{scale} \to 10^{-8}, Lalign103L_{align} \to 10^{-3}).

6. Contributions, Limitations, and Future Directions

Key Contributions

  1. Mutual Optimization: A bi-directional training loop where neural implicit surfaces and explicit 3D Gaussian splatting refine each other.
  2. Scale Regularization: Enforces thin ellipsoids, automating the process of aligning Gaussians with the true surface.
  3. Normal Alignment: Refines Gaussian orientations using SDF-derived normals, denoising guidance to the SDF.
  4. Computational Efficiency: Demonstrates robust, SOTA-quality reconstruction on large scenes within practical compute budgets.

Limitations

  • The reliance on dense, multi-view coverage renders the method less effective under sparse or single-view setups.
  • Though substantially more efficient than comparators, NeuSG still requires approximately 16 hours of GPU compute per scene.
  • In under-constrained regions with limited photometric or geometric information, both Gaussians and the SDF may drift, potentially producing artifacts.

This suggests avenues for future research could include reducing data requirements, accelerating convergence, or enabling real-time updates.

(Chen et al., 2023)

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to NeuSG Framework.