Papers
Topics
Authors
Recent
Search
2000 character limit reached

Restricted Boltzmann Machine Overview

Updated 20 January 2026
  • Restricted Boltzmann Machines are stochastic neural networks featuring a visible layer and a hidden layer with pairwise interactions but no intra-layer connections.
  • They are applied to model complex spectral densities and construct data-adaptive kernels for advanced regression tasks.
  • Integration with quantum annealing enables efficient sampling and gradient-based optimization to enhance kernel performance in real-world applications.

A Restricted Boltzmann Machine (RBM) is a stochastic neural network architecture consisting of a visible layer and a hidden layer with undirected, pairwise connections between them and no intra-layer connections. The RBM defines a joint Boltzmann distribution over binary-valued visible units v{1,+1}Nv\mathbf{v} \in \{-1,+1\}^{N_v} and binary-valued hidden units h{1,+1}Nh\mathbf{h} \in \{-1,+1\}^{N_h}, parameterized by biases for the visible (b\mathbf{b}) and hidden (c\mathbf{c}) layers and a weight matrix W\mathbf{W}.

1. Energy Function and Gibbs–Boltzmann Distribution

The RBM assigns an energy function

E(v,h)=i=1Nvbivij=1Nhcjhji,jWijvihjE(\mathbf{v}, \mathbf{h}) = -\sum_{i=1}^{N_v} b_i v_i - \sum_{j=1}^{N_h} c_j h_j - \sum_{i,j} W_{ij} v_i h_j

which defines the joint probability

P(v,hb,c,W)=exp[E(v,h)]ZP(\mathbf{v}, \mathbf{h} \mid \mathbf{b}, \mathbf{c}, \mathbf{W}) = \frac{\exp[-E(\mathbf{v}, \mathbf{h})]}{Z}

where ZZ denotes the partition function over all visible and hidden configurations.

The absence of intra-layer connections (i.e., restricted architecture) renders the conditional distributions P(hv)P(\mathbf{h} \mid \mathbf{v}) and P(vh)P(\mathbf{v} \mid \mathbf{h}) fully factorized, enabling efficient block Gibbs sampling for stochastic learning and inference (Hasegawa et al., 13 Jan 2026).

2. Role in Spectral Density Modeling

In kernel learning frameworks, an RBM can be deployed to model the spectral density p(ω)p(\omega) within Bochner's theorem for shift-invariant, positive-definite kernels: k(Δx)=Eωp(ω)[cos(ωΔx)]k(\Delta \mathbf{x}) = \mathbb{E}_{\omega \sim p(\omega)}[\cos(\omega^{\top} \Delta \mathbf{x})] where Δx=xx\Delta \mathbf{x} = \mathbf{x} - \mathbf{x}'. The RBM leverages its capacity to approximate complex, potentially multimodal distributions by learning p(ω)p(\omega) as an implicit spectral distribution over discrete binary vectors v\mathbf{v}. These are later mapped to continuous frequencies ω\omega through auxiliary conditional models (Hasegawa et al., 13 Jan 2026).

This approach supports the construction of data-adaptive kernels whose spectral properties transcend the limitations of fixed-form parametric distributions (e.g., a single Gaussian).

3. Quantum Annealing-Based Sampling and Frequency Mapping

RBM configurations (v,h)(\mathbf{v}, \mathbf{h}) are sampled using quantum annealing (QA) hardware, e.g., the D-Wave Advantage_system4.1, by compiling the energy function into an Ising Hamiltonian. Quantum annealing yields samples that approximate the Gibbs–Boltzmann distribution of the RBM for the device's temperature and noise regime.

Each sampled vector v\mathbf{v} is transformed into a dd-dimensional continuous frequency ω\omega (matching the data dimension) via a Gaussian–Bernoulli conditional: P(ωv,a,U,σ)=i=1dN(ωi|ai+jUijvj,σi2)P(\omega \mid \mathbf{v}, \mathbf{a}, \mathbf{U}, \boldsymbol{\sigma}) = \prod_{i=1}^d \mathcal{N}\left(\omega_i \,\middle|\, a_i + \sum_{j} U_{ij} v_j,\, \sigma_i^2 \right) with learnable offsets a\mathbf{a}, weights U\mathbf{U}, and standard deviations σ\boldsymbol{\sigma} (σi2=exp(zi)\sigma_i^2 = \exp(z_i)), ensuring a flexible mapping from binary RBM states to the continuous spectral domain (Hasegawa et al., 13 Jan 2026).

4. Integration with Random Fourier Features and Kernel Construction

Given SS frequency samples {ωs}s=1Sp(ω)\{\omega_s\}_{s=1}^S \sim p(\omega) constructed via the RBM-QA pipeline, the real-valued random Fourier feature (RFF) map

φθ(x)=1S[cos(ω1x),,cos(ωSx),sin(ω1x),,sin(ωSx)]\varphi_\theta(\mathbf{x}) = \frac{1}{\sqrt{S}}\left[ \cos(\omega_1^{\top} \mathbf{x}),\, \ldots,\, \cos(\omega_S^{\top} \mathbf{x}),\, \sin(\omega_1^{\top} \mathbf{x}),\, \ldots,\, \sin(\omega_S^{\top} \mathbf{x}) \right]^\top

enables kernel approximation

kθ(x,x)φθ(x)φθ(x)k_\theta(\mathbf{x}, \mathbf{x}') \approx \varphi_\theta(\mathbf{x})^\top \varphi_\theta(\mathbf{x}')

with the simplified cosine-only average also used: k(Δx)1Ss=1Scos(ωsΔx)k(\Delta \mathbf{x}) \approx \frac{1}{S} \sum_{s=1}^S \cos(\omega_s^\top \Delta \mathbf{x})

The resulting kernel is not a traditional fixed Gaussian kernel, but is data-adaptive as its spectral density is learned through the RBM and QA-driven sampling loop (Hasegawa et al., 13 Jan 2026).

5. Optimization within Kernel Regression Frameworks

RBM parameters, along with frequency mapping parameters, are trained end-to-end with respect to the leave-one-out Nadaraya–Watson (NW) mean squared error (MSE) objective under squared-kernel weights: wij=kij2w_{ij} = k_{ij}^2 to prevent denominator collapse and enhance the contrast of similar neighbors. The prediction for each training instance ii is computed as

y^i(i)=jiwijyjjiwij+ϵ\hat{y}_i^{(-i)} = \frac{\sum_{j\neq i} w_{ij} y_j}{\sum_{j\neq i} w_{ij} + \epsilon}

where ϵ1\epsilon \ll 1 stabilizes the denominator numerically.

Gradients with respect to kernel and hence RBM parameters are propagated through score-function identities for Boltzmann models, utilizing: logpθ(ω)θ=Eθ(ω)θEpθ[Eθ(ω)θ]\frac{\partial \log p_\theta(\omega)}{\partial \theta} = -\frac{\partial E_\theta(\omega)}{\partial \theta} - \mathbb{E}_{p_\theta}\left[-\frac{\partial E_\theta(\omega)}{\partial \theta}\right] and all parameter updates are handled via stochastic gradient steps (Hasegawa et al., 13 Jan 2026).

6. Performance Characteristics and Structural Insights

Empirical evaluation across standardized regression datasets demonstrates that the RBM-driven kernel learning framework (KLNW: Kernel Learning with Nadaraya–Watson) monotonically reduces training loss, induces structural modifications in the kernel matrix—specifically in off-diagonal entries reflecting learned neighbor relations—and systematically outperforms both fixed Gaussian kernel NW regression and Gaussian-kernel SVR in R2R^2 and RMSE.

Key performance results (test R2R^2 at S=2000S=2000 random features):

Dataset KLNW R2R^2 Baseline NW R2R^2
bodyfat 0.910 0.715
mg 0.666 0.643
energy 0.980 0.932
ccs 0.748 0.677

The learned spectral histograms are strongly non-Gaussian and often multimodal, confirming that the RBM–QA procedure discovers data-adaptive, complex frequency structures. Increasing SS during inference consistently improves prediction accuracy due to reduced Monte Carlo variance in the RFF approximation (Hasegawa et al., 13 Jan 2026).

7. Application-Specific Adjustments and Inference Procedures

At inference, endpoint queries (test samples with any coordinate in the lowest or highest 1% of the empirical training marginal) invoke a local linear regression (LLR) correction utilizing the squared-kernel weights: wj=kθ(x,xj)2w_{*j} = k_\theta(\mathbf{x}_*, \mathbf{x}_j)^2 A weighted least squares problem is solved to fit yβ0+β(xx)y \approx \beta_0 + \beta^\top (\mathbf{x} - \mathbf{x}_*); the intercept β^0\hat{\beta}_0 serves as the local linear prediction. For interior points, the NW prediction is used. The LLR correction further improves the metrics, including surpassing SVR in certain regimes.

A plausible implication of these findings is that the representational flexibility of an RBM, particularly when coupled with quantum annealing for efficient, hardware-based sampling, enhances adaptive kernel construction for regression beyond parametric spectral models and classical gradient-based estimators. The observed multimodal spectral densities learned by the RBM suggest its practical effectiveness in modeling structured frequency mixtures relevant for kernel-based machine learning (Hasegawa et al., 13 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Restricted Boltzmann Machine (RBM).