How Gravity Can Explain the Collapse of the Wavefunction
Abstract: I present a simple argument for why a fundamental theory that unifies matter and gravity gives rise to what seems to be a collapse of the wavefunction. The resulting model is local, parameter-free and makes testable predictions.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
Overview
This paper asks a big question in quantum physics: Why do we only see one outcome when we measure something, even though quantum math says it can be in many states at once? The author suggests that gravity—the force that shapes space and time—can explain why the “wavefunction” seems to collapse to a single result during measurements. The model aims to keep everything local (no spooky action at a distance), uses no adjustable knobs or free parameters, and suggests ways to test it.
What is the paper trying to figure out?
In simple terms, the paper wants to answer:
- Why do we never see “cat states” (huge objects being in two places at once)?
- How can measurements produce single outcomes without requiring nonlocal, instant changes?
- Can gravity naturally cause the wavefunction to collapse in a way that matches what we observe?
- Can this be done while keeping the predictions of quantum mechanics (especially Born’s rule, which tells us the probabilities of outcomes)?
How does the model work? (Methods explained simply)
Think of the wavefunction as a cloud of possibilities for a system. Normally, this cloud evolves smoothly according to the Schrödinger equation (the main rule for how quantum systems change over time). But when we measure, the cloud “collapses” to one outcome. The challenge is to explain this collapse using local physics.
Here are the main ideas, with everyday analogies:
- Matter and geometry are two views of the same thing: The paper assumes that the stuff in the universe (particles) and the shape of space-time (gravity) are two tightly linked descriptions of one underlying reality—like two synchronized versions of the same song. If one changes, the other changes in step.
- Product state constraint: Because matter and gravity carry the same information in this view, the combined state must look like a “product state”—meaning the matter part and the gravity part aren’t entangled with each other. That’s unusual in standard quantum gravity, which normally entangles them. So the model allows small, controlled deviations from the usual Schrödinger evolution to keep matter and geometry in sync.
- Residual and action principle: The paper defines a “residual” S (think of it as a measure of how much a path through time deviates from the usual rule). The system prefers the path that makes S as small as possible. This is like choosing the path of least effort.
- Teleology via superdeterminism: In this context, “superdeterminism” means the whole evolution depends on both the initial setup and the final measurement arrangement, so they fit together consistently. Technically, this violates “measurement independence” (a Bell-theorem assumption). Practically, it means the allowed paths are those that lead to local, consistent outcomes.
- Penrose-phase (gravity’s role): If a mass is in a superposition of two places, each branch has a slightly different gravitational potential. That difference causes the quantum phases to drift apart—the paper calls the accumulation of that drift the “Penrose-phase.” The bigger the mass and the stronger the gravity effect, the faster this drift builds up.
- Detectors amplify mass/energy: Real measurements involve detectors that move many particles (like a photomultiplier releasing lots of electrons). That means a huge amount of matter would be “in two places” if the superposition persisted, making the gravitational penalty (the residual S) grow very fast.
- Local collapse before the detector: To keep S small, the system picks a path that briefly and locally “rotates” into one branch (one outcome) before reaching the detector. That avoids a macroscopic superposition and keeps things consistent with local energy conservation.
- Born’s rule via hidden variables: To recover the usual quantum probabilities, the model adds hidden variables (unknown details we don’t track). It assigns a simple, maximum-entropy random process to these variables and shows that the chance of each outcome comes out proportional to the square of the amplitude, matching Born’s rule.
What did the paper find, and why does it matter?
Key results include:
- Gravity can explain collapse locally: When a measurement would create a large superposition of mass/energy, it becomes too “costly” (in the S residual sense). The system then chooses a local, brief collapse into one branch—before the detector—avoiding nonlocal jumps.
- The effect scales with total gravitational potential: The size of the gravity-related phase buildup scales like the sum of the gravitational potentials of the branches and the mass involved. Heavier systems collapse faster under measurement-like conditions.
- Product states as pointer states: The outcomes we actually see correspond to “pointer states” of the detector that are nearly product states of matter and geometry, so they build little to no gravitational phase penalty.
- Born’s rule emerges naturally: By adding hidden variables with a simple exponential distribution, the model reproduces the familiar squared-amplitude probabilities.
- Predictions differ from other gravity-collapse models (Penrose-Diósi): In this model, once the branches are well separated (don’t overlap), the effect is roughly independent of their distance apart and scales with the sum of potentials, not with their variance. That gives different experimental signatures.
How did the paper approach the technical parts?
- “Residual” S: This is defined as the time integral of how much the evolution deviates from the standard Schrödinger path. Minimizing S picks the “least-cost” way the system can evolve while keeping matter and gravity unentangled.
- Local rotations in Hilbert space: The model uses a simple, shortest-path “rotation” between two quantum branches. This is the most efficient way to switch from a superposition to a single branch locally.
- Hidden variables and rates: The model assigns a rate to each possible end-state based on how costly it is (via S), then uses a maximum-entropy probability distribution (an exponential) to select outcomes. This yields the same probabilities as standard quantum mechanics.
- Weak measurements: If a detector only sometimes amplifies a signal to macroscopic sizes, it sometimes builds a large gravitational phase (causing collapse) and sometimes doesn’t. That matches how weak measurements work in practice.
Why is this important?
- It keeps physics local: The collapse happens without instant, long-distance action. Energy doesn’t have to jump between far-apart places at measurement time.
- It uses gravity and has no free parameters: Gravity is a known interaction. The model doesn’t introduce new adjustable constants.
- It explains why we don’t see macroscopic superpositions: Detectors rapidly amplify tiny quantum events into big mass/energy changes. That makes superpositions too costly to maintain, so they collapse locally.
- It matches quantum probabilities: Despite proposing a new way to think about collapse, it recovers the same outcome statistics we observe in experiments.
- It suggests tests: Because its predictions differ in how the effect scales (sum of potentials, weak dependence on branch distance after separation), experiments with massive superpositions and carefully designed interferometers could check which model is right.
What could this mean for the future?
If this approach is correct:
- It links the measurement problem to gravity and might guide the search for a deeper theory where matter and geometry are truly unified.
- It could clarify why black holes don’t lose information: If gravity and matter are tightly connected, information may be preserved in geometry.
- It could reshape how we design quantum experiments with heavy objects: Heavier superpositions should collapse faster, and the distance between separated branches might matter less once they don’t overlap.
- It offers a path to local, realistic quantum models that still match observed probabilities.
In short, the paper proposes that gravity quietly nudges quantum systems to pick one outcome, especially when measurements would otherwise spread large amounts of matter across multiple places. This keeps physics local, recovers standard quantum statistics, and gives us new ways to test how quantum mechanics and gravity fit together.
Knowledge Gaps
Below is a concise, actionable list of knowledge gaps, limitations, and open questions that remain unresolved in the paper. Items are grouped thematically to aid follow-up research.
Foundational Assumptions and Scope
- Lack of an explicit underlying unifying theory: the “underlying Hamiltonian” and the unitary map that identify matter and geometry are not specified; no microphysical dynamics are provided from which the product-state constraint emerges.
- Justification and consistency of the core hypothesis: the postulated subset of product states is asserted rather than derived; it conflicts with standard canonical and perturbative quantum gravity where matter–geometry entanglement is generic—no proof of consistency with known constraints or limits.
- Relation to diffeomorphism invariance: the assumption that matter and geometry “are the same” requires a careful treatment of gauge redundancies and relational observables; how this identification respects diffeomorphism invariance remains unclarified.
- Scope of applicability: the model is developed in a Newtonian, first-quantized setting; a concrete, worked-out, fully relativistic, covariant quantum field-theoretic formulation is not provided.
Mathematical Formulation and Well-Posedness
- Stationarity and solution concept: the variation of the residual functional with endpoint variation is posited, but existence, uniqueness, and construction of stationary paths (beyond the two-branch toy rotation) are not established.
- Choice of functional and gauge: the use of the L2 norm of the residual (with absolute value) can be non-differentiable at zeros; whether an equivalent, smooth functional (e.g., squared norm or ) yields identical stationary paths is not demonstrated.
- Dependence on “energy gauge”: the minimization relies on choosing a time-dependent phase to set the parallel residual to zero; gauge independence (or controlled gauge dependence) of physical predictions is not shown.
- Endpoint variation (“teleology”): the precise mathematical implementation of varying both paths and endpoints is not formalized (e.g., admissible endpoint sets, measures, constraints); no general algorithm for computing optimal paths in realistic many-body settings.
- Definition and role of vs : the statistical functional is introduced ad hoc and distinct from ; no derivation from an underlying measure on histories is provided, nor proof that predictions are robust to alternative reasonable definitions of .
Locality, Relativity, and Constraints
- Proof of locality: while the model aims to be locally causal via superdeterminism (violating measurement independence), a rigorous demonstration that all dynamics respect microcausality and light-cone propagation (especially in spacelike-separated measurements) is missing.
- Compatibility with Hamiltonian and diffeomorphism constraints: in a canonical or covariant QG setting, it is unclear how the product-state restriction and residual minimization respect constraint algebras and local conservation laws.
- Energy–momentum conservation: local “rotations” off the Schrödinger evolution are claimed to preserve conservation laws, but a detailed demonstration (including gravitational constraints) is absent.
Probabilistic Element and Born’s Rule
- Hidden variables remain unspecified: the physical nature, dynamics, and distribution of the hidden variables are not identified; only a functional form for their statistics is chosen.
- Ad hoc choice of the exponential distribution: the choice of with is not derived; uniqueness, invariance under reparameterizations, and robustness w.r.t. alternative maximum-entropy assumptions need justification.
- Definition and interpretation of : the integrand involving is introduced without derivation from a more fundamental theory; dependence on the overlap and slicing choices needs scrutiny.
- Bell-type correlations: while measurement independence is violated, the model does not show explicitly that it reproduces full quantum correlations (e.g., CHSH/Tsirelson bound) in spacelike-separated Bell tests without enabling signaling.
Physical Consistency and Pointer States
- Identification of pointer states: “near-product” matter–geometry states are asserted as pointer states, but criteria for their emergence and stability (e.g., against environmental interactions) are not quantitatively developed.
- Interplay with decoherence: the model’s collapse trigger competes with environmental decoherence; quantitative conditions identifying when the “Penrose phase” effect dominates are not provided.
- Weak measurements: the threshold condition “∫dt ||R|| ≳ 1” is qualitative; no quantitative modeling of residual growth rates for realistic weak measurement devices or predictions beyond standard QM statistics.
Quantitative Predictions and Experimental Tests
- Missing concrete numbers: beyond order-of-magnitude heuristics (e.g., Penrose phase ∝ τ m Φ), the paper lacks explicit collapse times or thresholds for realistic systems (molecular interferometry, optomechanics, nanospheres).
- Distance and mass scaling: the claim that the effect is roughly independent of branch separation once packets are orthogonal and scales with the sum (not variance) of potentials needs quantitative benchmarking against existing interferometry data (fullerenes, clusters).
- Discriminating tests vs PD/CSL: no detailed proposals for experiments that distinguish this model from Penrose–Diòsi or CSL (e.g., distance dependence, heating, noise spectra) with predicted effect sizes and required sensitivities.
- Parameter-free claim vs smearing scale: predictions depend on how mass density is smeared (e.g., wavepacket width); this introduces an implicit parameter. A rule that fixes smearing uniquely is needed for falsifiability.
- Photons and relativistic particles: treatment of massless particles is not developed (phase estimates rely on mΦ); a consistent relativistic generalization that handles energy-based gravitational coupling is needed.
- Macroscopic quantum systems: potential conflicts with observed macroscopic superpositions (e.g., SQUID flux qubits) are not addressed; the model should predict whether and when collapse would suppress such phenomena.
Many-Body and Interaction Structure
- Scaling in complex entangled systems: beyond two-branch, two-subsystem examples, general scaling of the residual in highly entangled, interacting many-body states is not derived; impact of non-gravitational interactions on residual minimization is unclear.
- Mean-field approximations: the decomposition with mean-field potentials (V_A, V_B) is heuristic; accuracy and limits in strongly correlated systems remain to be analyzed.
Covariant QFT Generalization
- Concrete covariant formulation: the suggested replacement by constraint densities and lapse/shift is sketched, but a complete covariant construction, including handling of constraints, regularization/renormalization, and local commutativity, is missing.
- Frame dependence: the statement that detector rest frames “naturally” break covariance needs to be reconciled with a consistent covariant description that yields frame-independent observable predictions.
Edge Cases, Limits, and Conceptual Issues
- Asymptotic states: the requirement that asymptotic particles be “gravitationally coherent” is strong; a constructive definition and demonstration that standard scattering theory can be recovered is lacking.
- Special-phase cases and stationarity: cases where C(t)=0 (due to specific amplitudes/phases) are acknowledged as non-stationary, but a general proof that only physically acceptable paths are stationary is not provided.
- Teleology and boundary conditions: the model’s endpoint dependence raises questions about time-symmetry, thermodynamics, and operational meaning (e.g., “what if the system is never measured?”); a clear statement of boundary conditions in the far future is needed.
- Black hole information remark: the claim that the identification of matter and geometry “would solve” information loss is suggestive but undeveloped; a concrete argument or model is absent.
These gaps indicate clear next steps: specify the underlying dynamics and hidden variables, formalize the variational problem and its locality, develop a covariant QFT version, and deliver quantitative, discriminating experimental predictions with realistic systems and noise.
Practical Applications
Immediate Applications
The paper proposes a local, parameter-free, and testable model of wavefunction collapse driven by gravity via a product-state constraint between matter and geometry. Even before a complete underlying theory is known, the framework yields concrete design rules, analysis tools, and experimental protocols.
- Industry (quantum technologies, metrology, photonics, optomechanics)
- Design rules for macroscopic superposition experiments
- What: Use the model’s “residual action” S and “Penrose-phase” (∼τ m |Φ12|) to identify when superpositions will collapse locally (e.g., at a beam splitter) rather than at the detector.
- Why: Helps prevent false interpretations of decoherence sources and sets realistic performance ceilings for macroscopic superposition devices (levitated nanoparticles, nanomechanical resonators, superconducting cat states).
- Potential tools/products/workflows:
- Residual Action Calculator (plugins for QuTiP/Julia/Qiskit) to estimate ||R|| and A for given device geometry, mass distribution, timing, and detector amplification workflow.
- “GravCollapse Budget” templates integrated with standard decoherence budgets (gas, thermal, EM) in precision experiments.
- Assumptions/dependencies: Newtonian-gravity limit for potentials; accurate mass-density modeling; branch wave packets approximately orthogonal at macroscopic scales; detector amplification threshold modeled as time-integrated residual ≳ O(1).
- Metrology-grade bounds on gravity-induced collapse
- What: Add “gravitational collapse” as a modeled noise floor in macroscopic interference platforms and put upper limits from existing data (e.g., optomechanical resonators, molecule interferometry).
- Why: Parameter-free predictions let metrology labs set and report conservative ceilings for gravitational contributions to loss of coherence.
- Potential tools/products/workflows: Lab SOPs for reporting gravitational-collapse budgets alongside environmental decoherence; analysis modules for comparing PD (Penrose-Diósi) versus the paper’s constant-with-distance (post-orthogonality) scaling.
- Assumptions/dependencies: High vacuum, cryogenic operation, verified environmental noise characterization to avoid conflation.
- Academia (experimental and theoretical physics)
- Differential tests against PD-type models
- What: Protocols that vary (i) total mass in superposition and duration τ, and (ii) branch separation beyond the overlap length.
- Predicted signatures:
- This model: After branch orthogonality, the effective collapse-driving term becomes approximately distance-independent and scales with the sum of branch potentials; visibility should not recover or weaken with increased separation once orthogonality is reached.
- PD models: Decoherence strength declines roughly as 1/d for large separations; sensitivity to variance of the mass-density self-energy, not sum.
- Platforms:
- Levitated nanoparticles (107–1010 amu) in Talbot-Lau or double-path interferometry.
- Nanomechanical cat states (superconducting circuits coupled to membranes/beam resonators).
- High-mass molecule interferometry with tunable slit separations and post-selection.
- Potential tools/products/workflows: Experimental design toolkit to scan separation beyond overlap; fit libraries supplying both PD and “sum-of-potentials” models; visibility-vs-time dashboards with saturation tests.
- Assumptions/dependencies: Ability to ensure branch orthogonality; known detector amplification threshold; careful gauge fixing for potentials in analysis; isolation from technical decoherence to reveal gravitational contribution.
- Early-collapse versus late-collapse probes with nested interferometers
- What: Tests that check whether collapse occurs locally at the first interaction region (e.g., first beam splitter) rather than only at detection, by using reconfigurable nested Mach–Zehnder/Sagnac geometries and weak/pre-measurements near threshold.
- Why: The model predicts local, brief departures from Schrödinger evolution in interaction regions that minimize the integrated residual, leading to earlier collapse for would-be macroscopic branches.
- Potential tools/products/workflows: Programmable linear optics chips with variable amplification elements; weak-measurement threshold tuning to test when accumulated residual crosses O(1).
- Assumptions/dependencies: Superdeterminism implies correlations with measurement settings—experimental designs must focus on measurable rate scalings (mass, time, orthogonality, amplification), not on delayed-choice paradoxes.
- Theory development and benchmarking
- What: Develop general-covariant field-theoretic formulations of the residual functional; extend to interacting multipartite systems; refine the hidden-variable distribution underpinning Born-rule recovery.
- Why: To translate the phenomenology into covariant QFT and close gaps between canonical models and the product-state constraint.
- Potential tools/products/workflows: Open-source reference implementations for computing S and A in simplified field-theoretic models; benchmark suites for PD vs. local-collapse predictions.
- Assumptions/dependencies: Choice of lapse/shift (energy gauge), relational/gauge-fixing decisions, and mapping U between matter and geometry sectors.
- Policy (research strategy and funding)
- Targeted programs for macroscopic-quantum and gravity–quantum tests
- What: Calls that prioritize experiments varying mass, duration, and post-orthogonality separation to probe the model’s parameter-free predictions; support for microgravity testbeds.
- Why: Direct, discriminating tests between collapse paradigms need dedicated platforms and clean environments.
- Potential tools/products/workflows: Shared facilities for high-vacuum, cryogenic interferometry; standards for reporting gravitational-collapse budgets.
- Assumptions/dependencies: Cross-lab reproducibility; open datasets enabling joint analyses.
- Daily life (education and communication)
- Curricular and outreach material on local collapse and superdeterminism
- What: Modules illustrating how locality can be restored via superdeterministic constraints and how Born’s rule can emerge from path-selection statistics.
- Why: Clarifies the measurement problem with concrete, testable physics.
- Potential tools/products/workflows: Interactive notebooks that compute S and A for simple interferometers; visualizations of when amplification triggers collapse.
- Assumptions/dependencies: Careful presentation of limitations (phenomenological model; underlying unification still unknown).
Long-Term Applications
If confirmed experimentally, the framework would enable new engineering and scientific directions that purposefully leverage or mitigate gravity-driven collapse at macroscopic scales.
- Industry (sensing, devices)
- Gravity-assisted measurement engineering
- What: Engineer devices whose operation depends on being at (or just beyond) the amplification threshold where time-integrated residual crosses O(1), enabling ultra-sensitive mass or energy-difference sensing through controlled onset of collapse.
- Potential tools/products/workflows: Threshold “gravi-collapsometers” for inertial or mass sensing; adaptive control that tunes amplification to sit at the boundary of detection/no-detection for maximal sensitivity.
- Assumptions/dependencies: Ability to reliably create and hold macroscopic superpositions near threshold; accurate modeling of all non-gravitational decoherence channels.
- Collapse-aware architectures for macroscopic quantum technologies
- What: Design superposition-based devices (quantum memories, macroscopic qubits, mechanical processors) to minimize gravitational residual (e.g., symmetric mass motion, gravitationally coherent branches) and extend coherence times.
- Potential tools/products/workflows: CAD tools that co-optimize electromagnetic and gravitational coherence; co-designed mass distributions to suppress the Penrose-phase.
- Assumptions/dependencies: Large-scale, low-loss macroscopic superpositions; verified control over amplification pathways.
- Academia (space-based and extreme-environment tests; foundational physics)
- Space missions for macroscopic interferometry
- What: Leverage long free-fall times, ultralow noise, and tunable gravitational environments to map the dependence on mass, time, and orthogonality, and to explore the model’s near distance-independence post-orthogonality.
- Potential tools/products/workflows: CubeSat/ISS platforms for levitated nanoparticle interferometry; gravity-gradient tuning to probe sum-of-potentials sensitivity.
- Assumptions/dependencies: Robust particle control in microgravity; scalable cooling and isolation.
- Black-hole information and matter–geometry unification studies
- What: Use the “matter = geometry state” ansatz and product-state constraint to develop toy models addressing information retention in gravitational collapse and evaporation.
- Potential tools/products/workflows: Simulation frameworks for unitary matter–geometry evolution with enforced product subspaces; comparisons to holographic or spin-network models.
- Assumptions/dependencies: Existence of an underlying unifying dynamics Hu and mapping U compatible with low-energy phenomenology.
- Policy (standards, certification, strategy)
- Standards for macroscopic-superposition claims and reporting
- What: Guidelines requiring an explicit gravitational-collapse budget (residual S, A integrals, amplification thresholds) when claiming macroscopic superpositions or cat states.
- Potential tools/products/workflows: Certification protocols for reporting gravitational versus environmental decoherence; repositories for benchmark datasets across platforms.
- Assumptions/dependencies: Community consensus on reporting formats; validated modeling tools.
- Strategic roadmaps for gravity–quantum interfaces
- What: Long-term funding and infrastructure plans for facilities that can produce, control, and measure macroscopic superpositions under variable gravitational conditions (including space-based assets).
- Assumptions/dependencies: International collaboration and sustained investment.
- Daily life (long-horizon implications)
- Randomness, security, and philosophy-of-physics interfaces
- What: If superdeterminism is empirically supported, audit implications for certifications of physical randomness (while maintaining operational unpredictability in practice).
- Potential tools/products/workflows: Position papers and certification updates clarifying operational versus ontic randomness.
- Assumptions/dependencies: Strong experimental confirmation of superdeterministic correlations; community/legal acceptance of revised definitions.
Cross-cutting assumptions and dependencies (affecting feasibility across applications)
- The model’s key premises: a unification where matter and geometry are the same quantum state; enforced product-state subset; residual functional S measuring deviation from canonical Schrödinger evolution; path selection that minimizes S; probabilistic selection via hidden variables with exponential distribution recovering Born’s rule.
- Validity regime: Newtonian limit used for estimates; gauge fixing required to meaningfully compare branch potentials; amplification defined as time-integrated residual ≳ O(1).
- Experimental preconditions: Ability to produce near-orthogonal branches with large masses, suppress conventional decoherence below gravitational effects, and control amplification pathways (strong/weak measurement thresholds).
- Distinguishing predictions: Post-orthogonality distance-independence and scaling with sum of branch potentials (versus variance/1/d behavior in PD models); local, early collapse near interaction regions to minimize integrated residual.
- Open theoretical tasks: General-covariant field-theoretic formulation; precise characterization of U (mapping between matter and geometry dofs); derivation or justification of the hidden-variable distribution from a deeper theory.
Glossary
- Action principle: A variational method that selects physical evolutions by extremizing an action functional. Example: "the action principle is ideally suited to incorporate a superselection rule"
- All-at-once constraint: A global constraint on entire histories/evolutions rather than initial-value dynamics. Example: "(an all-at-once constraint in the terminology of \cite{adlam2024taxonomy})"
- Bell's condition of local causality: The requirement that influences be local in Bell’s framework, violated by nonlocal models. Example: "violating Bell's condition of local causality."
- Bell's theorem: A result showing that no locally causal theory can reproduce all quantum predictions without additional assumptions. Example: "We know from Bell's theorem \cite{Bell1964OnEPR,Bell2004Speakable}"
- Black hole information loss problem: The puzzle of whether and how information is preserved in black hole evaporation. Example: "this would solve the black hole information loss problem."
- Born's rule: The quantum rule that outcome probabilities equal squared amplitudes. Example: "which is what Born's rule requires."
- Braided spin networks: Graph-based structures in quantum gravity with braiding, proposed as fundamental spacetime degrees of freedom. Example: "braided spin networks \cite{bilson2012emergent}"
- Cauchy hypersurfaces: Spacelike slices that provide complete initial data for deterministic evolution in relativity. Example: " is a family of Cauchy hypersurfaces"
- Causal Fermion Systems: A proposed framework where spacetime and fields emerge from a measure on fermionic states. Example: "Causal Fermion Systems \cite{finster2015causal}"
- Collapse of the wavefunction: The apparent reduction from a superposed quantum state to a definite outcome upon measurement. Example: "collapse of the wavefunction"
- Coherent states: Quantum states closest to classical behavior, often eigenstates of annihilation operators. Example: "coherent states"
- Constraint operator density: The local (density) operators generating constraints in generally covariant quantum theories. Example: " is the constraint operator density"
- Decoherence: The loss of quantum coherence due to environment-induced entanglement, making superpositions effectively classical mixtures. Example: "decoherence-based approaches"
- Detector eigenstate: A measurement outcome state that is an eigenstate of the detector’s observable. Example: "the outcome of the time evolution is always a detector eigenstate"
- Energy gauge: A choice of time-dependent phase that nulls the parallel component of the residual to minimize the action. Example: "This is sometimes called the ``energy gauge.''"
- Euler-Lagrange equations: Differential equations derived from the action principle that determine classical dynamics. Example: "the Euler-Lagrange equations"
- Feynman diagram: A graphical representation of particle processes used to organize perturbative calculations. Example: "a virtual particle-antiparticle pair in a Feynman diagram."
- General covariance: Invariance of physical laws under arbitrary spacetime coordinate transformations. Example: "general covariance is naturally broken in a typical quantum experiment by the rest frame of the detector."
- Geon: A gravitational-electromagnetic (or gravitational) self-confined configuration proposed as a model for particles. Example: "the geon approach \cite{wheeler1955geons}"
- Geometric engineering: Constructing gauge theories from geometry in string theory setups. Example: "geometric engineering in string theory \cite{katz1997geometric}"
- Geometric Unity: A speculative unification framework proposing a geometric underpinning of physics. Example: "Geometric Unity \cite{Weinstein2021GUsite}"
- Ghirardi-Rimini-Weber (GRW) model: A spontaneous-collapse model modifying quantum dynamics to produce definite outcomes. Example: "Ghiradi-Rimini-Weber ({\sc GRW}) model \cite{ghirardi1986unified}"
- Gravitational self-energy: The self-energy associated with a mass distribution due to its own gravitational field. Example: "gravitational self-energy of the mass-density."
- Gravitons: Hypothetical quantum particles mediating the gravitational interaction. Example: "I am here including gravitons as a type of particle"
- Hilbert space: The complete vector space of quantum states with an inner product structure. Example: "Hilbert space of quantum gravity"
- Lapse/shift vector: The decomposition of spacetime evolution into normal (lapse) and tangential (shift) components in canonical gravity. Example: " is the lapse/shift vector"
- Locally causal model: A theory respecting locality in the sense used by Bell, i.e., no superluminal causal influences. Example: "any locally causal model that correctly describes observations needs to violate measurement independence."
- Mach-Zehnder interferometer: A two-path interferometric setup used to study quantum interference. Example: "consider a Mach-Zehnder interferometer."
- Mean field path: An approximate evolution where a subsystem feels only the average effect of others. Example: "the deviation of each subsystem from the mean field path"
- Measurement independence: The assumption that measurement settings are statistically independent of hidden variables. Example: "violate measurement independence."
- Newtonian potential: The classical gravitational potential used in the nonrelativistic limit. Example: "Newtonian potential"
- Penrose-phase: The phase accumulation associated with gravitational differences between branches, used here as an estimate of residual growth. Example: "I will in the following refer to this estimate as the ``Penrose-phase''."
- Penrose's model of gravitationally induced collapse: A proposal that gravity causes the collapse of quantum superpositions. Example: "Penrose's \cite{penrose1996gravity,penrose1998quantum} model of gravitationally induced collapse."
- Planck mass: The fundamental mass scale in quantum gravity, set by ħ, c, and G. Example: "Planck mass."
- Pointer states: Stable, effectively classical states selected by system–environment interactions (or detector amplification). Example: "pointer states of the measurement device."
- Product state: A separable state with no entanglement between subsystems (here, matter and geometry). Example: "our product state schematically has the form"
- Residual: The deviation from Schrödinger evolution used to define the action to be minimized. Example: "the residual "
- Schr\"odinger evolution: Unitary time evolution of quantum states governed by the Schrödinger equation. Example: "the Schr\"odinger evolution"
- Schr\"odinger's cat state: A macroscopic superposition illustrating quantum–classical tension. Example: "``Schr\"odinger's cat state\""
- Shape Dynamics: An alternative formulation of gravity emphasizing spatial conformal invariance. Example: "Shape Dynamics \cite{barbour2012shape}"
- Spinor Gravity: A proposal in which gravity emerges from spinor fields. Example: "Spinor Gravity \cite{hebecker2003spinor}"
- Superdeterminism: The class of theories in which hidden variables are correlated with measurement settings, violating measurement independence. Example: "Superdeterminism is mathematically defined as a correlation between the measurement settings and the presumed to exist hidden variables "
- Superposition: A quantum state formed by a linear combination of distinct states, potentially over macroscopically distinct configurations. Example: "macroscopic superpositions"
- Superselection rule: A rule forbidding certain superpositions or restricting allowable transitions/outcomes. Example: "as a superselection rule"
- Teleological (teleology): Depending on future boundary conditions or end-states in a way that appears goal-directed. Example: "only seemingly teleological"
- Unitary evolution: Evolution preserving state norm and inner products; in quantum theory, driven by a Hermitian Hamiltonian. Example: "allow the evolution of the state in Hilbert space to proceed unitarily"
- Virtual particle-antiparticle pair: Intermediate, off-shell excitations in perturbation theory that do not appear in final states. Example: "a virtual particle-antiparticle pair"
- Weak measurements: Measurements that only slightly disturb the system or only sometimes register, yielding limited information per trial. Example: "we can deal with weak measurements."
- Yang-Mills theory: A nonabelian gauge theory underlying the Standard Model’s strong and electroweak interactions. Example: "Yang-Mills theory"
Collections
Sign up for free to add this paper to one or more collections.