Jump to content

SergejMaterov

Members
  • Joined

  1. Abstract This work is inspired by the book  Lloyd’s “Computational capacity of the universe” [5] and reports an explicitly falsifiable (already today) discrete–quantum‐graph model of spacetime and noise in quantum processors. Rather than invoking Planck‐scale assumptions or ad hoc temperature thresholds, we derive a single measurable scale: [latex]k_BTc=Jz[/latex] where J is the qubit‐qubit coupling (noise) energy and z the average vertex degree. Below T_c long‐range correlations [latex]\Psi(r) = \left\langle \sigma_i^z \,\sigma_{i+r}^z \right\rangle[/latex] persist; above T_c they vanish. We introduce the microscopic noise Hamiltonian [latex]\widehat{H_{\mathrm{noise}}}= \sum_{\langle i,j \rangle} J_{ij}\, \sigma_i^z \sigma_j^z+ \sum_{i} h_i\, \sigma_i^x [/latex] allowing direct spectroscopy of [latex]\\ J_{ij} [/latex] and [latex] h_{i} [/latex] From this single relation we obtain multiple near‐term experimental tests—e.g. heat‐capacity and error‐rate crossovers at T≈T_c, correlation‐length collapse in small‐graph Monte Carlo, and spectral‐DOS corrections—all calibrated by measured J and z. Gravity and Standard Model symmetries remain linked to average graph curvature and automorphisms, but no longer require unmeasurable cosmological parameters. Appendix A presents the concrete protocol for extracting [latex] T_c [/latex] on existing QPU topologies. This work proposes a concrete route from discrete graph degrees of freedom to effective continuum gauge dynamics and testable experimental signatures. To help the reader follow the chain of logic, we summarize the main conceptual steps here before the detailed derivations: (i) start with local edge variables U_e on a finite graph and define a plaquette action [latex]S_{lat} = \kappa \sum_{p} \left( 1 - \Re \ \operatorname{Tr}(V_p) \right), \quad V_p = \prod_{e \in p} U_e [/latex] (ii) in the small-fluctuation, short-edge limit [latex]U_e\approx\exp{\left(iaA_\mu^aT^a\right)\ } [/latex] standard expansion of [latex]V_p [/latex] yields a leading [latex]F_{\mu\nu}^2 [/latex] term and an identification of the continuum coupling g with microscopic parameters [latex] \kappa,J,z [/latex],; (iii) block-averaging and RG flow control whether the low-energy theory is governed by a Yang–Mills action and determine the sign and rate of running via the effective low-mode count [latex] N_f [/latex]; and (iv) physical observables — critical crossover [latex] T_c [/latex], spectral signatures in Δtan δ , QPU error-rate crossovers, and topology sensitivity — follow from the same mapping and provide falsifiable tests.
  2. Everything is described and written down in the mathematical language I am accustomed to at the link.
  3. My theory is currently undergoing experimental testing. Or rather, it's undergoing falsification, according to Popper.
  4. This work aims, first and foremost, to construct a falsifiable model of quantum gravity, seamlessly linked to low-energy experiments. — On the one hand, it derives the classical Newton–Maxwell–Einstein (GR) equations from a single "noisy" Hamiltonian on a discrete quantum graph, so it is essentially a candidate for a theory of quantum gravity. — On the other hand, it also incorporates the symmetries of the Standard Model (U(1)×SU(2)×SU(3)), the spontaneous breaking of these symmetries, and even predictions about the "shor anomaly" and microwave resonances in cryogenic qubits. Thus, this is not just "yet another" model of gravity, but a genuine attempt at a unified description of: • quantum dynamics (via the σ-Hamiltonian), • SM symmetries (via graph automorphisms), • gravity (via discrete curvature and its continuous limit). In this sense, the author claims a "theory of everything" on a discrete graph. But the key distinguishing feature is that it is constructed to yield specific laboratory predictions already at energies ≲10⁻⁴ eV and can therefore be quickly refuted or confirmed in cryogenic qubits and microwave experiments. The theory itself is described in detail here: A Testable Quantum Graph Theory of Spacetime: Predictions for Cryogenic Qubits and Colliders
  5. The utility of a scientific theory resides in its empirical content—its capacity to be tested and potentially refuted. A theory that is logically unfalsifiable, or one whose falsification would require experiments that are infeasible in principle or practice, lacks operational significance. In either scenario the theory does not provide an empirically meaningful account of phenomena and therefore cannot play the role required of scientific theories in advancing understanding. Wolfram’s program has been criticized as practically unfalsifiable because, while it aims to derive physics from simple computational rules, it has not yet produced unique, precise, and independently testable quantitative predictions that differ from established models. In practice the framework is flexible (many rule choices and interpretive steps), and the proposed observational signatures are often qualitative or lie at unrealistic scales/conditions; that combination makes it easy to accommodate existing data but hard to subject the proposal to a decisive experimental refutation.

Important Information

We have placed cookies on your device to help make this website better. You can adjust your cookie settings, otherwise we'll assume you're okay to continue.

Configure browser push notifications

Chrome (Android)
  1. Tap the lock icon next to the address bar.
  2. Tap Permissions → Notifications.
  3. Adjust your preference.
Chrome (Desktop)
  1. Click the padlock icon in the address bar.
  2. Select Site settings.
  3. Find Notifications and adjust your preference.