Why Quantum?

Deriving the Born rule, measurement, and the end of the measurement problem

Guide
Why is probability the square of an amplitude? Why do experiments give definite outcomes? Why does observation seem to matter?

Questions this page addresses

  • Why is the probability rule |ψ|²?
  • What determines the measurement basis?
  • What physically happens during measurement?
  • Does the framework resolve the measurement problem?
  • How does it compare to existing interpretations?

Quantum mechanics is the most successful and least understood theory in physics. Its predictions are confirmed to extraordinary precision. Its interpretation has been debated for a century without consensus.

The debate centers on three questions. Why is probability given by the square of the amplitude (the Born rule)? What determines which physical quantity gets a definite value when you measure (the preferred basis)? And what, physically, happens during a measurement (the measurement problem)?

The standard approach treats these as independent puzzles requiring separate solutions — or, in some interpretations, as mysteries to be accepted rather than explained. The framework derives all three from the axioms and the interaction structure already established.

The Born Rule

In standard quantum mechanics, the probability of a measurement outcome is postulated: the probability of finding the system in state k|k\rangle is kψ2|\langle k|\psi\rangle|^2, the squared magnitude of the amplitude. This is the Born rule. It works perfectly, but it is put in by hand.

The framework derives it. The argument runs roughly as follows: given an observer with a coherence domain that assigns amplitudes to possible measurement outcomes, what probability measure on those outcomes is consistent with the three axioms?

The constraints are tight. Coherence conservation requires that probabilities sum to one (normalization). The U(1) phase structure of minimal observers requires that the probability be insensitive to overall phase (phase covariance). And the composition of independent systems requires that joint probabilities factor appropriately.

The result: kψ2|\langle k|\psi\rangle|^2 is the unique probability measure satisfying these constraints. Not one option among many — the only one. The Born rule is forced by the same axioms that produce everything else in the framework.

The Preferred Basis

When you measure the spin of an electron, you get “up” or “down” — not some superposition of up and down. But the electron’s quantum state, before measurement, can be expressed in any basis. What selects the specific basis in which measurement gives definite outcomes?

Standard quantum mechanics addresses this through decoherence and einselection — the environment selects the “pointer basis” by making certain states more robust against environmental noise. This works well quantitatively, but it relies on a prior notion of what counts as “the environment” and what counts as “the system,” which is not fully satisfying as a foundational explanation.

The framework’s answer is more direct. A measurement is a Type III interaction between the system and the measuring apparatus. This interaction generates a new relational invariant — a new conserved quantity living on the joint state space. The basis in which this new invariant is diagonal — the basis in which it takes definite values — is the preferred basis.

The preferred basis is not chosen by convention, by the environment, or by the physicist. It is determined by the specific relational invariant that the measurement interaction creates. Different measurements create different relational invariants and therefore select different bases. This is why measuring position gives definite positions and measuring momentum gives definite momenta — each creates a different relational invariant.

What Happens During Measurement

The measurement problem, stated precisely: quantum mechanics describes systems evolving smoothly and deterministically via the Schrödinger equation. Measurement outcomes, by contrast, are sudden, definite, and apparently random. What bridges the gap?

The framework’s answer: measurement is a Type III interaction. Before the interaction, the system and apparatus have independent invariants. The system can be in a superposition relative to the observer — its coherence state is genuinely underdetermined with respect to the relational invariant that will be created.

During the Type III interaction, a new relational invariant crystallizes. This crystallization is irreversible (relational invariants are conserved — they cannot be un-generated) and it is specific (the invariant takes a definite value in the preferred basis). After the interaction, the observer’s coherence domain includes the new invariant, and the outcome is definite.

No collapse postulate is needed. No branching of the universe is needed. No hidden variables are needed. The measurement is a specific, analyzable physical process — the same kind of interaction (Type III) that generates chemical bonds, causal relationships, and all the other relational invariants in the hierarchy. There is nothing mysterious about measurement except that it was previously treated as primitive.

The Single Outcome

One question remains: if the system is in a superposition of, say, spin-up and spin-down, and the Type III interaction produces a definite outcome — why this outcome and not the other?

The framework’s answer involves the atemporal coherence dependency graph — the block universe described in the holography chapter. In the complete graph, every possible measurement outcome exists as a node. Every outcome is real, in the graph. The observer’s coherence domain advances through the graph along one specific path.

From within the observer’s domain — advancing node by node, unable to see the full graph — the forward paths appear genuinely open. The outcome appears random. The Born rule gives the correct probabilities for which path is traversed. From the perspective of the complete graph, every path exists, and the observer’s trajectory is already determined.

Both perspectives are simultaneously valid at their respective levels of description. Quantum randomness is the inside view of a determined structure, experienced from within a bounded coherence domain.

Comparison with Interpretations

The framework agrees with Relational QM that quantum states are observer-relative, but goes further by deriving (rather than postulating) the relational structure and the Born rule.

It agrees with Many Worlds that all outcomes are structurally present, but denies that the universe literally branches — there is one coherence geometry, not many.

It agrees with Copenhagen that measurement produces definite outcomes, but disagrees that measurement is primitive and unanalyzable.

It disagrees with Bohmian mechanics that there are hidden definite values prior to measurement, and with GRW collapse that unitary evolution is modified. The framework predicts exact unitarity at all scales — coherence conservation is absolute.

On solid ground: The Born rule derivation is rigorous — ψ2|\psi|^2 is proved to be the unique probability measure consistent with the axioms. The preferred basis determination via relational invariants is formally established. The identification of measurement with Type III interaction is structurally precise.

Work in progress: The Born rule derivation requires the amplitude-coherence identification postulate, which bridges the formal coherence measure to quantum mechanical amplitudes. The single-outcome question involves interpretive elements — the relationship between the atemporal block universe and the observer’s temporal experience — that are honestly flagged as philosophical rather than mathematical in the derivation. These are not evasions; they are the points where physics meets interpretation, and the framework locates them precisely rather than ignoring them.

From quantum mechanics, the framework extends to the particle spectrum and to the deepest structures of spacetime — holography and black holes.