Chance, Calculation and Life. Группа авторов
Чтение книги онлайн.

Читать онлайн книгу Chance, Calculation and Life - Группа авторов страница 10

СКАЧАТЬ analysis is grounded in a mathematical “tour de force”. He proved the non-analyticity of the (apparently) simple system of nonlinear equations describing two planets and a Sun in their gravitational fields (three bodies). The planets disturb each other’s trajectories and this gives the formal divergence of the Lindstedt-Fourier series meant to give a linear approximation of the solution of the equations. More precisely, by using his notions of bifurcations and homoclinic orbit (the intersection of a stable and an unstable manifold), he showed that the “small divisors”, which make the series diverge, physically mean that an undetectable fluctuation or perturbation may be amplified to a measurable quantity by the choice of a branch, or another in a bifurcation, or a manifold along a homoclinic orbit. It is often difficult to give a physical meaning to the solution of a system of equations; it is particularly hard and inventive to make sense of the absence of a solution. Yet, this made us understand randomness as deterministic unpredictability and non-analyticity as a strong form of classical unpredictability2.

      In this classical framework, a random event has a cause, yet this cause is below measurement. Thus, Curie’s principle3 is preserved: “the asymmetries of the consequences are already present in the causes” or “symmetries are preserved” – the asymmetries in the causes are just hidden.

      For decades, Poincaré’s approach was quoted and developed by only a few, that is, until Kolmogorov’s work in the late 1950s and Lorentz in the 1960s. Turing is one of these few: he based his seminal paper on morphogenesis (Turing 1952) on the nonlinear dynamics of forms generated by chemical reactants. His “action/reaction/diffusion system” produced different forms by spontaneous symmetry breaking. An early hint of these ideas is given by him in Turing (1950, p. 440): “The displacement of a single electron by a billionth of a centimetre at one moment might make the difference between a man being killed by an avalanche a year later, or escaping”. This Poincarian remark by Turing preceded by the famous “Lorentz butterfly effect” (proposed in 1972) by 20 years on the grounds of Lorentz’s work from 1961.

      By “theory” we mean the equational or functional determination, possibly by a nonlinear system of equations or evolution functions.

      Quantum randomness is hailed to be more than “epistemic”, that is, “intrinsic” (to the theory). However, quantum randomness is not part of the standard mathematical model of the quantum which talks about probabilities, but is about the measurement of individual observables. So, to give more sense to the first statement we need to answer (at least) the following questions: (1) What is the source of quantum randomness? (2) What is the quality of quantum randomness? (3) Is quantum randomness different from classical randomness?

      A naive answer to (1) is to say that quantum mechanics has shown “without doubt” that microscopic phenomena are intrinsically random. For example, we cannot predict with certainty how long it will take for a single unstable atom in a controlled environment to decay, even if one has complete knowledge of the “laws of physics” and the atom’s initial conditions. One can only calculate the probability of decay in a given time, nothing more! This is intrinsic randomness guaranteed.

      Following Einstein’s approach (Einstein et al. 1935), quantum indeterminism corresponds to the absence of physical reality, if reality is what is made accessible by measurement: if no unique element of physical reality corresponding to a particular physical observable (thus, measurable) quantity exists, this is reflected by the physical quantity being indeterminate. This approach needs to be more precisely formalized. The notion of value indefiniteness, as it appears in the theorems of Bell (Bell 1966) and, particularly, Kochen and Specker (1967), has been used as a formal model of quantum indeterminism (Abbott et al. 2012). The model also has empirical support as these theorems have been experimentally tested via the violation of various inequalities (Weihs et al. 1998). We have to be aware that, going along this path, the “belief” in quantum indeterminism rests on the assumptions used by these theorems.

      An observable is value definite for a given quantum system in a particular state if the measurement of that observable is pre-determined to take a (potentially hidden) value. If no such pre-determined value exists, the observable is value indefinite. Formally, this notion can be represented by a (partial) value assignment function (see Abbott et al. (2012) for the complete formalism).

      When should we conclude that a physical quantity is value definite? Einstein, Podolsky and Rosen (EPR) defined physical reality in terms of certainty of predictability in Einstein et al. (1935, p. 777):

      If, without in any way disturbing a system, we can predict with certainty (i.e., with probability equal to unity) the value of a physical quantity, then there exists an element of reality corresponding to that quantity.

      EPR principle: If, without disturbing a system in any way, we can predict with certainty the value of a physical quantity, then there exists a definite value prior to the observation corresponding to this physical quantity.

      The EPR principle justifies:

      Eigenstate principle: a projection observable corresponding to the preparation basis of a quantum state is value definite.

      The СКАЧАТЬ