Many answers (which i still study) focus on the fact that the combined probability of 2 (assumed independent) events in QM is not equal to the sum of the probabilities of each event (sth that holds classicaly by definition). This fact (appears to) makes the formulation of another probability (a quantum one) a necessity.

Yet this again breaks down to assumed independent, if this is not so, the "classic probability" is applicable (as indeed in other areas).

Maybe it's because i am not a physicist, but maybe you should say what you think is different.
–
kutschkemJun 3 '14 at 8:07

3

I would definitely say that 'normal' probability theory is perfectly capable of encompassing QM. The physical meaning that we choose to attach to things in this context may be strange and counterintuitive, but that's a separate issue from the mathematics.
–
DanuJun 3 '14 at 8:08

4

The question is not clear to me. Would you provide some example which the probability is used different than QM?
–
user26143Jun 3 '14 at 8:17

2

In QM the 'or' is very different. The classical rule $P(A or B)=P(A)+P(B)$ for independent events $A$ or $B$ does not generically apply in QM (as e.g. very well illustrated by the famous two slits experiments and the interference pattern). The amplitudes sum up, but then you square it and a new mixed term responsible for the interference appears.
–
TwoBsJun 3 '14 at 8:56

7 Answers
7

Having given it some more thought, there is an unambiguous philosophical difference, with practical implications. The two-slit experiment provides a good example of this.

In a classical universe, any particular photon that hits the screen either went through slit A or slit B. Even if we didn't bother to measure this, one or the other still happened, and we can meaningfully define P(A) and P(B).

In a quantum universe, if we didn't bother to measure which slit a photon went through, then it isn't true that it went through one slit or the other. You might say it went through both, though even that isn't entirely true; all we can really say is that it "went though the slits".

(Asking which slit a photon went through in the two-slit experiment is like asking what the photon's religion is. It simply isn't a meaningful question.)

That means that P(A) and P(B) just don't exist. Here's where one of the practical implications comes in: if you don't understand QM properly [I'm lying a bit here; I'll come back to it] then you can still calculate a probability that the particle went through slit A and a probability that it went through slit B. And then when you try to apply the usual mathematics to those probabilities, it doesn't work, and then you start saying that quantum probability doesn't follow the same rules as classical probability.

(Actually what you're really doing is calculating what the probabilities for those events would have been if you had chosen to measure them. Since you didn't, they're meaningless, and the mathematics doesn't apply.)

So: the philosophical difference is that when studying quantum systems, unlike classical systems, the probability that something would have happened if you had measured it is not in general meaningful unless you actually did; the practical implication is that you have to keep track of what you did or did not measure in order to avoid doing an invalid calculation.

(In classical systems most syntactically valid questions are meaningful; it took me some time to come up with the counter-example given above. In quantum mechanics most questions are not meaningful and you have to know what you're doing to find the ones that are.)

Note that keeping track of whether you've measured something or not is not an abstract exercise restricted to cases where you are trying to apply probability theory. It has a direct and concrete impact on the experiment: in the case of the two-slit experiment, if you measure which slit each photon went through, the interference pattern disappears.

(Trickier still: if you measure which slit each photon went through, and then properly erase the results of that measurement before looking at the film, the interference pattern comes back again.)

PS: it may be unfair to say that calculating a "would-have" probability means that you don't understand QM properly. It may simply mean that you're consciously choosing to use a different interpretation of it, and prefer to modify or generalize your conception of probability as necessary. V. Moretti's answer goes into some detail about how you might go about doing this. However, while this sort of thing is interesting, it does not appear to me to be of any obvious use. (It isn't clear that it gives any insight into the disappearance and reappearance of the interference pattern as described above, for example.) Addendum: that has become clearer following the discussion in the comments. It seems that it is thought that the alternative formulation may have advantages when dealing with more complicated scenarios (QFT on curved spacetime was mentioned as one example). That is entirely plausible, and I certainly don't mean to imply that the work lacks value; however, it is still not clear to me that it is pedagogically useful as an alternative to the conventional approach when learning basic QM.

PPS: depending on interpretation, there may be other philosophical differences related to the nature or origin of randomness. Bayesian statistics is broad enough, I believe, that these differences are not of any great importance, and even from a frequentist viewpoint I don't think they have any practical implications.

The application of probability in areas other than quantum mechanics is a clever way to model situations that are complex enough so that the exact analysis is non feasible, or at least highly tedious.

On the other hand in QM nature is inherently probabilistic. When you make an observation the quantum state your system is in has a probability for each possible outcome. It is no more a trick to make calculations. It is a feature of nature. That is the difference.

+1, thanks, will wait for other answers also
–
Nikos M.Jun 3 '14 at 8:55

should i assume that your answer selects the 2nd option? (from the options given in question)
–
Nikos M.Jun 3 '14 at 8:57

2

@TwoBs: that isn't due to the nature of quantum probability being fundamentally different, it's because the supposedly "mutually exclusive" events aren't. When you consider properly formed questions such as "what is the probability of measurement A returning value B" the probability has the same practical meaning as that in classical physics; whether the philosophical meaning is different depends on your assumptions.
–
Harry JohnstonJun 3 '14 at 21:26

2

@TwoBs: the rule P(A or B) = P(A) + P(B) only applies to independent events. In the two slit experiment, "the particle went through slit A" and "the particle went through slit B" aren't even events, let alone independent ones! The amplitude only predicts what the probability of a given result will be if you do the relevant measurement; since you aren't measuring which slit the particle went through, it is incorrect to interpret it as a probability. P(A) and P(B) are meaningless, they don't exist.
–
Harry JohnstonJun 4 '14 at 20:43

1

@TwoBs: furthermore, if you do measure which slit the particle goes through, so that P(A) and P(B) do have a meaning, then P(A or B) does equal P(A) + P(B) (and the interference pattern disappears). In fact, all the classical probability equations are still true in QM whenever they are meaningful, so I conclude that there is no practical distinction between classical and QM probability.
–
Harry JohnstonJun 4 '14 at 20:51

The probabilities in QM are given by the square amplitudes of the relevant terms in the wavefunction, or by by the expectation value of the relevant projector or POVM. However, it is not the case that those numbers always act in a way that is consistent with the calculus of probability.

For example, if there are two mutually exclusive ways for an event to happen then the calculus of probability would say that the probability for that event is the sum of the probabilities of it happening in each of those ways. But in single photon interference experiments this doesn't seem to work. There are two routes through the interferometer, the photon cannot be detected on both routes at once, so they are mutually exclusive, right? So then to get the probability of the photon emerging from a particular port on the other end you should just add the probability of it going along each route. But that calculation gives the wrong answer: you can get any probability you like by changing the path lengths see:

I think these arguments don't work because the frequency approach itself doesn't work. Why would the relative frequency over an infinite number of samples have anything to do with what is observed in a laboratory? And if there is some explanation, then why are we bothering with this relative frequency stuff rather than using the actual explanation? The best explanation of why it is applicable is the decision theoretic approach:

thanks for the papers, havent seen this approach, will read them through
–
Nikos M.Jun 3 '14 at 9:04

as my comments in other answers, the whole thing breaks down to the fact that the prob of 2 events (assumed independent) is not equal to the sum. Yet this again breaks down to "assumed independent" as a result if not assumed independent (a matter of interpretation), "classic probability" is still as good. This is the heart of the question. Still studying the references on your answer (and others)
–
Nikos M.Jun 4 '14 at 19:00

In this paper it is
shown that quantum theory can be derived from five
very reasonable axioms. The first four of these axioms
are obviously consistent with both quantum theory
and classical probability theory. Axiom 5 (which
requires that there exist continuous reversible transformations
between pure states) rules out classical
probability theory.

good one, will check it out, very interesting i actually learn sth more
–
Nikos M.Jun 3 '14 at 9:05

Not a good one. There are serious problems with that unjustifiably famous essay. The first one is, he does not get involved with the physics of a measurement process, so he never connects the observables with the Hamiltonian of an amplifier. The second one is, his axioms do not lead to any recognised mathematical theory of probability, in fact, his axioms violate the law of large numbers, etc. You can refer to my critique, "Remarks on an attempted axiomatisation of QM by Lucien Hardy", arXiv:quant-ph/0606038 just google on Axiomatisation of Physics (spell it with an esss, not a zed).
–
joseph f. johnsonJun 5 '14 at 2:27

i am still studying the references but violation of the law of large numbers might be a serious issue, nevertheless as i pointed out elsewhere there can be alternative probability theories (or calculi if you like), utilising different numbers or arithmetic (and still have results meaningful in the same way as the "classic" probability calculus)
–
Nikos M.Jun 5 '14 at 16:55

The theory of probability used in QM is intrinsically different from the one commonly used for the following reason: The space of events is non-commutative (more properly non-Boolean) and this fact deeply affects the conditional probability theory. The probability that A happens if B happened is computed differently in classical probability theory and in quantum theory, when A and B are quantistically incompatible events. In both cases probability is a measure on a lattice, but, in the classical case, the lattice is a Boolean one (a $\sigma$-algebra), in the quantum case it is not.

To be clearer, classical probability is a map $\mu: \Sigma(X) \to [0,1]$ such that $\Sigma(X)$ is a class of subsets of the set $X$ including $\emptyset$, closed with respect to the complement and the countable union, and such that $\mu(X)=1$ and:
$$\mu(\cup_{n\in \mathbb N}E_n) = \sum_n \mu(E_n)\quad \mbox{if $E_k \in \Sigma(X)$ with $E_p\cap E_q= \emptyset$ for $p\neq q$.}$$
The elements of $\Sigma(X)$ are the events whose probability is $\mu$. In this view, for instance, if $E,F \in \Sigma(X)$, $E\cap F$ is logically interpreted as the event "$E$ AND $F$".
Similarly $E\cup F$ corresponds to "$E$ OR $F$" and $X\setminus F$ has the meaning of "NOT $F$" and so on.
The probability of $P$ when $Q$ is given verifies $$\mu(P|Q) = \frac{\mu(P \cap Q)}{\mu(Q)}\:.\tag{1}$$

If you instead consider a quantum system, there are "events", i.e. elementary "yes/no" propositions experimentally testable, that cannot by joined by logical operators AND and OR.

An example is $P=$"the $x$ component of this electron is $1/2$" and $Q=$"the $y$ component of this electron is $1/2$". There is no experimental device able to assign a truth value to $P$ and $Q$ simultaneously, so that elementary propositions as "$P$ and $Q$" make no sense. Pairs of propositions like $P$ and $Q$ above are physically incompatible.

In quantum theories (the most elementary version due to von Neumann), the events of a physical system are represented by the orthogonal projectors of a separable Hilbert space $H$. The set ${\cal P}(H)$ of those operators replaces the classical $\Sigma(X)$.

In general, the meaning of $P\in {\cal P}(H)$ is something like
"the value of the observable $Z$ belongs to the subset $I \subset \mathbb R$" for some observable $Z$ and some set $I$. There is a procedure to integrate such a class of projectors labeled on real subsets to construct a self-adjoint operator $\hat{Z}$ associated to the observable $Z$, and this is nothing but the physical meaning of the spectral decomposition theorem.

If $P, Q \in {\cal P}(H)$, there are two possibilities: $P$ and $Q$ commute or they do not.

Von Neumann's fundamental axiom states that commutability is the mathematically corresponding of physical compatibility.

When $P$ and $Q$ commutes, $PQ$ and $P+Q-PQ$ still are orthogonal projectors, that is elements of ${\cal P}(H)$.

In this situation, $PQ$ corresponds to "$P$ AND $Q$", whereas $P+Q-PQ$ corresponds to "$P$ OR $Q$" and so on, in particular "NOT $P$" is always interpreted as the orthogonal projector onto $P(H)^\perp$ (the orthogonal subspace of $P(H)$), and all classical formalism holds true this way.
As a matter of fact, a maximal set of pairwise commuting projectors has formal properties identical to those of classical logic: is a Boolean $\sigma$-algebra.

In this picture, a quantum state is a map assigning the probability $\mu(P)$ that $P$ is experimentally verified to every $P\in {\cal P}(H)$.
It has to satisfy: $\mu(I)=1$ and
$$\mu\left(\sum_{n\in \mathbb N}P_n\right) = \sum_n \mu(P_n)\quad \mbox{if $P_k \in {\cal P}(H)$ with $P_p P_q= P_qP_p =0$ for $p\neq q$.}$$

Celebrated Gleason's Theorem, establishes that, if $\text{dim}(H)>2$, the measures $\mu$ are all of the form $\mu(P)= \text{tr}(\rho_\mu P)$ for some mixed state $\rho_\mu$ (a positive trace-class operator with unit trace), biunivocally determined by $\mu$.
In the convex set of states, the extremal elements are the standard pure states. They are determined, up to a phase, by unit vectors $\psi \in H$, so that, with some trivial computation (completing $\psi_\mu$ to an orthonormal basis of $H$ and using that basis to compute the trace),
$$\mu(P) = \langle \psi_\mu | P \psi_\mu \rangle = ||P \psi_\mu||^2\:.$$

(Nowadays, there is a generalized version of this picture, where the set ${\cal P}(H)$ is replaced by the class of bounded positive operators in $H$ (the so-called "effects") and Gleason's theorem is replaced by Busch's theorem with a very similar statement.)

Quantum probability is therefore given by the map, for a given generally mixed state $\rho$,
$${\cal P}(H) \ni P \mapsto \mu(P) =\text{tr}(\rho_\mu P) $$

It is clear that, as soon as one deals with physically incompatible
propositions, (1) cannot hold just because there is nothing like $P \cap Q$ in the set of physically sensible quantum propositions.
All that is due to the fact that the space of events ${\cal P}(H)$ is now a noncommutative set of projectors, giving rise to a non-Boolean lattice.

Therein, $QPQ$ is an orthogonal projector and can be interpreted as "$P$ AND $Q$" (i.e., $P\cap Q$) when $P$ and $Q$ are compatible. In this case (1) holds true again. (2) gives rise to all "strange things" showing up in quantum experiments (as in the double slit one).

(2) just relies upon the von Neumann-L\"uders reduction postulate stating that, if the outcome of the measurement of $P\in {\cal P}(H)$ is YES when the state was $\mu$ (i.e., $\rho_\mu$), the the state immediately after the measurement is $\mu'$ associated to $\rho_{\mu'}$ with

$$\rho_{\mu'} := \frac{P\rho_\mu P}{\text{tr}(\rho_\mu P)}\:.$$

ADDENDUM. Actually, it is possible to extend the notion of logical operators AND and OR for all pairs of elements in ${\cal P}(H)$ and that was the program of von Neumann and Birkhoff (the quantum logic). In fact just the lattice structure of ${\cal P}(H)$ permits it, or better is it. With this extended notion of AND and OR, "$P$ AND $Q$" is the orthogonal projector onto $P(H)\cap Q(H)$ whereas "$P$ OR $Q$" is the orthogonal projector onto the closure of the space $P(H)+Q(H)$. When $P$ and $Q$ commute these notions of AND and OR reduce to the standard ones. However, with the extended definitions, ${\cal P}(H)$ becomes a lattice in the proper mathematical sense, where the partial order relation is given by the standard inclusion of closed subspaces ($P \geq Q$ means
$P(H) \supset Q(H)$).
The point is that the physical interpretation of this extension of AND and OR is not clear. The resulting lattice is however non-Boolean. In other words, for instance, these extended AND and OR are not distributive as the standard AND and OR are (this reveals their quantum nature). However, also keeping the definition of "NOT $P$" as the orthogonal projector onto $P(H)^\perp$, the found structure of ${\cal P}(H)$ is well known: A $\sigma$-complete, orthomodular, separable, atomic, irreducible and verifying the covering property, lattice. At the end of 1900 it was definitely proved, by Solér, a conjecture due to von Neumann stating that there are only three possibilities for practically realizing such lattices: The lattice of orthogonal projectors in a separable complex Hilbert space, the lattice of orthogonal projectors in a separable real Hilbert space, the lattice of orthogonal projectors in a separable quaternionic Hilbert space. Requiring the existence of time reversal symmetry introduces a complex structure in a real Hilbert space giving rise to a complex Hilbert space. Conversely, still nowadays, it is not obvious if it is possible to rule out quaternionic Hilbert spaces to describe quantum physics.

ADDENDUM2. After a discussion with Harry Johnston, I think that an interpretative remark is worth to be mentioned about the probabilistic content of the state $\mu$ within the picture I illustrated above. In QM $\mu(P)$ is the probability that, if I performed a certain experiment (in order to check $P$), $P$ would turn out to be true.
It seems that there is here a difference with respect to the classical notion of probability applied to classical systems. There, probability mainly refers to something already existent (and to our incomplete knowledge of it). In the formulation of QM I presented above, probability instead refers to that which will happen if...

All probabilities arise from quantum mechanics. As pointed out here, there are no known examples of classical probabilities that don't have a quantum mechanical origin. Here we're not nitpicking about some small quantum effects, whether you consider coin throws, betting on the digits of pi etc., the probabilities can always be shown to be purely of quantum mechanical origin, and therefore the treatment according to quantum mechanics applies. Classical probability theory is thus not fundamental, it should be derived as an appropriate approximation from quantum mechanics.

Absolutely not. Probability theory has a much broader scope than the "physical" probability theory you seem to have in mind. Probability theory can really be seen as an extension of logic, and as such applies to all situations in which only partial information is available, or in which the probability represents subjective assessments of likelihood. You should have a look, for example, at Jaynes' book "Probability Theory: The Logic of Science".
–
Yvan VelenikJun 3 '14 at 19:09

Real randomness (inasmuch as such a thing exist) might indeed have quantum origin, but one does not need anything truly random to be able to apply probability theory successfully, just situations of sufficient complexity that renouncing a precise description is the only way to proceed.
–
Yvan VelenikJun 3 '14 at 19:15

But then for any mathematical theory to be relevant to the real world, it has to be compatible with it, so arguments in the article I linked to will apply.
–
Count IblisJun 3 '14 at 19:31

1

I do not believe the arguments in that article are philosophically sound. It certainly does not describe any of the mainstream views on the nature of probability.
–
Harry JohnstonJun 3 '14 at 21:20

1

@CountIblis: You're missing my point. There is no need to have any randomness to apply probability theory. So there is absolutely no need of quantum mechanics to derive or justify classical probability theory. To say that it requires quantum mechanics because the universe is fundamentally quantum, would be the same as saying that, say, literature is reducible to quantum mechanics because the brains of writers are quantum objects... Maybe true, but utterly useless.
–
Yvan VelenikJun 4 '14 at 16:31

In both cases, probability arises from the need to compare the results from two incompatible models operating at different scales, the microscopic and the macroscopic.

Darwin and Fowler long ago showed how to derive Classical Statistical Mechanics, the main place in classical physics where probabilities occur, from Quantum Mechanics. So in a sense, Quantum Mechanics is fundamental and there is no problem deriving the Classical case from it.
Fowler, Statistical Mechanics

But I will present them in the other order, anyway.
In Classical physics, if one is analysing, say, an ideal gas, the system of 10^23 particles is deterministic. And the number of variables is 6 times 10^23. This is the microscopic view of the system as a whole. But one can also study certain properties of this gas in terms of a very few thermodynamic variables, temperature pressure and volume, which describe a macro-state. But in terms of this description, the system is probabilistic: one only knows the probabilities with which its molecules will possess a given energy, etc. Furthermore, the connection between the two levels of description of the system, the micro-level and the macro-level, is via measurement. The measurement of the velocity of a molecule is modelled by the long-time average over its trajectory of its velocity. Then it turns out that for all normal molecules, this procedure, provided the system is in equilibrium, yields the same answer almost without regard to which molecule or which trajectory you study, and Einstein defined this as the probabilistic expectation of the energy of a molecule. See Jan von Plato, Creating Modern Probability.
So only the results of measurements are assigned probabilities.

Now, according to Feynman and others, something parallel is true in Quantum Mechanics. The probabilities arise from the necessity to amplify micro-phenomena up to the macro-level where we can see the measuring apparatus, see a needle on a dial pointing to a number on the dial. (Schroedinger's equation is itself a deterministic equation and probabilities only come in to the measurement axioms.) The only "events" in the sense of mathematical probability theory, i.e., things which are assigned probabilities, are the results of measurements. And here, too, the measurement has something to do with describing in a reduced fashion the state of a micro-system in terms of macro-states instead of its micro-states. The needle on the dial really obeys the laws of quantum mechanics: it has a wave function, it is in an entangled state, etc., but when we say "the result of the measurement ws that the needle pointed to 3" we are describing the measurement apparatus in classical terms, which are macro-terms. The passage from the micro-description of the particle in terms of quantum concepts to this reduced description brings in probabilities.

====== what probabilities are not

It is a myth that the probabilities in classical statistical mechanics are due to ignorance or are subjective. They come about only because one restricts one's attention
to the normal cell of micro-states (normal cell in the sense of Darwin and Fowler) and
ignores exceptional states. The definition of "normal" is an objective one: states can be grouped into cells of states: all those states which possess the same time-average properties as each other. The normal cell is the largest cell. In the thermodynamic limit, the normal cell is not only the largest, it is is the only one with positive volume, all the other cells are mere boundaries with lower dimension.

It is a myth that the probabilities in Quantum Mechanics are somehow "non-commutative". The problem is not that there are non-commuting observables. If you are measuring momentum, the experimental setup is quite definite, and the space of events depends on the physical setup, and only has the results of measuring momentum. If the measuring apparatus is one suitable to measure momentum, then results for position are not events. The setup excludes measuring position, so measurements of position are impossible in this setup. And conversely. There is no one, over-arching, probability space
with both kinds of events in it, as mathematicians who study so-called "Quantum Probability" or "Non-commutative Probability" naively suppose. Bohr taught us that if you setup the apparatus for one type of measurement (e.g., momentum), you physically exclude the possibility of the other type of complementary measurement (e.g., position). That means that you either work in one probability space with events and normal measures of their probability, or you are in a totally other probability space with its own events and its own measure. Now, no one would say that an operator on space A either commuted with or didnot commute with an operator on a totally different space B, and that is what we have here.

nice explanation of the probability spaces for a quantum experiment, i assume the answer implies the difference is not in applying probability theory but in physical interpretation?
–
Nikos M.Jun 5 '14 at 16:35

The applications are very similar, except that amplification is not involved per se in Classical Stat Mech. By ignoring the Heisenberg uncertainty principle, Classical Stat MEch assumes that you really could track the trajectory of an individual molecule and calculate its time average. That does not make sense quantum mechanically, so that is a difference. But the meaning of probability is the same in both cases, the one not involving amplificaiton and the one involving amplification.
–
joseph f. johnsonJun 6 '14 at 3:04