Quantum field theory

"Relativistic quantum field theory" redirects here. For other uses, see Relativity.

In theoretical physics, quantum field theory (QFT) is a theoretical framework for constructing quantum mechanical models of subatomic particles in particle physics and quasiparticles in condensed matter physics. A QFT treats particles as excited states of an underlying physical field, so these are called field quanta.

In quantum field theory, quantum mechanical interactions between particles are described by interaction terms between the corresponding underlying quantum fields. These interactions are conveniently visualized by Feynman diagrams, that also serve as a formal tool to evaluate various processes.

Historically the development began in the 1920s with the quantization of the electromagnetic field, the quantization being based on an analogy with the eigenmode expansion of a vibrating string with fixed endpoints. In Weinberg (2005), QFT is brought forward as an unavoidable consequence of the reconciliation of quantum mechanics with special relativity.

History

Early development

Max Born (18821970), one of the founders of quantum field theory.

He is also known for the Born rule that introduced the probabilistic interpretation in quantum mechanics. He received the 1954 Nobel Prize in Physics together with Walther Bothe.

The first achievement of quantum field theory, namely quantum electrodynamics (QED), is “still the paradigmatic example of a successful quantum field theory” according to Weinberg (2005). Ordinary QM cannot give an account of photons, which constitute the prime case of relativistic ‘particles’. Since photons have rest mass zero, and correspondingly travel in the vacuum at the speed c, a non-relativistic theory such as ordinary QM cannot give even an approximate description. Photons are implicit in the emission and absorption processes which have to be postulated, for instance, when one of an atom's electrons makes a transition between energy levels. The formalism of QFT is needed for an explicit description of photons. In fact most topics in the early development of quantum theory (the so-called old quantum theory, 1900–25) were related to the interaction of radiation and matter and thus should be treated by quantum field theoretical methods. However, quantum mechanics as formulated by Dirac, Heisenberg, and Schrödinger in 1926–27 started from atomic spectra and did not focus much on problems of radiation.

As soon as the conceptual framework of quantum mechanics was developed, a small group of theoreticians tried to extend quantum methods to electromagnetic fields. A good example is the famous paper by Born, Jordan & Heisenberg (1926). P. Jordan was especially acquainted with the literature on light quanta and made important contributions to QFT. The basic idea was that in QFT the electromagnetic field should be represented by matrices in the same way that position and momentum were represented in QM by matrices in matrix mechanics. The ideas of QM were extended to systems having an infinite number of degrees of freedom.

The inception of QFT is usually considered to be Dirac's famous 1927 paper on “The quantum theory of the emission and absorption of radiation”.[1] Here Dirac coined the name “quantum electrodynamics” (QED) for the part of QFT that was developed first. Dirac supplied a systematic procedure for transferring the characteristic quantum phenomenon of discreteness of physical quantities from the quantum-mechanical treatment of particles to a corresponding treatment of fields. Employing the theory of the quantum harmonic oscillator, Dirac gave a theoretical description of how photons appear in the quantization of the electromagnetic radiation field. Later, Dirac's procedure became a model for the quantization of other fields as well. These first approaches to QFT were further developed during the following three years. P. Jordan introduced creation and annihilation operators for fields obeying Fermi–Dirac statistics. These differ from the corresponding operators for Bose–Einstein statistics in that the former satisfy anti-commutation relations while the latter satisfy commutation relations.

The methods of QFT could be applied to derive equations resulting from the quantum-mechanical (field-like) treatment of particles, e.g. the Dirac equation, the Klein–Gordon equation and the Maxwell equations. Schweber points out[2] that the idea and procedure of second quantization goes back to Jordan, in a number of papers from 1927,[3] while the expression itself was coined by Dirac. Some difficult problems concerning commutation relations, statistics, and Lorentz invariance were eventually solved. The first comprehensive account of a general theory of quantum fields, in particular the method of canonical quantization, was presented by Heisenberg & Pauli in 1929. Whereas Jordan's second quantization procedure applied to the coefficients of the normal modes of the field, Heisenberg & Pauli started with the fields themselves and subjected them to the canonical procedure. Heisenberg and Pauli thus established the basic structure of QFT as presented in modern introductions to QFT. Fermi and Dirac, as well as Fock and Podolsky, presented different formulations which played a heuristic role in the following years.

Quantum electrodynamics rests on two pillars, see e.g., the short and lucid “Historical Introduction” of Scharf (2014). The first pillar is the quantization of the electromagnetic field, i.e., it is about photons as the quantized excitations or 'quanta' of the electromagnetic field. This procedure will be described in some more detail in the section on the particle interpretation. As Weinberg points out the “photon is the only particle that was known as a field before it was detected as a particle” so that it is natural that QED began with the analysis of the radiation field.[4] The second pillar of QED consists of the relativistic theory of the electron, centered on the Dirac equation.

The problem of infinities

The emergence of infinities

Pascual Jordan (19021980), doctoral student of Max Born, was a pioneer in quantum field theory, coauthoring a number of seminal papers with Born and Heisenberg.

Jordan algebras were introduced by him to formalize the notion of an algebra of observables in quantum mechanics. He was awarded the Max Planck medal 1954.

Quantum field theory started with a theoretical framework that was built in analogy to quantum mechanics. Although there was no unique and fully developed theory, quantum field theoretical tools could be applied to concrete processes. Examples are the scattering of radiation by free electrons, Compton scattering, the collision between relativistic electrons or the production of electron-positron pairs by photons. Calculations to the first order of approximation were quite successful, but most people working in the field thought that QFT still had to undergo a major change. On the one side some calculations of effects for cosmic rays clearly differed from measurements. On the other side and, from a theoretical point of view more threatening, calculations of higher orders of the perturbation series led to infinite results. The self-energy of the electron as well as vacuum fluctuations of the electromagnetic field seemed to be infinite. The perturbation expansions did not converge to a finite sum and even most individual terms were divergent.

The various forms of infinities suggested that the divergences were more than failures of specific calculations. Many physicists tried to avoid the divergences by formal tricks (truncating the integrals at some value of momentum, or even ignoring infinite terms) but such rules were not reliable, violated the requirements of relativity and were not considered as satisfactory. Others came up with first ideas of coping with infinities by a redefinition of the parameters of the theory and using a measured finite value, for example of the charge of the electron, instead of the infinite ‘bare’ value. This process is called renormalization.

From the point of view of philosophy of science it is remarkable that these divergences did not give enough reason to discard the theory. The years from 1930 to the beginning of World War II were characterized by a variety of attitudes towards QFT. Some physicists tried to circumvent the infinities by more-or-less arbitrary prescriptions, others worked on transformations and improvements of the theoretical framework. Most of the theoreticians believed that QED would break down at high energies. There was also a considerable number of proposals in favour of alternative approaches. These proposals included changes in the basic concepts e.g. negative probabilities and interactions at a distance instead of a field theoretical approach, and a methodological change to phenomenological methods that focusses on relations between observable quantities without an analysis of the microphysical details of the interaction, the so-called S-matrix theory where the basic elements are amplitudes for various scattering processes.

Despite the feeling that QFT was imperfect and lacking rigour, its methods were extended to new areas of applications. In 1933 Fermi's theory of the beta decay started with conceptions describing the emission and absorption of photons, transferred them to beta radiation and analyzed the creation and annihilation of electrons and neutrinos described by the weak interaction. Further applications of QFT outside of quantum electrodynamics succeeded in nuclear physics with the strong interaction. In 1934 Pauli & Weisskopf showed that a new type of fields (scalar fields), described by the Klein–Gordon equation, could be quantized. This is another example of second quantization. This new theory for matter fields could be applied a decade later when new particles, pions, were detected.

The taming of infinities

Werner Heisenberg (19011976), doctoral student of Arnold Sommerfeld, was one of the founding fathers of quantum mechanics.

In particular, he introduced the version of quantum mechanics known as matrix mechanics, but is now more known for the Heisenberg uncertainty relations. He was awarded the Nobel prize in physics 1932 together with Erwin Schrödinger and Paul Dirac.

After the end of World War II more reliable and effective methods for dealing with infinities in QFT were developed, namely coherent and systematic rules for performing relativistic field theoretical calculations, and a general renormalization theory. On three famous conferences, the Shelter Island Conference 1947, the Pocono Conference 1948, and the 1949 Oldstone Conference, developments in theoretical physics were confronted with relevant new experimental results. In the late forties there were two different ways to address the problem of divergences. One of these was discovered by Richard Feynman, the other one (based on an operator formalism) by Julian Schwinger and independently by Sin-Itiro Tomonaga. In 1949 Freeman Dyson showed that the two approaches are in fact equivalent. Thus, Freeman Dyson, Feynman, Schwinger and Tomonaga became the inventors of renormalization theory. The most spectacular experimental successes of renormalization theory were the calculations of the anomalous magnetic moment of electron and the Lamb shift in the spectrum of hydrogen. These successes were so outstanding because the theoretical results were in better agreement with high precision experiments than anything in physics before. Nevertheless, mathematical problems lingered on and prompted a search for rigorous formulations (to be discussed in the main article).

The basic idea of renormalization is to avoid divergences that appear in physical predictions by shifting them into a part of the theory where they do not influence empirical propositions. Dyson could show that a rescaling of charge and mass (‘renormalization’) is sufficient to remove all divergences in QED to all orders of perturbation theory. In general, a QFT is called renormalizable, if all infinities can be absorbed into a redefinition of a finite number of coupling constants and masses. A consequence is that the physical charge and mass of the electron must be measured and cannot be computed from first principles. Perturbation theory gives well defined predictions only in renormalizable quantum field theories, and luckily QED, the first fully developed QFT, belonged to this class of renormalizable theories. There are various technical procedures to renormalize a theory. One way is to cut off the integrals in the calculations at a certain value Λ of the momentum which is large but finite. This cut-off procedure is successful if, after taking the limit Λ → ∞, the resulting quantities are independent of Λ. Part II of Peskin & Schroeder (1995) gives an extensive description of renormalization.

Richard Feynman (19181988)
His 1945 PhD thesis developed the path integral formulation of ordinary quantum mechanics. This was later generalized to field theory.

Feynman's formulation of QED is of special interest from a philosophical point of view. His so-called space-time approach is visualized by the famous Feynman diagrams that look like depicting paths of particles. Feynman's method of calculating scattering amplitudes is based on the functional integral formulation of field theory.[5] A set of graphical rules can be derived so that the probability of a specific scattering process can be calculated by drawing a diagram of that process and then using the diagram to write down the mathematical expressions for calculating its amplitude. The diagrams provide an effective way to organize and visualize the various terms in the perturbation series, and they seem to display the flow of electrons and photons during the scattering process. External lines in the diagrams represent incoming and outgoing particles, internal lines are connected with virtual particles and vertices with interactions. Each of these graphical elements is associated with mathematical expressions that contribute to the amplitude of the respective process. The diagrams are part of Feynman's very efficient and elegant algorithm for computing the probability of scattering processes. The idea of particles travelling from one point to another was heuristically useful in constructing the theory. This heuristics, based on Huygen's principle, is useful for concrete calculations and actually give the correct particle propagators as derived more rigorously.[6] Nevertheless, an analysis of the theoretical justification of the space-time approach shows that its success does not imply that particle paths have to be taken seriously. General arguments against a particle interpretation of QFT clearly exclude that the diagrams represent paths of particles in the interaction area. Feynman himself was not particularly interested in ontological questions.

Gauge theory and the standard model

In the beginning of the 1950s QED had become a reliable theory which no longer counted as preliminary. It took two decades from writing down the first equations until QFT could be applied to interesting physical problems in a systematic way. The new developments made it possible to apply QFT to new particles and new interactions. In the following decades QFT was extended to describe not only the electromagnetic force, but also weak and strong interaction so that new Lagrangians had to be found which contain new classes of ‘particles’ or quantum fields. The research aimed at a more comprehensive theory of matter and in the end at a unified theory of all interactions.

New theoretical concepts had to be introduced, mainly connected with non-Abelian gauge theories (the effort of developing such theories started in 1954 with the work of Yang and Mills) and spontaneous symmetry breaking. Today there are trustworthy theories of the strong, weak, and electromagnetic interactions of elementary particles which have a similar structure as QED. A combined theory associated with the gauge group SU(3) × SU(2) × U(1) is considered as the standard model of elementary particle physics which was achieved by Sheldon Glashow, Steven Weinberg and Abdul Salam in 1968, and Frank Wilczek, David Gross and David Politzer in 1973.

According to the standard model there are, on the one side, six types of leptons (e.g. the electron and its neutrino) and six types of quarks, where the members of both group are all fermions with spin 1/2. On the other side, there are spin 1 particles (thus bosons) that mediate the interaction between elementary particles and the fundamental forces, namely the photon for electromagnetic interaction, two W and one Z-boson for weak interaction, and the gluon for strong interaction. Altogether there is good agreement with experimental data, for example the masses of W+ and W bosons (detected in 1983) confirmed the theoretical prediction within one per cent deviation.

Common trends in particle, condensed matter and statistical physics

Renormalization group theory

Parallel developments in the understanding of phase transitions in condensed matter physics led to the study of the renormalization group. This involved the work of Leo Kadanoff (1966) and Michael Fisher (1973), which led to the seminal reformulation of quantum field theory by Kenneth G. Wilson in 1975.

Conformal field theory

During the same period, Kadanoff (1969) introduced an operator algebra formalism for the two-dimensional Ising model, a widely studied mathematical model of ferromagnetism in statistical physics. This development suggested that quantum field theory describes its scaling limit. Later, there developed the idea that a finite number of generating operators could represent all the correlation functions of the Ising model. In the 1980s, the existence of a much stronger symmetry for the scaling limit of two-dimensional critical systems was suggested by Alexander Belavin, Alexander Polyakov and Alexander Zamolodchikov in 1984, which eventually led to the development of conformal field theory,[7] a special case of quantum field theory, which is presently employed successfully in different areas of particle physics and condensed matter physics.

Historiography

The first chapter in Weinberg (1995) is a very good short description of the earlier history of QFT. Detailed accounts of the historical development of QFT can be found, e.g., in Darrigol 1986, Schweber (1994) and Cao 1997a. Various historical and conceptual studies of the standard model are gathered in Hoddeson et al. 1997 and of renormalization theory in Brown 1993.

Definition

Quantum electrodynamics (QED) has one electron field and one photon field; quantum chromodynamics (QCD) has one field for each type of quark; and, in condensed matter, there is an atomic displacement field that gives rise to phonon particles.

Dynamics

Ordinary quantum mechanical systems have a fixed number of particles, with each particle having a finite number of degrees of freedom. In contrast, the excited states of a quantum field can represent any number of particles. This makes quantum field theories especially useful for describing systems where the particle count/number may change over time, a crucial feature of relativistic dynamics.

States

QFT interaction terms are similar in spirit to those between charges with electric and magnetic fields in Maxwell's equations. However, unlike the classical fields of Maxwell's theory, fields in QFT generally exist in quantum superpositions of states and are subject to the laws of quantum mechanics.

Because the fields are continuous quantities over space, there exist excited states with arbitrarily large numbers of particles in them, providing QFT systems with an effectively infinite number of degrees of freedom. Infinite degrees of freedom can easily lead to divergences of calculated quantities (e.g., the quantities become infinite). Techniques such as renormalization of QFT parameters or discretization of spacetime, as in lattice QCD, are often used to avoid such infinities so as to yield physically plausible results.

Fields and radiation

The gravitational field and the electromagnetic field are the only two fundamental fields in nature that have infinite range and a corresponding classical low-energy limit, which greatly diminishes and hides their "particle-like" excitations. Albert Einstein in 1905, attributed "particle-like" and discrete exchanges of momenta and energy, characteristic of "field quanta", to the electromagnetic field. Originally, his principal motivation was to explain the thermodynamics of radiation. Although the photoelectric effect and Compton scattering strongly suggest the existence of the photon, it might alternately be explained by a mere quantization of emission; more definitive evidence of the quantum nature of radiation is now taken up into modern quantum optics as in the antibunching effect.[8]

Varieties of approaches

There is currently no complete quantum theory of the remaining fundamental force, gravity. Many of the proposed theories to describe gravity as a QFT postulate the existence of a graviton particle that mediates the gravitational force. Presumably, the as yet unknown correct quantum field-theoretic treatment of the gravitational field will behave like Einstein's general theory of relativity in the low-energy limit. Quantum field theory of the fundamental forces itself has been postulated to be the low-energy effective field theory limit of a more fundamental theory such as superstring theory.

Most theories in standard particle physics are formulated as relativistic quantum field theories, such as QED, QCD, and the Standard Model. QED, the quantum field-theoretic description of the electromagnetic field, approximately reproduces Maxwell's theory of electrodynamics in the low-energy limit, with small non-linear corrections to the Maxwell equations required due to virtual electron–positron pairs.

In the perturbative approach to quantum field theory, the full field interaction terms are approximated as a perturbative expansion in the number of particles involved. Each term in the expansion can be thought of as forces between particles being mediated by other particles. In QED, the electromagnetic force between two electrons is caused by an exchange of photons. Similarly, intermediate vector bosons mediate the weak force and gluons mediate the strong force in QCD. The notion of a force-mediating particle comes from perturbation theory, and does not make sense in the context of non-perturbative approaches to QFT, such as with bound states.

Principles

Classical and quantum fields

A classical field is a function defined over some region of space and time.[9] Two physical phenomena which are described by classical fields are Newtonian gravitation, described by Newtonian gravitational field g(x, t), and classical electromagnetism, described by the electric and magnetic fields E(x, t) and B(x, t). Because such fields can in principle take on distinct values at each point in space, they are said to have infinite degrees of freedom.[9]

Classical field theory does not, however, account for the quantum-mechanical aspects of such physical phenomena. For instance, it is known from quantum mechanics that certain aspects of electromagnetism involve discrete particlesphotonsrather than continuous fields. The business of quantum field theory is to write down a field that is, like a classical field, a function defined over space and time, but which also accommodates the observations of quantum mechanics. This is a quantum field.

It is not immediately clear how to write down such a quantum field, since quantum mechanics has a structure very unlike a field theory. In its most general formulation, quantum mechanics is a theory of abstract operators (observables) acting on an abstract state space (Hilbert space), where the observables represent physically observable quantities and the state space represents the possible states of the system under study.[10] For instance, the fundamental observables associated with the motion of a single quantum mechanical particle are the position and momentum operators \hat{x} and \hat{p}. Field theory, in contrast, treats x as a way to index the field rather than as an operator.[11]

There are two common ways of developing a quantum field: the path integral formalism and canonical quantization.[12] The latter of these is pursued in this article.

Lagrangian formalism

Quantum field theory frequently makes use of the Lagrangian formalism from classical field theory. This formalism is analogous to the Lagrangian formalism used in classical mechanics to solve for the motion of a particle under the influence of a field. In classical field theory, one writes down a Lagrangian density, \mathcal{L}, involving a field, φ(x,t), and possibly its first derivatives (∂φ/∂t and ∇φ), and then applies a field-theoretic form of the Euler–Lagrange equation. Writing coordinates (t, x) = (x0, x1, x2, x3) = xμ, this form of the Euler–Lagrange equation is[9]

\frac{\partial}{\partial x^\mu} \left[\frac{\partial\mathcal{L}}{\partial(\partial\varphi/\partial x^\mu)}\right] - \frac{\partial\mathcal{L}}{\partial\varphi} = 0,

where a sum over μ is performed according to the rules of Einstein notation.

By solving this equation, one arrives at the "equations of motion" of the field.[9] For example, if one begins with the Lagrangian density

 \mathcal{L}(\varphi,\nabla\varphi) = -\rho(t,\mathbf{x})\,\varphi(t,\mathbf{x}) - \frac{1}{8\pi G}|\nabla\varphi|^2,

and then applies the Euler–Lagrange equation, one obtains the equation of motion

 4\pi G \rho(t,\mathbf{x}) = \nabla^2 \varphi.

This equation is Newton's law of universal gravitation, expressed in differential form in terms of the gravitational potential φ(t, x) and the mass density ρ(t, x). Despite the nomenclature, the "field" under study is the gravitational potential, φ, rather than the gravitational field, g. Similarly, when classical field theory is used to study electromagnetism, the "field" of interest is the electromagnetic four-potential (V/c, A), rather than the electric and magnetic fields E and B.

Quantum field theory uses this same Lagrangian procedure to determine the equations of motion for quantum fields. These equations of motion are then supplemented by commutation relations derived from the canonical quantization procedure described below, thereby incorporating quantum mechanical effects into the behavior of the field.

Single- and many-particle quantum mechanics

In non-relativistic quantum mechanics, a particle (such as an electron or proton) is described by a complex wavefunction, ψ(x, t), whose time-evolution is governed by the Schrödinger equation:

-\frac{{\hbar}^2}{2m}\frac{{\partial}^2}{\partial x^2}\psi(x,t) + V(x)\psi(x,t) = i \hbar \frac{\partial}{\partial t} \psi(x,t).

Here m is the particle's mass and V(x) is the applied potential. Physical information about the behavior of the particle is extracted from the wavefunction by constructing expected values for various quantities; for example, the expected value of the particle's position is given by integrating ψ*(x) x ψ(x) over all space, and the expected value of the particle's momentum is found by integrating ψ*(x)dψ/dx. The quantity ψ*(x)ψ(x) is itself in the Copenhagen interpretation of quantum mechanics interpreted as a probability density function. This treatment of quantum mechanics, where a particle's wavefunction evolves against a classical background potential V(x), is sometimes called first quantization.

This description of quantum mechanics can be extended to describe the behavior of multiple particles, so long as the number and the type of particles remain fixed. The particles are described by a wavefunction ψ(x1, x2, , xN, t), which is governed by an extended version of the Schrödinger equation.

Often one is interested in the case where N particles are all of the same type (for example, the 18 electrons orbiting a neutral argon nucleus). As described in the article on identical particles, this implies that the state of the entire system must be either symmetric (bosons) or antisymmetric (fermions) when the coordinates of its constituent particles are exchanged. This is achieved by using a Slater determinant as the wavefunction of a fermionic system (and a Slater permanent for a bosonic system), which is equivalent to an element of the symmetric or antisymmetric subspace of a tensor product.

For example, the general quantum state of a system of N bosons is written as

 |\phi_1 \cdots \phi_N \rang = \sqrt{\frac{\prod_j N_j!}{N!}} \sum_{p\in S_N} |\phi_{p(1)}\rang \otimes \cdots \otimes |\phi_{p(N)} \rang,

where |\phi_i\rang are the single-particle states, Nj is the number of particles occupying state j, and the sum is taken over all possible permutations p acting on N elements. In general, this is a sum of N! (N factorial) distinct terms. \sqrt{\frac{\prod_j N_j!}{N!}} is a normalizing factor.

There are several shortcomings to the above description of quantum mechanics, which are addressed by quantum field theory. First, it is unclear how to extend quantum mechanics to include the effects of special relativity.[13] Attempted replacements for the Schrödinger equation, such as the Klein–Gordon equation or the Dirac equation, have many unsatisfactory qualities; for instance, they possess energy eigenvalues that extend to ∞, so that there seems to be no easy definition of a ground state. It turns out that such inconsistencies arise from relativistic wavefunctions not having a well-defined probabilistic interpretation in position space, as probability conservation is not a relativistically covariant concept. The second shortcoming, related to the first, is that in quantum mechanics there is no mechanism to describe particle creation and annihilation;[14] this is crucial for describing phenomena such as pair production, which result from the conversion between mass and energy according to the relativistic relation E = mc2.

Second quantization

Main article: Second quantization

In this section, we will describe a method for constructing a quantum field theory called second quantization. This basically involves choosing a way to index the quantum mechanical degrees of freedom in the space of multiple identical-particle states. It is based on the Hamiltonian formulation of quantum mechanics.

Several other approaches exist, such as the Feynman path integral,[15] which uses a Lagrangian formulation. For an overview of some of these approaches, see the article on quantization.

Bosons

For simplicity, we will first discuss second quantization for bosons, which form perfectly symmetric quantum states. Let us denote the mutually orthogonal single-particle states which are possible in the system by |\phi_1\rang, |\phi_2\rang, |\phi_3\rang, and so on. For example, the 3-particle state with one particle in state |\phi_1\rang and two in state |\phi_2\rang is

 \frac{1}{\sqrt{3}} \left[ |\phi_1\rang |\phi_2\rang
|\phi_2\rang + |\phi_2\rang |\phi_1\rang |\phi_2\rang + |\phi_2\rang
|\phi_2\rang |\phi_1\rang \right].

The first step in second quantization is to express such quantum states in terms of occupation numbers, by listing the number of particles occupying each of the single-particle states |\phi_1\rang, |\phi_2\rang, etc. This is simply another way of labelling the states. For instance, the above 3-particle state is denoted as

 |1, 2, 0, 0, 0, \dots \rangle.

An N-particle state belongs to a space of states describing systems of N particles. The next step is to combine the individual N-particle state spaces into an extended state space, known as Fock space, which can describe systems of any number of particles. This is composed of the state space of a system with no particles (the so-called vacuum state, written as |0\rang), plus the state space of a 1-particle system, plus the state space of a 2-particle system, and so forth. States describing a definite number of particles are known as Fock states: a general element of Fock space will be a linear combination of Fock states. There is a one-to-one correspondence between the occupation number representation and valid boson states in the Fock space.

At this point, the quantum mechanical system has become a quantum field in the sense we described above. The field's elementary degrees of freedom are the occupation numbers, and each occupation number is indexed by a number j indicating which of the single-particle states |\phi_1\rang, |\phi_2\rang,\dots,|\phi_j\rang,\dots it refers to:

 | N_1, N_2, N_3, \dots, N_j, \dots \rang .

The properties of this quantum field can be explored by defining creation and annihilation operators, which add and subtract particles. They are analogous to ladder operators in the quantum harmonic oscillator problem, which added and subtracted energy quanta. However, these operators literally create and annihilate particles of a given quantum state. The bosonic annihilation operator a_2 and creation operator a_2^\dagger are easily defined in the occupation number representation as having the following effects:

 a_2 | N_1, N_2, N_3, \dots \rang = \sqrt{N_2} \mid N_1, (N_2 - 1), N_3, \dots \rang,
 a_2^\dagger | N_1, N_2, N_3, \dots \rang = \sqrt{N_2 + 1} \mid N_1, (N_2 + 1), N_3, \dots \rang.

It can be shown that these are operators in the usual quantum mechanical sense, i.e. linear operators acting on the Fock space. Furthermore, they are indeed Hermitian conjugates, which justifies the way we have written them. They can be shown to obey the commutation relation


\left[a_i , a_j \right] = 0 \quad,\quad
\left[a_i^\dagger , a_j^\dagger \right] = 0 \quad,\quad
\left[a_i , a_j^\dagger \right] = \delta_{ij},

where \delta stands for the Kronecker delta. These are precisely the relations obeyed by the ladder operators for an infinite set of independent quantum harmonic oscillators, one for each single-particle state. Adding or removing bosons from each state is therefore analogous to exciting or de-exciting a quantum of energy in a harmonic oscillator.

Applying an annihilation operator a_k followed by its corresponding creation operator a_k^\dagger returns the number N_k of particles in the kth single-particle eigenstate:

a_k^\dagger\,a_k|\dots, N_k, \dots \rangle=N_k| \dots, N_k, \dots \rangle.

The combination of operators a_k^\dagger a_k is known as the number operator for the kth eigenstate.

The Hamiltonian operator of the quantum field (which, through the Schrödinger equation, determines its dynamics) can be written in terms of creation and annihilation operators. For instance, for a field of free (non-interacting) bosons, the total energy of the field is found by summing the energies of the bosons in each energy eigenstate. If the kth single-particle energy eigenstate has energy E_k and there are N_k bosons in this state, then the total energy of these bosons is E_k N_k. The energy in the entire field is then a sum over k:

 E_\mathrm{tot} = \sum_k E_k N_k

This can be turned into the Hamiltonian operator of the field by replacing N_k with the corresponding number operator, a_k^\dagger a_k. This yields

H = \sum_k E_k \, a^\dagger_k \,a_k.

Fermions

It turns out that a different definition of creation and annihilation must be used for describing fermions. According to the Pauli exclusion principle, fermions cannot share quantum states, so their occupation numbers Ni can only take on the value 0 or 1. The fermionic annihilation operators c and creation operators c^\dagger are defined by their actions on a Fock state thus

 c_j | N_1, N_2, \dots, N_j = 0, \dots \rangle = 0
 c_j | N_1, N_2, \dots, N_j = 1, \dots \rangle = (-1)^{(N_1 + \cdots + N_{j-1})} | N_1, N_2, \dots, N_j = 0, \dots \rangle
 c_j^\dagger | N_1, N_2, \dots, N_j = 0, \dots \rangle = (-1)^{(N_1 + \cdots + N_{j-1})} | N_1, N_2, \dots, N_j = 1, \dots \rangle
 c_j^\dagger | N_1, N_2, \dots, N_j = 1, \dots \rangle = 0.

These obey an anticommutation relation:


\left\{c_i , c_j \right\} = 0 \quad,\quad
\left\{c_i^\dagger , c_j^\dagger \right\} = 0 \quad,\quad
\left\{c_i , c_j^\dagger \right\} = \delta_{ij}.

One may notice from this that applying a fermionic creation operator twice gives zero, so it is impossible for the particles to share single-particle states, in accordance with the exclusion principle.

Field operators

We have previously mentioned that there can be more than one way of indexing the degrees of freedom in a quantum field. Second quantization indexes the field by enumerating the single-particle quantum states. However, as we have discussed, it is more natural to think about a "field", such as the electromagnetic field, as a set of degrees of freedom indexed by position.

To this end, we can define field operators that create or destroy a particle at a particular point in space. In particle physics, these operators turn out to be more convenient to work with, because they make it easier to formulate theories that satisfy the demands of relativity.

Single-particle states are usually enumerated in terms of their momenta (as in the particle in a box problem.) We can construct field operators by applying the Fourier transform to the creation and annihilation operators for these states. For example, the bosonic field annihilation operator \phi(\mathbf{r}) is

\phi(\mathbf{r}) \ \stackrel{\mathrm{def}}{=}\   \sum_{j} e^{i\mathbf{k}_j\cdot \mathbf{r}} a_{j}.

The bosonic field operators obey the commutation relation


\left[\phi(\mathbf{r}) , \phi(\mathbf{r'}) \right] = 0 \quad,\quad
\left[\phi^\dagger(\mathbf{r}) , \phi^\dagger(\mathbf{r'}) \right] = 0 \quad,\quad
\left[\phi(\mathbf{r}) , \phi^\dagger(\mathbf{r'}) \right] = \delta^3(\mathbf{r} - \mathbf{r'})

where \delta(x) stands for the Dirac delta function. As before, the fermionic relations are the same, with the commutators replaced by anticommutators.

The field operator is not the same thing as a single-particle wavefunction. The former is an operator acting on the Fock space, and the latter is a quantum-mechanical amplitude for finding a particle in some position. However, they are closely related, and are indeed commonly denoted with the same symbol. If we have a Hamiltonian with a space representation, say

H = - \frac{\hbar^2}{2m} \sum_i \nabla_i^2 + \sum_{i < j} U(|\mathbf{r}_i - \mathbf{r}_j|)

where the indices i and j run over all particles, then the field theory Hamiltonian (in the non-relativistic limit and for negligible self-interactions) is

H = - \frac{\hbar^2}{2m} \int d^3\!r \ \phi^\dagger(\mathbf{r}) \nabla^2 \phi(\mathbf{r}) + \frac{1}{2}\int\!d^3\!r \int\!d^3\!r' \; \phi^\dagger(\mathbf{r}) \phi^\dagger(\mathbf{r}') U(|\mathbf{r} - \mathbf{r}'|) \phi(\mathbf{r'}) \phi(\mathbf{r}).

This looks remarkably like an expression for the expectation value of the energy, with \phi playing the role of the wavefunction. This relationship between the field operators and wavefunctions makes it very easy to formulate field theories starting from space-projected Hamiltonians.

Dynamics

Once the Hamiltonian operator is obtained as part of the canonical quantization process, the time dependence of the state is described with the Schrödinger equation, just as with other quantum theories. Alternatively, the Heisenberg picture can be used where the time dependence is in the operators rather than in the states.

Implications

Unification of fields and particles

The "second quantization" procedure that we have outlined in the previous section takes a set of single-particle quantum states as a starting point. Sometimes, it is impossible to define such single-particle states, and one must proceed directly to quantum field theory. For example, a quantum theory of the electromagnetic field must be a quantum field theory, because it is impossible (for various reasons) to define a wavefunction for a single photon.[16] In such situations, the quantum field theory can be constructed by examining the mechanical properties of the classical field and guessing the corresponding quantum theory. For free (non-interacting) quantum fields, the quantum field theories obtained in this way have the same properties as those obtained using second quantization, such as well-defined creation and annihilation operators obeying commutation or anticommutation relations.

Quantum field theory thus provides a unified framework for describing "field-like" objects (such as the electromagnetic field, whose excitations are photons) and "particle-like" objects (such as electrons, which are treated as excitations of an underlying electron field), so long as one can treat interactions as "perturbations" of free fields. There are still unsolved problems relating to the more general case of interacting fields that may or may not be adequately described by perturbation theory. For more on this topic, see Haag's theorem.

Physical meaning of particle indistinguishability

The second quantization procedure relies crucially on the particles being identical. We would not have been able to construct a quantum field theory from a distinguishable many-particle system, because there would have been no way of separating and indexing the degrees of freedom.

Many physicists prefer to take the converse interpretation, which is that quantum field theory explains what identical particles are. In ordinary quantum mechanics, there is not much theoretical motivation for using symmetric (bosonic) or antisymmetric (fermionic) states, and the need for such states is simply regarded as an empirical fact. From the point of view of quantum field theory, particles are identical if and only if they are excitations of the same underlying quantum field. Thus, the question "why are all electrons identical?" arises from mistakenly regarding individual electrons as fundamental objects, when in fact it is only the electron field that is fundamental.

Particle conservation and non-conservation

During second quantization, we started with a Hamiltonian and state space describing a fixed number of particles (N), and ended with a Hamiltonian and state space for an arbitrary number of particles. Of course, in many common situations N is an important and perfectly well-defined quantity, e.g. if we are describing a gas of atoms sealed in a box. From the point of view of quantum field theory, such situations are described by quantum states that are eigenstates of the number operator \hat{N}, which measures the total number of particles present. As with any quantum mechanical observable, \hat{N} is conserved if it commutes with the Hamiltonian. In that case, the quantum state is trapped in the N-particle subspace of the total Fock space, and the situation could equally well be described by ordinary N-particle quantum mechanics. (Strictly speaking, this is only true in the noninteracting case or in the low energy density limit of renormalized quantum field theories)

For example, we can see that the free-boson Hamiltonian described above conserves particle number. Whenever the Hamiltonian operates on a state, each particle destroyed by an annihilation operator a_k is immediately put back by the creation operator a_k^\dagger.

On the other hand, it is possible, and indeed common, to encounter quantum states that are not eigenstates of \hat{N}, which do not have well-defined particle numbers. Such states are difficult or impossible to handle using ordinary quantum mechanics, but they can be easily described in quantum field theory as quantum superpositions of states having different values of N. For example, suppose we have a bosonic field whose particles can be created or destroyed by interactions with a fermionic field. The Hamiltonian of the combined system would be given by the Hamiltonians of the free boson and free fermion fields, plus a "potential energy" term such as

H_I = \sum_{k,q} V_q (a_q + a_{-q}^\dagger) c_{k+q}^\dagger c_k,

where a_k^\dagger and a_k denotes the bosonic creation and annihilation operators, c_k^\dagger and c_k denotes the fermionic creation and annihilation operators, and V_q is a parameter that describes the strength of the interaction. This "interaction term" describes processes in which a fermion in state k either absorbs or emits a boson, thereby being kicked into a different eigenstate k+q. (In fact, this type of Hamiltonian is used to describe interaction between conduction electrons and phonons in metals. The interaction between electrons and photons is treated in a similar way, but is a little more complicated because the role of spin must be taken into account.) One thing to notice here is that even if we start out with a fixed number of bosons, we will typically end up with a superposition of states with different numbers of bosons at later times. The number of fermions, however, is conserved in this case.

In condensed matter physics, states with ill-defined particle numbers are particularly important for describing the various superfluids. Many of the defining characteristics of a superfluid arise from the notion that its quantum state is a superposition of states with different particle numbers. In addition, the concept of a coherent state (used to model the laser and the BCS ground state) refers to a state with an ill-defined particle number but a well-defined phase.

Axiomatic approaches

The preceding description of quantum field theory follows the spirit in which most physicists approach the subject. However, it is not mathematically rigorous. Over the past several decades, there have been many attempts to put quantum field theory on a firm mathematical footing by formulating a set of axioms for it. These attempts fall into two broad classes.

The first class of axioms, first proposed during the 1950s, include the Wightman, Osterwalder–Schrader, and Haag–Kastler systems. They attempted to formalize the physicists' notion of an "operator-valued field" within the context of functional analysis, and enjoyed limited success. It was possible to prove that any quantum field theory satisfying these axioms satisfied certain general theorems, such as the spin-statistics theorem and the CPT theorem. Unfortunately, it proved extraordinarily difficult to show that any realistic field theory, including the Standard Model, satisfied these axioms. Most of the theories that could be treated with these analytic axioms were physically trivial, being restricted to low-dimensions and lacking interesting dynamics. The construction of theories satisfying one of these sets of axioms falls in the field of constructive quantum field theory. Important work was done in this area in the 1970s by Segal, Glimm, Jaffe and others.

During the 1980s, a second set of axioms based on geometric ideas was proposed. This line of investigation, which restricts its attention to a particular class of quantum field theories known as topological quantum field theories, is associated most closely with Michael Atiyah and Graeme Segal, and was notably expanded upon by Edward Witten, Richard Borcherds, and Maxim Kontsevich. However, most of the physically relevant quantum field theories, such as the Standard Model, are not topological quantum field theories; the quantum field theory of the fractional quantum Hall effect is a notable exception. The main impact of axiomatic topological quantum field theory has been on mathematics, with important applications in representation theory, algebraic topology, and differential geometry.

Finding the proper axioms for quantum field theory is still an open and difficult problem in mathematics. One of the Millennium Prize Problemsproving the existence of a mass gap in Yang–Mills theoryis linked to this issue.

Associated phenomena

In the previous part of the article, we described the most general features of quantum field theories. Some of the quantum field theories studied in various fields of theoretical physics involve additional special ideas, such as renormalizability, gauge symmetry, and supersymmetry. These are described in the following sections.

Renormalization

Main article: Renormalization

Early in the history of quantum field theory, it was found that many seemingly innocuous calculations, such as the perturbative shift in the energy of an electron due to the presence of the electromagnetic field, give infinite results. The reason is that the perturbation theory for the shift in an energy involves a sum over all other energy levels, and there are infinitely many levels at short distances that each give a finite contribution which results in a divergent series.

Many of these problems are related to failures in classical electrodynamics that were identified but unsolved in the 19th century, and they basically stem from the fact that many of the supposedly "intrinsic" properties of an electron are tied to the electromagnetic field that it carries around with it. The energy carried by a single electronits self energyis not simply the bare value, but also includes the energy contained in its electromagnetic field, its attendant cloud of photons. The energy in a field of a spherical source diverges in both classical and quantum mechanics, but as discovered by Weisskopf with help from Furry, in quantum mechanics the divergence is much milder, going only as the logarithm of the radius of the sphere.

The solution to the problem, presciently suggested by Stueckelberg, independently by Bethe after the crucial experiment by Lamb, implemented at one loop by Schwinger, and systematically extended to all loops by Feynman and Dyson, with converging work by Tomonaga in isolated postwar Japan, comes from recognizing that all the infinities in the interactions of photons and electrons can be isolated into redefining a finite number of quantities in the equations by replacing them with the observed values: specifically the electron's mass and charge: this is called renormalization. The technique of renormalization recognizes that the problem is essentially purely mathematical, that extremely short distances are at fault. In order to define a theory on a continuum, first place a cutoff on the fields, by postulating that quanta cannot have energies above some extremely high value. This has the effect of replacing continuous space by a structure where very short wavelengths do not exist, as on a lattice. Lattices break rotational symmetry, and one of the crucial contributions made by Feynman, Pauli and Villars, and modernized by 't Hooft and Veltman, is a symmetry-preserving cutoff for perturbation theory (this process is called regularization). There is no known symmetrical cutoff outside of perturbation theory, so for rigorous or numerical work people often use an actual lattice.

On a lattice, every quantity is finite but depends on the spacing. When taking the limit of zero spacing, we make sure that the physically observable quantities like the observed electron mass stay fixed, which means that the constants in the Lagrangian defining the theory depend on the spacing. Hopefully, by allowing the constants to vary with the lattice spacing, all the results at long distances become insensitive to the lattice, defining a continuum limit.

The renormalization procedure only works for a certain class of quantum field theories, called renormalizable quantum field theories. A theory is perturbatively renormalizable when the constants in the Lagrangian only diverge at worst as logarithms of the lattice spacing for very short spacings. The continuum limit is then well defined in perturbation theory, and even if it is not fully well defined non-perturbatively, the problems only show up at distance scales that are exponentially small in the inverse coupling for weak couplings. The Standard Model of particle physics is perturbatively renormalizable, and so are its component theories (quantum electrodynamics/electroweak theory and quantum chromodynamics). Of the three components, quantum electrodynamics is believed to not have a continuum limit, while the asymptotically free SU(2) and SU(3) weak hypercharge and strong color interactions are nonperturbatively well defined.

The renormalization group describes how renormalizable theories emerge as the long distance low-energy effective field theory for any given high-energy theory. Because of this, renormalizable theories are insensitive to the precise nature of the underlying high-energy short-distance phenomena. This is a blessing because it allows physicists to formulate low energy theories without knowing the details of high energy phenomenon. It is also a curse, because once a renormalizable theory like the standard model is found to work, it gives very few clues to higher energy processes. The only way high energy processes can be seen in the standard model is when they allow otherwise forbidden events, or if they predict quantitative relations between the coupling constants.

Haag's theorem

See also: Haag's theorem

From a mathematically rigorous perspective, there exists no interaction picture in a Lorentz-covariant quantum field theory. This implies that the perturbative approach of Feynman diagrams in QFT is not strictly justified, despite producing vastly precise predictions validated by experiment. This is called Haag's theorem, but most particle physicists relying on QFT largely shrug it off.

Gauge freedom

A gauge theory is a theory that admits a symmetry with a local parameter. For example, in every quantum theory the global phase of the wave function is arbitrary and does not represent something physical. Consequently, the theory is invariant under a global change of phases (adding a constant to the phase of all wave functions, everywhere); this is a global symmetry. In quantum electrodynamics, the theory is also invariant under a local change of phase, that is – one may shift the phase of all wave functions so that the shift may be different at every point in space-time. This is a local symmetry. However, in order for a well-defined derivative operator to exist, one must introduce a new field, the gauge field, which also transforms in order for the local change of variables (the phase in our example) not to affect the derivative. In quantum electrodynamics this gauge field is the electromagnetic field. The change of local gauge of variables is termed gauge transformation. It is worth noting that by Noether's theorem, for every such symmetry there exists an associated conserved current. The aforementioned symmetry of the wavefunction under global phase changes implies the conservation of electric charge.

In quantum field theory the excitations of fields represent particles. The particle associated with excitations of the gauge field is the gauge boson, which is the photon in the case of quantum electrodynamics.

The degrees of freedom in quantum field theory are local fluctuations of the fields. The existence of a gauge symmetry reduces the number of degrees of freedom, simply because some fluctuations of the fields can be transformed to zero by gauge transformations, so they are equivalent to having no fluctuations at all, and they therefore have no physical meaning. Such fluctuations are usually called "non-physical degrees of freedom" or gauge artifacts; usually some of them have a negative norm, making them inadequate for a consistent theory. Therefore, if a classical field theory has a gauge symmetry, then its quantized version (i.e. the corresponding quantum field theory) will have this symmetry as well. In other words, a gauge symmetry cannot have a quantum anomaly. If a gauge symmetry is anomalous (i.e. not kept in the quantum theory) then the theory is non-consistent: for example, in quantum electrodynamics, had there been a gauge anomaly, this would require the appearance of photons with longitudinal polarization and polarization in the time direction, the latter having a negative norm, rendering the theory inconsistent; another possibility would be for these photons to appear only in intermediate processes but not in the final products of any interaction, making the theory non-unitary and again inconsistent (see optical theorem).

In general, the gauge transformations of a theory consist of several different transformations, which may not be commutative. These transformations are together described by a mathematical object known as a gauge group. Infinitesimal gauge transformations are the gauge group generators. Therefore, the number of gauge bosons is the group dimension (i.e. number of generators forming a basis).

All the fundamental interactions in nature are described by gauge theories. These are:

Multivalued gauge transformations

The gauge transformations which leave the theory invariant involve, by definition, only single-valued gauge functions \Lambda(x_i) which satisfy the Schwarz integrability criterion

\partial_{x_i x_j} \Lambda  = \partial_{x_jx_i} \Lambda.

An interesting extension of gauge transformations arises if the gauge functions \Lambda(x_i) are allowed to be multivalued functions which violate the integrability criterion. These are capable of changing the physical field strengths and are therefore not proper symmetry transformations. Nevertheless, the transformed field equations describe correctly the physical laws in the presence of the newly generated field strengths. See the textbook by H. Kleinert cited below for the applications to phenomena in physics.

Supersymmetry

Main article: Supersymmetry

Supersymmetry assumes that every fundamental fermion has a superpartner that is a boson and vice versa. It was introduced in order to solve the so-called Hierarchy Problem, that is, to explain why particles not protected by any symmetry (like the Higgs boson) do not receive radiative corrections to its mass driving it to the larger scales (GUT, Planck...). It was soon realized that supersymmetry has other interesting properties: its gauged version is an extension of general relativity (Supergravity), and it is a key ingredient for the consistency of string theory.

The way supersymmetry protects the hierarchies is the following: since for every particle there is a superpartner with the same mass, any loop in a radiative correction is cancelled by the loop corresponding to its superpartner, rendering the theory UV finite.

Since no superpartners have yet been observed, if supersymmetry exists it must be broken (through a so-called soft term, which breaks supersymmetry without ruining its helpful features). The simplest models of this breaking require that the energy of the superpartners not be too high; in these cases, supersymmetry is expected to be observed by experiments at the Large Hadron Collider. The Higgs particle has been detected at the LHC, and no such superparticles have been discovered.

See also

Notes

  1. Dirac 1927
  2. Schweber 1994, p. 28
  3. See references in Schweber (1994, pp. 695f)
  4. Weinberg 2005, p. 15
  5. Peskin & Schroeder (1995, Chapter4)
  6. Greiner & Reinhardt 1996
  7. Clément Hongler, "Conformal invariance of Ising model correlations", Ph.D. thesis, Université of Geneva, 2010, p. 9.
  8. Thorn et al. 2004
  9. 1 2 3 4 Tong 2015, Chapter 1
  10. Srednicki 2007, p. 19
  11. Srednicki 2007, pp. 2526
  12. Zee 2010, p. 61
  13. Tong 2015, Introduction
  14. Zee 2010, p. 3
  15. Pais 1994. Pais recounts how his astonishment at the rapidity with which Feynman could calculate using his method. Feynman's method is now part of the standard methods for physicists.
  16. Newton & Wigner 1949, pp. 400406

References

Historical references
General reader level
Articles
Introductory texts
Advanced texts

Further reading

General readers
Introductory texts
Advanced texts

Articles

External links

One-dimensional quantum field theory on Wikiversity
This article is issued from Wikipedia - version of the Wednesday, May 04, 2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.