Jump to content

Byers–Yang theorem

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Brienanni (talk | contribs) at 14:10, 3 September 2014 (Brienanni moved page Byers-yang theorem to Byers-Yang theorem: capitalization). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In quantum mechanics, the Byers-Yang theorem states that all physical properties of a doubly-connected system (a ring) enclosing a magnetic flux Φ through the opening are periodic in the flux with period Failed to parse (syntax error): {\displaystyle \Phi_0=h/e$ (the socalled [[flux quantum]]). The theorem was first stated and proven by [[Byers]] and [[Chen-Ning Yang]] (1961) <ref>{{cite journal|last=Feynman|first=R. P.|year=1939|title=Forces in Molecules|journal=Phys. Rev.|volume=56|issue=4|pages=340|doi=10.1103/PhysRev.56.340 |bibcode = 1939PhRv...56..340F }}</ref>, and later by [[Felix Bloch]] <ref>{{cite journal|last=Feynman|first=R. P.|year=1939|title=Forces in Molecules|journal=Phys. Rev.|volume=56|issue=4|pages=340|doi=10.1103/PhysRev.56.340 |bibcode = 1939PhRv...56..340F }}</ref>. relates the derivative of the total energy with respect to a parameter, to the [[Expectation value (quantum mechanics)|expectation value]] of the derivative of the [[Hamiltonian (quantum mechanics)|Hamiltonian]] with respect to that same parameter. According to the theorem, once the spatial distribution of the electrons has been determined by solving the [[Schrödinger equation]], all the forces in the system can be calculated using [[Classical electromagnetism|classical electrostatics]]. ==Proof== The flux Φ can be eliminated by a gauge transformation This proof of the Hellmann–Feynman theorem requires that the wavefunction be an eigenfunction of the Hamiltonian under consideration; however, one can also prove more generally that the theorem holds for non-eigenfunction wavefunctions which are stationary (partial derivative is zero) for all relevant variables (such as orbital rotations). The [[Hartree–Fock]] wavefunction is an important example of an approximate eigenfunction that still satisfies the Hellmann–Feynman theorem. Notable example of where the Hellmann–Feynman is not applicable is for example finite-order [[Møller–Plesset perturbation theory]], which is not variational.<ref>{{cite book|last=Jensen|first=Frank|title=Introduction to Computational Chemistry|publisher=John Wiley & Sons|location=West Sussex|year=2007|isbn=0-470-01186-6|page=322}}</ref> The proof also employs an identity of normalized wavefunctions&nbsp;– that derivatives of the overlap of a wavefunction with itself must be zero. Using Dirac's [[bra–ket notation]] these two conditions are written as :<math>\hat{H}_{\lambda}|\psi_\lambda\rangle = E_{\lambda}|\psi_\lambda\rangle,}

The proof then follows through an application of the derivative product rule to the expectation value of the Hamiltonian viewed as a function of λ:

For a deep critical view of the proof see[1]

Alternate proof

The Hellmann–Feynman theorem is actually a direct, and to some extent trivial, consequence of the variational principle (the Rayleigh-Ritz variational principle) from which the Schrödinger equation can be made to derive. This is why the Hellmann–Feynman theorem holds for wave-functions (such as the Hartree–Fock wave-function) that, though not eigenfunctions of the Hamiltonian, do derive from a variational principle. This is also why it holds, e.g., in density functional theory, which is not wave-function based and for which the standard derivation does not apply.

According to the Rayleigh–Ritz variational principle, the eigenfunctions of the Schrödinger equation are stationary points of the functional (which we nickname Schrödinger functional for brevity):

The eigenvalues are the values that the Schrödinger functional takes at the stationary points:

where satisfies the variational condition:

Let us differentiate Eq. (3) using the chain rule:

Due to the variational condition, Eq. (4), the second term in Eq. (5) vanishes. In one sentence, the Hellmann–Feynman theorem states that the derivative of the stationary values of a function(al) with respect to a parameter on which it may depend, can be computed from the explicit dependence only, disregarding the implicit one. On account of the fact that the Schrödinger functional can only depend explicitly on an external parameter through the Hamiltonian, Eq. (1) trivially follows. As simple as that.

Example applications

Molecular forces

The most common application of the Hellmann–Feynman theorem is to the calculation of intramolecular forces in molecules. This allows for the calculation of equilibrium geometries – the nuclear coordinates where the forces acting upon the nuclei, due to the electrons and other nuclei, vanish. The parameter λ corresponds to the coordinates of the nuclei. For a molecule with 1 ≤ iN electrons with coordinates {ri}, and 1 ≤ α ≤ M nuclei, each located at a specified point {Rα={Xα,Yα,Zα)} and with nuclear charge Zα, the clamped nucleus Hamiltonian is

The force acting on the x-component of a given nucleus is equal to the negative of the derivative of the total energy with respect to that coordinate. Employing the Hellmann–Feynman theorem this is equal to

Only two components of the Hamiltonian contribute to the required derivative – the electron-nucleus and nucleus-nucleus terms. Differentiating the Hamiltonian yields[2]

Insertion of this in to the Hellmann–Feynman theorem returns the force on the x-component of the given nucleus in terms of the electronic density (ρ(r)) and the atomic coordinates and nuclear charges:

Expectation values

An alternative approach for applying the Hellmann–Feynman theorem is to promote a fixed or discrete parameter which appears in a Hamiltonian to be a continuous variable solely for the mathematical purpose of taking a derivative. Possible parameters are physical constants or discrete quantum numbers. As an example, the radial Schrödinger equation for a hydrogen-like atom is

which depends upon the discrete azimuthal quantum number l. Promoting l to be a continuous parameter allows for the derivative of the Hamiltonian to be taken:

The Hellmann–Feynman theorem then allows for the determination of the expectation value of for hydrogen-like atoms:[3]

Van der Waals forces

In the end of Feynman's paper, he states that, "Van der Waals's forces can also be interpreted as arising from charge distributions with higher concentration between the nuclei. The Schrödinger perturbation theory for two interacting atoms at a separation R, large compared to the radii of the atoms, leads to the result that the charge distribution of each is distorted from central symmetry, a dipole moment of order 1/R7 being induced in each atom. The negative charge distribution of each atom has its center of gravity moved slightly toward the other. It is not the interaction of these dipoles which leads to van der Waals's force, but rather the attraction of each nucleus for the distorted charge distribution of its own electrons that gives the attractive 1/R7 force."

Hellmann–Feynman theorem for time-dependent wavefunctions

For a general time-dependent wavefunction satisfying the time-dependent Schrödinger equation, the Hellmann–Feynman theorem is not valid. However, the following identity holds:

For

Proof

The proof only relies on the Schrödinger equation and the assumption that partial derivatives with respect to λ and t can be interchanged.

Notes

  1. ^ Carfì, David (2010). "The pointwise Hellmann–Feynman theorem". AAPP Physical, Mathematical, and Natural Sciences. 88 (1). no. C1A1001004. doi:10.1478/C1A1001004. ISSN 1825–1242. {{cite journal}}: Check |issn= value (help)
  2. ^ Piela, Lucjan (2006). Ideas of Quantum Chemistry. Amsterdam: Elsevier Science. p. 620. ISBN 0-444-52227-1.
  3. ^ Fitts, Donald D. (2002). Principles of Quantum Mechanics : as Applied to Chemistry and Chemical Physics. Cambridge: Cambridge University Press. p. 186. ISBN 0-521-65124-7.