Super cool atom thermometer

RAM

The southern Man
New Member
Joined
Jul 15, 2009
Messages
2,289
Likes
455
Country flag
Super cool atom thermometer

Physicists face the daunting task of developing new, reliable ways of measuring extreme low temperatures






The conquest of the cold has become one the most exciting goals in modern atomic physics. The development of laser and evaporative cooling techniques have been among the greatest achievements in the second half of last century, as recognized by two Nobel prizes on these topics (1997 and 2001). These methods made it possible to reach temperatures of the order of a few nK and led to the first experimental realization of a Bose-Einstein condensate [1, 2] and a quantum degenerate Fermi gas [3] in dilute atomic vapors.

Even though nK temperatures may seem extremely low, they are still too high for various potential applications, and the battle for reaching lower and lower temperatures in ultracold atoms continues. One of the driving forces for achieving even lower temperatures is Richard Feynman’s pioneering idea of a quantum simulator [4] wherein the behavior of a complex quantum system can be simulated by another quantum system. For instance, we want to load atoms in optical lattices (light shift potentials created by the interference of multiple laser beams) and use them to mimic the physics of electrons in solid-state crystals [5]. The optical lattice supplies the periodic potential in which atoms move. Bosonic (Fermionic) atoms in optical lattices are actually a perfect implementation of Bose (Fermi) Hubbard Hamiltonians [6], which describe particles hopping on a lattice with onsite interactions [7]. It is believed that these are the simplest models that contain the fundamental ingredients required to describe the behavior of strongly correlated materials, including, for example, quantum magnets or high-temperature superconductors. However, since atoms are much heavier than electrons, and since typical optical lattice interwell spacings are of the order of 104 times the ionic lattice spacings, temperatures below 10-2 nK are required in cold-atom laboratories in order to probe the same physics that occur at Kelvin temperatures in solid-state systems. The capability of reaching such low temperatures has to be accompanied by the development of new thermometers capable of measuring them. In a recent paper published in Physical Review Letters, David Weld and colleagues at the MIT-Harvard center for ultracold atoms [8] now report a new thermometry method for ultracold atoms in lattices, with the potential to measure temperatures as low as tens of pK.
To date, typical bosonic cold-atom experiments in a single trap without an optical lattice have determined temperature from fits to absorption images of the expanded gas [9]. The data are fit to a bimodal density distribution. The area under the central peak, linked to the Bose-condensed fraction, and a Gaussian fit to the wings, which assumes noninteracting thermal atoms are combined to infer temperature. This technique has been very successful in a broad range of experimentally relevant temperatures, however, when atoms are loaded into optical lattices its regime of applicability reduces considerably.

When atoms are loaded into a lattice they acquire an effective mass, which exponentially increases with the depth of lattice potential. As a consequence, the diluteness condition, which requires a small ratio between the mean interaction energy per particle to its kinetic energy, becomes invalid as the lattice potential depth is ramped up and the system enters the strongly correlated regime. Beyond a critical lattice potential depth, the average kinetic energy required for an atom to hop from one site to the next, J, becomes insufficient to overcome the interaction energy cost, U, and atoms tend to get localized at individual lattice sites, forming the so-called Mott insulator [10].

In a homogeneous system, Mott insulating phases occur only at integer densities; noninteger density contours lie entirely in the superfluid phase because there is always an extra particle that can hop without energy cost. In the presence of an additional parabolic trapping potential, both phases can coexist and the system exhibits a shell structure in which Mott insulating domains with n atoms per lattice site are separated by a superfluid region from other Mott domains with n-1 atoms per site.

Simple, direct thermometry of systems in the Mott insulating state has remained a challenge since methods relying on the noninteracting approximation fail. One of the most common techniques to estimate the temperature of a lattice gas is to measure the initial temperature of the gas before the lattice is switched on, and then determine the final temperature assuming the lattice is turned on adiabatically: One equates the entropy of the final state to that of the initial state and then deduces the final temperature from the initial temperature. However, the errors of this method are uncontrolled, not only because of the adiabatic approximation assumption, but also because of the difficulty of an accurate determination of the temperature dependence of entropy for the many-body system of interest.

In their work, Weld et al. [8] have experimentally demonstrated a solution to this problem in a two-component Bose gas in the Mott insulator regime. In the experiment, the two spin states used were the F=|1,-1〉 and F=|2,-2〉 hyperfine states of 87Rb atoms. Their strategy was to impose a magnetic field gradient and to use the mean magnetization of the gas as a thermometer (see Fig. 1). The basic idea can be understood by considering a spin-1/2 atom with magnetic moment μσ, σ=±1/2. In the present of a magnetic field B(x), its mean magnetization is just 〈σ〉=1/2 tanh[-βΔμσ⋅B(x)/2] with Δμσ the difference in magnetic moment between the two spin states, β=1/(kBT) the inverse temperature and kB the Boltzmann constant. For an incoherent mixture of atoms with spin independent interactions (condition well satisfied in 87Rb atoms) the spatial and spin degrees of freedom can be factored out and the spin distribution becomes just ρ(x)〈σ〉 with ρ(x) the density distribution. By imaging the spin and density distributions, and knowing Δμσ⋅B(x), they were able to determine the temperature of the atoms by a standard fitting procedure.
The spin gradient thermometer was demonstrated to work in a range of temperatures not accessible to prior thermometry schemes. In the high-temperature regime, where interactions can be safety neglected, it agreed well with other methods relying on the noninteracting approximation, and was even shown to operate at temperatures high enough that no condensate existed before the lattice was ramped up. In the low-temperature regime it was used to measure a temperature as low as 1 nK, providing a direct demonstration that current experiments are able to operate deep in the quantum regime where the Mott insulator shell structure is well resolved. The lowest temperature measurable with the method is limited only by the experimental ability to resolve the domain wall (~50 pK for current imaging resolution of a few μm) and more fundamentally by kBTs~J2/U, the onset of super-exchange interactions (spin-spin interactions mediated by virtual tunneling process [11]), whichever is higher. Below Ts the incoherent mixture approximation becomes invalid.

The development of experimental methods, such as the one reported in Ref. [8], capable of measuring subnanokelvin temperatures brings us one step forward towards the dream of using cold atoms to simulate and manipulate different strongly correlated many-body systems, which appear in various fields in physics, ranging from condensed matter to subatomic physics.



Physics - The super cool atom thermometer

Super cool atom thermometer
 

Articles

Top