Heat capacity The electronic heat capacity C e can be found by taking the derivative of Equation (19.18): For temperatures that are small compared with.

Similar presentations

Presentation on theme: "Heat capacity The electronic heat capacity C e can be found by taking the derivative of Equation (19.18): For temperatures that are small compared with."— Presentation transcript:

1
Heat capacity The electronic heat capacity C e can be found by taking the derivative of Equation (19.18): For temperatures that are small compared with the Fermi temperature, we can neglect the second term in the expansion compared with the first and obtain

2
Thus the electronic specific heat capacity is 2.2 x R. This small value explains why metals have a specific heat capacity of about 3R, the same as for other solids. It was originally believed that their free electrons should contribute an additional (3/2) R associated with their three translational degrees of freedom. Our last calculation shows that the contribution is negligible. The energy of the electrons changes only slightly with temperature (dU/dT is small) because only those electrons near the Fermi level can increase their energies as the temperature is raised, and there are precious few of them.

3
At very low temperatures the picture is different. From the Debye theory, C v is proportional to T 3 and so the heat capacity of a metal takes the form C v = AT + BT 3, where the first term is the electronic contribution and the second is associated with the crystal lattice. At sufficiently low temperatures, the AT term can dominate, as the sketch of Figure 19.9 indicates. Figure 19.9 Sketch of the heat capacity of a metal as a function of temperature showing the electronic and lattice contributions.

4
S = 0 at T = 0, as it must be. The Helmholtz function F = U -TS is The fermion gas pressure is found from

5
For silver we found that N/V = 5.9 x10 28 m -3 and T F = 65,000K. Thus P = 2/5 *5.9*10 28 *(1.38* ) (6.5*10 4 ) = 2.1*10 10 Pa = 2.1*10 5 atm. Given this tremendous pressure, we can appreciate the role of the surface potential barrier in keeping the electrons from evaporating from the metal.

6
19.5 Applications to White Dwarf Stars The temperature inside the core of a typical star is at the order of 10 7 K. The atoms are completely ionized at such a high T, which creates a hugh electron gas The loss of gravitational energy balances with an increase in the kinetic energy of the electrons and ions, which prevent the collapse of star!

16
(b) What is the Fermi temperature? (c) What is the average speed of the electrons in the fermion gas (see problem 19-4). Compare your answer with the speed of light.

17
Chapter 20 Information Theory

18
20.1 Introduction Statistical thermodynamics provides the tool of calculating entropy. Entropy is a measure of the degree of randomness or disorder of a system. Disorder implies a lack of information regarding the exact state of the system. A disordered system is one about which we lack complete information.

19
20.2 Uncertainty and Information Claude Shannon laid down the foundations of the information theory. Further developed by Leon Brillouin. Applied to Statistical thermodynamics by E. T. Jaynes. The cornerstone of the Shannon theory is the observation that information is a combination of the certain and the uncertain, of the expected and the unexpected.

20
The degree of surprise generated by a certain event – one that has already occurred – is zero. If a less probable event is reported, the information conveyed is greater. The information should increase as the probability decreases.

21
For a given experiment, consider a set of possible outcomes whose probabilities are p 1, p 2, … p n. It is possible to find a quantity H(p 1... p n ) that measures the amount of uncertainty represented by the given set of probabilities. Only three conditions are needed to specify the function H(p 1... p n ) to within a constant factor. They are: 1. H is a continuous function of the p. 2. If all the p i ’s are equal, p i = 1/n; then H(1/n,…, 1/n) is a monotonic increasing function of n. 3. If the possible outcomes of a particular experiment depend on the possible outcomes of n subsidiary experiments, then H is the sum of the uncertainties of the subsidiary experiments.

22

23

24
The above discussion leads to g(R) + g(S) = g(RS), where one can expect that the function g( ) shall be a logarithm function. In a general format, the function can be written as g(x) = A ln(x) + C, where A and C are constants. From the earlier transformation g(x) = x f(1/x), one gets that the uncertainty quantity, H, shall be (1/p)f(p) = A ln(p) + C, where p = 1/n (n is the total number of event). Therefore, f(p) = A*p*ln(p) + C Given that if the probability is 1, the uncertainty H must be zero, the constant C should be equal to ZERO. Thus, f(p) = A*p*ln(p). Since p is smaller than 1, ln(p) shall be minus and thus the constant A is inherently negative

27
20.3 Unit of Information Choosing 2 as the basis of the logarithm and take K = 1, one gets H = 1 We call the unit of information a bit for binary event. Decimal digit, H = log 2 (10) = 3.32, thus a decimal digit contains about 3 and 1/3 bits of information.

28
Linguistics A more refined analysis works in terms of component syllables. One can test what is significant in a syllable in speech by swapping syllables and seeing if meaning or tense is changed or lost. The table gives some examples of the application of this statistical approach to some works of literature.

29
Linguistics The type of interesting results that arise from such studies include: (a) English has the lowest entropy of any major language, and (b) Shakespeare’s work has the lowest entropy of any author studied. These ideas are now progressing beyond the scientific level and are impinging on new ideas of criticism. Here as in biology, the thermodynamic notions can be helpful though they must be applied with caution because concepts such as ‘quality’ cannot be measured as they are purely subjective