Franck-Hertz Experiment
Some things to ponder
Thermal Energy
The fundamental insight from the Franck-Hertz experiment is that of discrete energy levels. When electrons that are accelerated across a potential difference attains a “critical” energy and “collide” with Hg atoms, something magical happens– a sudden transfer of energy! This is precisely what you’re observing when you increase your accelerating voltage $V_a$ and obtain these peaks in the collector current (although with many many electrons). Recall that an electron moving across a potential difference of $V_a$ attains an energy of $e V_a$. For example: if $V_a=5V$, then the electron gains an energy of $5eV$ after crossing over the potential difference.
Simultaneously, we know that turning up the temperature makes atoms jiggle, i.e, they gain energy. Imagine setting $V_a=5V$, you would then naturally conclude that the energy gained by an electron is $5 eV$. But wait! if these electrons also move due to temperature, surely they must gain some thermal energy right? Even though your power supply reads $5.0 V$, how sure are you that the thermal energy doesn’t change the expected energy to, say $(5\pm 2)eV$ instead of $5 eV$?
In other words, how relevant are these temperature effects to the energy?
What's happening?
As a physicist, a way to think about the question of “relevance” is that of scales. In nature, there are time scales, length scales and energy scales. What therefore matters is not the minute details, but the order of magnitude. The difference in these scales generally requires different models and, in a nutshell, is also what makes physics hard!
In our problem, let’s compare the energy scales of thermal noise with the energy supplied (i.e electron volts). Whatever sources you look up for the energy due to thermal noise, you will find that the thermal energy $E_T$ at temperature $T$ is roughly \(E_T \sim k_B T.\)
Given that your operating temperature is $T=180^\circ C$, an order of magnitude estimation gives \(T\sim 10^2 K.\)
The Boltzmann constant $k_B$ has units of $\frac{\rm Energy}{\rm Temperature}$, and can be written in terms of $eV/K$ as
\[k_B= 8.6 \times 10^{-5}eV/K\sim 10^{-4} eV/JK\]i.e the thermal noise is roughly \(E_T\sim 10^{-2} eV.\)
What this means is that any fluctuations you see will be on the 2nd decimal place to your right which is beyond the range that the FH power supply can yield!
Energy Levels?
You’re diligently doing your Franck-Hertz experiment, imagining the electrons colliding with the mercury atoms and transferring their energy to the atoms by exciting them to the “next energy level” until you ask yourself “what are these energy levels and what do they correspond to?”
You’ve learnt about the Bohr model and decide to estimate this difference in the energy between these energy levels–surely the difference should be the same energy that the electrons transfer right? So you pull up the atomic configuration of mercury and find that last electron is at the $n=6$ energy level. The next state is at $n=7$, so the difference in the energy should be
\[E_{7}-E_{6}= -13.6 eV Z^2 \cdot \left(\frac{1}{7^2}-\frac{1}{6^2}\right)\approx 641 eV,\]where you have used the atomic number of mercury $Z=80$ and therefore the nucleus has a positive charge of approximately $80 e$.
Then you realize that something is wrong with your calculation: It is two orders of magnitude bigger than what you expect! What’s wrong? Could it be that there are missing energy levels? if so, where or what are they?
What's happening?
The Bohr model, that’s the problem! Specifically, the Bohr model is only valid for hydrogen-like atoms and our Hg atom is the last thing that is close to being hydrogen-like. However, most of chemists uses hydrogen-like orbitals to do chemistry; the “s,p,d,f” orbitals we talk about are all solutions to the hydrogen atom shoe horned onto heavier atoms, so what gives? Start with the hydrogen atom which has one electron and a proton, the Bohr model tells us that the ground state corresponds to \(n=1\). The index “n” are what we call quantum numbers and describes a quantum state. What’s interesting is that the solutions of the Schrödinger equation produces new quantum numbers \(\ell\) and \(m\) from conservation of angular momentum. The states with \(\ell=0\) is what we call the “s” orbitals and the other funny letters are for \(\ell \geq 1\). So in the atomic configuration of H, we find
\[[{\rm H}]= 1 s^1,\]which is meant to be read as a state which has one electron in the so-called “s” shell in the \(n=1\) state. So if there are states labelled by different quantum numbers, then shouldn’t they have different energies contradicting the formula \(E_{n}= -13.6 {\rm eV}/n^2\)? It turns out that for the hydrogen atom, we’ve got a degeneracy! the energy levels only depend on the quantum number \(n\).
The story is different in the case of the atomic configuration of Hg in which you’ll find
\[[{\rm Hg}]= [{\rm Xe}]\ 4f^{14}\ 5d^{10}\ 6s^2.\]The configuration tell us that that the outermost shell \(n=6\) contains two electrons in the “s” shell and is described succinctly as \(6s^2\). For non hydrogen-like atoms, the energy degeneracy is lifted due to the simple fact that the energetics depend on the electron-electron interaction! Coulomb repulsion forces electrons to optimize between the attractive force of the nucleus and the repulsive force from their neighbors making them occupy these other orbitals between \(n=6\) and \(n=7\) shells. These orbitals in between, too, are an approximation. Atomic physicists use something called term symbols to describe the states which you can find if you search energy level diagrams for your atom, but the main takeaway here is that “missing” different energy levels that were always present in the background for the hydrogen atom but washed away due to degeneracy arise from the coulomb repulsion between many electrons. This many-body interaction is precisely what makes Quantum mechanics so hard!
Fitting data
You’re finishing up the Franck-Hertz experiment, almost done with this hectic lab. By now you’ve got a comfortable grip on your apparatus and you’ve seen the troughs and peaks at roughly \(5 {\rm eV}\) intervals–congrats on doing some cool science! Now that the experimental aspects are over, you move to your analysis. Earlier you observed “roughly” \(5 {\rm eV}\), but can you do better? You want to find the exact location (value of \(V_a\)) of the troughs (or peaks) to see the difference clearly. As a first order approximation, you might look for the smallest (biggest) current or amplitude from Logger Pro within a certain range and identify that as the trough (peak).
But, as you zoom into the trough (peak), the illusion of a smooth continuous curve vanishes, you’re left with discrete points. In the very unlikely case that your smallest (bigger) current amplitude is what you measured, you’re golden. However, the most likely scenario is that the resolution you have wasn’t fine enough to obtain the exact trough (peak). For example, two points were recorded with some width \(\Delta V\) such that the real trough (peak) is in between them.
In these cases when you’ve got missing data points, you need to perform a fit. A fit extrapolates the data you measured using a particular model (linear, exponential etc) and outputs the parameters used in the model. In your case, the parameters you’re looking for are the location (\(V_a\)) of the trough (peak) and a characteristic width. The question is what fit do you use?
Note: As part of the analysis, you need to perform a fit and a \(\chi^2\) test for your fits. Refer lecture notes for the details on how to do both. In addition, feel free to check out my notes on fitting Least-squares and fitting curves through points and \(\chi^2\) What the heck is \(\chi^2\) anyway? if you want a more intuitive perspective.
What's happening?
From class, you know two “peak” shaped functions (a trough is just an inverted peak): The Gaussian and Lorentzian, so you need to
- Isolate the troughs (peaks)
- Find the mean of fit for each trough (peak) which is what you’re looking for and its associated error.
You are probably asking yourself the very natural question: which one do I use? This is a great question, and I will let you justify the fit based on your data.
Here are some pointers: recall that the Gaussian is defined as
\[G(x; A, \mu, \sigma):= \frac{A}{\sqrt{2\sigma^2 \pi}} e^{-\frac{(x-\mu)^2}{2\sigma^2}}\]where \(A\) is the amplitude (overall scaling), \(\mu\) is the mean (center) and \(\sigma\) is the characteristic width of the Gaussian.
The Lorentzian is defined as
\[L(x; A,\mu,\Gamma):=\frac{A}{\pi} \frac{\Gamma/2}{(x-\mu)^2+(\Gamma/2)^2}\]where \(A\) is the amplitude, \(\mu\) the center and \(\Gamma\) is what’s called “Full width-half maximum”.
Remember that in experiments, a measured value is always associated with an error. Therefore, your center (location of the trough or peak) is given by \(\mu \pm \sigma\) if you’re using a Gaussian distribution and \(\mu\pm \Gamma\) if you are using a Lorentzian.
Aside: To see why \(\Gamma\) is the natural equivalent of \(\sigma\) which is what you’d use as an error, observe that for the standard Gaussian distribution that is normalized \(A=1\) (called the Normal distribution), the area under the curve between \(x=\mu\pm \sigma\) is
\[\int_{\mu-\sigma}^{\mu+\sigma} G(x;1,\mu,\sigma)= 0.68,\]meaning that most (\(\approx 68\%\)) of the data from a random variable that follows a Normal distribution with mean \(\mu\) lies within the range \((\mu-\sigma,\mu+\sigma)\).
You might wonder what would happen if you do the same with the normalized (\(A=1\)) Lorentzian distribution and you find that (exercise, try it!)
\[\int_{\mu-\Gamma}^{\mu+\Gamma} L(x;a,\mu,\Gamma)=\frac{2}{\pi} {\rm arctan}(2)\approx 0.70,\]i.e this area is \(70 \%\) of the whole! This is why the \(\Gamma\) of a Lorentzian distribution is the natural equivalent of \(\sigma\) even though it does not have a well defined notation of a variance!