Consider a simple experiment in which the goal is to measure what is presumed to be a constant signal in the presence of a constant background. For this example, assume that there is Gaussian noise on both the signal and the background. In the first part of this notebook assume that you can measure the 'total' of 'background + noise', and then you can turn off the signal, and measure the 'noise' alone. The 'signal' will then be the difference between the 'total' and the 'noise.' For class on Tuesday execute the cells in this notebook, and then determine the value of the signal, with an uncertainty. (There is nothing new here -- I just want you to become familiar with the 'experiment.' In the final cells of this notebook I propose a small variation to the experiment that I want to ponder. This will serve as an introduction to Tuesday's class.
import numpy as np
from scipy import stats
import matplotlib as mpl # As of July 2017 Bucknell computers use v. 2.x
import matplotlib.pyplot as plt
# Following is an Ipython magic command that puts figures in the notebook.
%matplotlib notebook
# As of Aug. 2017 reverting to 1.x defaults.
# In 2.x text.ustex requires dvipng, texlive-latex-extra, and texlive-fonts-recommended,
# which don't seem to be universal
# See https://stackoverflow.com/questions/38906356/error-running-matplotlib-in-latex-type1cm?
mpl.style.use('classic')
# M.L. modifications of matplotlib defaults using syntax of v.2.0
# More info at http://matplotlib.org/2.0.0/users/deflt_style_changes.html
# Changes can also be put in matplotlibrc file, or effected using mpl.rcParams[]
plt.rc('figure', figsize = (6, 4.5)) # Reduces overall size of figures
plt.rc('axes', labelsize=16, titlesize=14)
plt.rc('figure', autolayout = True) # Adjusts supblot parameters for new size
n_trials = 100
s = 20
alpha_s_gen = 1
b = 10
alpha_b_gen = 2
values_sig = stats.norm.rvs(s, alpha_s_gen, size=n_trials)
values_bg = stats.norm.rvs(b, alpha_b_gen, size=n_trials)
values_tot = values_sig + values_bg
np.savetxt('correlation_1.dat', values_tot)
np.savetxt('correlation_2.dat', values_bg)
#values_tot=np.loadtxt('http://www.eg.bucknell.edu:/~phys310/skills/data_analysis/correlation_1.dat')http://www.eg.bucknell.edu/~phys310/skills/data_analysis/correlation_1.dat
values_tot = np.loadtxt('http://www.eg.bucknell.edu/~phys310/skills/data_analysis/correlation_1.dat')
values_bg = np.loadtxt('http://www.eg.bucknell.edu/~phys310/skills/data_analysis/correlation_2.dat')
n_trials = len(values_tot)
plt.figure()
x = np.linspace(1,n_trials,n_trials)
plt.scatter(x, values_tot, color='r', label='signal + bg')
plt.scatter(x, values_bg, label='background only')
plt.xlabel('trial')
plt.legend()
plt.ylim(0,45)
plt.grid();
mean_tot = np.mean(values_tot)
std_tot = np.std(values_tot)
mean_bg = np.mean(values_bg)
std_bg = np.std(values_bg)
alpha_tot = std_tot/np.sqrt(n_trials - 1)
alpha_bg = std_bg/np.sqrt(n_trials - 1)
signal = mean_tot - mean_bg
alpha_sig = np.sqrt(alpha_tot**2 + alpha_bg**2)
print('signal =',signal)
print('uncertainty in total = ', alpha_tot)
print('uncertainty in background =', alpha_bg)
print('uncertainty in signal =', alpha_sig)
$ \mbox{signal} = 19.9 \pm 0.3 $
We will discuss this quantitatively tomorrow, but I want you to think about some of the issues that are involved before Tuesday's discssion.
plt.figure()
x = np.linspace(1,n_trials,n_trials)
plt.scatter(x, values_tot, color='r', label='signal + bg')
plt.scatter(x, values_bg, label='background only')
plt.xlabel('trial')
plt.legend()
plt.ylim(0,50)
plt.grid();
values_sig = values_tot - values_bg
mean_sig = np.mean(values_sig)
std_sig = np.std(values_sig)
alpha_sig = std_sig/np.sqrt(n_trials -1)
print('signal mean =', mean_sig)
print('uncertainty = ', alpha_sig)
In this scenario, I find that the signal is
$\mbox{signal} = 20.0 \pm 0.1$
Notice that the uncertainty is reduced by a factor of 3!
As in H&H 4.2, consider a derived quantity $Z$ that depends on measured values of variables $A$ and $B$:
$$ Z = f(A,B). $$If we have determined that
$$ A = \bar{A} \pm \alpha_A\quad\quad\mbox{and}\quad\quad B = \bar{B} \pm \alpha_B, $$we say that
$$ \bar Z = f(\bar A, \bar B), $$but what is the value of $\alpha_Z$?
Up until now we have used the rule given in H&H Eq.(4.10) on p. 41 (and also in the PHYS 211/212 lab manual) that
\begin{eqnarray*} \alpha_Z^2 &=& \left(\alpha_Z^A\right)^2 + \left(\alpha_Z^{B}\right)^2\\ \end{eqnarray*}When there are correlations in the measured quantities, we should be using
$$ \alpha_Z^2 = \left(\alpha_Z^A\right)^2 + \left(\alpha_Z^B\right)^2 + 2\left.\frac{\partial f}{\partial A}\right|_{\bar A}\, \left.\frac{\partial f}{\partial B}\right|_{\bar B} \left[\frac{1}{N-1}\sum_i (A_i - \bar A)(B_i - \bar B)\right] . $$The new term in this equation can be written in terms the sample covariance, $\sigma_{AB}$, where
$$ \left(\alpha_Z^{AB}\right)^2 = 2\left.\frac{\partial f}{\partial A}\right|_{\bar A}\, \left.\frac{\partial f}{\partial B}\right|_{\bar B} \sigma^2_{AB} . $$Rewriting the expression for the standard error in $Z$ in terms of the covariance gives
$$ \alpha_Z^2 =\left(\alpha_Z^A\right)^2 + \left(\alpha_Z^B\right)^2 + \left(\alpha_Z^{AB}\right)^2. $$In the example in this notebook, the signal (i.e., $Z$) is just the difference between the total and the background (i.e., $Z = A-B$), so
$$ \left.\frac{\partial f}{\partial A}\right|_{\bar A}\, \left.\frac{\partial f}{\partial B}\right|_{\bar B} = -1. $$cov = np.sum((values_tot - mean_tot)*(values_bg - mean_bg))/(n_trials - 1)
print(cov)
std_sig = cor = np.sqrt(std_tot**2 + std_bg**2 - 2*cov)
print(std_sig)
print(std_sig/np.sqrt(n_trials-1))
$\mbox{Uncertainty} = 0.1$
std_tot**2 + std_bg + 2*cov
np.std(values_tot), np.std(values_bg)
std_tot**2 + std_bg**2 + 2*cov_s_cor
mean_s_2 = sp.mean()
np.sqrt(461 + 127- 2*245)
np.sqrt(1.9**2 + 0.99**2)
(values_t - mean_t)*(values_b - mean_b)
np.sum(_)