We consider a random variable $x$ on $]-\infty ,+\infty [$ following a Gauss’s law with $\mu $ and $\sigma $ parameters.
- Give the expression of Gauss’s law. - Give the conditions on parameters to make this law be a probability distribution
function (pdf). - What is the Characteristic function (definition and expression)? - Do the same for cumulative function.
We do another substitution variable $y=exp(x)$, where $x$ is defined above and follows Gauss’s
law.
- How will you calculate the pdf of $y$? - Give the expression of pdf($y$). - Calculate the expectation $E(y)$; first write its definition. - Calculate the expression of $V(y)$ (variance of $y$).
Correction:
We have to find the distribution function of $y$ recorded $g(y)$. Then,
One starts from 2 random variables $x$ and $y$ with respectively $\sigma _{x}^{2}$ and $\sigma _{y}^{2}$ standard deviations and $\rho _{xy}$
correlation factor.
We do the following substitution of variable:
$u=x+y\sigma _{x}/\sigma _{y}$
$v=x-y\sigma _{x}/\sigma _{y}$
Calculate $V(u)$ and $V(v)$ variances and covariance $\text{cov}(u,v)$. What can you conclude? What is the
correlation factor $\rho _{uv}$?
We see that $u$ and $v$ are not correlated: conclusion is that we can made 2 no-correlated
variables from 2 correlated ones.
Exercise 2.2:
Both $x$ and $y$ random variables follow a Gauss’s law (pdf$= f(x,y)$) with $\mu _{x}=0$ and $\mu _{y}=0$.
- Write its expression. - Calculate Expectations $E(x)$ and $E(y)$. - Calculate the distribution function for $u$ and $v$ (pdf$=h(u,v)$). - What do you conclude about these two random variables $u$ and $v$?
Correction:
General formula of 2D Gaussian distribution function:
We can see that $u$ and $v$ are independant because of pdf factorization $h(u,v)=\text {pdf}(u) \cdot \text {pdf}(v)$. If two variables $X$ and $Y$
are statiscally independent, then they are not correlated. The reverse is not necessarily
true. Indeed, no-correlation implies independence only in particular cases and Gaussian random variable is one of them.
Exercise 3.1:
One considers a continous and positive random variable $t$, following an exponential law $\propto exp(-t/\tau )$.
Give the expression of its pdf $f(t)$ and calculate its expectation $E(t)$ and variance $V(t)$. We have $N$
measures $t_{i}$ of this variable and choose the maximum likelihood method to determine $\tau $
parameter. Describe the principle of the maximum likelihood method and apply it to
calculate this parameter $\tau$ and its variance. Now, we have 2 independants samples of
$N_{1}$and $N_{2}$ measures so that total sample is consisted of $N=N_{1}+N_{2}$ measures. Calculate the
estimation of $\tau $ from these N measures while making appear the contribution of $\tau _{1}$ and
$\tau _{2}$.
Correction:
Let $t$ be a random variable $>0$ following an exponential law; pdf is given by:
We own $N$ measures $t_{i}, i=1,..,N$. Maximum likelihood allows to compute an estimation of $\tau $ parameter
which is solution of the maximum or minimum for likelihood function $\mathcal{L}$ defined
as:
This result is the weighted average from these 2 samples.
Exercise 3.2:
Two independant experiments have measured ($\tau _{1},\sigma _{1}$) and ($\tau _{2},\sigma _{2}$) with $\sigma _{i}$ representing errors on
measures.
(1) From these two measures, assuming errors are gaussian, we want to get the estimation
of $\tau $ and its error (i.e with a combination of two measures). - Which method do you use? - Calculate the estimation of $\tau $ and its error.
(2) From these two measures ($\tau _{1},\sigma _{1}$) and ($\tau _{2},\sigma _{2}$): Define the equivalent number $\tilde {N}_{1}$ and $\tilde {N}_{2}$ of each measure; give the relations defining them.
We use the maximum likelihood method to find the estimation of $\tau $ from the
definition of these 2 equivalent numbers. Calculate this estimation of $\tau $ in this case
(Making appear $\tau _{1},\sigma _{1}$ et $\tau _{2},\sigma _{2}$ in the expression). Compare it to previous expression in
(1).
Correction:
As the previous Exercise, we choose the maximum likelihood method with the pdf of 2
measures:
Principle is to generate a sample $(x_{1},x_{2},...,x_{N})$ with law of $X$ (so we calculate an estimator called "the
estimator of Monte-Carlo" from this sample).
Law of large numbers assumes to build this estimator from the empirical average:
which is, by the way, an unbiased estimator of expectation. This is the estimator of
Monte-Carlo. We can clearly see that, by replacing sample by a set of values located on an
integral support, and the function to integrate, we can build an approximation of its
value, statistically made.
Thanks to the "uniform generator", Monte-Carlo method gives a numerical value
of the integral noticed $I$. Area of integration is represented on the figure below :
So we can distinguish two cases:
case (1) - We do 2 random sampling, one for $x$ and the other for $y$:
If $x_{i}>y_{i}$, then we increment $I$ in the following way: $I=I+f(x_{i},y_{i})$, else we redo a random sampling. We have
done 2 random sampling and the average to lose one is 1/2.
Then, as a function of results, we increment $I$ in the following way:
\begin{equation}
\textrm{Incrementation}\left\{
\begin{array}{lcc}
if &x_{i} > y_{i} & I=I+f(x_{i},y_{i}) \\
if &x_{i} < y_{i} & I=I+f(y_{i},x_{i}) \\
\end{array}
\right.
\end{equation}
The advantage here is that we use all the random sampling values unlike case
(1). Finally, one has to multiply the $I$ quantity by $(b-a)$ interval and divide by the size of random
sampling $N$, so we get the numerical value of integral.
Exercise 4.2:
A point source emits isotropically and covers an angle of $\theta _{0}$. A disk detector is positioned
perpendicular to this source. So we have a cylindrical symetry with 2 angles: $\varphi $ between
[0,2$\pi $] and $\theta $ as $\text {cos}(\theta _{0})<\text {cos}(\theta )<1$.
Calculate the pdf of $(\varphi ,\text {cos}\,\theta )$.
Having a generator of random values on [0,1], how will you random sample in acceptance
disk a couple $(\varphi ,\text {cos}\,\theta )$? With this detector, we want to do counting during equal interval time $\Delta t$. Express the
distribution function for the number of recorded hits.
Correction:
Since point source emits isotropically, random variables $\varphi $ and $\text {cos}\,\theta $ follow a uniform law,
respectively on [0,2$\pi $] et [$\text {cos}\,\theta _{0},1$].
The number of hits recorded during interval $\Delta t$ will follow a Poisson law.
Exercise 5.1:
we have a set of measures $y_{i}\,\,i=1,...,n$ depending from coordinates $x_{i}$ and whose theorical
model is linear $y=ax+b$. Thanks to these data, we look for determining values of $a$ and
$b$.
Measures $y_{i}$ have an error $\sigma _{i}$. Firstly, coordinates $x_{i}$ are considered being without
error.
- Express the $\chi ^{2}$ you have to use. - Express the 2 equations from which you can deduce estimations for $a$ and $b$.
Correction:
With $n$ independant measures $y_{i}\,i=1,...,n$ and $n$ coordinates $x_{i}$ in a linear model $y=ax+b$, with $\sigma _{i}$ errors on $y_{i}$ and no
errors on $x_{i}$, one can write the $\chi ^{2}$ as:
Now, $x_{i}$ coordinates have $\delta _{i}$ errors.
Express the $\chi ^{2}$ which has to be used in this case. Write the 2 equations from which you calculate $a$ and $b$ estimation. What’s the difference
with previous case?
Correction:
$\chi ^{2}$formula must be modified because we take into account of $\delta _{i}$ errors on coordinates $x_{i}$. Indeed,
variance of $(y_{i}-a\,x_{i}-b)$ is not only equal to $V(y_{i})=\sigma _{i}^{2}$:
But we notice that all these equations are not linear since the second one which
minimizes the $\chi ^{2}$ with $b$ depends on powered $a$: we have not analytical solution in this
case.
Exercise 6.1:
$\chi ^{2}$method can give estimations on parameters, $a\pm \sigma _{a}$ and $b\pm \sigma _{b}$, so a minimum value $\chi _{min}^{2}$.
We want to draw in $a,b$ plane the contour related to a given confidence level.
Express the distribution function that you use.
In this case, by fixing the confidence level, express the variation of $\chi ^{2}$ compared to $\chi _{min}^{2}$, so we
could write: $\chi ^{2}(CL)=\chi _{min}^{2}+\Delta\chi_{CL}^{2}$
What value do you get with $CL=0.68$?
Correction:
$\chi ^{2}(CL)$term only contains second derivatives of $\chi ^{2}$ at lowest level: