<< Chapter < Page Chapter >> Page >

Dc level in white guassian noise

Suppose we observe an unknown amplitude in white Gaussian noise with unknown variance: x n A w n n 0 1 N 1 , where w n 0 2 are independent and identically distributed. We would like to estimate A 2 by computing the MLE. Differentiating the log-likelihood gives A p x 1 2 n 1 N x n A 2 p x N 2 1 2 4 n 1 N x n A 2 Equating with zero and solving gives us our MLEs: A 1 N n 1 N x n and 2 1 N n 1 N x n A 2

2 is biased!

As an exercise, try the following problem:

Suppose we observe a random sample x x 1 x N of Poisson measurements with intensity : x i n n n , n 0 1 2 . Find the MLE for .

Unfortunately, this approach is only feasible for the most elementarypdfs and pmfs. In general, we may have to resort to more advanced numerical maximization techniques:
  • Newton-Raphson iteration
  • Iteration by the Scoring Method
  • Expectation-Maximization Algorithm
All of these are iterative techniques which posit some initial guess at the MLE, and then incrementally update thatguess. The iteration procedes until a local maximum of the likelihood is attained, although in the case of the first twomethods, such convergence is not guaranteed. The EM algorithm has the advantage that the likelihood is always increased ateach iteration, and so convergence to at least a local maximum is guaranteed (assuming a bounded likelihood). For eachalgorithm, the final estimate is highly dependent on the initial guess, and so it is customary to try several differentstarting values. For details on these algorithms, see Kay, Vol. I .

Asymptotic properties of the mle

Let x x 1 x N denote an IID sample of size N , and each sample is distributed according to p x . Let N denote the MLE based on a sample x .

Asymptotic properties of mle

If the likelihood x p x satisfies certain "regularity" conditions

The regularity conditions are essentially the same as those assumed for the Cramer-Rao lower bound : the log-likelihood must be twice differentiable, and theexpected value of the first derivative of the log-likelihood must be zero.
, then the MLE N is consistent , and moreover, N converges in probability to , where I where I is the Fisher Information matrix evaluated at the true value of .

Since the mean of the MLE tends to the true parameter value, we say the MLE is asymptotically unbiased . Since the covariance tends to the inverse Fisher information matrix, we saythe MLE is asymptotically efficient .

In general, the rate at which the mean-squared error converges to zero is not known. It is possible that for small samplesizes, some other estimator may have a smaller MSE.The proof of consistency is an application of the weak law of largenumbers. Derivation of the asymptotic distribution relies on the central limit theorem. The theorem is also true in moregeneral settings (e.g., dependent samples). See, Kay, Vol. I, Ch. 7 for further discussion.

The mle and efficiency

In some cases, the MLE is efficient, not just asymptotically efficient. In fact, when an efficient estimator exists, itmust be the MLE, as described by the following result:

If is an efficient estimator, and the Fisher information matrix I is positive definite for all , then maximizes the likelihood.

Recall the is efficient (meaning it is unbiased and achieves the Cramer-Rao lower bound) if and only if p x I for all and x . Since is assumed to be efficient, this equation holds, and in particular it holds when x . But then the derivative of the log-likelihood is zero at x . Thus, is a critical point of the likelihood. Since the Fisher information matrix, which is the negative ofthe matrix of second order derivatives of the log-likelihood, is positive definite, must be a maximum of the likelihood.

An important case where this happens is described in the following subsection.

Optimality of mle for linear statistical model

If the observed data x are described by x H w where H is N p with full rank, is p 1 , and w 0 C , then the MLE of is H C H H C x This can be established in two ways. The first is to compute the CRLB for . It turns out that the condition for equality in the bound is satisfied, and can be read off from that condition.

The second way is to maximize the likelihood directly. Equivalently, we must minimize x H C x H with respect to . Since C is positive definite, we can write C U U D D , where D 1 2 U , where U is an orthogonal matrix whose columns are eigenvectors of C , and is a diagonal matrix with positive diagonal entries. Thus, we must minimize D x D H D x D H But this is a linear least squares problem, so the solution is given by the pseudoinverse of D H :

D H D H D H D x H C H H C x

Consider X 1 , , X N s 2 I , where s is a p 1 unknown signal, and 2 is known. Express the data in the linear model and find the MLE s for the signal.

Invariance of mle

Suppose we wish to estimate the function w W and not itself. To use the maximum likelihood approach for estimating w , we need an expression for the likelihood x w p w x . In other words, we would need to be able to parameterize thedistribution of the data by w . If W is not a one-to-one function, however, this may not be possible. Therefore, we define the induced likelihood x w W w x The MLE w is defined to be the value of w that maximizes the induced likelihood. With this definition, the following invarianceprinciple is immediate.

Let denote the MLE of . Then w W is the MLE of w W .

The proof follows directly from the definitions of and w . As an exercise, work through the logical steps of the proof on your own.

Let x x 1 x N where x i Poisson Given x , find the MLE of the probability that x Poisson exceeds the mean .

W x n 1 n n where z largest integer z . The MLE of w is w n 1 n n where is the MLE of : 1 N n 1 N x n

Be aware that the MLE of a transformed parameter does not necessarily satisfy the asymptotic properties discussed earlier.

Consider observations x 1 ,, x N , where x i is a p -dimensional vector of the form x i s w i where s is an unknown signal and w i are independent realizations of white Gaussian noise: w i 0 2 I p p Find the maximum likelihood estimate of the energy E s s of the unknown signal.

Summary of mle

The likelihood principle states that information broughtby an observation x about is entirely contained in the likelihood function p x . The maximum likelihood estimator is one effective implementation of the likelihood principle. In some cases, the MLE can be computedexactly, using calculus and linear algebra, but at other times iterative numerical algorithms are needed. The MLE has severaldesireable properties:

  • It is consistent and asymptotically efficient (as N we are doing as well as MVUE).
  • When an efficient estimator exists, it is the MLE.
  • The MLE is invariant to reparameterization.

Questions & Answers

A golfer on a fairway is 70 m away from the green, which sits below the level of the fairway by 20 m. If the golfer hits the ball at an angle of 40° with an initial speed of 20 m/s, how close to the green does she come?
Aislinn Reply
cm
tijani
what is titration
John Reply
what is physics
Siyaka Reply
A mouse of mass 200 g falls 100 m down a vertical mine shaft and lands at the bottom with a speed of 8.0 m/s. During its fall, how much work is done on the mouse by air resistance
Jude Reply
Can you compute that for me. Ty
Jude
what is the dimension formula of energy?
David Reply
what is viscosity?
David
what is inorganic
emma Reply
what is chemistry
Youesf Reply
what is inorganic
emma
Chemistry is a branch of science that deals with the study of matter,it composition,it structure and the changes it undergoes
Adjei
please, I'm a physics student and I need help in physics
Adjanou
chemistry could also be understood like the sexual attraction/repulsion of the male and female elements. the reaction varies depending on the energy differences of each given gender. + masculine -female.
Pedro
A ball is thrown straight up.it passes a 2.0m high window 7.50 m off the ground on it path up and takes 1.30 s to go past the window.what was the ball initial velocity
Krampah Reply
2. A sled plus passenger with total mass 50 kg is pulled 20 m across the snow (0.20) at constant velocity by a force directed 25° above the horizontal. Calculate (a) the work of the applied force, (b) the work of friction, and (c) the total work.
Sahid Reply
you have been hired as an espert witness in a court case involving an automobile accident. the accident involved car A of mass 1500kg which crashed into stationary car B of mass 1100kg. the driver of car A applied his brakes 15 m before he skidded and crashed into car B. after the collision, car A s
Samuel Reply
can someone explain to me, an ignorant high school student, why the trend of the graph doesn't follow the fact that the higher frequency a sound wave is, the more power it is, hence, making me think the phons output would follow this general trend?
Joseph Reply
Nevermind i just realied that the graph is the phons output for a person with normal hearing and not just the phons output of the sound waves power, I should read the entire thing next time
Joseph
Follow up question, does anyone know where I can find a graph that accuretly depicts the actual relative "power" output of sound over its frequency instead of just humans hearing
Joseph
"Generation of electrical energy from sound energy | IEEE Conference Publication | IEEE Xplore" ***ieeexplore.ieee.org/document/7150687?reload=true
Ryan
what's motion
Maurice Reply
what are the types of wave
Maurice
answer
Magreth
progressive wave
Magreth
hello friend how are you
Muhammad Reply
fine, how about you?
Mohammed
hi
Mujahid
A string is 3.00 m long with a mass of 5.00 g. The string is held taut with a tension of 500.00 N applied to the string. A pulse is sent down the string. How long does it take the pulse to travel the 3.00 m of the string?
yasuo Reply
Who can show me the full solution in this problem?
Reofrir Reply
Got questions? Join the online conversation and get instant answers!
Jobilize.com Reply

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Statistical signal processing. OpenStax CNX. Jun 14, 2004 Download for free at http://cnx.org/content/col10232/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Statistical signal processing' conversation and receive update notifications?

Ask