I was trying to resolve this exercise:
This exercise is from the book "Statistical Inference, Second Edition" by Casella and Berger. Checking the solutions manual, I was understanding the solution but I can't figure out why the operation that is hightlighted in yelow it's equal to one. My possition is that the value hightlighted depends on the sample, but the solution is hardly wrong. Can someone explain that part?. Here's the solution:
Best Answer
Set everything into an exponential:
begin{align*} left(prod_{i=1}^nx_iright)^{hattheta_0-hatmu}left(prod_{i=1}^my_iright)^{hattheta_0-hattheta}&=expleft{ (hattheta_0-hatmu)sum_ilog(x_i)+(hattheta_0-hattheta)sum_ilog(y_i)right}\ &=expleft{ dfrac{-(n+m)sum_ilog(x_i)}{sum_ilog(x_i)+sum_ilog(y_i)}+n+dfrac{-(n+m)sum_ilog(y_i)}{sum_ilog(x_i)+sum_ilog(y_i)}+mright}\ &=expleft{ dfrac{-(n+m)left[sum_ilog(x_i)+sum_ilog(y_i)right]}{sum_ilog(x_i)+sum_ilog(y_i)}+n+mright}\ &=expleft{-(n+m)+n+mright}\ &=1 end{align*}
Similar Posts:
- Solved – Bayesian Estimation: Bernoulli and Quadratic Loss Function
- Solved – Does the biased estimator always have less variance than unbiased one
- Solved – the BIC prior for Bayesian linear regression
- Solved – Maximum likelihood estimator, exact distribution
- Solved – an example of a transformation on a posterior distribution such that the MAP estimate will be non-invariant