Skip to main content

Statistical Inference ( Unit 2: Cramer-Rao Inequality, Method Of Moment, Maximum Likelihood Estimator)

Changing Color Blog Name

Statistical Inference I: (Cramer-Rao Inequality, Method Of Moment, Maximum Likelihood Estimator)

I. Introduction

We see in unbiased estimator from two distinct unbiased estimators give infinitely many unbiased estimators of θ, among these estimators we find the best estimator for parameter θ by comparing their variance or mean square errors. But in some examples, we see that the number of estimators is possible as.

For Normal distribution:

If X1, X2, ........Xn. random sample from a normal distribution with mean 𝛍 and variance 𝛔², then T1 = x̄, T2 = Sample median, both are unbiased estimators for parameter 𝛍. Now we find a sufficient estimator; therefore, T1 is a sufficient estimator for 𝛍, hence it is the best estimator for parameter 𝛍.

Thus, for finding the best estimator, we check if the estimator is sufficient or not.

Now we are interested in finding the variance of the best estimator. This can be discussed in this article.

II. Properties of Probability Mass Function (p.m.f.) or Probability Density Function (p.d.f)

If X1, X2, ........Xn. are from any p.d.f or p.m.f f(x,θ), θ ∈ Θ, following are the properties of P.D.F. OR P.M.F:

Properties of Fisher Information Function

1. \( \int_{-\infty}^{\infty} f(x, \theta) \, dx = 1 \)

Proof: We have

\( \int_{-\infty}^{\infty} f(x, \theta) \, dx = 1 \) from 1

2. \( \frac{\partial}{\partial\theta} \int_{-\infty}^{\infty} f(x, \theta) \, dx = 0 \)

Proof: We have

\( \frac{\partial}{\partial\theta} \int_{-\infty}^{\infty} f(x, \theta) \, dx = 0 \) from 2

3. \( E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right] = 0 \)

Proof: We have

\( E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right] = 0 \) from 3

4. \( E\left[\frac{\partial^2}{\partial\theta^2} \log f(x, \theta)\right] = -E\left[\frac{\partial\log f(x, \theta)}{\partial\theta}\right]^2 \)

Proof: We have

\( E\left[\frac{\partial^2}{\partial\theta^2} \log f(x, \theta)\right] = -E\left[\frac{\partial\log f(x, \theta)}{\partial\theta}\right]^2 \) from 4

5. \( \text{Var}\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right] = -E\left[\frac{\partial^2}{\partial\theta^2} \log f(x, \theta)\right] = E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right]^2 \)

Proof: We have

Var(\( \frac{\partial}{\partial\theta} \log f(x, \theta) \)) = \( E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right]^2 - \left(E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right]\right)^2 \)

Var(\( \frac{\partial}{\partial\theta} \log f(x, \theta) \)) = \( E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right]^2 - 0 \) from 3

Var(\( \frac{\partial}{\partial\theta} \log f(x, \theta) \)) = \( E\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right]^2 = -E\left[\frac{\partial^2}{\partial\theta^2} \log f(x, \theta)\right] \) from 4

Fisher Information Function:

Definition:

  1. Fisher Information Measure (or the amount of information) about parameter \(\theta\) obtained in the random variable \(x\) is denoted as \(I(\theta)\) and is defined as
  2. \(I(\theta) = \text{Var}\left[\frac{\partial}{\partial\theta} \log f(x, \theta)\right] = \text{E}\left[\left(\frac{\partial}{\partial\theta} \log f(x, \theta)\right)^2\right] = -\text{E}\left[\frac{\partial^2}{\partial\theta^2} \log f(x, \theta)\right]\)

  3. Fisher Information Measure (or the amount of information) about parameter \(\theta\) obtained in the random variable \(X_1, X_2, \ldots, X_n\) of size \(n\) is denoted as \(I_n(\theta)\) and is defined as
  4. \(I_n(\theta) = \text{Var}\left[\frac{\partial}{\partial\theta} \log L(\theta)\right] = \text{E}\left[\left(\frac{\partial}{\partial\theta} \log L(\theta)\right)^2\right] = -\text{E}\left[\frac{\partial^2}{\partial\theta^2} \log L(\theta)\right]\)

  5. Let \(X_1, X_2, \ldots, X_n\) be a random sample from the distribution of random variable \(x\), and \(T = T(X_1, X_2, \ldots, X_n)\) be any statistic for the parameter \(\theta\), and \(g(t, \theta)\) is the probability function, then the Fisher information function (or the amount of information) about parameter \(\theta\) contained in statistic \(T\) is given by \(I_T(\theta)\) and is given as
  6. \(I_T(\theta) = \text{Var}\left[\frac{\partial}{\partial\theta} \log g(t, \theta)\right] = \text{E}\left[\left(\frac{\partial}{\partial\theta} \log g(t, \theta)\right)^2\right] = -\text{E}\left[\frac{\partial^2}{\partial\theta^2} \log g(t, \theta)\right]\)

Properties of Fisher Information Function

Result 1:

Let \(X_1, X_2, \ldots, X_n\) be a random sample from the distribution \(f(x,\theta)\), then \(I_n(\theta) = n I(\theta)\).

Proof:

Let \(X_1, X_2, \ldots, X_n\) be a random sample from the distribution \(f(x,\theta)\), \(\theta \in \Theta\).

The Fisher Information Measure (or the amount of information) about parameter \(\theta\) obtained in random variable \(x\) is denoted as \(I(\theta)\) and is defined as:

\[ I(\theta) = \text{Var}\left[\frac{\partial}{\partial \theta} \log f(x,\theta)\right] = \text{E}\left[\left(\frac{\partial}{\partial \theta} \log f(x,\theta)\right)^2\right] = -\text{E}\left[\frac{\partial^2}{\partial \theta^2} \log f(x,\theta)\right] \]

Now consider the joint probability function or the likelihood function of random variables \(X_1, X_2, \ldots, X_n\), then \(L(\theta) = f(X_1, X_2, \ldots, X_n,\theta)\).

\[ L(\theta) = \prod f(x,\theta) \]

Taking the logarithm on both sides:

\[ \log L(\theta) = \log\left(\prod f(x,\theta)\right) \]

\[ \log L(\theta) = \sum \log f(x,\theta) \]

Differentiating with respect to \(\theta\):

\[ \frac{\partial}{\partial \theta} \log L(\theta) = \sum \frac{\partial}{\partial \theta} \log f(x,\theta) \quad \text{(i)} \]

We have Fisher Information Measure (or the amount of information) about parameter \(\theta\) obtained in random variable \(X_1, X_2, \ldots, X_n\) is denoted as \(I_n(\theta)\) and it is defined as:

\[ I_n(\theta) = \text{Var}\left[\frac{\partial}{\partial \theta} \log L(\theta)\right] = \text{E}\left[\left(\frac{\partial}{\partial \theta} \log L(\theta)\right)^2\right] = -\text{E}\left[\frac{\partial^2}{\partial \theta^2} \log L(\theta)\right] \]

So, \(I_n(\theta) = n I(\theta)\), hence proved.

Result 2:

Show that for any statistic \(T\), \(I_n(\theta) \geq I_T(\theta)\).

Proof:

Let \(X_1, X_2, \ldots, X_n\) be a random sample from the distribution of random variable \(x\), and \(T = T(X_1, X_2, \ldots, X_n)\) be any statistic for the parameter \(\theta\), and \(g(t,\theta)\) is the probability function.

The Fisher information function (or the amount of information) about parameter \(\theta\) contained in statistic \(T\) is given by \(I_T(\theta)\) and is given as:

\[ I_T(\theta) = \text{Var}\left[\frac{\partial}{\partial \theta} \log g(t,\theta)\right] = \text{E}\left[\left(\frac{\partial}{\partial \theta} \log g(t,\theta)\right)^2\right] = -\text{E}\left[\frac{\partial^2}{\partial \theta^2} \log g(t,\theta)\right] \]

Now, consider the joint probability function or the likelihood function of random variables \(X_1, X_2, \ldots, X_n\).

\[ L(\theta) = f(X_1, X_2, \ldots, X_n,\theta) \]

\[ L(\theta) = g(t,\theta) \cdot h(x) \]

Taking the logarithm of \(L(\theta)\):

\[ \log L(\theta) = \log (g(t,\theta) \cdot h(x)) \]

\[ \log L(\theta) = \log (g(t,\theta)) + \log (h(x)) \]

Differentiating with respect to \(\theta\):

\[ \frac{\partial}{\partial \theta} \log L(\theta) = \frac{\partial}{\partial \theta} \log (g(t,\theta)) + \frac{\partial}{\partial \theta} \log (h(x)) \]

\[ \frac{\partial}{\partial \theta} \log L(\theta) \geq \frac{\partial}{\partial \theta} \log (g(t,\theta)) + 0 \quad \text{(ii)} \]

\[ \text{Var}\left[\frac{\partial}{\partial \theta} \log L(\theta)\right] \geq \text{Var}\left[\frac{\partial}{\partial \theta} \log (g(t,\theta))\right] \]

\[ I_n(\theta) \geq I_T(\theta) \]

Remark:

If \(T\) is a sufficient statistic for \(\theta\), then \(I_n(\theta) = I_T(\theta)\).

Example 1: Fisher Information Function for Exponential Distribution

Let \(X_1, X_2, \ldots, X_n\) be a random sample from the Exponential distribution with parameter \(\frac{1}{\theta}\). The probability density function is:

\[ f(x,\theta) = \begin{cases} \frac{1}{\theta} e^{-\frac{x}{\theta}}, & x \geq 0, \theta > 0 \\ 0, & \text{otherwise} \end{cases} \]

Likelihood Function

The likelihood function of the sample \(X_1, X_2, \ldots, X_n\) is given by:

\[ L(\theta) = \prod f(x,\theta) = \left(\frac{1}{\theta}\right)^n e^{-\sum \frac{x}{\theta}} \]

Taking the logarithm on both sides:

\[ \log L(\theta) = -n \log(\theta) - \sum \frac{x}{\theta} \]

Derivative with Respect to \(\theta\)

Differentiating with respect to \(\theta\):

\[ \frac{\partial}{\partial \theta} \log L(\theta) = -\frac{n}{\theta} + \frac{\sum x}{\theta^2} \]

Second derivative with respect to \(\theta\):

\[ \frac{\partial^2}{\partial \theta^2} \log L(\theta) = \frac{n}{\theta^2} - \frac{2 \sum x}{\theta^3} \]

Fisher Information Function

By definition of the Fisher information function:

\[ I_n(\theta) = -E\left[\frac{\partial^2}{\partial \theta^2} \log L(\theta)\right] = -E\left[\frac{n}{\theta^2} - \frac{2 \sum x}{\theta^3}\right] = -\frac{E(n)}{\theta^2} + \frac{2E(\sum x)}{\theta^3} \]

Since \(E(n) = n\) and \(E(\sum x) = n\theta\), we have:

\[ I_n(\theta) = -\frac{n}{\theta^2} + \frac{2n\theta}{\theta^3} = -\frac{n}{\theta^2} + \frac{2n}{\theta^2} = \frac{n}{\theta^2}\]

Answer:

So, \(I_n(\theta) = \frac{n}{\theta^2} \).

Example 2: Fisher Information Function for Poisson Distribution

Let \(X_1, X_2, \ldots, X_n\) be a random sample from the Poisson distribution with parameter \(\lambda\). The probability mass function is:

\[ f(x,\theta) = \begin{cases} \frac{e^{-\lambda} \lambda^x}{x!}, & x = 0, 1, 2, \ldots, \lambda > 0 \\ 0, & \text{otherwise} \end{cases} \]

Likelihood Function

The likelihood function is given by:

\[ L(\theta) = \prod f(x,\theta) = \left(\frac{e^{-n\lambda} \lambda^{\sum x}}{\prod x!}\right) \]

Taking the logarithm on both sides:

\[ \log L(\theta) = -n\lambda + \sum x \log \lambda - \sum \log x \]

Derivative with Respect to \(\lambda\)

Differentiating with respect to \(\lambda\):

\[ \frac{\partial}{\partial \lambda} \log L(\theta) = -n + \frac{\sum x}{\lambda} \]

Second derivative with respect to \(\lambda\):

\[ \frac{\partial^2}{\partial \lambda^2} \log L(\theta) = -\frac{\sum x}{\lambda^2} \]

Fisher Information Function

By definition of the Fisher information function:

\[ I_n(\theta) = -E\left[\frac{\partial^2}{\partial \lambda^2} \log L(\theta)\right] = -E\left[-\frac{\sum x}{\lambda^2}\right] = \frac{E(\sum x)}{\lambda^2} \]

Since \(E(\sum x) = n\lambda\), we have:

\[ I_n(\theta) = \frac{n\lambda}{\lambda^2} = \frac{n}{\lambda} \]

Answer:

So, \(I_n(\theta) = \frac{n}{\lambda}\).

Example3:

Let \(X_1, X_2, \ldots, X_n\) be a random sample from the Normal distribution with \(\mu\) and \(\sigma^2\), then the probability density function is:

\[ f(x, \theta) = \frac{1}{\sigma\sqrt{2\pi}}e^{-\frac{1}{2\sigma^2}(x-\mu)^2}, \quad -\infty \leq x, \mu \leq \infty, \sigma^2 > 0 \]

\[ f(x, \theta) = \begin{cases} \frac{1}{\sigma\sqrt{2\pi}}e^{-\frac{1}{2\sigma^2}(x-\mu)^2}, & \text{-∞≤x,μ≥∞,σ^2>0} \\ 0, & \text{otherwise} \end{cases} \]

1. \(I(\mu)\)

We have:

\[ \begin{align*} \log f(x, \theta) & = \log \left(\frac{1}{\sigma\sqrt{2\pi}}e^{-\frac{1}{2\sigma^2}(x-\mu)^2}\right) \\ & = -\log(\sqrt{2\pi}) - \log(\sigma) - \frac{1}{2\sigma^2}(x-\mu)^2 \quad \text{(i)} \end{align*} \]

Differentiating with respect to \(\mu\):

\[ \begin{align*} \frac{\partial}{\partial \mu} \log f(x, \theta) & = \frac{\partial}{\partial \mu} \left(-\log(\sqrt{2\pi}) - \log(\sigma) - \frac{1}{2\sigma^2}(x-\mu)^2\right) \\ & = 0 - 0 - \frac{2}{2\sigma^2}(x-\mu)(-1) \\ & = \frac{x-\mu}{\sigma^2} \end{align*} \]

Second derivative with respect to \(\mu\):

\[ \frac{\partial^2}{\partial \mu^2} \log f(x, \theta) = -\frac{1}{\sigma^2} \]

So, \(I(\mu) = -E\left[\frac{\partial^2}{\partial \mu^2} \log f(x, \theta)\right] = -E\left[-\frac{1}{\sigma^2}\right] = \frac{1}{\sigma^2}\).

2. \(I(\sigma)\)

Differentiating equation (i) with respect to \(\sigma\):

\[ \begin{align*} \frac{\partial}{\partial \sigma} \log f(x, \theta) & = \frac{\partial}{\partial \sigma} \left(-\log(\sqrt{2\pi}) - \log(\sigma) - \frac{1}{2\sigma^2}(x-\mu)^2\right) \\ & = 0 - \frac{1}{\sigma} - \frac{-2}{2\sigma^3}(x-\mu)^2 \\ & = \frac{-1}{\sigma} + \frac{(x-\mu)^2}{\sigma^3} \end{align*} \]

Second derivative with respect to \(\sigma\):

\[ \frac{\partial^2}{\partial \sigma^2} \log f(x, \theta) = \frac{1}{\sigma^2} - \frac{3(x-\mu)^2}{\sigma^4} \]

So, \(I(\sigma) = -E\left[\frac{\partial^2}{\partial \sigma^2} \log f(x, \theta)\right] = -E\left[\frac{1}{\sigma^2} - \frac{3(x-\mu)^2}{\sigma^4}\right] = \frac{2}{\sigma^2}\).

3. \(I(\sigma^2)\)

Let \(\theta = \sigma^2\), and rewrite equation (i):

\[ \log f(x, \theta) = -\log(\sqrt{2\pi}) - \frac{1}{2}\log(\theta) - \frac{1}{2\theta}(x-\mu)^2 \]

Differentiating with respect to \(\theta\):

\[ \begin{align*} \frac{\partial}{\partial \theta} \log f(x, \theta) & = \frac{\partial}{\partial \theta} \left(-\log(\sqrt{2\pi}) - \frac{1}{2}\log(\theta) - \frac{1}{2\theta}(x-\mu)^2\right) \\ & = 0 - \frac{1}{2\theta} - \frac{-1}{2\theta^2}(x-\mu)^2 \\ & = \frac{1}{2\theta} - \frac{(x-\mu)^2}{2\theta^2} \end{align*} \]

Second derivative with respect to \(\theta\):

\[ \frac{\partial^2}{\partial \theta^2} \log f(x, \theta) = \frac{1}{2\theta^2} - \frac{(x-\mu)^2}{\theta^3} \]

So, \(I(\sigma^2) = -E\left[\frac{\partial^2}{\partial \theta^2} \log f(x, \theta)\right] = -E\left[\frac{1}{2\theta^2} - \frac{(x-\mu)^2}{\theta^3}\right] = \frac{1}{2\theta^2}\).

\[ I(\mu) = \frac{1}{\sigma^2} \]

\[ I(\sigma) = \frac{2}{\sigma^2} \]

\[ I(\sigma^2) = \frac{1}{2\sigma^4} \]

Cramer-Rao Inequality

Regularity Conditions:

  • The parameter space \(\Theta\) is an open interval.
  • The support or range of the distribution is independent of \(\theta\).
  • For every \(x\) and \(\theta\), \(\frac{\partial}{\partial\theta} f(x,\theta)\) and \(\frac{\partial^2}{\partial\theta^2} f(x,\theta)\) exist and are finite.
  • The statistic \(T\) has finite mean and variance.
  • Differentiation and integration are permissible, i.e., \(\frac{\partial}{\partial\theta} \int T L(x,\theta) \, dx = \int \frac{\partial}{\partial\theta} T L(x,\theta) \, dx\).

Cramer-Rao Inequality

Regularity Conditions:

  • The parameter space \(\Theta\) is an open interval.
  • The support or range of the distribution is independent of \(\theta\).
  • For every \(x\) and \(\theta\), \(\frac{\partial}{\partial\theta} f(x,\theta)\) and \(\frac{\partial^2}{\partial\theta^2} f(x,\theta)\) exist and are finite.
  • The statistic \(T\) has finite mean and variance.
  • Differentiation and integration are permissible, i.e., \(\frac{\partial}{\partial\theta} \int T L(x,\theta) \, dx = \int \frac{\partial}{\partial\theta} T L(x,\theta) \, dx\).

Cramer-Rao Inequality Statement:

Let \(X_1, X_2, \ldots, X_n\) be a random sample from any probability density function (p.d.f.) or probability mass function (p.m.f.) \(f(x, \theta)\), where \(\theta \in \Theta\). If \(T = T(X_1, X_2, \ldots, X_n)\) is an unbiased estimator of \(\phi(\theta)\) under regularity conditions, then:

\[ \text{Var}(T) \geq \frac{(\frac{\partial}{\partial\theta}\phi(\theta))^2}{I_n(\theta)} \quad \text{or} \quad \text{Var}(T) \geq \frac{(\phi'(\theta))^2}{nI(\theta)} \]

Proof:

Let \(x\) be a random variable following the p.d.f. or p.m.f. \(f(x,\theta)\), \(\theta \in \Theta\), and \(L(\theta)\) is the likelihood function of a random sample \(X_1, X_2, \ldots, X_n\) from the distribution. Then:

\[L(\theta) = \prod f(x, \theta) = f(X_1, X_2, \ldots, X_n, \theta)\]

\(\int L(\theta) \, dx = 1\)

\(\frac{\partial}{\partial\theta} \int L(\theta) \, dx = 0\)

\(\int \frac{\partial}{\partial\theta} L(\theta) \, dx = 0\)

\(\int \frac{1}{L} \cdot \frac{\partial}{\partial\theta} L(\theta) \, L \, dx = 0\)

\(\int \frac{\partial}{\partial\theta} \log L(\theta) \, L \, dx = 0\)

\[E\left[\frac{\partial}{\partial\theta}\log L(\theta)\right] = 0\ldots \text{ii}\]

And we know that \(T\) is an unbiased estimator of \(\phi(\theta)\), such that:

\[E(T(X)) = \phi(\theta)\]

\(\int T \cdot L(\theta) \, dx = \phi(\theta)\)

Now, differentiating with respect to \(\theta\):

\(\frac{\partial}{\partial\theta} \int T \cdot L(\theta) \, dx = \frac{\partial}{\partial\theta} \phi(\theta)\)

\(\int \frac{\partial}{\partial\theta} T \cdot L(\theta) \, dx = \phi'(\theta)\)

\(\int \frac{1}{L} \cdot \frac{\partial}{\partial\theta} L(\theta) \cdot T \, L \, dx = \phi'(\theta)\)

\(\int \frac{\partial}{\partial\theta} \log L(\theta) \cdot T \cdot L \, dx = \phi'(\theta)\)

E\(\left[\frac{\partial}{\partial\theta} \log L(\theta) \cdot T\right] = \phi'(\theta) \quad \ldots \text{iii}\)

And we have:

COV\(\left(\frac{\partial}{\partial\theta} \log L(\theta) \cdot T\right) = E\left(\frac{\partial}{\partial\theta} \log L(\theta) \cdot T\right) - 0 \cdot E(T)\)

COV\(\left(\frac{\partial}{\partial\theta} \log L(\theta) \cdot T\right) = \phi'(\theta) \quad \ldots \text{iv}\)

By Cauchy-Schwarz inequality for covariance, we have:

\[\left(\text{COV}\left(\frac{\partial}{\partial\theta} \log L(\theta) \cdot T\right)\right)^2 \leq \text{Var}\left(\frac{\partial}{\partial\theta} \log L(\theta)\right) \cdot \text{Var}(T)\]

\[\left(\phi'(\theta)\right)^2 \leq I_n(\theta) \cdot \text{Var}(T)\]

\[\text{Var}(T) \geq \frac{\left(\phi'(\theta)\right)^2}{I_n(\theta)}\]

This is the lower bound given by Cramer-Rao inequality, known as Cramer-Rao Lower Bound.

Remark: If the estimator \(T\) is unbiased, then \(\phi(\theta) = \theta\) and \(\phi'(\theta) = 1\). In this case, Cramer-Rao Lower Bound is:

\[\text{Var}(T) \geq \frac{1}{I_n(\theta)}\]

Comments

Popular posts from this blog

MCQ'S based on Basic Statistics (For B. Com. II Business Statistics)

    (MCQ Based on Probability, Index Number, Time Series   and Statistical Quality Control Sem - IV)                                                            1.The control chart were developed by ……         A) Karl Pearson B) R.A. fisher C) W.A. Shewhart D) B. Benjamin   2.the mean = 4 and variance = 2 for binomial r.v. x then value of n is….. A) 7 B) 10 C) 8 D)9   3.the mean = 3 and variance = 2 for binomial r.v. x then value of n is….. A) 7 B) 10 C) 8 D)9 4. If sampl...

Basic Concepts of Probability and Binomial Distribution , Poisson Distribution.

 Probability:  Basic concepts of Probability:  Probability is a way to measure hoe likely something is to happen. Probability is number between 0 and 1, where probability is 0 means is not happen at all and probability is 1 means it will be definitely happen, e.g. if we tossed coin there is a 50% chance to get head and 50% chance to get tail, it can be represented in probability as 0.5 for each outcome to get head and tail. Probability is used to help us taking decision and predicting the likelihood of the event in many areas, that are science, finance and Statistics.  Now we learn the some basic concepts that used in Probability:  i) Random Experiment OR Trail: A Random Experiment is an process that get one or more possible outcomes. examples of random experiment include tossing a coin, rolling a die, drawing  a card from pack of card etc. using this we specify the possible outcomes known as sample pace.  ii)Outcome: An outcome is a result of experi...

Statistical Inference: Basic Terms and Definitions.

  📚📖 Statistical Inference: Basic Terms. The theory of estimation is of paramount importance in statistics for several reasons. Firstly, it allows researchers to make informed inferences about population characteristics based on limited sample data. Since it is often impractical or impossible to measure an entire population, estimation provides a framework to generalize findings from a sample to the larger population. By employing various estimation methods, statisticians can estimate population parameters such as means, proportions, and variances, providing valuable insights into the population's characteristics. Second, the theory of estimating aids in quantifying the estimates' inherent uncertainty. Measures like standard errors, confidence intervals, and p-values are included with estimators to provide  an idea of how accurate and reliable the estimates are. The range of possible values for the population characteristics and the degree of confidence attached to those est...

Index Number

 Index Number      Introduction  We seen in measures of central tendency the data can be reduced to a single figure by calculating an average and two series can be compared by their averages. But the data are homogeneous then the average is meaningful. (Data is homogeneous means data in same type). If the two series of the price of commodity for two years. It is clear that we cannot compare the cost of living for two years by using simple average of the price of the commodities. For that type of problem we need type of average is called Index number. Index number firstly defined or developed to study the effect of price change on the cost of living. But now days the theory of index number is extended to the field of wholesale price, industrial production, agricultural production etc. Index number is like barometers to measure the change in change in economics activities.   An index may be defined as a " specialized  average designed to measure the...

B. Com. -I Statistics Practical No. 1 Classification, tabulation and frequency distribution –I: Qualitative data.

  Shree GaneshA B. Com. Part – I: Semester – I OE–I    Semester – I (BASIC STATISTICS PRACTICAL-I) Practical: 60 Hrs. Marks: 50 (Credits: 02) Course Outcomes: After completion of this practical course, the student will be able to: i) apply sampling techniques in real life. ii) perform classification and tabulation of primary data. iii) represent the data by means of simple diagrams and graphs. iv) summarize data by computing measures of central tendency.   LIST OF PRACTICALS: 1. Classification, tabulation and frequency distribution –I: Qualitative data. 2. Classification, tabulation and frequency distribution –II : Quantitative data. 3. Diagrammatic representation of data by using Pie Diagram and Bar Diagrams. 4. Graphical representation of data by using Histogram, Frequency Polygon, Frequency Curve and     Locating Modal Value. 5. Graphical representation of data by using Ogive Curves and Locating Quartile Values....

B. Com. I Practical No. 3 :Diagrammatic representation of data by using Pie Diagram and Bar Diagrams.

Practical No. 3 :Diagrammatic representation of data by using Pie Diagram and Bar Diagrams. Diagrammatic Presentation. We have observed the classification and tabulation method. We use this method to take a lot of information and make it fit into a small table. The reason we do this is to make the information more organized and easier to understand. Tabulation helps us arrange data neatly so that it's not messy and confusing. tabulation is a way to make big files of information look neat and tidy in a table.  but better and beautiful way to represent data using diagrams and graphs. the diagram and graph have some advantages because that used to visualise the data. that helps to understand and give information easily to any common man or any one, following are the some  advantages of diagram and graph.  I. Advantages i. Data Representation: Diagrams and graphs are excellent for presenting data visually, making trends, comparisons, and statistical information easier to...

Method of Moment & Maximum Likelihood Estimator: Method, Properties and Examples.

 Statistical Inference I: Method Of Moment:   One of the oldest method of finding estimator is Method of Moment, it was discovered by Karl Pearson in 1884.  Method of Moment Estimator Let X1, X2, ........Xn be a random sample from a population with probability density function (pdf) f(x, θ) or probability mass function (pmf) p(x) with parameters θ1, θ2,……..θk. If μ r ' (r-th raw moment about the origin) then μ r ' = ∫ -∞ ∞ x r f(x,θ) dx for r=1,2,3,….k .........Equation i In general, μ 1 ' , μ 2 ' ,…..μ k ' will be functions of parameters θ 1 , θ 2 ,……..θ k . Let X 1 , X 2 ,……X n be the random sample of size n from the population. The method of moments consists of solving "k" equations (in Equation i) for θ 1 , θ 2 ,……..θ k to obtain estimators for the parameters by equating μ 1 ' , μ 2 ' ,…..μ k ' with the corresponding sample moments m 1 ' , m 2 ' ,…..m k ' . Where m r ' = sample m...

Time Series

 Time series  Introduction:-         We see the many variables are changes over period of time that are population (I.e. population are changes over time means population increase day by day), monthly demand of commodity, food production, agriculture production increases and that can be observed over period of times known as time series. Time series is defined as a set of observation arranged according to time is called time series. Or a time Series is a set of statistical observation arnging chronological order. ( Chronological order means it is arrangements of variable according to time) and it gives information about variable.  Also we draw the graph of time series to see the behaviour of variable over time. It can be used of forecasting. The analysis of time series is helpful to economist, business men, also for scientist etc. Because it used to forecasting the future, observing the past behaviour of that variable or items. Also planning for future...

Non- Parametric Test: Run Test

Non- Parametric Test  A Non-Parametric tests is a one of the part of Statistical tests that non-parametric test does not assume any particular distribution for analyzing the variable. unlike the parametric test are based on the assumption like normality or other specific distribution  of the variable. Non-parametric test is based on the rank, order, signs, or other non-numerical data. we know both test parametric and non-parametric, but when use particular test? answer is that if the assumption of parametric test are violated such as data is not normally distributed or sample size is small. then we use Non-parametric test they can used to analyse categorical data  or ordinal data and data are obtained form in field like psychology, sociology and biology. For the analysis use the  some non-parametric test that are Wilcoxon signed-ranked test, mann-whiteny U test, sign test, Run test, Kruskal-wallis test. but the non-parametric test have lower statistical power than ...