intermediate 18 minutes

Point Estimation

Learn about point estimators and their properties. Understand bias, efficiency, consistency, and choosing the best estimator.

On This Page
Advertisement

What is Point Estimation?

A point estimate is a single value used to estimate an unknown population parameter.

ParameterSymbolPoint EstimatorSymbol
Population meanμSample mean
Population proportionpSample proportion
Population varianceσ²Sample variance
Population SDσSample SDs

Properties of Good Estimators

What makes one estimator better than another? Three key properties:

1. Unbiasedness

An estimator is unbiased if its expected value equals the parameter being estimated.

Unbiasedness

An estimator θ^\hat{\theta} is unbiased for θ if:

E(θ^)=θE(\hat{\theta}) = \theta

The estimator is “correct on average.”

Unbiased vs Biased

Unbiased: Sample mean x̄ for estimating μ

  • E(x̄) = μ ✓

Biased: Sample range for estimating population range

  • Sample range tends to underestimate population range

Bias of an estimator: Bias(θ^)=E(θ^)θBias(\hat{\theta}) = E(\hat{\theta}) - \theta

If Bias = 0, the estimator is unbiased.

2. Efficiency

Among unbiased estimators, the one with smallest variance is most efficient.

Relative Efficiency

Comparing estimators θ^1\hat{\theta}_1 and θ^2\hat{\theta}_2:

Relative Efficiency=Var(θ^2)Var(θ^1)\text{Relative Efficiency} = \frac{Var(\hat{\theta}_2)}{Var(\hat{\theta}_1)}

If this is greater than 1, θ^1\hat{\theta}_1 is more efficient.

Efficiency

For estimating μ of a normal distribution:

EstimatorVarianceEfficiency
Sample mean x̄σ²/nMost efficient
Sample median1.57σ²/nLess efficient

The mean uses all the data; the median ignores some information.

3. Consistency

An estimator is consistent if it converges to the true parameter as sample size increases.

Consistency

θ^\hat{\theta} is consistent for θ if:

As n → ∞, θ^\hat{\theta} → θ (in probability)


The Sample Mean

Sample Mean

xˉ=1ni=1nxi\bar{x} = \frac{1}{n}\sum_{i=1}^{n} x_i

Properties:

  • E(x̄) = μ (unbiased)
  • Var(x̄) = σ²/n
  • Consistent (variance → 0 as n → ∞)

The sample mean is the minimum variance unbiased estimator (MVUE) of μ for normal populations.


The Sample Variance

Sample Variance

s2=1n1i=1n(xixˉ)2s^2 = \frac{1}{n-1}\sum_{i=1}^{n}(x_i - \bar{x})^2

Why n-1? Division by n-1 makes s² an unbiased estimator of σ².


Sample Proportion

Sample Proportion

p^=Xn=number of successesn\hat{p} = \frac{X}{n} = \frac{\text{number of successes}}{n}

Properties:

  • E(p̂) = p (unbiased)
  • Var(p̂) = p(1-p)/n
  • Consistent

Mean Squared Error (MSE)

Sometimes we accept a small bias in exchange for lower variance. The MSE balances both:

Mean Squared Error

MSE(θ^)=E[(θ^θ)2]=Var(θ^)+[Bias(θ^)]2MSE(\hat{\theta}) = E[(\hat{\theta} - \theta)^2] = Var(\hat{\theta}) + [Bias(\hat{\theta})]^2


Maximum Likelihood Estimation (MLE)

Maximum Likelihood Estimation finds the parameter value that makes the observed data most likely.

MLE Concept

Given data x₁, x₂, …, xₙ and parameter θ:

Likelihood: L(θ)=P(dataθ)L(\theta) = P(\text{data} | \theta)

MLE: Find θ that maximizes L(θ)

MLE for Proportion

Flip a coin 100 times, get 60 heads.

What value of p makes this most likely?

L(p) = P(60 heads in 100 flips | p) = (10060)p60(1p)40\binom{100}{60}p^{60}(1-p)^{40}

Taking the derivative and setting to 0:

MLE: p̂ = 60/100 = 0.60

(This matches our intuition!)

Properties of MLEs

  • Asymptotically unbiased: Bias → 0 as n → ∞
  • Consistent: Converges to true value
  • Asymptotically efficient: Minimum variance as n → ∞
  • Invariant: MLE of g(θ) = g(MLE of θ)

Comparing Estimators

Choosing an Estimator

Estimating population center μ:

EstimatorUnbiased?EfficiencyRobust?
Sample meanYesBest for normalNo (sensitive to outliers)
Sample medianNo (for normal)Less efficientYes (robust to outliers)
Trimmed meanSlightly biasedModerateModerately robust

Choice depends on:

  • Distribution shape (normal? skewed? outliers?)
  • Sample size
  • Research goals

Standard Error

The standard error is the standard deviation of an estimator.

Standard Error

For sample mean: SE(xˉ)=σnSE(\bar{x}) = \frac{\sigma}{\sqrt{n}}

Estimated by: SE(xˉ)=snSE(\bar{x}) = \frac{s}{\sqrt{n}}

EstimatorStandard Error
Sample meanσ/√n (or s/√n)
Sample proportion√[p(1-p)/n] (or √[p̂(1-p̂)/n])
Difference in means√(σ₁²/n₁ + σ₂²/n₂)

Summary

In this lesson, you learned:

  • Point estimate: Single value to estimate a parameter
  • Unbiased: E(estimator) = parameter
  • Efficient: Smallest variance among unbiased estimators
  • Consistent: Converges to true value as n → ∞
  • MSE = Variance + Bias²: Overall accuracy measure
  • Sample mean x̄ is unbiased for μ
  • Sample variance s² uses n-1 to be unbiased for σ²
  • MLE maximizes the likelihood of observed data
  • Standard error measures variability of an estimator

Practice Problems

1. A sample of 64 has mean 50 and standard deviation 8. What is the standard error of the sample mean?

2. An estimator has E(θ̂) = θ + 2 and Var(θ̂) = 9. a) Is it unbiased? b) What is its MSE?

3. Why do we divide by n-1 instead of n when calculating sample variance?

4. You flip a coin 80 times and get 52 heads. What is the MLE for the probability of heads?

Click to see answers

1. SE(x̄) = s/√n = 8/√64 = 8/8 = 1

2a. No, it is biased. E(θ̂) = θ + 2 ≠ θ, so Bias = 2

2b. MSE = Var(θ̂) + Bias² MSE = 9 + 2² = 9 + 4 = 13

3. Dividing by n produces a biased estimator that underestimates σ².

When we calculate s², we use x̄ instead of the true μ. Since x̄ is calculated from the same data, the deviations from x̄ are artificially small (x̄ minimizes Σ(xᵢ - x̄)²).

Dividing by n-1 corrects for this, producing an unbiased estimator. (n-1 = degrees of freedom)

4. MLE for proportion: p̂ = successes/trials p̂ = 52/80 = 0.65

Next Steps

Build on your estimation knowledge:

Advertisement

Was this lesson helpful?

Help us improve by sharing your feedback or spreading the word.