Σ
SDCalc
AdvancedAdvanced·14 min

Hypothesis Testing with Standard Deviation

Learn how standard deviation is used in hypothesis testing. Understand t-tests, z-tests, and how to determine statistical significance.

Overview

Hypothesis testing is a statistical method for making decisions about populations based on sample data. Standard deviation plays a crucial role in determining whether observed differences are statistically significant or just due to random chance.

1

State Hypotheses

State null hypothesis (H₀) and alternative hypothesis (H₁)
2

Choose Significance Level

Choose significance level (α), typically 0.05
3

Calculate Test Statistic

Calculate test statistic using standard deviation
4

Compare to Critical Value

Compare to critical value or calculate p-value
5

Make Decision

Make decision: reject or fail to reject H₀

Z-Test

Use a Z-test when you know the population standard deviation (σ) and have a large sample size (n ≥ 30).

Z-Test Statistic

z = (x̄ - μ₀) / (σ / √n)

Example

A manufacturer claims batteries last 100 hours on average (μ₀ = 100). You test 36 batteries and find x̄ = 98 hours. If σ = 12 hours: z = (98 - 100) / (12 / √36) = -2 / 2 = -1 With z = -1 and α = 0.05 (two-tailed), we fail to reject H₀. The difference is not statistically significant.

T-Test

Use a t-test when you don't know the population standard deviation and must estimate it from the sample (using s instead of σ).

T-Test Statistic

t = (x̄ - μ₀) / (s / √n)

When to Use T-Test vs Z-Test

- Z-test: σ is known, n ≥ 30 - T-test: σ is unknown (use s), any sample size In practice, t-tests are much more common because we rarely know the true population σ.

Standard Error

The standard error (SE) measures how much sample means vary from the population mean. It's the key link between standard deviation and hypothesis testing.

Standard Error of the Mean

SE = σ / √n (or s / √n when using sample SD)

The standard error decreases as sample size increases. Larger samples give more precise estimates and make it easier to detect true differences.

Statistical Significance

A result is statistically significant when the probability of observing it by chance (p-value) is below your chosen threshold (α).

If p-value < α

Reject H₀. The result is statistically significant.

If p-value ≥ α

Fail to reject H₀. Result could be due to chance.

Statistical vs Practical Significance

A statistically significant result isn't necessarily practically important. With very large samples, tiny differences can be "significant" but meaningless in practice. Always consider effect size alongside p-values.