Skip to main content
Stats for Scholars
Concepts Decision Tree Reporting Calculators Blog Software Cheat Sheets
Concepts Decision Tree Reporting Calculators Blog Software Cheat Sheets
Home Concepts Cronbach's Alpha

Descriptive Statistics

  • Descriptive Statistics

Inferential Statistics

  • Chi-Square Test of Independence
  • Independent Samples t-Test
  • One-Way ANOVA
  • Paired Samples t-Test
  • Pearson Correlation
  • Simple Linear Regression

Effect Size & Power

  • Effect Size
  • Sample Size Determination
  • Statistical Power & Power Analysis

Reliability & Validity

  • Cronbach's Alpha
  • Inter-Rater Reliability

Cronbach's Alpha

beginner Reliability & Validity

Cronbach's Alpha

Purpose
Measures the internal consistency of a scale — the degree to which a set of items measures a single underlying construct.
When to Use
When you have a multi-item scale or questionnaire and want to demonstrate that the items consistently measure the same construct.
Data Type
Continuous or ordinal item scores (e.g., Likert-scale responses) from a single administration
Key Assumptions
Items measure a single unidimensional construct, items are scored in the same direction (reverse-code where needed), tau-equivalence (equal true-score variances across items).
Tools
Reliability Calculator on Subthesis →

What Is Cronbach's Alpha?

Cronbach's alpha (α\alphaα) is the most widely used measure of internal consistency reliability in the social and behavioral sciences. It estimates how well a set of items on a questionnaire or scale "hang together" — that is, how consistently they measure the same underlying construct.

When you create a scale with multiple items (e.g., a 10-item anxiety questionnaire), you expect that someone who scores high on one item will tend to score high on the others. Cronbach's alpha quantifies this expectation. A high alpha means the items are intercorrelated and likely measuring the same thing; a low alpha means the items may be measuring different constructs or contain too much measurement error.

Alpha is expressed as a number between 0 and 1. Technically, alpha can be negative if items are negatively correlated (usually indicating a scoring error), but in practice you should see values between 0 and 1.

Interpretation Benchmarks

Cronbach's α\alphaα Interpretation
≥.90\geq .90≥.90 Excellent
.80.80.80 -- .89.89.89 Good
.70.70.70 -- .79.79.79 Acceptable
.60.60.60 -- .69.69.69 Questionable
.50.50.50 -- .59.59.59 Poor
<.50< .50<.50 Unacceptable

The threshold of α≥.70\alpha \geq .70α≥.70 is the most commonly cited minimum for research purposes (Nunnally, 1978). For high-stakes decisions (e.g., clinical diagnosis), α≥.90\alpha \geq .90α≥.90 is recommended.

When to Use It

Use Cronbach's alpha when:

  • You have a multi-item scale (e.g., a 15-item self-esteem questionnaire) and want to show the items are internally consistent.
  • You are developing or validating a new instrument and need to report reliability.
  • You want to justify combining multiple items into a single composite score (e.g., summing or averaging items).

Do not use Cronbach's alpha when:

  • You have a single-item measure — alpha requires at least two items.
  • Your items measure different constructs by design (e.g., a comprehensive exam covering distinct topics). Use subscale alphas instead.
  • You want to assess test-retest reliability (stability over time) or inter-rater reliability (agreement between raters). Those require different methods.

Assumptions

  1. Unidimensionality. The items should measure a single underlying construct. If your scale has multiple subscales (e.g., cognitive anxiety and somatic anxiety), compute alpha separately for each subscale. Running a factor analysis first can confirm dimensionality.

  2. Tau-equivalence. Strictly, alpha assumes that each item contributes equally to the total score (equal factor loadings). When items have very different loadings, alpha can underestimate true reliability. In such cases, consider McDonald's omega (ω\omegaω) as an alternative.

  3. Items scored in the same direction. If some items are positively worded and others negatively worded, you must reverse-code the negatively worded items before computing alpha. Failing to do so will artificially deflate alpha.

  4. Continuous or ordinal item data. Alpha is designed for items with continuous or quasi-continuous scores (e.g., Likert scales with 5+ points). For truly dichotomous items (yes/no), the Kuder-Richardson 20 (KR-20) formula is the appropriate equivalent.

Formula

Cronbach's alpha is defined as:

α=kk−1(1−∑i=1ksi2st2)\alpha = \frac{k}{k - 1}\left(1 - \frac{\sum_{i=1}^{k} s_i^2}{s_t^2}\right) α=k−1k​(1−st2​∑i=1k​si2​​)

Where:

  • kkk = the number of items on the scale
  • si2s_i^2si2​ = the variance of item iii
  • st2s_t^2st2​ = the variance of the total score (sum of all items)
  • ∑i=1ksi2\sum_{i=1}^{k} s_i^2∑i=1k​si2​ = the sum of all individual item variances

Intuition Behind the Formula

The ratio ∑si2st2\frac{\sum s_i^2}{s_t^2}st2​∑si2​​ compares the variability within individual items to the total variability. If the items are perfectly correlated, the item variances will be small relative to the total variance, and alpha will approach 1. If items are uncorrelated (pure noise), the sum of item variances will equal the total variance, and alpha will approach 0.

The term kk−1\frac{k}{k-1}k−1k​ is a correction factor that adjusts for the number of items. More items generally produce a higher alpha, even if inter-item correlations stay the same.

Standardized Alpha

When items are on different scales, the standardized alpha uses the average inter-item correlation (rˉ\bar{r}rˉ):

αstandardized=krˉ1+(k−1)rˉ\alpha_{standardized} = \frac{k \bar{r}}{1 + (k - 1)\bar{r}} αstandardized​=1+(k−1)rˉkrˉ​

This form makes clear that alpha depends on two things: the number of items (kkk) and the average correlation among items (rˉ\bar{r}rˉ).

Worked Example

Scenario: A researcher develops a 4-item scale measuring academic motivation. Five students respond on a 1--7 scale.

Student Item 1 Item 2 Item 3 Item 4 Total
A 6 5 6 7 24
B 4 4 3 4 15
C 5 5 5 6 21
D 3 2 3 3 11
E 7 6 7 7 27

Step 1: Compute item variances.

For each item, calculate s2=∑(X−Xˉ)2n−1s^2 = \frac{\sum(X - \bar{X})^2}{n - 1}s2=n−1∑(X−Xˉ)2​ (using n−1n - 1n−1 for sample variance).

  • Item 1: Xˉ1=5.0\bar{X}_1 = 5.0Xˉ1​=5.0, s12=(1+1+0+4+4)4=2.50s_1^2 = \frac{(1+1+0+4+4)}{4} = 2.50s12​=4(1+1+0+4+4)​=2.50
  • Item 2: Xˉ2=4.4\bar{X}_2 = 4.4Xˉ2​=4.4, s22=(0.36+0.16+0.36+5.76+2.56)4=2.30s_2^2 = \frac{(0.36+0.16+0.36+5.76+2.56)}{4} = 2.30s22​=4(0.36+0.16+0.36+5.76+2.56)​=2.30
  • Item 3: Xˉ3=4.8\bar{X}_3 = 4.8Xˉ3​=4.8, s32=(1.44+3.24+0.04+3.24+4.84)4=3.20s_3^2 = \frac{(1.44+3.24+0.04+3.24+4.84)}{4} = 3.20s32​=4(1.44+3.24+0.04+3.24+4.84)​=3.20
  • Item 4: Xˉ4=5.4\bar{X}_4 = 5.4Xˉ4​=5.4, s42=(2.56+1.96+0.36+5.76+2.56)4=3.30s_4^2 = \frac{(2.56+1.96+0.36+5.76+2.56)}{4} = 3.30s42​=4(2.56+1.96+0.36+5.76+2.56)​=3.30

∑si2=2.50+2.30+3.20+3.30=11.30\sum s_i^2 = 2.50 + 2.30 + 3.20 + 3.30 = 11.30 ∑si2​=2.50+2.30+3.20+3.30=11.30

Step 2: Compute total score variance.

Total scores: 24, 15, 21, 11, 27. Mean total =19.6= 19.6=19.6.

st2=(4.42+4.62+1.42+8.62+7.42)4=(19.36+21.16+1.96+73.96+54.76)4=171.204=42.80s_t^2 = \frac{(4.4^2 + 4.6^2 + 1.4^2 + 8.6^2 + 7.4^2)}{4} = \frac{(19.36 + 21.16 + 1.96 + 73.96 + 54.76)}{4} = \frac{171.20}{4} = 42.80 st2​=4(4.42+4.62+1.42+8.62+7.42)​=4(19.36+21.16+1.96+73.96+54.76)​=4171.20​=42.80

Step 3: Apply the formula.

α=44−1(1−11.3042.80)=43(1−0.264)=1.333×0.736=0.981\alpha = \frac{4}{4-1}\left(1 - \frac{11.30}{42.80}\right) = \frac{4}{3}\left(1 - 0.264\right) = 1.333 \times 0.736 = 0.981 α=4−14​(1−42.8011.30​)=34​(1−0.264)=1.333×0.736=0.981

Wait — this is suspiciously high for a 4-item scale with n=5n = 5n=5. Let us double-check. Actually, for a small sample with highly correlated items, α>.90\alpha > .90α>.90 is plausible. The items covary strongly: students who score high on one item score high on all items.

Result: α=.98\alpha = .98α=.98, indicating excellent internal consistency.

Interpretation

  • α=.98\alpha = .98α=.98 suggests the four items are very highly intercorrelated and appear to measure the same construct consistently.
  • In practice, an alpha this high might indicate item redundancy — the items may be so similar that some could be removed without losing information.
  • Remember that alpha is inflated by the number of items. A 50-item scale can achieve α>.90\alpha > .90α>.90 even with modest inter-item correlations.

What to Do if Alpha Is Low

If α<.70\alpha < .70α<.70, consider the following steps:

  1. Check for reverse-coded items that were not recoded. This is the most common cause of low alpha.
  2. Examine the "alpha if item deleted" column. Most statistics software (SPSS, R, jamovi) provides this. If removing a particular item increases alpha substantially, consider dropping it.
  3. Check item-total correlations. Items with corrected item-total correlations below .30.30.30 may not belong to the scale.
  4. Run a factor analysis. Low alpha sometimes means your items are actually measuring two or more factors. Split them into subscales and compute alpha for each.
  5. Rewrite or replace poor items. Items that do not correlate well with the total may be ambiguous or off-topic.
  6. Add more items. The Spearman-Brown prophecy formula can estimate how many items you need to reach a target alpha.

Common Mistakes

  1. Reporting alpha without checking dimensionality. A high alpha does not prove unidimensionality. Items measuring two correlated factors can still produce a high alpha. Always run a factor analysis alongside reliability analysis.

  2. Forgetting to reverse-code items. If item 3 is "I rarely feel motivated" while the other items are positively worded, you must reverse-score item 3 before computing alpha. Failing to do so will drastically lower alpha.

  3. Assuming higher is always better. Very high alpha (>.95> .95>.95) often signals item redundancy rather than excellent measurement. Aim for .80.80.80 -- .90.90.90 in most research contexts.

  4. Treating alpha as a fixed property of the scale. Alpha is a property of the scores in your sample, not the instrument itself. Always compute and report alpha for your own data, even if the original scale developers reported α=.90\alpha = .90α=.90.

  5. Using alpha for multidimensional scales. If your scale has subscales (e.g., verbal and quantitative sections of a test), report alpha for each subscale separately. A total-scale alpha is misleading if the construct is not unidimensional.

  6. Confusing reliability with validity. High reliability does not mean you are measuring what you intend to measure. A scale can be internally consistent yet measure the wrong construct entirely.

How to Report in APA Format

Internal consistency for the 4-item Academic Motivation Scale was excellent (Cronbach's α=.98\alpha = .98α=.98).

For a more detailed report:

The 10-item perceived stress scale demonstrated good internal consistency in the current sample (Cronbach's α=.84\alpha = .84α=.84). Item-total correlations ranged from .41 to .68, and no item deletion would have substantially improved alpha.

If alpha is below the threshold:

Internal consistency for the 6-item scale was questionable (α=.63\alpha = .63α=.63). Examination of item-total correlations revealed that item 4 (corrected r=.11r = .11r=.11) contributed poorly. Removing this item improved alpha to .74.74.74.

Key elements:

  • Name of the scale and number of items
  • The alpha value
  • Note on item-total correlations or items deleted if relevant
  • Use the word "current sample" to emphasize that reliability is sample-dependent

Ready to calculate?

Now that you understand the concept, use the free Reliability Calculator on Subthesis to run your own analysis.

Calculate Reliability on Subthesis

Related Concepts

Inter-Rater Reliability

Learn how to assess inter-rater reliability using Cohen's Kappa and ICC. Understand percent agreement, chance correction, and interpretation benchmarks.

Descriptive Statistics

Master descriptive statistics: learn about mean, median, mode, standard deviation, variance, and range. Know when to use each measure for your research data.

Effect Size

Learn what effect size is, why it matters more than p-values alone, and how to calculate and interpret Cohen's d, Hedges' g, and eta-squared for your research.

Stats for Scholars

Statistics for Researchers, Not Statisticians

A Subthesis Resource

Learn

  • Statistical Concepts
  • Choose a Test
  • APA Reporting
  • Blog

Resources

  • Calculators
  • Cheat Sheets
  • About
  • FAQ
  • Privacy
  • Terms

© 2026 Angel Reyes / Subthesis. All rights reserved.

Privacy Policy Terms of Use