New Methodology Published Mar 30, 2026
Confidence Interval
A confidence interval is the blurry margin around a study’s estimate that shows how much the result could reasonably wobble if the study were repeated.
Also known as
CI · 95% CI · confidence limits · interval estimate
Why this matters
This is the number range that tells you whether a study result is precise or shaky. If you only read the single headline number and ignore the interval, you can mistake a rough guess for a solid finding, especially in supplement studies with small sample sizes.
4 min read · 807 words · 4 sources · evidence: robust
Deep dive
How it works
For many common estimates, a confidence interval is built from three ingredients: the estimate itself, a measure of how much estimates would vary across repeated samples, and a critical value from a probability distribution. As sample size grows, that sampling variability usually shrinks, which is why confidence intervals often get narrower in larger studies. But narrowing also depends on noise in the data and on whether the model assumptions are reasonable.
When you'll see this
The term in the wild
Scenario
You read a creatine trial abstract reporting: lean mass gain 1.1 kg, 95% CI 0.2 to 2.0.
What to notice
The study’s best estimate is 1.1 kg, but the plausible range is much wider. The result supports a gain, yet the exact size is still somewhat fuzzy.
Why it matters
This keeps you from treating 1.1 kg as a guaranteed outcome for every person who takes creatine.
Scenario
A forest plot in a fish oil meta-analysis shows several horizontal lines crossing the vertical no-effect line.
What to notice
Each line is a confidence interval graph for one study. If a line crosses the no-effect marker, that study is compatible with little or no clear effect on that outcome.
Why it matters
You learn to read the picture, not just the bold pooled estimate at the bottom.
Scenario
A paper table lists blood pressure change as -2.5 mmHg (95% CI -7.0 to 2.0).
What to notice
Because the interval crosses 0, the data fit both a modest drop and almost no real change. The point estimate alone would sound more decisive than the full result.
Why it matters
This helps you avoid overclaiming from underpowered studies.
Key takeaways
- A confidence interval is a range around an estimate, not a second estimate.
- A 95% confidence interval describes the long-run performance of the method, not a 95% probability for one finished interval.
- Narrow intervals mean more precision; wide intervals mean more uncertainty.
- Whether an interval crosses the no-effect value often changes how a result should be interpreted.
- In research papers, CIs appear in tables, parentheses, brackets, forest plots, and error-bar graphs.
The full picture
The number that pretends to be exact
A paper says magnesium improved sleep score by 3.2 points. That looks crisp, almost surgical. Then you notice the line beside it: 95% CI, 0.4 to 6.0. Suddenly the result stops being a pin and starts looking more like a flashlight beam. That is the trap with confidence intervals: readers remember the center number and skim past the spread, even though the spread often tells the more honest story.
Why 95% does not mean “95% chance this study is right”
Picture an archer shooting at the same target again and again in gusty wind. Each arrow lands in a slightly different place. A study estimate works the same way: if you repeated the whole study many times, the estimate would wander because samples differ by luck alone. A confidence interval is the range produced by a method designed so that, over many repeats, 95% of those ranges would capture the true value when the method’s assumptions hold.
That is why the common phrase “there is a 95% chance the true value is inside this interval” is not quite right. After you calculate one interval, it either did or did not catch the truth; the 95% belongs to the procedure, not to your single finished interval.
What the notation is showing you
In papers, confidence interval notation often appears as estimate (95% CI 1.2 to 4.8), 95% CI [1.2, 4.8], or as horizontal bars on a confidence interval graph or forest plot. The center is the estimate. The ends are the lower and upper confidence limits. A confidence interval table usually lists the estimate, standard error, and the two limits. In many intro classes, the confidence interval formula for a mean is taught as:
estimate ± critical value × standard error
For a 95% confidence interval using the normal distribution, that critical value is often about 1.96, which is why people search for the 95% confidence interval z score. But that shortcut is not universal; different data types and smaller samples may use different distributions and assumptions.
Width matters more than the badge
The most useful question is usually not “Is it 95%?” but “How wide is it?” A narrow interval means the estimate is tightly pinned down. A wide one means the study is telling you, “somewhere in this neighborhood.” If the interval crosses the “no effect” point—often 0 for mean differences or 1 for ratios like risk ratios—the data are compatible with little or no effect as well as benefit or harm, depending on the measure.
One practical decision: when comparing two supplement studies, trust the one with the narrower interval around a meaningful effect, not just the one with the more exciting headline number. That single habit will make you a better reader of evidence immediately.
Myths vs reality
What people get wrong
Myth
A 95% confidence interval means there is a 95% chance the true value lies inside it.
Reality
Not exactly. The 95% belongs to the method: if you repeated the study many times, about 95% of those intervals would catch the true value.
Why people believe this
Everyday language treats probability as a property of one event, so the shorthand sounds natural even though the statistical meaning is different.
Myth
If two 95% confidence intervals overlap, the groups are definitely not different.
Reality
Overlap does not automatically mean “no difference.” Two estimates can have overlapping intervals and still differ statistically, depending on the comparison being tested.
Why people believe this
People learn a quick visual rule from charts and then apply it too broadly to every situation.
Myth
A 95% CI always means estimate ± 1.96 standard errors.
Reality
That is a common intro-class shortcut, not a universal law. The exact interval depends on the model, sample size, outcome type, and assumptions.
Why people believe this
The named cause is the standard normal-distribution classroom formula, where the 95% confidence interval z score is taught as 1.96 and later remembered as if it applied everywhere.
Myth
Error bars on a graph are always confidence intervals.
Reality
They might be standard deviations, standard errors, or confidence intervals. Those are different things and tell different stories.
Why people believe this
Journal figure conventions often label bars vaguely, and many papers use mean ± SEM without readers noticing that SEM bars are not confidence intervals.
How to use this knowledge
Specific failure mode: do not compare supplement results by point estimate alone when one study has a much wider interval. A flashy estimate from a tiny study can look better than a modest estimate from a precise study, even when the precise study is more trustworthy.
Frequently asked
Common questions
What does a 95% confidence interval actually tell you?
What is meant by confidence interval in plain English?
What is the so-called 95% confidence interval rule?
What are the four steps for building a confidence interval?
How do I interpret a confidence interval that crosses zero?
Related
Where this term shows up
Evidence guides and other glossary entries that touch this concept.
Concept
Concept
NewHeterogeneity (I²)
I² is the percent of study-to-study disagreement in a meta-analysis that likely reflects real differences, not just random noise.
Apr 29, 2026
Concept
Concept
NewMeta-Analysis
A meta-analysis is a way of mathematically combining similar studies so the overall pattern is easier to see than it is in any one study alone.
Apr 1, 2026
Concept
Concept
NewBlinding (Single, Double, Triple)
Blinding is the study design trick that keeps expectations from smudging the result before anyone even reads the data.
Mar 15, 2026
Concept
Concept
NewCohen's d
Cohen’s d tells you how far apart two group averages are in real-world spread, not just whether a difference technically exists.
Apr 22, 2026
Concept
Concept
NewPublication Bias
Publication bias is what happens when the studies that get published are the shiny winners, while the quiet null results stay backstage and the whole evidence picture looks better than reality.
Apr 13, 2026
Concept
Concept
NewConfounding
Confounding is when a hidden third factor makes one thing look like it caused another, the way a tilted stage can make the wrong actor seem center spotlighted.
Apr 27, 2026
Sources