Skip to main content
Guide

How to Read Supplement Research: The 5-Step Claim Audit

The evidence hierarchy, four red flags in any supplement study, how to skim an abstract in 60 seconds, and the three research sources actually worth trusting.

·10 min read
By Formulate Team · Independent supplement research
Key Takeaways
10 min read
  • Not all evidence is equal — the hierarchy runs meta-analyses > RCTs > cohorts > case reports > in-vitro / animal > expert opinion. Supplement marketing routinely cites the bottom of that stack
  • Four red flags that flip a study's credibility: industry funding + no disclosure, surrogate endpoints dressed as outcomes, sample sizes under ~30, and duration too short for the claimed effect
  • PubMed, Cochrane, and Examine.com are the three sources worth trusting. 'Dr.' influencer blog posts, podcast transcripts, and press releases are not evidence
  • Effect sizes matter more than p-values. A study can show 'statistically significant' 2% HbA1c improvement at p<0.001 and still be clinically meaningless

The supplement industry routinely cites “studies show” and “research proves” with a straight face. Most of those claims are true in a narrow, technical sense — thereis a study — and completely wrong in the practical sense that matters to a user deciding whether to spend $40 on a bottle. Learning to read supplement research isn’t about becoming a researcher. It’s about knowing which four questions to ask every time someone quotes a study at you.

This guide covers the evidence hierarchy, common traps, how to skim an abstract in 60 seconds, and the three research sources actually worth using. Skip the jargon — the goal is a checklist you can carry into your next supplement purchase.

The Evidence Hierarchy

Not all research is equal. In descending order of reliability for answering “does this supplement work?”:

  1. Meta-analyses of RCTs. Pool data from multiple randomized trials. The top of the pyramid — closest thing to certainty in biology. Look for Cochrane reviews specifically, which have the most rigorous methodology.
  2. Randomized controlled trials (RCTs). Participants randomly assigned to supplement or placebo. Blinded (ideally double-blinded). Measure a defined outcome. The workhorse of actual evidence.
  3. Prospective cohort studies. Follow a population forward over years, measure who takes what, correlate with outcomes. Can show association but not cause — confounders are everywhere.
  4. Case-control studies. Retrospectively compare affected vs unaffected users. Useful for rare outcomes but recall bias is a real problem.
  5. Case reports / case series. Individual or small- group reports of notable events. Useful for signal detection (“this happened, we should study it”), not for treatment decisions.
  6. Animal studies. Mouse biology is not human biology. A compound that cures cancer in mice ~90% of the time fails in human trials. Treat animal data as suggestive only.
  7. In vitro studies. Compound + cells in a dish. Even weaker than animal data for predicting human outcomes. Most “curcumin kills cancer cells” marketing comes from in vitro work — bleach kills cancer cells too.
  8. Expert opinion / mechanism / theory. Plausible biological reasoning. Important for hypothesis generation, not conclusion. “It should work because X” is where research starts, not ends.
📊Formulate's evidence grades use this hierarchy
Every ingredient on the site carries an evidence grade (A, B, C, or D) that maps directly to where its strongest available evidence sits on this pyramid. Grade A means multiple RCTs or meta-analyses converge on the same effect. Grade D means we have a biological theory and maybe an animal study — the compound is in the encyclopedia for completeness, not endorsement.

Four Red Flags in Any Supplement Study

1. Industry Funding Without Disclosure

A study funded by the supplement’s manufacturer isn’t automatically wrong — but it’s measurably more likely to produce favorable results than independent research. Check the “Funding” or “Conflict of Interest” statement at the bottom of the paper. Undisclosed industry funding is a major red flag. The FDA and most journals require disclosure now, but older papers often slipped through.

2. Surrogate Endpoints Dressed as Outcomes

A supplement lowers LDL cholesterol (measurable in blood) — but does it reduce actual heart attacks? LDL is a surrogate endpoint, a stand-in for what you actually care about. Most supplements are tested on surrogates because the real outcome (heart attack, death, disability) requires long, expensive trials. When you read “improved biomarker X,” ask whether that biomarker reliably predicts the real-world outcome. Often it doesn’t.

3. Tiny Sample Sizes

RCTs with fewer than 30 participants per arm are underpowered for most useful claims. They can detect massive effects (say, a drug that cures a disease outright), but they regularly miss real modest effects (false negatives) and produce apparent effects that don’t replicate (false positives). When a single small trial generates headlines, wait for replication.

4. Duration Too Short for the Claimed Effect

A 4-week study of a supplement claimed to prevent age-related cognitive decline can’t actually measure that claim — the effect takes years to materialize. Similarly, a 12-week weight loss study may catch early novelty effects that vanish by month 6. Match the duration to what’s being claimed.

How to Skim an Abstract in 60 Seconds

Most studies have a structured abstract. Here’s the order to read it and what to ask:

  1. Design / Methods. RCT? Crossover? Open-label? (Open-label = participants know what they’re taking = placebo effect unmeasured.) How many people? How long?
  2. Primary outcome. What was the prespecified main measure? This is the one that counts. Studies often report secondary outcomes more prominently — those are exploratory and prone to multiple-testing artifacts.
  3. Results: effect size AND statistical significance. “Significant at p<0.05” tells you the effect probably isn’t zero. It doesn’t tell you whether the effect matters. Look for the magnitude — a mean difference of 2 mg/dL on LDL with p<0.001 is statistically significant but clinically trivial.
  4. Limitations. Good papers list their own weaknesses. A paper with no limitations section is a warning sign.

Effect Size vs Statistical Significance

This is where most readers get fooled. Two concepts:

  • Statistical significance (p-value): How likely is it we’d see this result if the supplement had no effect? p<0.05 means “less than 5% likely.”
  • Effect size: How big is the effect? Expressed as mean difference, relative risk, Cohen’s d, etc.

A big enough study can produce “p<0.001” on a meaningless effect. For example: a study of 10,000 people showing a supplement reduces LDL by 1 mg/dL (statistically bulletproof) tells you the supplement does close to nothing clinically — you’d need a 20–40 mg/dL drop for meaningful cardiovascular benefit. Always ask: if this effect were true, would I act on it?

Common Traps Specific to Supplement Research

The “In Vitro to Headline” Pipeline

Journalist gets a press release: “X supplement kills cancer cells in lab study.” Cells in a dish, direct chemical contact, concentrations unreachable in human bloodstream. The headline writes itself; the real-world relevance is near zero. Curcumin, resveratrol, green tea extract, CBD, and nearly every “anti- cancer” supplement has this pedigree.

Extract Potency vs Label Potency

Studies often use specific standardized extracts at specific doses (Ashwagandha KSM-66 at 600 mg, Turmeric Meriva at 1 g). The product on the shelf may use a generic extract at half the dose in a form that doesn’t match. “Backed by research” is technically true but the product you’re buying wasn’t.

Single-Study Hype

One positive trial isn’t science — it’s a data point. True effects replicate. Wait for at least 2–3 independent positive trials before treating a claim as established. The file-drawer effect — unpublished negative trials — means single positive studies overstate effect sizes by about 30–50% in supplement literature.

Post-Hoc Subgroup Analyses

A trial failed on its primary outcome, so the authors dig through subgroups until they find one where the supplement appears to work (“effective in postmenopausal women over 60 with elevated homocysteine”). Treat post-hoc findings as hypothesis-generating, not confirmatory. They frequently fail to replicate in follow-up trials.

Where to Find Good Supplement Research

  • PubMed — the NIH’s database of biomedical literature. Free. Search with specific terms (“ashwagandha” + “randomized controlled trial”). Filter by “Meta- Analysis” or “Randomized Controlled Trial” for higher-quality results.
  • Cochrane Library — gold-standard systematic reviews. Smaller selection, but what they cover is cover is done to the highest methodological standard available.
  • Examine.com — independent research aggregator. Not peer-reviewed but methodologically rigorous. Paywalled for full detail; free for basic summaries.

Sources to skip:

  • Supplement manufacturer websites and blog posts
  • “Dr.”-branded influencer sites without cited studies
  • Podcast transcripts and YouTube videos as primary sources
  • Press releases (read the underlying paper instead)
  • Wikipedia (decent for overview, not for evidence weight)

Putting It Together: The 2-Minute Claim Audit

You see a supplement making a specific health claim. Run this checklist:

  1. What’s the evidence? Click through to the study. If there’s no link, stop here.
  2. Where does it sit on the hierarchy? Meta- analysis / RCT / cohort / animal / in-vitro? If it’s below RCT, the claim is weak regardless of how confident the marketing sounds.
  3. Was the tested product the same form and dose? Standardized extracts, specific forms, specific doses matter enormously. A study on Meriva curcumin doesn’t validate a plain turmeric powder.
  4. Was the outcome the one you care about? Did they measure a real endpoint (reduced joint pain, fewer heart attacks) or a surrogate (a biomarker that’s loosely associated with the real endpoint)?
  5. Has it replicated? One positive trial, especially industry-funded, isn’t enough. Check if 2+ independent studies converge.

If the claim survives all five, it’s probably real. Most supplement claims fail at step 2 or step 3.

How Formulate Applies This

Every supplement product and ingredient on Formulate is scored against this evidence framework. The rubric weighs evidence quality explicitly — a supplement with a plausible mechanism but zero human trials (Grade D) can’t score higher than a ceiling set by its evidence tier, no matter how good the sourcing or transparency is otherwise. The methodology page has the full scoring breakdown. The ingredient encyclopedia surfaces the evidence grade per substance so you can see the tier at a glance.

Frequently Asked Questions

What’s the difference between “significant” and “meaningful”?

Statistical significance (p<0.05) tells you the effect is probably real — non-zero. Clinical meaningfulness tells you whether the effect is big enough to matter. A supplement can be statistically significant and clinically useless at the same time. Always ask about magnitude, not just whether there was “an effect.”

Why can’t I just trust a positive study?

Because of the replication crisis and publication bias. Roughly half of initial positive findings in biomedical research fail to replicate in follow-up studies. Supplement research is worse than average due to industry funding and smaller sample sizes. Wait for independent replication before treating a single positive trial as established.

Is a systematic review better than an RCT?

Usually yes, when it’s done well. A Cochrane-style systematic review pools data from multiple RCTs and accounts for quality differences. But a bad systematic review can be worse than a good RCT — check the methodology section, not just the title.

What if there’s no research on a supplement?

Then the claim that it “works” is conjecture. Some supplements are effective and simply haven’t been trial- tested because there’s no financial incentive. But “no research” isn’t evidence of efficacy — it’s the absence of evidence. Proceed cautiously and weight the risk/benefit without a data anchor.

How do I find the original research on a supplement?

PubMed is your starting point. Search for the supplement name + “randomized.” Read the abstracts first, then focus on the most-cited recent meta-analyses. The Formulate ingredient encyclopedia also cites primary research per substance so you can skip the search step for common supplements.

See full scores in Formulate

Every product scored 50–100 against clinical research. Compare brands, check dose safety, and build your stack — free, no account required.

📬 Get new guides in your inbox

Evidence-based supplement guides, no spam

We send a short digest when new guides drop — nothing else. No affiliate pushes, no sponsored content, unsubscribe anytime.