Survey Sample Size Guide

Updated February 12, 2026

You don’t need to survey everyone to get reliable results. Sample size is the number of completed survey responses needed to represent a larger population within a chosen margin of error and confidence level. A properly sized sample can reflect thousands - or even millions - of people with known accuracy.

This guide explains how sample size works, why it matters, and how to avoid common mistakes. Use the Sample Size Calculator to calculate your required sample size.

Most business surveys don’t fail because of bad questions. They fail because of bad sampling.

Sample size

Why Sample Size Matters

Sample size affects two key factors:

  • Margin of error: How close your sample results are to the true population value
  • Confidence level: How certain you can be that your results fall within the margin of error

Get the sample size wrong, and you’ll either waste resources surveying too many people or get unreliable results from too few.

Quick Reference: How Many Responses Do You Need?

All values assume 95% confidence level and maximum variability (p = 0.5). Use the Sample Size Calculator for custom values.

Your Population±3% Margin±5% Margin±10% Margin
50484534
100928050
20016913266
50034121881
1,00051727888
5,00088035795
10,00096537096
50,000+1,04538296

For populations above 50,000, sample size barely changes - you’re already past the point of diminishing returns.

Key Concepts

Population Size

The total number of people who could take your survey. This might be:

  • All your customers (e.g., 50,000)
  • Employees in your company (e.g., 500)
  • Users of a specific feature (e.g., 10,000)
  • Visitors to your website last month (e.g., 100,000)

For very large populations (100,000+), sample size barely changes because you’re already capturing statistical significance.

Confidence Level

How certain you want to be that your sample results reflect the true population.

Confidence LevelInterpretation
90%You’re 90% sure the true value falls within your margin of error
95%Industry standard - 95% certainty
99%Very high certainty, requires larger sample

Use 95% unless you have a specific reason to change it. It’s the standard for business decisions.

Margin of Error

The acceptable range of error in your results. If your survey shows 60% satisfaction with a ±5% margin of error, the true value is between 55% and 65%.

Margin of ErrorUse Case
±3%High-stakes decisions requiring precision
±5%Standard for most business surveys
±10%Exploratory research or resource-constrained surveys

Smaller margins require larger samples. A ±3% margin needs roughly 2.5× more responses than ±5%.

The Sample Size Formula

This guide uses Cochran’s formula adjusted for finite populations.

Step 1: Calculate Initial Sample Size (Infinite Population)

n₀ = (Z² × p × (1-p)) / e²

Where:

  • Z = Z-score for your confidence level (1.96 for 95%)
  • p = Expected proportion (0.5 for maximum variability)
  • e = Margin of error as decimal (0.05 for 5%)

Step 2: Adjust for Finite Population

n = n₀ / (1 + (n₀ - 1) / N)

Where:

  • N = Your population size
  • n₀ = Initial sample size from step 1

You don’t need to solve this manually-this is what the calculator does.

Why We Use p = 0.5

The proportion (p) represents expected response distribution. If you’re asking a yes/no question, p = 0.5 means you expect a 50/50 split.

We use 0.5 because:

  • It produces the largest (most conservative) sample size
  • You don’t need to guess what results you’ll get
  • It’s the safest assumption when you don’t have prior data

If you know your proportion is extreme (e.g., 90% yes), you can use a smaller sample-but only when you have strong prior data.

Response Rates & Invitations

Your required sample size is not the number of people to invite. You need to account for response rates.

Typical Survey Response Rates

Survey TypeResponse Rate
In-app/embedded20-40%
Email (customers)10-30%
Email (employees)30-50%
Post-transaction15-25%
NPS relationship surveys10-20%

Formula: People to invite = Sample size ÷ Expected response rate

If you need 400 responses and expect 20% response rate:

  • 400 ÷ 0.20 = 2,000 people to invite

Common Mistakes

Ignoring response rates

Calculating you need 400 responses but only inviting 400 people guarantees insufficient data.

Using population size for sample size

“We have 10,000 customers, so we need 10,000 responses.” You don’t. Around 370 responses give you ±5% accuracy.

Forgetting about subgroups

If you plan to analyze by segment (region, product, customer type), each subgroup needs adequate sample size. You might need 400 responses per segment, not 400 total.

Not considering non-response bias

If only happy customers respond, your results won’t represent everyone. Aim for response rates above 20% to minimize bias.

Over-precision

A ±3% margin sounds better than ±5%, but it requires 2.5× more responses. For most business decisions, ±5% is sufficient.

When Sample Size Rules Don’t Apply

Standard sample size calculations assume:

  • Random sampling: Everyone has equal chance of being selected
  • Representative responses: Respondents are similar to non-respondents
  • Single population: You’re measuring one group, not comparing groups

These assumptions break down when:

  • You’re surveying self-selected volunteers
  • Response rates are very low (under 10%)
  • You’re running a longitudinal study
  • You have highly niche populations

In these cases, statistical accuracy depends more on sampling method than sample size.

Best Practices

Define your population clearly

Who exactly are you surveying? “Our customers” is vague. “Active customers who made a purchase in the last 12 months” is specific.

Use stratified sampling for diverse populations

If your population has distinct segments, sample proportionally from each to ensure representation.

Plan for analysis upfront

If you’ll break down results by region, ensure each region has enough responses. Three respondents from Asia isn’t statistically meaningful.

Document your methodology

Note your population size, sampling method, response rate, and margin of error. This adds credibility to your findings.

Consider the decision at stake

A ±10% margin might be fine for exploratory research. For decisions affecting millions of dollars, invest in ±3% precision.

Sample Size Examples by Use Case

Employee Engagement Survey

Scenario: Company with 500 employees
Required responses: 217 (±5%, 95% confidence)
Expected response rate: 40%
Action: Invite all 500 employees, expect 200-250 responses

Customer Satisfaction Survey

Scenario: 10,000 active customers
Required responses: 370 (±5%, 95% confidence)
Expected response rate: 15%
Action: Invite 2,500 customers via email. Use the CSAT Calculator to measure satisfaction from collected responses.

Product Feedback Survey

Scenario: 100,000+ users
Required responses: 384 (±5%, 95% confidence)
Expected response rate: 10%
Action: Show in-app survey to 4,000 users

Small Business Survey

Scenario: 50 customers
Required responses: 44 (±5%, 95% confidence)
Expected response rate: 30%
Action: Invite all 50 customers - at 30% you’ll get ~15 responses, which is below target. Send a follow-up reminder (typically doubles response rate) or accept ±10% margin (only 33 needed).

Next Steps

Use the Sample Size Calculator to determine your target, then check out our survey templates to get started.

Frequently Asked Questions

Estimate conservatively. If you think you have 5,000-10,000 customers, use 5,000. For very large populations (50,000+), the exact number barely matters - the required sample size plateaus around 380 responses for ±5% accuracy.

Not necessarily. Beyond your minimum sample size, additional responses have diminishing returns. Going from 400 to 4,000 responses changes margin of error from ±5% to ±1.5% - rarely worth the effort for most business decisions.

Sample size calculations apply to quantitative data (numbers, percentages). For open-ended questions, you'll often reach "saturation" (no new themes emerging) with 20-30 thoughtful responses.

Keep surveys short (5 minutes max), send at optimal times (Tuesday-Thursday, 10am-2pm), use personalization in invitations, explain why feedback matters, and offer incentives if appropriate. Follow-up reminders typically double your initial response rate.

It depends on your population size. For populations over 10,000, 100 responses gives you roughly ±10% margin of error at 95% confidence. That's acceptable for exploratory research but not for important business decisions. For ±5% accuracy, you need at least 370 responses.

The "rule of 30" applies to statistical tests like t-tests, not to survey sampling. For surveys, 30 responses rarely provides meaningful accuracy. Even with a small population of 100, you need about 80 responses for ±5% margin of error. Use the sample size calculator to find the right number for your situation.

It varies by channel. Employee surveys typically get 30-50%, customer email surveys 10-30%, and in-app surveys 20-40%. Anything above 20% is generally considered adequate for minimizing non-response bias. Below 10%, your results may not represent the full population regardless of sample size.

For 95% confidence with ±5% margin of error: about 370 responses for large populations (10,000+), 278 for 1,000 people, or 217 for 500 people. The smaller your population, the larger the proportion you need to survey. Use the sample size calculator for exact numbers.

With 50 responses from a large population, your margin of error is roughly ±14% at 95% confidence. That's imprecise but still useful for directional insights - identifying major trends or severe problems. Consider accepting a wider margin, or try follow-up reminders, shorter surveys, and in-app distribution to increase your response count.

Calculate the margin of error for your actual response count. If your survey got 50 responses from 5,000 people, your margin of error is about ±14% - meaning a result of 60% could actually be anywhere from 46% to 74%. If that range is too wide to make a confident decision, your sample is too small. Use the sample size calculator to check.

No. Sample size depends on population size, margin of error, and confidence level - not on the number of questions. However, longer surveys reduce response rates, which means you may need to invite more people.