Learn how to choose the right number of responses to make your survey results meaningful and statistically reliable.

Man in wheelchair working on laptop, next to screenshot of survey setup


Survey sample size is ​​the number of completed responses you need for your findings to reliably represent your population. In research, the definition of sample size also reflects how much precision you can expect from your estimates, since factors like sample size vs. population, margin of error, and confidence level all shape statistical significance.

A clear sample size helps you understand how confidently you can interpret patterns in your data and whether your results generalize to the broader group you want to study.

In this guide, you’ll learn how sample size works, how to determine survey sample size, and how to calculate your sample size—whether you’re running a quick feedback study or planning formal survey research.

Survey sample size is ​​the number of completed responses you need for your findings to reliably represent your population. Sample size reflects both the scope of your target population and the level of precision you want from your estimates. A right-sized sample helps ensure your findings are statistically significant and not due to random variation.

Your ideal sample size depends on factors such as the population size, the margin of error, and the confidence level. Most teams use a few standard settings and confirm the number of completed responses they’ll need with a sample size calculator. Clear inputs and a consistent approach set the foundation for reliable survey research across audiences and use cases.

Before calculating your sample size, it helps to understand how it relates to response rate and completion rate.

  • Sample size is the number of completed responses you need for your findings to reliably represent your population.
  • Response rate is the percentage of invited people who submit a complete response.
  • Completion rate is the percentage of people who start your survey and finish it.

For example, you might send your survey to a sample size of 1,000 people. If 400 people open your survey, your response rate is 40%. In most cases, not everyone who opens your survey will complete it. If 400 people open your survey but only 200 submit responses, your completion rate is 50%.

While response and completion rates do not set your sample size directly, they influence how many people you need to invite to reach your desired number of completed responses. Accounting for these rates helps ensure your final sample is representative and supports the reliability and statistical significance of your results.

Your survey sample size is determined by three inputs: the size of the population you want to represent, the margin of error you can tolerate, and the confidence level you need. These settings work together to show the sample size required to detect statistically significant differences. The clearer your inputs, the easier it is to confirm the required number of completions in a sample size calculator.

Population is the full group you want your results to reflect. Required sample sizes level off for large populations; after roughly 100,000 people, you typically need about the same number of completes, no matter how big the audience gets. When your planned sample is a sizable share of the population (about 5% or more), you may apply the finite population correction (FPC) to slightly reduce the required number of completes.

Margin of error sets how precise your estimates must be. Teams often start with ±5%; higher-stakes decisions usually tighten to ±3% to reduce variability. Smaller margins of error require larger sample sizes because they allow for less uncertainty in your results.

Confidence level shows how often you'd expect similar results if you repeated the survey with another sample. 95% confidence is a strong default, 90% confidence requires fewer completes when speed matters, and 99% confidence increases certainty but typically needs a larger sample. These inputs work together to determine a statistically significant survey sample size.

The fastest way to calculate survey sample size is to use a sample size calculator. Once you enter your population, margin of error, and confidence level, the calculator shows how many completed responses you need to reach a statistically reliable result.

If you want to see the math behind the calculation, here’s the standard sample size calculation formula for estimating a proportion:

Sample size formula
  • n = required sample size (completed responses)
  • z = z-score for your chosen confidence level (for 95% confidence, z = 1.96)
  • p = expected proportion selecting a given answer (use 0.5 if you’re not sure)
  • E = margin of error in decimal form (e.g., 5% = 0.05)

Example:

If your population is 10,000,000 and you choose 95% confidence (z = 1.96), a ±5% margin of error (E = 0.05), and p = 0.5, the formula gives:

n≈384.16n \approx 384.16n≈384.16

Rounded up, you need 385 completed responses to estimate results for a 10-million-person population at those settings. You can confirm this number or test different scenarios in the sample size calculator, then adjust your inputs based on the precision you need in your estimates.

After you determine your target sample size, you’ll want a method for randomly selecting participants. A straightforward approach is to generate a random sample in Excel to select survey recipients without introducing selection bias.

Want a quick estimate of how many people need to take your survey? Here’s a handy table to help you decide. All you need is the number of people in your target population and your desired margin of error (3-10%). These minimum survey completions assume a 95% confidence level and a 50% response distribution (p = 0.5) unless stated otherwise.

Population±3%±5%±10%
50034522080
1,00052528590
3,000810350100
5,000910370100
10,0001,000385100
100,0001,100400100
1,000,0001,100400100
10,000,0001,100400100

*Percentages represent 3%, 5%, and 10% margin of error

For example, if you’re running medical research, you might stick to a lower margin of error, like 3%. But if you’re choosing a customer satisfaction survey sample size, you might be more comfortable with a higher margin of error, like 10%.

A good sample size depends on how precise your results need to be, how diverse your population is, and how much risk is tied to the decision you’re making. Larger samples support tighter margins of error, higher confidence, and stronger subgroup analysis—but bigger isn’t always better. In some surveys, rich qualitative feedback or overall participation matters more than hitting a specific sample size target.

Below are the scenarios where a larger sample truly makes a difference—and where it doesn’t.

When accuracy is critical, such as medical studies, policy research, or safety-related changes, you’ll generally want a larger sample with a tighter margin of error (often ±3%) and at least a 95% confidence level.

These settings reduce variability and help ensure your estimates reflect the full population. Probability sampling is ideal here, since each person has a known chance of being selected and results support more rigorous inference.

For decisions like new market entry, pricing work, or brand research, a moderate-to-large sample size lets you segment results by region, demographic groups, or customer type.

As population diversity increases—differences in age, location, income, or other traits—you’ll typically need more completed responses to ensure each subgroup is represented. A margin of error around ±3–5% usually supports this level of analysis.

When your audience is small and well-defined (such as 20–100 event attendees or usability testers), you don’t need a large sample to learn what matters. Written feedback, open-ended responses, and engagement patterns often provide more value than achieving a specific sample size.

A full census of the group or a pragmatic nonprobability approach is usually sufficient—as long as you stay mindful of potential biases.

Across all of these scenarios, remember that sample size has diminishing returns. Once your population reaches a certain scale, required sample sizes level off, and you usually only adjust with a finite population correction when your achieved sample is a sizable share of the total.

Sampling design determines how you select people for your survey, and it shapes whether you end up with a representative sample and reliable results. A strong design helps reduce sampling bias and ensures your sample reflects the population you want to understand.

Sampling methods fit into two groups: probability sampling and non-probability sampling, each with a set of common techniques.

Probability sampling gives every person in your target population a known chance of being selected. This approach is the foundation for producing a representative sample, supporting stronger inferences, and reducing sampling bias. 

There are four widely used probability sampling methods, each suited to different research needs.

In simple random sampling, every member of your population has an equal chance of being chosen, often through a lottery process or random number generator. This method minimizes bias because selection is truly random, but it requires a complete and accurate list of the entire population, which can be challenging to obtain for large or decentralized groups.

Systematic sampling selects people at regular intervals, such as every 50th name on a list. It’s fast, efficient, and easy to implement, especially when working with ordered data. However, if the list contains underlying patterns that align with the interval (for example, if every 50th entry shares a similar characteristic), it can unintentionally introduce bias.

Stratified sampling divides your population into meaningful subgroups—such as age ranges, regions, or customer types—and then randomly samples within each group. This method is especially valuable when subgroup representation matters, because it ensures that your sample reflects the diversity of your audience. The trade-off is that it requires accurate information about your strata upfront and takes more planning to implement.

Cluster sampling groups your population into clusters, such as neighborhoods, schools, or store locations. You then randomly select a set of clusters and survey everyone within them. This approach is practical and cost-effective for large, geographically dispersed populations, but it can increase sampling error if the clusters differ significantly

Non-probability sampling is used when people in your population do not have an equal or known chance of being selected. These approaches are often faster, more flexible, and less expensive than probability methods, but they carry a higher risk of sampling bias. Because selection isn’t random, the results may not represent the full population as reliably. 

Here are the five common non-probability sampling methods and when they’re typically used.

Convenience sampling collects responses from the people who are easiest to reach—such as customers already on your email list or visitors to your website. It’s quick and cost-effective, making it useful for exploratory work or early feedback. However, because participants self-select based on accessibility, their views may not reflect the broader population.

Purposive sampling selects individuals because they have specific characteristics or expertise relevant to your research. This method is commonly used in qualitative studies where you want depth over breadth—for example, interviewing product experts or longtime customers. The trade-off is that researcher judgment influences who is included, which can introduce bias.

Voluntary response sampling relies on people choosing to take your survey on their own, such as opting in from a link on a website or social post. It’s simple to deploy and can generate responses quickly. At the same time, it often overrepresents people who are highly motivated, either very satisfied or dissatisfied, which can skew results.

Snowball sampling begins with a small group of participants who then recruit others from their networks. This method is helpful when you’re studying niche or hard-to-reach populations, because it expands your reach organically. The limitation is that social networks tend to include people with similar traits, which can narrow the range of perspectives in your sample.

Quota sampling divides your population into categories—such as age, gender, or location—and then fills a preset number of responses in each category. This can help you quickly achieve demographic balance. However, because participants within each quota aren’t selected randomly, results may still be influenced by selection bias.

  • Is sample size the number invited or the number who answered?
  • Do I always need 1,000+?
  • Which is better: 90% or 95% confidence?
  • Do I have to set p?

Need help choosing your sample size? From expert sampling design to reaching your target market, pick the path that fits your next step.

You don’t need advanced statistics to set the right sample size. A few clear inputs and a quick check are all it takes. Start with common defaults like 95% confidence and a ±5% margin of error, then tighten or relax those settings based on how precise your results need to be. Once you’ve set your population, confidence level, and margin of error, use the sample size calculator to confirm the minimum number of completed responses your study requires.

When you’re ready to field your survey, you can get started for free and build it quickly. If your study calls for verified respondents or specific demographic profiles, SurveyMonkey Audience can help you reach targeted participants, with built-in quality controls to support reliable results.