Published on December 11, 2020 by Pritha Bhandari. Revised on July 6, 2022. The standard error of the mean, or simply standard error, indicates how different the population mean is
likely to be from a sample mean. It tells you how much the sample mean would vary if you were to repeat a study using new samples from within a single population. The standard error of the mean (SE or SEM) is the most commonly reported type of standard error. But you can also find the standard error for other statistics, like
medians or proportions. The standard error is a common measure of sampling error—the difference between a population parameter and a sample statistic. In statistics, data from
samples is used to understand larger populations. Standard error matters because it helps you estimate how well your sample data represents the whole population. With probability sampling, where elements of a sample are randomly selected, you can collect data that
is likely to be representative of the population. However, even with probability samples, some sampling error will remain. That’s because a sample will never perfectly match the population it comes from in terms of measures like means and standard deviations. By calculating standard error, you can estimate how representative your sample is of your population and make valid conclusions. A high standard error shows that sample means are widely spread around the population
mean—your sample may not closely represent your population. A low standard error shows that sample means are closely distributed around the population mean—your sample is representative of your population. You can decrease standard error by increasing sample size. Using a large, random sample is the best way to minimize sampling bias. Standard error and standard deviation are both measures of variability: The standard deviation is a descriptive statistic that can be calculated from sample data. In contrast, the standard error is an inferential statistic that can only be estimated (unless the real population parameter is known). The standard deviation of the math scores is 180. This number reflects on average how much each score differs from the sample mean score of 550. The standard error of the math scores, on the other hand, tells you how much the sample mean score of 550 differs from other sample mean scores, in samples of equal size, in the population of all test takers in the region. What can proofreading do for your paper?Scribbr editors not only correct grammar and spelling mistakes, but also strengthen your writing by making sure your paper is free of vague language, redundant words and awkward phrasing. See editing example Standard error formulaThe standard error of the mean is calculated using the standard deviation and the sample size. From the formula, you’ll see that the sample size is inversely proportional to the standard error. This means that the larger the sample, the smaller the standard error, because the sample statistic will be closer to approaching the population parameter. Different formulas are used depending on whether the population standard deviation is known. These formulas work for samples with more than 20 elements (n > 20). When population parameters are knownWhen the population standard deviation is known, you can use it in the below formula to calculate standard error precisely.
When population parameters are unknownWhen the population standard deviation is unknown, you can use the below formula to only estimate standard error. This formula takes the sample standard deviation as a point estimate for the population standard deviation.
First, find the square root of your sample size (n).
Next, divide the sample standard deviation by the number you found in step one.
The standard error of math SAT scores is 12.8. How should you report the standard error?You can report the standard error alongside the mean or in a confidence interval to communicate the uncertainty around the mean. Example: Reporting the mean and standard error The mean math SAT score of a random sample of test takers is 550 ± 12.8 (SE).The best way to report the standard error is in a confidence interval because readers won’t have to do any additional math to come up with a meaningful interval. A confidence interval is a range of values where an unknown population parameter is expected to lie most of the time, if you were to repeat your study with new random samples. With a 95% confidence level, 95% of all sample means will be expected to lie within a confidence interval of ± 1.96 standard errors of the sample mean. Based on random sampling, the true population parameter is also estimated to lie within this range with 95% confidence. For a normally distributed characteristic, like SAT scores, 95% of all sample means fall within roughly 4 standard errors of the sample mean.
With random sampling, a 95% CI [525 575] tells you that there is a 0.95 probability that the population mean math SAT score is between 525 and 575. Other standard errorsAside from the standard error of the mean (and other statistics), there are two other standard errors you might come across: the standard error of the estimate and the standard error of measurement. The standard error of the estimate is related to regression analysis. This reflects the variability around the estimated regression line and the accuracy of the regression model. Using the standard error of the estimate, you can construct a confidence interval for the true regression coefficient. The standard error of measurement is about the reliability of a measure. It indicates how variable the measurement error of a test is, and it’s often reported in standardized testing. The standard error of measurement can be used to create a confidence interval for the true score of an element or an individual. Frequently asked questions about standard errorWhat is standard error? The standard error of the mean, or simply standard error, indicates how different the population mean is likely to be from a sample mean. It tells you how much the sample mean would vary if you were to repeat a study using new samples from within a single population. Sources in this articleWe strongly encourage students to use sources in their work. You can cite our article (APA Style) or take a deep dive into the articles below. This Scribbr article
Is this article helpful?You have already voted. Thanks :-) Your vote is saved :-) Processing your vote... What is the relationship between standard error and standard deviation?The standard error of the sample mean depends on both the standard deviation and the sample size, by the simple relation SE = SD/√(sample size).
What is the relationship between standard error and sample size?The standard error is also inversely proportional to the sample size; the larger the sample size, the smaller the standard error because the statistic will approach the actual value. The standard error is considered part of inferential statistics. It represents the standard deviation of the mean within a dataset.
What is the relationship between the sample size and the coefficient of standard error and why?The standard error measures the dispersion of the distribution. As the sample size gets larger, the dispersion gets smaller, and the mean of the distribution is closer to the population mean (Central Limit Theory). Thus, the sample size is negatively correlated with the standard error of a sample.
How does sample size affect mean standard deviation and standard error of the mean?Standard error increases when standard deviation, i.e. the variance of the population, increases. Standard error decreases when sample size increases – as the sample size gets closer to the true size of the population, the sample means cluster more and more around the true population mean.
|