What are 'Nonparametric Statistics'

Nonparametric statistics refer to a statistical method in which the data is not required to fit a normal distribution. Nonparametric statistics uses data that is often ordinal, meaning it does not rely on numbers, but rather a ranking or order of sorts. For example, a survey conveying consumer preferences ranging from like to dislike would be considered ordinal data.

BREAKING DOWN 'Nonparametric Statistics'

In statistics, parametric statistics includes parameters such as the mean, median, standard deviation, variance, etc. This form of statistics uses the observed data to estimate parameters of the distribution. Under parametric statistics, data is assumed to fit a normal distribution with unknown parameters μ (population mean) and σ2 (population variance), which are then estimated using the sample mean and sample variance. For example, a researcher that wants an estimate of the number of babies in North America born with brown eyes in 2017 may decide to take a sample of 150,000 babies and run an analysis on the data set. The measurement that s/he derives will be used as an estimate of the entire population of babies with brown eyes born in 2017.

Nonparametric statistics does not assume that data is drawn from a normal distribution. Instead, the shape of the distribution is estimated under this form of statistical measurement. While there are many situations in which a normal distribution can be assumed, there are also some scenarios in which it will not be possible to determine whether the data will be normally distributed. For example, consider a researcher who wants to know whether going to bed early or late is linked to how frequently one falls ill. Assuming the sample is chosen randomly from the population, the sample size distribution of illness frequency can be assumed to be normal. However, an experiment that measures the resistance of the human body to a strain of bacteria cannot be assumed have a normal distribution. This is because a randomly selected sample data may be resistance to the strain. On the other hand, if the researcher considers factors such as genetic make up and ethnicity, he may find that a sample size selected using these characteristics may not be resistant to the strain. Hence, one cannot assume a normal distribution.

Nonparametric statistics includes nonparametric descriptive statistics, statistical models, inference, and statistical tests. The model structure of nonparametric models is not specified a priori but is instead determined from data. The term 'non-parametric' is not meant to imply that such models completely lack parameters, but that the number and nature of the parameters are flexible and not fixed in advance. A histogram is an example of a nonparametric estimate of a probability distribution.

Nonparametric statistics makes no assumption about the sample size or whether the observed data is quantitative. This method is useful when the data has no clear numerical interpretation, and is best to use with data that has a ranking of sorts. For example, a personality assessment test may have a ranking of its metrics set as strongly disagree, disagree, indifferent, agree, and strongly agree. In this case, nonparametric methods should be used.

Nonparametric statistics have gained appreciation due to their ease of use. As the need for parameters is relieved, the data becomes more applicable to a larger variety of tests. This type of statistics can be used without the mean, sample size, standard deviation, or the estimation of any other related parameters when none of that information is available. Since nonparametric statistics makes fewer assumptions about the sample data, its application is wider in scope than parametric statistics.

In cases where parametric testing is more appropriate, nonparametric methods will be less efficient. This is because the results obtained from nonparametric statistics have a lower degree of confidence than if the results were obtained using parametric statistics.

 

RELATED TERMS
  1. Sampling Distribution

    A sampling distribution is a probability distribution of a statistic ...
  2. Sampling

    Sampling is a process used in statistical analysis in which a ...
  3. T Distribution

    A T distribution is a type of probability function that is appropriate ...
  4. Simple Random Sample

    A simple random sample is a subset of a statistical population ...
  5. T-Test

    A t-test is an analysis framework used to determine the difference ...
  6. Frequency Of Exclusion

    Frequency of exclusion refers to the rate of occurrence of a ...
Related Articles
  1. Investing

    How to Use Stratified Random Sampling

    Stratified random sampling is a technique best used with a sample population easily broken into distinct subgroups. Samples are then taken from each subgroup based on the ratio of the subgroup’s ...
  2. Investing

    Most Common Probability Distributions

    In this article, we'll go over a few of the most popular probability distributions and show you how to calculate them.
  3. Investing

    Lognormal and normal distribution

    When and why do you use lognormal distribution or normal distribution for analyzing securities? Lognormal for stocks, normal for portfolio returns.
  4. Investing

    Calculating volatility: A simplified approach

    Though most investors use standard deviation to determine volatility, there's an easier and more accurate way of doing it: the historical method.
  5. Personal Finance

    Birch Box Review: Is It Worth It?

    Learn more about the convenience of the subscription beauty box industry, and discover why the Birchbox company in particular has become so popular.
  6. Trading

    Using Genetic Algorithms to Forecast Financial Markets

    Genetic algorithms are problem-solving methods that mimic the process of natural evolution and can be applied to predicting security prices.
  7. Investing

    Bet Smarter With the Monte Carlo Simulation

    This technique can reduce uncertainty in estimating future outcomes.
  8. Tech

    How Big Data Has Changed Finance

    The vast proliferation of data and increasing technological complexities continues to transform the way industries operate and compete.
RELATED FAQS
  1. When is it better to use systematic over simple random sampling?

    Learn when systematic sampling is better than simple random sampling, such as in the absence of data patterns and when there ... Read Answer >>
  2. What is the difference between systematic sampling and cluster sampling?

    Learn about the differences between systematic sampling and cluster sampling, including how the samples are created for each ... Read Answer >>
  3. What are the disadvantages of using a simple random sample to approximate a larger ...

    Learn here what a simple random sample is, how researchers use it as a statistical tool and the disadvantages it carries ... Read Answer >>
  4. What assumptions are made when conducting a t-test?

    Learn what a t-test is and discover the five standard assumptions made regarding the validity of sampling and data used in ... Read Answer >>
  5. How do I calculate the standard error using Matlab?

    Learn how to calculate the standard error for a sample statistical measure, such as the sample mean, using standard Matlab ... Read Answer >>
  6. What's an example of stratified random sampling?

    Stratified random sampling divides a population into subgroups or strata, whereby the members in each of the stratum formed ... Read Answer >>
Trading Center