Non-Parametric Test, Characteristics, Types

Non-Parametric Tests, also known as distribution free tests, are statistical hypothesis tests that do not assume the underlying data follow any specific probability distribution such as the normal distribution. Non-parametric tests work with ordinal or nominal data and are valid even with small samples or severe violations of assumptions. In Indian business research, non parametric tests are essential when analyzing Likert scale data (treated as ordinal), ranked preferences, or highly skewed variables like income or expenditure. Common non parametric tests include Mann Whitney U (alternative to independent t test), Kruskal Wallis H (alternative to ANOVA), Wilcoxon signed rank (alternative to paired t test), Spearman rank correlation (alternative to Pearson), and chi square test. While less powerful than parametric tests when assumptions hold, they are safer and more widely applicable.

Characteristics of Non parametric Test:

1. No Normality Assumption

Non parametric tests do not require the data to follow a normal distribution. This is their most important characteristic. While parametric tests (t test, ANOVA) assume that data are drawn from normally distributed populations, non parametric tests make no such assumption. They work equally well with skewed, flat, peaked, or multimodal distributions. In Indian business research, many variables are naturally non normal: income (right skewed), number of purchases (discrete with many zeros), or satisfaction ratings (bunched at high values). Non parametric tests handle these without transformation or data cleaning. This characteristic makes them robust and widely applicable. However, if data are normal, non parametric tests are slightly less powerful (require larger samples to detect the same effect).

2. Applicable to Ordinal and Nominal Data

Non parametric tests work with ordinal data (rankings, Likert scales) and nominal data (categories) where parametric tests require interval or ratio measurement. For example, Mann Whitney U test compares two groups on an ordinal outcome like “satisfaction rank.” Spearman correlation measures association between two ranked variables. Chi square test analyzes nominal frequency data. In Indian business research, much data are ordinal: customer satisfaction on 1 to 5 scale, employee performance rankings, or product preference orders. Parametric tests would be inappropriate because means and standard deviations are not meaningful for ordinal data. Non parametric tests use medians, ranks, or frequencies instead of means. This characteristic matches the measurement level of most social and business research data.

3. Fewer Assumptions

Non parametric tests have minimal assumptions compared to parametric alternatives. They do not assume normality, homogeneity of variance, linearity, or interval level measurement. The primary assumptions are: independent observations (except for paired designs) and that the underlying distributions are continuous (for rank based tests). Some tests assume that the two groups being compared have similarly shaped distributions. In Indian business research, this characteristic is valuable because real world data rarely meet parametric assumptions. Researchers need not spend time on transformation, outlier removal, or complex assumption checks. However, fewer assumptions come with a cost: lower statistical power when parametric assumptions are actually met. The trade off is simplicity and safety versus power. Choose based on your data characteristics.

4. Based on Ranks or Signs

Most non parametric tests convert raw data into ranks or signs before analysis, discarding the actual numerical values. For example, Mann Whitney U test combines two groups, ranks all observations together from smallest to largest, then analyzes the sum of ranks in each group. Wilcoxon signed rank test uses the signs and magnitudes of differences between paired observations. This rank based approach makes the tests resistant to outliers because extreme values become ranks 1, 2, or N instead of pulling means. In Indian business research, outliers are common in expenditure data, transaction values, or response times. Ranking neutralizes their influence. However, ranking loses information about the actual distances between observations. Two observations with a tiny difference receive consecutive ranks, as do two with a huge difference.

5. Lower Power (When Assumptions Met)

Non parametric tests have lower statistical power than their parametric counterparts when the data meet parametric assumptions (normality, equal variance). Power is the probability of correctly rejecting a false null hypothesis. For normally distributed data, a parametric test (e.g., t test) can detect a true effect with a smaller sample size than a non parametric test (e.g., Mann Whitney). The loss of power is typically 5 to 15 percent, meaning a non parametric test requires about 10 to 20 percent more observations to achieve the same power. In Indian business research, if you have large samples (n > 100 per group) or are confident of normality, parametric tests are preferred. For small samples (n < 30) or non normal data, non parametric tests may actually have higher power because parametric assumptions are violated.

6. Robust to Outliers

Non parametric tests are highly robust to outliers because they use ranks or signs rather than actual values. An outlier that is extremely large becomes simply the highest rank, not a value that pulls the mean dramatically. For example, in a t test, a single income of ₹10,00,000 in a sample otherwise around ₹50,000 would distort the mean and variance, potentially causing a false significant result or masking a real difference. In Mann Whitney U test, that same outlier becomes the largest rank (e.g., rank 50) with no disproportionate influence. In Indian business research, outliers are common in financial data, transaction values, and survey response times. Non parametric tests provide safety without requiring outlier removal (which is subjective and reduces sample size). This robustness is a major practical advantage for messy real world data.

7. Valid for Small Samples

Non parametric tests are valid even with very small sample sizes (n < 30) where parametric tests may be unreliable due to unknown distribution shape. For example, Mann Whitney U test can be used with as few as 3 or 4 observations per group, using exact probability tables rather than normal approximations. Parametric t tests with such small samples require normality, which cannot be verified. In Indian business research, pilot studies, case studies, or research on rare populations (e.g., women entrepreneurs in a small village) often yield small samples. Non parametric tests allow valid inference without pretending that small samples are normal. However, power is very low with tiny samples; non significant results should be interpreted cautiously. For very small samples (n < 5 per group), consider descriptive analysis only.

8. No Homogeneity of Variance Requirement

Non-parametric tests do not require equal variances across groups. Levene’s test and similar assumption checks are unnecessary. Mann Whitney U test and Kruskal Wallis H test work correctly whether variances are equal or unequal. This characteristic is valuable when comparing groups with naturally different spreads. For example, in Indian business research, income variance among high income professionals may be much larger than among low income workers. A t test would require variance correction (Welch’s test) or transformation. Mann Whitney U handles it without adjustment. However, if the groups have different distribution shapes (not just different spreads), the interpretation of Mann Whitney U shifts from testing medians to testing stochastic dominance. Consult statistical guidance for such cases.

9. Can Be Used with Count Data

Non parametric tests work appropriately with count data (number of purchases, website visits, complaints) that often violate parametric assumptions. Count data are typically non negative integers with many zeros, right skewed, and heteroscedastic (variance increases with mean). Parametric tests on count data often require Poisson or negative binomial regression. Non parametric tests like Mann Whitney U or Kruskal Wallis H provide a simpler alternative for comparing counts across groups. In Indian business research, comparing number of ecommerce transactions between two cities or number of customer complaints before and after training can use non parametric tests. However, if you need to model the effect of multiple predictors on count outcomes, non parametric tests cannot handle this; use specialized count regression (Poisson, negative binomial) instead.

10. Results Are Medians or Ranks, Not Means

Non parametric tests produce results based on medians or rank sums, not means and standard deviations. For example, Mann Whitney U tests whether one group tends to have larger values than the other (stochastic superiority), often reported as comparing medians. The test statistic (U) represents the number of times observations from one group exceed observations from the other. In Indian business research, reporting “median satisfaction was 4 for Group A and 3 for Group B” is often more appropriate than “mean satisfaction 3.8 vs 3.2” when data are ordinal or skewed. However, some readers expect mean comparisons. Explain why you chose non parametric methods. Effect sizes for non parametric tests include rank biserial correlation (for Mann Whitney) or epsilon squared (for Kruskal Wallis). Report these alongside p values.

Types of Non Parametric Tests:

1. Mann Whitney U-Test

The Mann Whitney U test (also called Wilcoxon rank sum test) is the non parametric alternative to the independent samples t test. It compares two independent groups on a continuous or ordinal dependent variable. The test ranks all observations from both groups together, then sums the ranks within each group. The U statistic measures whether one group’s ranks are systematically higher than the other’s. In Indian business research, use Mann Whitney when comparing ecommerce satisfaction between men and women with non normal data or small samples. Assumptions: independent observations, ordinal or continuous data, and similar distribution shapes between groups. Report U value, sample sizes, and p value. For effect size, calculate rank biserial correlation (r = 1 2U/n1n2). This test is robust, widely applicable, and available in all statistical software.

2. Wilcoxon Signed Rank Test

The Wilcoxon signed rank test is the non parametric alternative to the paired samples t test. It compares two related (paired) measurements from the same subjects, such as before and after scores, or two conditions. The test calculates differences between each pair, ignores zero differences, ranks the absolute differences, then sums ranks separately for positive and negative differences. In Indian business research, use Wilcoxon when comparing customer satisfaction before and after a website redesign with non normal difference scores. Assumptions: paired observations, ordinal or continuous data, and symmetry of differences (not normality). Report the test statistic (W or T), sample size (number of non zero differences), and p value. For effect size, calculate r = Z/√N. Wilcoxon is more powerful than the sign test because it considers magnitude of differences.

3. Kruskal Wallis H Test

The Kruskal Wallis H test is the non parametric alternative to one way ANOVA. It compares three or more independent groups on a continuous or ordinal dependent variable. The test ranks all observations across all groups together, then compares the sum of ranks between groups. The H statistic follows a chi square distribution with k 1 degrees of freedom. In Indian business research, use Kruskal Wallis when comparing ecommerce satisfaction across five cities or three income groups with non normal data. If significant, follow with post hoc pairwise comparisons using Dunn’s test with Bonferroni correction. Assumptions: independent observations, ordinal or continuous data, and similarly shaped distributions across groups. Report H value, degrees of freedom, and p value. Effect size: epsilon squared (ε²) or eta squared based on ranks. Do not use multiple Mann Whitney tests without correction.

4. Friedman Test

The Friedman test is the non parametric alternative to repeated measures ANOVA. It compares three or more related (paired) conditions measured on the same subjects, such as satisfaction ratings for three different ecommerce platforms rated by the same customers. The test ranks observations within each subject (row), then compares column rank sums. The Friedman statistic follows a chi square distribution with k 1 degrees of freedom. In Indian business research, use Friedman when customers rate multiple brands or employees rate different training programs. If significant, follow with post hoc comparisons using Wilcoxon signed rank tests with Bonferroni correction. Assumptions: ordinal or continuous data, and rows (subjects) are independent. Report chi square value, degrees of freedom, and p value. Effect size: Kendall’s W (coefficient of concordance). Friedman is powerful for repeated measures designs without normality.

5. Spearman Rank Correlation

Spearman rank correlation (Spearman’s rho, ρ) is the non parametric alternative to Pearson correlation. It measures the strength and direction of monotonic association between two ordinal or continuous variables. The test converts both variables to ranks, then calculates Pearson correlation on those ranks. Values range from 1 (perfect negative monotonic) to +1 (perfect positive monotonic). In Indian business research, use Spearman when correlating customer satisfaction rank with loyalty rank, or when variables are skewed or contain outliers. Unlike Pearson, Spearman detects any monotonic relationship (not just linear). Assumptions: paired observations, ordinal or continuous data, and monotonic relationship. Report rho value, sample size, and p value. For interpretation: 0.10 small, 0.30 medium, 0.50 large effect. Spearman is robust, widely used, and available in all software.

6. Chi Square Test of Independence

The chi square test of independence determines whether two categorical variables are associated or independent. It compares observed frequencies in a contingency table to expected frequencies under the null hypothesis of no association. The test statistic χ² = Σ (O E)²/E follows a chi square distribution with (rows 1)(columns 1) degrees of freedom. In Indian business research, use chi square to test association between gender (male, female) and ecommerce platform preference (Amazon, Flipkart, Meesho). Assumptions: random sample, independent observations, mutually exclusive categories, and expected frequency at least 5 in 80 percent of cells. Report χ² value, degrees of freedom, sample size, and p value. For significant results, examine standardized residuals. Effect sizes: phi coefficient (2×2 tables) or Cramer’s V (larger tables). Chi square does not assume normality and works with nominal data.

7. Chi Square Goodness of Fit Test

The chi square goodness of fit test determines whether an observed frequency distribution matches an expected theoretical distribution for a single categorical variable. For example, does the proportion of customers using UPI (observed 60 percent), credit card (25 percent), and cash on delivery (15 percent) match the company’s expected distribution (50, 30, 20)? The test calculates χ² = Σ (O E)²/E with k 1 degrees of freedom, where k is number of categories. In Indian business research, use goodness of fit to test if customer preferences are equally distributed across brands or if sample demographics match population demographics. Assumptions: random sample, independent observations, mutually exclusive categories, and expected frequency at least 5 per category. Report χ² value, degrees of freedom, and p value. Effect size: Cohen’s w. This test has no parametric equivalent.

8. Fisher’s Exact Test

Fisher’s exact test is used for 2×2 contingency tables when chi square assumptions are violated (expected frequency < 5 in any cell). It calculates the exact probability of observing the obtained table or more extreme tables given the marginal totals, using the hypergeometric distribution. Unlike chi square which approximates, Fisher’s test provides exact p values. In Indian business research, use Fisher’s test when comparing two binary outcomes (e.g., purchase yes/no) between two groups (e.g., coupon received yes/no) with small sample sizes. For larger tables (R x C), use Fisher Freeman Halton extension. Assumptions: random sample, independent observations, fixed marginal totals. Report exact p value (one tailed or two tailed) and odds ratio with confidence interval. Fisher’s test is conservative and valid for any sample size. It is the gold standard for small sample categorical analysis.

9. McNemar’s Test

McNemar’s test is used for paired nominal data, specifically 2×2 contingency tables where the same subjects are measured twice (before after). It tests whether the proportion of discordant pairs (changed from Yes to No versus No to Yes) is equal. For example, testing whether a training program changed employees’ attitude toward ecommerce adoption (favorable vs unfavorable) measured before and after. The test statistic follows a chi square distribution with 1 degree of freedom. In Indian business research, use McNemar when you have paired binary outcomes. Assumptions: paired observations, nominal binary data, and the sample of discordant pairs is not too small (correction available). Report chi square value and p value. Effect size: odds ratio for discordant pairs. McNemar has no parametric equivalent and is specifically designed for repeated binary measurements.

10. Kolmogorov Smirnov Test

The Kolmogorov Smirnov (K S) test compares an observed sample distribution to a theoretical distribution (one sample) or compares two sample distributions (two sample). It is based on the maximum absolute difference between cumulative distribution functions. Unlike other non parametric tests that compare central tendency (medians), the K S test detects any differences in shape, spread, or central tendency. In Indian business research, use one sample K S to test normality (though Shapiro Wilk is more powerful for small samples). Use two sample K S to compare whether two groups (e.g., urban vs rural ecommerce spending distributions) come from the same population distribution. Assumptions: continuous data, independent observations. Report D statistic and p value. The K S test is sensitive to differences anywhere in the distribution, not just medians.

One thought on “Non-Parametric Test, Characteristics, Types

Leave a Reply

error: Content is protected !!