Welcome to the critical value calculator! Here you can quickly determine the critical value(s) for **two-tailed** tests, as well as for **one-tailed** tests. It works for most common distributions in statistical testing: the standard normal distribution **N(0,1)** (that is when you have a Z-score), **t-Student, chi-square, and F-distribution**.

What is a critical value? And what is the critical value formula? Scroll down – we provide you with the critical value definition and explain how to calculate critical values in order to use them to construct rejection regions (also known as critical regions).

## How to use critical value calculator

The critical value calculator is your go-to tool for swiftly determining critical values in statistical tests, be it one-tailed or two-tailed. To effectively use the calculator, follow these steps:

In the first field, input the

**distribution of your test statistic**under the null hypothesis: is it a standard normal N (0,1), t-Student, chi-squared, or Snedecor's F? If you are not sure, check the sections below devoted to those distributions, and try to localize the test you need to perform.In the field

`What type of test?`

choose the**alternative hypothesis**: two-tailed, right-tailed, or left-tailed.If needed, specify the

**degrees of freedom**of the test statistic's distribution. If you need more clarification, check the description of the test you are performing. You can learn more about the meaning of this quantity in statistics from the degrees of freedom calculator.**Set the significance level,**$\alpha$α. By default, we pre-set it to the most common value, 0.05, but you can adjust it to your needs.The critical value calculator will display your critical value(s) and the rejection region(s).

Click the

`advanced mode`

if you need to**increase the precision**with which the critical values are computed.

**For example,** let's envision a scenario where you are conducting a one-tailed hypothesis test using a t-Student distribution with 15 degrees of freedom. You have opted for a right-tailed test and set a significance level (α) of 0.05. The results indicate that the **critical value is 1.7531,** and the critical region is (1.7531, ∞). This implies that if your test statistic exceeds 1.7531, you will reject the null hypothesis at the 0.05 significance level.

👩🏫 Want to learn more about critical values? Keep reading!

## What is a critical value?

In hypothesis testing, critical values are one of the two approaches which allow you to decide whether to retain or reject the null hypothesis. The other approach is to calculate the p-value (for example, using the p-value calculator).

The critical value approach consists of checking if the value of the test statistic generated by your sample belongs to the so-called *rejection region*, or *critical region*, which is the **region where the test statistic is highly improbable to lie**. A critical value is a cut-off value (or two cut-off values in the case of a two-tailed test) that constitutes the boundary of the rejection region(s). In other words, critical values divide the scale of your test statistic into the rejection region and the non-rejection region.

Once you have found the rejection region, **check if the value of the test statistic generated by your sample belongs to it**:

- If so, it means that you can reject the null hypothesis and accept the alternative hypothesis; and
- If not, then there is not enough evidence to reject H
_{0}.

But how to calculate critical values? First of all, you need to **set a significance level**, $\alpha$α, which quantifies the probability of rejecting the null hypothesis when it is actually correct. The choice of α is arbitrary; in practice, we most often use a value of 0.05 or 0.01. **Critical values also depend on the alternative hypothesis you choose for your test**, elucidated in the next section.

## Critical value definition

To determine critical values, you need to know the distribution of your test statistic under the assumption that the null hypothesis holds. **Critical values are then points with the property that the probability of your test statistic assuming values at least as extreme at those critical values is equal to the significance level α**. Wow, quite a definition, isn't it? Don't worry, we'll explain what it all means.

First, let us point out it is the alternative hypothesis that determines what "extreme" means. In particular, if the test is one-sided, then there will be just one critical value; if it is two-sided, then there will be two of them: one to the left and the other to the right of the median value of the distribution.

Critical values can be conveniently depicted as the **points with the property that the area under the density curve of the test statistic from those points to the tails is equal to** $\alpha$α:

**Left-tailed test:**the area under the density curve from the critical value to the left is equal to $\alpha$α;**Right-tailed test:**the area under the density curve from the critical value to the right is equal to $\alpha$α; and**Two-tailed test:**the area under the density curve from the left critical value to the left is equal to $\alpha/2$α/2, and the area under the curve from the right critical value to the right is equal to $\alpha/2$α/2 as well; thus, total area equals $\alpha$α.

As you can see, finding the critical values for a two-tailed test with significance $\alpha$α boils down to finding both one-tailed critical values with a significance level of $\alpha/2$α/2.

## How to calculate critical values?

The formulae for the critical values involve the **quantile function**, $Q$Q, which is the inverse of the cumulative distribution function ($\mathrm{cdf}$cdf) for the test statistic distribution (calculated under the assumption that H_{0} holds!): $Q = \mathrm{cdf}^{-1}$Q=cdf−1.

Once we have agreed upon the value of $\alpha$α, the critical value formulae are the following:

**Left-tailed test**:

$\qquad \footnotesize(-\infty, Q(\alpha)]$(−∞,Q(α)]

**Right-tailed test**:

$\qquad \footnotesize[Q(1-\alpha), \infty)$[Q(1−α),∞)

**Two-tailed test**:

$\qquad \footnotesize(-\infty, Q(\frac{\alpha}{2})] \ \cup \ [Q(1 - \frac{\alpha}{2}), \infty)$(−∞,Q(2α)]∪[Q(1−2α),∞)

In the case of a **distribution symmetric about 0**, the critical values for the two-tailed test are symmetric as well:

$\small Q\left(1 - \frac{\alpha}{2}\right) = -Q\left(\frac{\alpha}{2}\right)$Q(1−2α)=−Q(2α)

Unfortunately, the probability distributions that are the most widespread in hypothesis testing have somewhat complicated $\mathrm{cdf}$cdf formulae. To find critical values by hand, you would need to use specialized software or statistical tables. In these cases, the best option is, of course, our critical value calculator! 😁

## Z critical values

Use the `Z (standard normal)`

option if your test statistic follows (at least approximately) the **standard normal distribution N(0,1)**.

In the formulae below, $u$u denotes the quantile function of the standard normal distribution N(0,1):

Left-tailed Z critical value: $u(\alpha)$u(α)

Right-tailed Z critical value: $u(1-\alpha)$u(1−α)

Two-tailed Z critical value: $\pm u(1- \alpha/2)$±u(1−α/2)

Check out Z-test calculator to learn more about the most common Z-test used on the population mean. There are also Z-tests for the difference between two population means, in particular, one between two proportions.

## t critical values

Use the `t-Student`

option if your test statistic follows the **t-Student distribution**. This distribution is **similar to N(0,1)**, but **its tails are fatter** – the exact shape depends on the number of *degrees of freedom*. If this number is large (>30), which generically happens for large samples, then the t-Student distribution is practically indistinguishable from N(0,1). Check our t-statistic calculator to compute the related test statistic.

In the formulae below, $Q_{\text{t}, d}$Qt,d is the quantile function of the t-Student distribution with $d$d degrees of freedom:

Left-tailed t critical value: $Q_{\text{t}, d}(\alpha)$Qt,d(α)

Right-tailed t critical value: $Q_{\text{t}, d}(1 - \alpha)$Qt,d(1−α)

Two-tailed t critical values: $\pm Q_{\text{t}, d}(1 - \alpha/2)$±Qt,d(1−α/2)

Visit the t-test calculator to learn more about various t-tests: the one for a **population mean with an unknown population standard deviation**, those for the **difference between the means of two populations** (with either equal or unequal population standard deviations), as well as about the **t-test for paired samples**.

## chi-square critical values (χ²)

Use the `χ² (chi-square)`

option when performing a test in which the test statistic follows the **χ²-distribution**.

You need to determine the number of degrees of freedom of the χ²-distribution of your test statistic – below, we list them for the most commonly used χ²-tests.

Here we give the formulae for chi square critical values; $Q_{\chi^2, d}$Qχ2,d is the quantile function of the χ²-distribution with $d$d degrees of freedom:

Left-tailed χ² critical value: $Q_{\chi^2, d}(\alpha)$Qχ2,d(α)

Right-tailed χ² critical value: $Q_{\chi^2, d}(1 - \alpha)$Qχ2,d(1−α)

Two-tailed χ² critical values: $Q_{\chi^2, d}(\alpha/2)$Qχ2,d(α/2) and $Q_{\chi^2, d}(1 - \alpha/2)$Qχ2,d(1−α/2)

Several different tests lead to a χ²-score:

**Goodness-of-fit test**: does the empirical distribution agree with the expected distribution?This test is

**right-tailed**. Its test statistic follows the χ²-distribution with $k - 1$k−1 degrees of freedom, where $k$k is the number of classes into which the sample is divided.**Independence test**: is there a statistically significant relationship between two variables?This test is also

**right-tailed**, and its test statistic is computed from the contingency table. There are $(r - 1)(c - 1)$(r−1)(c−1) degrees of freedom, where $r$r is the number of rows, and $c$c is the number of columns in the contingency table.**Test for the variance of normally distributed data**: does this variance have some pre-determined value?This test can be

**one- or two-tailed!**Its test statistic has the χ²-distribution with $n - 1$n−1 degrees of freedom, where $n$n is the sample size.

## F critical values

Finally, choose `F (Fisher-Snedecor)`

if your test statistic follows the **F-distribution**. This distribution has a **pair of degrees of freedom**.

Let us see how those degrees of freedom arise. Assume that you have two independent random variables, $X$X and $Y$Y, that follow χ²-distributions with $d_1$d1 and $d_2$d2 degrees of freedom, respectively. If you now consider the ratio $(\frac{X}{d_1}):(\frac{Y}{d_2})$(d1X):(d2Y), it turns out it follows the F-distribution with $(d_1, d_2)$(d1,d2) degrees of freedom. That's the reason why we call $d_1$d1 and $d_2$d2 the **numerator and denominator degrees of freedom**, respectively.

In the formulae below, $Q_{\text{F}, d_1, d_2}$QF,d1,d2 stands for the quantile function of the F-distribution with $(d_1, d_2)$(d1,d2) degrees of freedom:

Left-tailed F critical value: $Q_{\text{F}, d_1, d_2}(\alpha)$QF,d1,d2(α)

Right-tailed F critical value: $Q_{\text{F}, d_1, d_2}(1 - \alpha)$QF,d1,d2(1−α)

Two-tailed F critical values: $Q_{\text{F}, d_1, d_2}(\alpha/2)$QF,d1,d2(α/2) and $Q_{\text{F}, d_1, d_2}(1 -\alpha/2)$QF,d1,d2(1−α/2)

Here we list the **most important tests that produce F-scores: each of them is right-tailed**.

**ANOVA**: tests the equality of**means in three or more groups**that come from normally distributed populations with equal variances. There are $(k - 1, n - k)$(k−1,n−k) degrees of freedom, where $k$k is the number of groups, and $n$n is the total sample size (across every group).**Overall significance in regression analysis**. The test statistic has $(k - 1, n - k)$(k−1,n−k) degrees of freedom, where $n$n is the sample size, and $k$k is the number of variables (including the intercept).**Compare two nested regression models**. The test statistic follows the F-distribution with $(k_2 - k_1, n - k_2)$(k2−k1,n−k2) degrees of freedom, where $k_1$k1 and $k_2$k2 are the number of variables in the smaller and bigger models, respectively, and $n$n is the sample size.**The equality of variances in two normally distributed populations**. There are $(n - 1, m - 1)$(n−1,m−1) degrees of freedom, where $n$n and $m$m are the respective sample sizes.

## Behind the scenes of the critical value calculator

I'm Anna, the mastermind behind the critical value calculator and a **PhD in mathematics from Jagiellonian University**.

The idea for creating the tool originated from my experiences in teaching and research. Recognizing the need for a **tool that simplifies the critical value determination process** across various statistical distributions, I built a user-friendly calculator accessible to both students and professionals. After publishing the tool, I soon found myself using the calculator in my research and as a teaching aid.

Trust in this calculator is paramount to me. Each tool undergoes a **rigorous review process**, with peer-reviewed insights from experts and meticulous proofreading by native speakers. This commitment to accuracy and reliability ensures that users can be confident in the content. Please check the Editorial Policies page for more details on our standards.

## FAQ

### What is a Z critical value?

A Z critical value is the value that defines the **critical region in hypothesis testing** when the test statistic follows the **standard normal distribution**. If the value of the test statistic falls into the critical region, you should reject the null hypothesis and accept the alternative hypothesis.

### How do I calculate Z critical value?

To find a Z critical value for a given confidence level `α`

:

Check if you perform a

**one- or two-tailed test**.For a one-tailed test:

**Left**-tailed: critical value is thequantile of the standard normal distribution N(0,1).`α`

-th**Right**-tailed: critical value is thequantile.`(1-α)`

-th

Two-tailed test: critical value equals

quantile of N(0,1).`±(1-α/2)`

-thNo

**quantile tables**? Use CDF tables! (The quantile function is the inverse of the CDF.)Verify your answer with an online critical value calculator.

### Is a t critical value the same as Z critical value?

In theory, **no**. In practice, very often, **yes**. The t-Student distribution is **similar** to the standard normal distribution, but it is **not the same**. However, if the number of **degrees of freedom** (which is, roughly speaking, the size of your sample) is large enough (>30), then the two distributions are **practically indistinguishable**, and so the t critical value has practically the same value as the Z critical value.

### What is the Z critical value for 95% confidence?

The Z critical value for a 95% confidence interval is:

**1.96**for a two-tailed test;**1.64**for a right-tailed test; and**-1.64**for a left-tailed test.