VisualLab is a powerful tool for probability and statistics anlysis of data. Descriptive statistics and linear or nonlinear regression are simply a matter of click. When t-tests and ANOVA tests are involved, however, it would be ideal to have some extra utilities to visualize the processes how p-value is worked out from t-value or f-value through the t-distribution probability density curve and f-distribution curve.
A t-test, also known as Student's t-test, is a test to determine if the difference between the means of two groups is statistically significant, or to determine if the mean of a group is statistically significant from a given value. It was developed by William Sealy Gosset under the pseudonym Student. The probability density function of t-test is symmetric, and its overall shape resembles the bell shape of a normally distributed variable with mean 0 and variance 1. As the number of d.o.f. grows, t-distribution approaches the normal distribution with μ = 0 and σ = 1.
A t-test can be a one sample test or two sample test (paired or unpaired). The above animation shows if you want to visualize how p-value is worked out from t-value, you simply input the d.o.f. value to display its corresponding t-distribution probability density curve. You then integrate the area size under the t-curve over the t-value (or under the t-value if negative) to obtain one tailed p-value. Two tailed p-value is the double of one tailed p-value, representating the area size of two tails on the both sides of the t-distribution curve.
ANOVA (analysis of variance):
ANOVA (analysis of variance) is a technique in statistics developed by tatistician Ronald
Fisher. It is used to analyse the variation among and between groups to check if the means of
two or more groups are significantly different from each other. Several screen shots are shown below
One way ANOVA: Screen shot for one way ANOVA, click the button "Run".
Two way ANOVA: Screen shot for two way ANOVA, check the factor box "2" and click the button
From F-value to p-value:
F-value is the ratio of between-group variability over within-group variability, a.k.a. the
ratio of explained variance over unexplained variance. There is a mini calculator to allow one
to compute p-value from F-value and d.o.f between groups and d.o.f. within groups.
An F-test is a statistical test in which the test statistic has an F-distribution under the
null hypothesis. The name was in honour of Ronald Fisher who initially developed the statistic
as the variance ratio around a century ago.
The F-distribution probability density function, or F-curve in short is needed to compute
p-value from F-value. Type the value of d.o.f. into the box, and click the button "F-distribution
curve", you'll see the corresponding F curve displaying on screen. You then integrate
the area size under the F-curve over the F-value to obtain the p-value. Unlike a t-test, there
is no "two tailed" p-value in an F-test.
F-distribution probability density function: Screen shot for the F-curve, F-value and p-value.
Normal distribution and Binomial distribution
Normal distribution (a.k.a. Gaussian or Gauss or Laplace - Gauss distribution) is a type of
continuous probability density distribution for a real-valued random variable. The general
form of its probability density function is:
where, parameter μ is the mean or expectation of the distribution; and σ is its
standard deviation. Variance of the distribution is σ2. A random variable with
a Gaussian distribution is said to be normally distributed and is called a normal deviate.
The normal distribution is the most important probability distribution in statistics.
Normal probability: Screen shot for a normal probability distribution, where assuming students' average score mark is μ = 70, with standard deviation σ = 10. The screen shot shows 5.48% students having score mark higher than 86 (standard normal deviate = 1.6), 94.5% students below.
Binomial distribution is a discrete probability distribution. The outcome of a trial is always binary: yes or no; positive or negative, true or false. If the probability for one outcome is described as p, the probability for the other outcome would be q = 1 - p. The binomial distribution function shows that for a sequence of n independent trials (a.k.a. Bernoulli trial or Bernoulli experiment), the probability that k trials turn out positive outcome, Pr(k). The binomial distribution is the basis for the popular binomial test of statistical significance.
Binomial probability: Screen shot for a binomial probability distribution, where n = 50, and p = 0.52.
To visit DSC Curve Solutions (DCS ®) and non-isothemal kinetics, click links shown in naviation bar.