1 / 15

The t test

The t test. Peter Shaw RIl. "Small samples are slippery customers whose word is not to be taken as gospel" (Moroney). Introduction. Last week we met the Mann-Whitney U test, a non-parametric test to examine how likely it is that 2 samples could have come from the same population.

valarie
Download Presentation

The t test

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The t test Peter Shaw RIl "Small samples are slippery customers whose word is not to be taken as gospel" (Moroney).

  2. Introduction • Last week we met the Mann-Whitney U test, a non-parametric test to examine how likely it is that 2 samples could have come from the same population. • This week we explore other approaches to this and related situations.

  3. Student’s t test • This test was invented by a statistician working for the brewer Guinness. He was called WS Gosset (1867-1937), but preferred to keep anonymous so wrote under the name “Student”. • Hence we have Student’s t test, the Studentised range, etc - in memory of Mr Gosset (!).

  4. T vs U? • These 2 tests are identical in hypothesis formulation. • They require 2 samples which may be from the same population. These samples need not be of equal #, nor are they paired. • H0: The 2 samples are from the same population - any differences are due to chance • H1: The 2 samples come from different populations.

  5. 1 big difference (+ a few small ones): • The t test is a parametric test - it assumes the data are normally distributed. • There are several different versions of the t test, depending on exactly what assumptions you make about the data. I’ll stick to the simplest.

  6. The basic idea Remember Z scores? These apply to the idealised normal distribution How many s.d.s is this data point from the mean? Zi = (Xi- μ)/σ We can look up Z in tables, but these assume that the values of μ and σ are known perfectly. σ μ

  7. Gosset’s discovery: • Was the formulae appropriate to Z when the sample is small, so that μ and S are based on inadequate data. • To distinguish this distribution from the idealised normal distribution, Gosset named the function the “t statistic”, and the value of (Xi- μ)/S when μ and S are estimates was renamed from Z to t. • Hence t is really just a special, unreliable Z score. To identify a t score you must also specify how many data points it comes from: a value based on 6 observations is FAR less reliable than one based on 6000.

  8. The theory... You have 2 samples which may be from 1 distribution or 2. To assess the likelihood, find how many s.d.s the means of the 2 populations are apart: How many S.D.’s? Calculate t = (μ1 - μ2) / pooled sd μ1 μ2

  9. The details are slightly more messy.. • Because of the question “How do we calculate the pooled sd?” • There are several ways of doing this which make different assumptions, and give slightly different answers. • The simplest model assumes that the 2 samples have a common variance, and gives t as follows: • Given data X1, X2 which have N1, N2 datapoints each, and sums of squares SSx1, SSx2 • t = (μ1 - μ2) with N1+n2-1 df • __________ • sq.root[(SSx1 + SSx2)*(1/Nx+1/Ny) / (n1+n2-2)]

  10. Beware! • I spent an afternoon in the library once checking ways to calculate t. • I found 3 different formulae, plus several confusing ways to express the relationship I just showed you. • Another one widely used differs in assuming that the 2 samples have unequal variance. This gives a messier formula, plus another even messier formula for the df. • The third approach assumes that samples are accurately paired - the paired samples t test.

  11. x1 x2 • 57.8 64.2 • 56.2 58.7 • 61.9 63.1 • 54.4 62.5 • 53.6 59.8 • 56.4 59.2 • 53.2 • n 7 6 • Sum x 393.5 367.5 • sumx*2 22174.41 22535.87 • mean 56.21% 61.25% • ss 54.089 26.495

  12. So you know what to do to compare 2 groups! X1 X2 X3 • You have the choice of M-W U, or Student’s t test. • But what if there are 3 groups, or 4, or 5? • You may work out the following routine: • Test group 1 vs group 2, then 2 vs 3, etc. • Clever, but WRONG! (The danger with multiple tests is that you will get a “p=0.05” significant result more often than 1:20).

  13. Multiple groups can be compared.. • With a suitable multiple test. • There are 2 options here, both of which are usually run on PCs. • Parametric data: Analysis of variance ANOVA • Non-Parametric data: Kruskal-Wallis ANOVA. • I make M.Sc. students run ANOVA calculations by hand, but K-W ANOVA is PC only.

  14. Type of data Number of groups: Parametric Non-Parametric 2 T test Mann-Whitney U test >=2 Analysis of variance (ANOVA) Kruskal_Wallis ANOVA

  15. n 7 6 • Sum x 393.5 367.5 • sumx*2 22174.41 22535.87 • mean 56.21% 61.25% • SS 54.089 26.495 ┌─ ─┐ • Sediffere = sqrt│(SSxx + SSyy)*(1/Nx+1/Ny)│ • │──────────────── │ • │ Nx+Ny-2 │ • └─ ─┘ Example: SEdiff= sqrt[(26.495+54.089)*(1/6+1/7)/(6+7-2)] = sqrt[2.2675] = 1.506 Hence t = (61.25 - 56.21)/1.506 = 3.35 with 12df This is significant at p<0.05

More Related