Answers to Application Activities in Chapter 9

Size: px
Start display at page:

Download "Answers to Application Activities in Chapter 9"

Transcription

1 Answers to Application Activities in Chapter Application Activity with One Way ANOVAs 1 Ellis and Yuan (2004) Use the dataset EllisYuan.sav. Import into R as EllisYuan. a. Check Assumptions We assume that the variables were independently gathered. Use boxplots and histograms as well as numerical values for skewness and kurtosis to check on normality assumptions and homogeneity of variances. SPSS Instructions: For boxplots, histograms and numerical values from one command, choose ANALYZE > DESCRIPTIVE STATISTICS > EXPLORE. Put the three variables of ERRORFREECLAUSES, MSTTR and SPM in the Dependent List. Put Group in the Factor List. Open the Plots button and tick on Histogram and Normality plots with tests. Press Continue and then OK. The Explore command produces a lot of output, so the results won't list all of it, just my overall summary of results. R Instructions: To look at normality graphically and numerically, first make sure EllisYuan is the active dataset. First let s look at boxplots. In R Commander, choose GRAPHS > BOXPLOT, and then choose errorfreeclauses. Open the Plot by groups button and group will already be chosen, so just

2 press OK. You can press the Apply button instead of OK in the main dialogue box, which will let you then choose the other two variables (msttr, spm) in turn. Press OK when you are finished. The R code for the first variable is: Boxplot(errorfreeclauses~group, data=ellisyuan, id.method="y") Now look at histograms. In R Commander, choose GRAPHS > HISTOGRAM, and then choose errorfreeclauses. Open the Plot by groups button and group will already be chosen, so just press OK. You can press the Apply button instead of OK in the main dialogue box, which will let you then choose the other two variables (msttr, spm) in turn. Press OK when you are finished. The R code for the first variable is: with(ellisyuan, Hist(errorfreeclauses, groups=group, scale="frequency", breaks="sturges", col="darkgray")) Last, to look at numbers like skewness and kurtosis the fbasics package provides a lot of numbers, but the basicstats( ) command does not provide a way to break up the data by group, so we will have to do that manually by examining what rows each group is found in. I opened up the button on the R Commander GUI View dataset and found that the NP group used rows 1-14, PTP rows 15 28, and OLP rows By the way, the R code to do this if you do not use R Commander is: showdata(ellisyuan, placement=' ', font=getrcmdr('logfont'),

3 maxwidth=80, maxheight=30) Now knowing these rows, I can implement the command for each of the three groups for each of the three variables I want to look at: library(fbasics) #open package attach(ellisyuan) #lets me type the variable without specifying the dataset basicstats(errorfreeclauses[1:14]) basicstats(errorfreeclauses[15:28]) basicstats(errorfreeclauses[29:42]) Repeat with the other variables and at the end: detach(ellisyuan) Results: I won t list all of it, but my general summary for the normality assumptions for each variable is given below: 1 ERRORFREECLAUSES: the NP group has non-normal distribution (skewness and kurtosis numbers are high), and both NP and PTP have outliers (as seen in the boxplot). Variances seem to be unequal (as seen in the boxplot).

4 2 MSTTR: the PTP group has outliers (as seen in the boxplot). Variances seem to be unequal (as seen in the boxplot). Variances seem to be unequal (as seen in the boxplot). 3 SPM: the PTP group is skewed (skewness and kurtosis numbers are high, histogram is skewed, boxplot not symmetrically distributed) and has an outlier (as seen in the boxplot). The OLP group is also skewed (histogram and boxplot). Variances seem to be unequal (as seen in the boxplot). b. Run the One way ANOVA SPSS Instructions: Open ANALYZE > COMPARE MEANS > ONE-WAY ANOVA. Put SPM, LEXICAL VARIETY OR MSTTR, and ERROR-FREE CLAUSES in the Dependent List box and GROUP in the Factor box. In the POST HOC area you can choose which post-hoc tests to use, if you want to use any at all (if using the old statistics approach, my recommendation is to open the POST HOC button and tick LSD, which does not adjust means, and also tick Games-Howell because variances are not equal. Press Continue). Open the Options button and tick Descriptive and Homogeneity of variances test. Press CONTINUE. Press OK and run the test. R Instructions: In R Commander, make sure EllisYuan is the active dataset. Choose STATISTICS > MEANS > ONE-WAY ANOVA. Choose group (=independent variable) and errorfreeclauses (=dependent variable). Tick the box for pairwise comparison of means. Change the name of the regression model if you like. You can press the Apply button instead of OK in the main dialogue box,

5 which will let you then choose the other two variables (msttr, spm) as the response variables in turn. I have suggested the following R code to obtain mostly the same results as R Commander, but sometimes with an improvement: AnovaModel.1= aov(errorfreeclauses ~ group, data=ellisyuan) #create model Anova(AnovaModel.1) #omnibus test levenetest(ellisyuan$errorfreeclauses, EllisYuan$group) #test homogeneity of variances numsummary(ellisyuan [,"errorfreeclauses"], groups = EllisYuan$group) #descriptives library(multcomp)# If necessary.pairs=glht(anovamodel.1,linfct=mcp(group="tukey")) #model for multiple comparisons confint(.pairs) #returns confidence intervals for comparisons summary(.pairs, adjusted(type=c(p.adjust.methods=c("none")))) library(hh)# If necessary ellisyuan.mmc=mmc(anovamodel.1,linfct=mcp(group="tukey")) #model for MMC plot(ellisyuan.mmc) #MMC plot

6 Results: Result are taken from SPSS (confidence intervals differ slightly from R) Omnibus F test SPM F 2,39 =11.19, p=.000 MSTTR F 2,39 =.18, (lexical p=.84 variety) Error-free clauses F 2,39 =3.04, p=.06 Mean difference for NP-PTP + CI of mean difference (using LSD post-hoc, in other words, no adjustment) (-5.83, -1.70) Mean difference for NP-OLP + CI of mean difference (LSD) Mean difference for OLP-PTP + CI of mean difference (LSD).73 (-1.34, 2.79) (-6.56, -2.43) (-.02,.02).005 (-.02,.03) (-.03,.02) -.03 (-.11,.04).09 (-.17, -.2).06 (-.02,.13) Effect sizes are calculated using means and pooled standard deviations for each group with an online calculator. Mean NP N=14 Mean PTP N=14 Mean OLP N=14 ES for NP- PTP ES for NP-OLP ES for PTP- OLP ES for NP-PTP SPM 12.5 (2.0) 16.3 (3.3) 11.8 (2.7) MSTTR.88 (.03).88 (.02).87 (.03) Error-free clauses.77 (.10).81 (.12).86 (.07) Remarks: It s interesting that planning time had no effect on lexical variety (MSTTR). It s interesting to look at effect sizes here and see what effect sizes are found even when differences between groups are very small. For speed, the difference between the No Planning and Online Planning group has a large effect size and is statistical (even though the omnibus is not quite under p=.05!). For error-free clauses, it seems that pre-task planning had a large effect on improving the number of these.

7 2 Pandey (2000) Use the Pandey2000.sav file. Import into R as Pandey. a. Check Assumptions We assume that the variables were independently gathered. Use histograms as well as numerical values for skewness and kurtosis to check on normality assumptions and homogeneity of variances. SPSS Instructions: For boxplots, histograms and numerical values from one command, choose ANALYZE > DESCRIPTIVE STATISTICS > EXPLORE. Put the variable of GAIN1in the Dependent List. Put Group in the Factor List. Open the Plots button and tick on Histogram and Normality plots with tests. Press Continue and then OK. The Explore command produces a lot of output, so the results won't list all of it, just my overall summary of results. R Instructions: To look at normality graphically and numerically, first make sure Pandey is the active dataset. First let s look at boxplots. In R Commander, choose GRAPHS > BOXPLOT, and then choose gain1. Open the Plot by groups button and group will already be chosen, so just press OK. Press OK again to run the command. The R code is: Boxplot(gain1~group, data=pandey, id.method="y")

8 Now look at histograms. In R Commander, choose GRAPHS > HISTOGRAM, and then choose gain1. Open the Plot by groups button and group will already be chosen, so just press OK. Press OK to run the command. The R code is: with(pandey, Hist(gain1, groups=group, scale="frequency", breaks="sturges", col="darkgray")) Last, to look at numbers like skewness and kurtosis the fbasics package provides a lot of numbers, but the basicstats( ) command does not provide a way to break up the data by group, so we will have to do that manually by examining what rows each group is found in. I opened up the button on the R Commander GUI View dataset and found that the Focus A group used rows 1 11, Focus B rows 12 23, and ControlA rows By the way, the R code to do this if you do not use R Commander is: showdata(pandey, placement=' ', font=getrcmdr('logfont'), maxwidth=80, maxheight=30) Now knowing these rows, I can implement the command for each of the three groups for each of the three variables I want to look at: library(fbasics)# open package attach(pandey)# lets me type the variable without specifying the dataset basicstats(gain1[1:11])

9 basicstats(gain1[12:23]) basicstats(gain1[24:45]) detach(pandey) Results: In the first gain score, Focus group B is a little bit skewed, and the control group has many outliers, judging by the boxplot. The histograms from Focus A and Focus B could be normal distributions, but the Control A histogram looks very non-normal. The skewness and kurtosis numbers for the groups are OK for Focus A and Focus B but large for Control A. From the boxplot we see that the control group s variance is certainly much different from either Focs A or Focus B, and the variance of Focus A looks different from Focus B. b. Run the One way ANOVA SPSS Instructions: Open ANALYZE > COMPARE MEANS > ONE-WAY ANOVA. Put GAIN1 in the Dependent List box and GROUP in the Factor box. For planned comparisons, open the CONTRASTS button. Our first contrast will look at just Group A compared to Group B, so the coefficients to enter will be: 1, -1, 0. After entering each number, press the Add button. After entering all three numbers, press the Next button. The next contrast will be Group A and Group B contrasted against Control, so enter: 1, 1, -2. Press CONTINUE. Open the Options button and tick Descriptive and Homogeneity of variances test. Press CONTINUE. Press OK to run the analysis. R Instructions: Refer to Section of the book. First, I want to set up the contrasts:

10 levels(pandey$group) [1] "Focus A" "Focus B" "Control A" contr=rbind("focus A-Focus B"=c(1,-1,0), "Focus A&B - ControlA"=c(1,1,-2)) In other words, the first contrast compares just Focus A against Focus B, while the second contrast puts A and B together against Control A. Now I ll set up the model: AnovaModel.1 <- aov(gain1 ~ group, data=pandey) library(multcomp) Pandey.Pairs=glht(AnovaModel.1,linfct=mcp(group=contr)) summary(pandey.pairs) confint(pandey.pairs) Results: Ignore the omnibus test and look at the Contrast Tests. Since the boxplots showed variances that did not seem equal, use the Does not assume equal variances line. The first contrast between Group A and Group B is statistical, mean difference = 29.5, 95% CI [21.09, 37.90]. The second contrast between Groups A & B against the control is also statistical, mean difference = 54.0, 95% CI [41.95, 65.98].

11 3 Thought Question The main problem with this design is that it is repeated measures, so data is not independent. If the researcher were to ask whether there were any difference between groups for only ONE of the three discourse completion tasks, assuming that each of the 3 classes received different treatments, this would be a valid one-way ANOVA. You might think this experiment looks a lot like the Pandey experiment in #2, and that is right, but in that case we looked at a gain score from Time 2 to Time 1. Doing a one-way ANOVA in this case on one gain score would be fine (say, the gain from DCT Time 2 minus DCT Time 1) but to try to put the scores of all three discourse completion tasks together and then perform a one-way ANOVA would compromise the fundamental assumption of independence of groups in the one-way ANOVA. 4 Inagaki and Long (1999) Use the InagakiLong.sav file. Import into R as InagakiLong. a. Check Assumptions We assume that the variables were independently gathered. Use histograms as well as numerical values for skewness and kurtosis to check on normality assumptions and homogeneity of variances. SPSS Instructions: For boxplots, histograms and numerical values from one command, choose ANALYZE > DESCRIPTIVE STATISTICS > EXPLORE. Put the two variables of GAINADJ and GAINLOC in the Dependent List. Put GROUP in the Factor List. Open the Plots button and tick on

12 Histogram and Normality plots with tests. Press Continue and then OK. The Explore command produces a lot of output, so the results won t list all of it, just my overall summary of results. R Instructions: To look at normality graphically and numerically, first make sure InagakiLong is the active dataset. First let s look at boxplots. In R Commander, choose GRAPHS > BOXPLOT, and then choose gainadj. Open the Plot by groups button and group will already be chosen, so just press OK. You can press the Apply button instead of OK in the main dialogue box, which will let you then choose the other variable (gainloc) in turn. Press OK when you are finished. The R code for the first variable is: Boxplot(gainadj~group, data=inagakilong, id.method="y") Now look at histograms. In R Commander, choose GRAPHS > HISTOGRAM, and then choose gainadj. Open the Plot by groups button and group will already be chosen, so just press OK. You can press the Apply button instead of OK in the main dialogue box, which will let you then choose the other variable (gainloc) in turn. Press OK when you are finished. The R code for the first variable is: with(inagakilong, Hist(gainadj, groups=group, scale="frequency", breaks="sturges", col="darkgray"))

13 Last, to look at numbers like skewness and kurtosis the fbasics package provides a lot of numbers, but the basicstats( ) command does not provide a way to break up the data by group, so we will have to do that manually by examining what rows each group is found in. I opened up the button on the R Commander GUI View dataset and found that the Model group used rows 1 8, Recast rows 9 16, and Control rows (the sample size is rather small). By the way, the R code to do this if you do not use R Commander is: showdata(inagakilong, placement=' ', font=getrcmdr('logfont'), maxwidth=80, maxheight=30) Now knowing these rows, I can implement the command for each of the three groups for each of the three variables I want to look at: library(fbasics)# open package attach(inagakilong)# lets me type the variable without specifying the dataset basicstats(gainadj[1:8]) basicstats(gainadj [9:16]) basicstats(gainadj [17:24]) Repeat with the other variable and at the end: detach(inagakilong)

14 Results: GainAdj: The boxplots and histograms show that all groups are skewed, and the control group did not make any gain, so they have no variance (there is just one outlier though). They should not be compared to the other two groups for adjectives. The skewness numbers are below 1 except for the Control group, which has a skewness of 1.9. GainLoc: The boxplots and histograms show that all groups are skewed. The variance of the Model group is much larger than that of the Recast or Control group. The skewness numbers are slightly above 1 for all the groups, so we can say these variables are not normally distributed. b. Run the One way ANOVA The previous look at the variables show that we cannot run a one-way ANOVA on the gainscore on adjectives since there is no variation in the Control group. We would be able to run a t-test on this data, comparing the Model and Recast groups; however, I will not do that here. SPSS Instructions: Open ANALYZE > COMPARE MEANS > ONE-WAY ANOVA. Put GAINLOC in the Dependent List box and GROUP in the Factor box. In the POST HOC area you can choose which post-hoc tests to use, if you want to use any at all (if using the old statistics approach, my recommendation is to open the POST HOC button and tick LSD, which does not adjust means, and also tick Games- Howell because variances are not equal. Press Continue). Open the Options button and tick Descriptive and Homogeneity of variances test. Press CONTINUE. We want robust tests, so open the BOOTSTRAP button and check the Perform bootstrapping box, change the number of samples to 10,000 (or possibly less if you don t have

15 much time to wait), and change the Confidence Intervals to BCa. Press CONTINUE then OK and run the test. To calculate effect sizes, use an online calculator with the mean scores and standard deviations. R Instructions: In R Commander, make sure InagakiLong is the active dataset. Choose STATISTICS > MEANS > ONE-WAY ANOVA. Choose group (=independent variable) and gainloc (=dependent variable). Tick the box for pairwise comparison of means. Change the name of the regression model if you like. Press OK. I have suggested the following R code to obtain mostly the same results as R Commander, but sometimes with an improvement: AnovaModel.1= aov(gainloc ~ group, data=inagakilong) #create model Anova(AnovaModel.1) #omnibus test levenetest(inagakilong$gainloc, InagakiLong$group) #test homogeneity of variances numsummary(inagakilong [,"gainloc"], groups = InagakiLong$group) #descriptives library(multcomp) #If necessary.pairs=glht(anovamodel.1,linfct=mcp(group="tukey")) #model for multiple comparisons confint(.pairs) #returns confidence intervals for comparisons summary(.pairs, adjusted(type=c(p.adjust.methods=c("none")))) library(hh) #if necessary

16 ellisyuan.mmc=mmc(anovamodel.1,linfct=mcp(group="tukey")) #model for MMC plot(ellisyuan.mmc) #MMC plot For a Bootstrapped One-way ANOVA Note that the only place where this syntax would change is in the underlined and red portions library(boot)# open the package MeanDifference <- function (data,i){ temp<-data[i,] aov.temp<-aov(gainloc ~ group, data=temp) Tuk <-TukeyHSD(aov.temp) return(tuk$group[,1]) } MeanDifferenceBoot <- boot(inagakilong, MeanDifference, 10000) #if takes too long you can change the number of #replicates to a smaller number (press the Escape key to stop the command) PAIRci.s <- NULL PAIRt0 <- as.data.frame(meandifferenceboot[1]) for(i in 1:length(MeanDifferenceBoot[[1]])){ CI <- boot.ci(meandifferenceboot, conf =.95, type = "bca", t0 = MeanDifferenceBoot$t0[i],

17 t = MeanDifferenceBoot$t[,i]) PAIRci.s[[paste("lwr",i,sep=".")]]<- CI$bca[c(4)] PAIRci.s[[paste("upr",i,sep=".")]]<- CI$bca[c(5)] } Pci <- matrix(pairci.s, ncol = 2, nrow = length(pairt0[[1]]), byrow=true) PAIRboot.ci <- data.frame(pairt0[1], lwr = Pci[, 1], upr = Pci[, 2]) print(pairboot.ci) To calculate Cohen s d effect size (and associated confidence intervals around this effect size), use the following code: bootes(inagakilong, R=2000, data.col="gainloc", group.col="group", contrast=c(recast=1, Model=-1),effect.type=c("cohens.d"), ci.type=c("bca")) Repeat with the other pairings of variables: contrast=c(recast=1, Control=-1) contrast=c(model=1, Control=-1)

18 Results: For subject placement after a locative phrase, the following results were found using R: Omnibus F test GainLoc F 2,21 =0.37, p=.695 Bootstrapped CIs Mean difference for Recast-Model + CI of mean difference (using LSD post-hoc, in other words, no adjustment) 0.38 [-1.64, 0.89] Mean difference for Control-Model + CI of mean difference (LSD) Mean difference for Control-Recast + CI of mean difference (LSD) [-1.64, 0.89] [-1.27, 1.27] [-1.52, 0.56] [-1.56, 0.50] [-0.70, 0.73] Mean Recast N=8 Mean Model N=8 Mean Control N=8 ES for Recast- Model (CI) GainLoc 0.38 (.74) 0.75 (1.39) 0.38 (.74).34 [-1.32, 0.76] ES for Control- Model (CI).34 [-0.94, 1.32] ES for Control- Recast (CI) 0.00 [-1.20, 0.84] The overall omnibus test is not statistical, F2,21 =.37, p =.69. In the old statistics this would be the end and we wouldn t look any further. However, we are interested in reporting confidence intervals, so we ll continue on to look at the pairings between groups. The confidence intervals all pass through zero and are pretty much centered around zero. It is true that sample sizes are small, but the confidence intervals show that even with repeated testing, no effect would be likely to be found for differences between the experimental groups. I note that one possible problem with this study was that only 3 items per structure were used. There might have been more differentiation with a larger number of items! Lastly, effect sizes for all pairings are small, with confidence intervals for the effect sizes going through zero and showing that the estimate is not very precise.

19 5 Dewaele and Pavlenko ( ) Use the BEQ.Context.sav file. Import into R as beq.context. a. Check Assumptions We assume that the variables were independently gathered. Use histograms as well as numerical values for skewness and kurtosis to check on normality assumptions and homogeneity of variances. See Exercise #1 and Exercise #4 for similar datasets. Since you have detailed instructions in those exercises, I will only give bare bones instructions for this dataset. SPSS Instructions: In the EXPLORE option, put the two variables of L2SPEAK and L2_READ in the Dependent List. Put L2CONTEXT in the Factor List. R Instructions: Boxplot(l2speak~l2context, data=beq.context, id.method="y") Boxplot(l2_read~l2context, data=beq.context, id.method="y") with(beq.context, Hist(l2speak, groups=l2context, scale="frequency", breaks="sturges", col="darkgray")) with(beq.context, Hist(l2_read, groups=l2context, scale="frequency", breaks="sturges", col="darkgray")) I need to break up the dataset into groups, but they are not all in 3 easy divisions, so I ll use code to do it: levels(beq.context$l2context)

20 [1] "Instructed" "Naturalistic" [3] "Both instructed and naturalistic" beqnatural <- subset(beq.context, subset=l2context=="naturalistic") beqinstr <- subset(beq.context, subset=l2context=="instructed") beqboth <- subset(beq.context, subset=l2context=="both instructed and naturalistic") Having these divisions I can implement the command for each of the three groups for each of the three variables I want to look at: library(fbasics) #open package basicstats(beqnatural$l2speak) basicstats(beqinstr$l2speak) basicstats(beqboth$l2speak) Repeat with the other variable (l2_read). Results: For speaking and reading, looking at boxplots and histograms, all data are skewed, with more results toward 5 than would be predicted by a normal distribution, and there are outliers. For reading, all groups have equal variances (looking at boxplots). For speaking, the instructed learners have a bigger variance than the other two groups (looking at boxplots). For speaking, every group except instructed learners have skewness numbers over 1. For reading, all groups have skewness numbers over 1. Variables are certainly not normally distributed.

21 b. Run the One way ANOVA SPSS Instructions: In the ONE-WAY ANOVA box, put L2SPEAK and L2_READ in the Dependent List box and L2CONTEXT in the Factor box. To calculate effect sizes, use an online calculator with the mean scores and standard deviations. R Instructions: AnovaModel.1= aov(l2speak ~ l2context, data=beq.context) Anova(AnovaModel.1) Repeat with other variable (l2_read). numsummary(beq.context [,"l2speak"], groups = beq.context$l2context) Repeat with other variable (l2_read)..pairs=glht(anovamodel.1,linfct=mcp(l2context="tukey")) confint(.pairs) Repeat with other variable (l2_read). For a Bootstrapped One-way ANOVA

22 MeanDifference <- function (data,i){ temp<-data[i,] aov.temp<-aov(l2speak ~ l2context, data=temp) Tuk <-TukeyHSD(aov.temp) return(tuk$l2context[,1]) } MeanDifferenceBoot <- boot(beq.context, MeanDifference, 2000) PAIRci.s <- NULL PAIRt0 <- as.data.frame(meandifferenceboot[1]) for(i in 1:length(MeanDifferenceBoot[[1]])){ CI <- boot.ci(meandifferenceboot, conf =.95, type = "bca", t0 = MeanDifferenceBoot$t0[i], t = MeanDifferenceBoot$t[,i]) PAIRci.s[[paste("lwr",i,sep=".")]]<- CI$bca[c(4)] PAIRci.s[[paste("upr",i,sep=".")]]<- CI$bca[c(5)] } Pci <- matrix(pairci.s, ncol = 2, nrow = length(pairt0[[1]]), byrow=true) PAIRboot.ci <- data.frame(pairt0[1], lwr = Pci[, 1], upr = Pci[, 2]) print(pairboot.ci)

23 To calculate Cohen s d effect size (and associated confidence intervals around this effect size), the following code should work: bootes(beq.context, R=500, data.col="l2speak", group.col="l2context", contrast=c("instructed"=1,"naturalistic"=-1),effect.type=c("cohens.d"), ci.type=c("bca")) However, I am getting a warning that w is infinite and I cannot get any results. I tried substituting in different types of CIs ("norm", "basic", "perc", "stud", "none") but still couldn t get any estimates. I will give up on the bootstrapped effect size CIs and just use an online calculated to calculate effect sizes from the means and pooled standard deviations of the groups. Results: Results from SPSS L2 Speaking L2 Reading Omnibus F test F 2,1010 =43.4, p<.0005 F 2,1010 =15.21, p<.0005 Mean difference for Instr-Natl + CI (using Tukey numbers) Mean difference for Instr-Both + CI Mean difference for Natl-Both -.54 (-.73, -.36) -.62 (-.75, -.48) -.07 (-.26,.11) -.16 (-.33, 01) -.38 (-.50, -.25) -.22 (-.38, -.05) CI (Bootstrapped) Mean difference for Instr- Natl CI (Bootstrapped) Mean difference for Instr- Both CI (Bootstrapped) Mean difference for Natl- Both L2 Speaking L2 Reading -.73, , , , , , -.04

24 L2 Speaking L2 Reading Mean Instr N=399 Mean Natl N=158 Mean Both N=456 ES for Natl- Instruct (Cohen s d) ES for Instr- Both 3.77 (1.18) 4.32 (.97) 4.39 (.85) (.98) 4.37 (1.10) 4.59 (.76) ES for Natl- Both Looking at the descriptive stats, we see that the highest scores for both speaking and reading were by those who reported learning the L2 both naturalistically and in an instructed way (the Both choice). The Levene s test has a p-value below p =.05, which means that we should reject the hypothesis that variances are equal, but we will ignore that because a look at the standard deviations does not reveal an extremely large difference in the standard deviations (the test is too sensitive because of the large sample size). Owing to the fact that we have extremely large sample sizes here the omnibus test for both ANOVAs is statistical. What we should be more concerned about are the comparison CIs and their associated effect sizes. For L2 speaking, differences are statistical for the comparison between Instruction and the other two conditions (but not between Natural and Both). There is some modest effect for the comparisons between those who learned Naturally and those who were Instructed, and also for those who were Instructed versus those who learned Both ways, but the effect size for those who learned Naturally versus Both is basically zero. There are basically no differences between the parametric and bootstrapped CIs. For L2 reading, differences are statistical for all three conditions, with mean scores showing that those who learned Both ways scored the highest. Effect sizes are quite small, however, for this

25 area and there aren t very big differences between groups. There are also basically no differences between the parametric and bootstrapped CIs. 6 Jensen & Vinther (2003) Use the Jensen&Vinther2003.sav file. Import into R as jensen. a. Check Assumptions We assume that the variables were independently gathered. Use histograms as well as numerical values for skewness and kurtosis to check on normality assumptions and homogeneity of variances. See Exercise #1 and Exercise #4 for similar datasets. Since you have detailed instructions in those exercises, I will only give bare bones instructions for this dataset. SPSS Instructions: In the EXPLORE option, put the variable of GAINSCOR in the Dependent List. Put GROUP in the Factor List. R Instructions: Boxplot(gainscore~group, data=jensen, id.method="y") with(jensen, Hist(gainscore, groups=group, scale="frequency", breaks="sturges", col="darkgray")) Control group is in rows 1:20, FSF in rows 21:41, and FSS in rows 42:62. basicstats(jensen$gainscore[1:20])

26 basicstats(jensen$gainscore[21:41]) basicstats(jensen$gainscore[42:62]) Results: Boxplots and histograms show that the three groups are not exactly symmetrically distributed but do not have a lot of non-normalities (no outliers, no extreme skewing). Skewness numbers are all well under 1. Variances are all about the same as well. As far as variables go, these look pretty good they could be considered normally distributed and as having equal variances! b. Run the One way ANOVA SPSS Instructions: In the ONE-WAY ANOVA box, put GAINSCORE in the Dependent List box and GROUP in the Factor box. To calculate effect sizes, use an online calculator with the mean scores and standard deviations. R Instructions: AnovaModel.1= aov(gainscore ~ group, data=jensen) Anova(AnovaModel.1) numsummary(jensen [,"gainscore"], groups = jensen$group).pairs=glht(anovamodel.1,linfct=mcp(group="tukey"))

27 confint(.pairs) For a Bootstrapped One-way ANOVA MeanDifference <- function (data,i){ temp<-data[i,] aov.temp<-aov(gainscore ~ group, data=temp) Tuk <-TukeyHSD(aov.temp) return(tuk$group[,1]) } MeanDifferenceBoot <- boot(jensen, MeanDifference, 2000) PAIRci.s <- NULL PAIRt0 <- as.data.frame(meandifferenceboot[1]) for(i in 1:length(MeanDifferenceBoot[[1]])){ CI <- boot.ci(meandifferenceboot, conf =.95, type = "bca", t0 = MeanDifferenceBoot$t0[i], t = MeanDifferenceBoot$t[,i]) PAIRci.s[[paste("lwr",i,sep=".")]]<- CI$bca[c(4)] PAIRci.s[[paste("upr",i,sep=".")]]<- CI$bca[c(5)] } Pci <- matrix(pairci.s, ncol = 2, nrow = length(pairt0[[1]]), byrow=true)

28 PAIRboot.ci <- data.frame(pairt0[1], lwr = Pci[, 1], upr = Pci[, 2]) print(pairboot.ci) To calculate Cohen s d effect size (and associated confidence intervals around this effect size), the following code should work: bootes(jensen, R=500, data.col="gainscore", group.col="group", contrast=c(fsf=1,control =-1),effect.type=c("cohens.d"), ci.type=c("bca")) bootes(jensen, R=500, data.col="gainscore", group.col="group", contrast=c(fss=1,control =-1),effect.type=c("cohens.d"), ci.type=c("bca")) bootes(jensen, R=500, data.col="gainscore", group.col="group", contrast=c(fss=1,fsf =-1),effect.type=c("cohens.d"), ci.type=c("bca")) Results: Results from R Omnibus F test Gainscore F 2,59 =5.20, p=.008 Bootstrapped CIs CI for mean difference FSF- Control CI for mean difference FSS-Control CI for mean difference FSS-FSF -.65, , , , , , 31.11

29 Mean Control N=20 Mean FSF N=21 Gainscore 11.6 (38.75) (37.37) Mean FSS N= (40.36) ES for FSF- Control (Cohen s d) 0.78 [.07, 1.38] ES for FSS- Control 0.94 [.37, 1.60] ES for FSS-FSF.19 [-.40,.83] The mean scores show clearly that the control group fared far worse than either experimental group. The parametric confidence interval for the difference between the FSF (fast-slow-fast) group and the control, however, went through zero, although the lower level was quite close to zero and with more precision would probably be farther away from zero. In fact, the bootstrapped confidence interval did not go through zero and showed a minimum of about 7 points of difference in the CI, with up to 52 points of difference between the groups. So here is one case where bootstrapping made quite a difference. The effect size estimate for this comparison was fairly large, although the CI of the Cohen s d effect size was also very wide and shows that the estimate is not very precise. So we may hope and suspect there is a difference between the control group and the FSF group, but there is some evidence for caution in making that assumption! There is clearly no difference between the FSS and FSF groups, with both the parametric and bootstrapped CIs going through zero and the effect size being quite small, with a CI very close to zero. On the other hand, there is a clear difference between the Control group and the FSS group, although the CIs and confidence interval of the effect size shows it could be a fairly modest effect.

Answers to Application Activities in Chapter 6

Answers to Application Activities in Chapter 6 Answers to Application Activities in Chapter 6 6.3.3 Application Activities with Scatterplots (Answers for both SPSS and R given for each item) 1 DeKeyser (2000) Use the data file Dekeyser2000.sav. SPSS

More information

SPSS t tests (and NP Equivalent)

SPSS t tests (and NP Equivalent) SPSS t tests (and NP Equivalent) Descriptive Statistics To get all the descriptive statistics you need: Analyze > Descriptive Statistics>Explore. Enter the IV into the Factor list and the DV into the Dependent

More information

starting on 5/1/1953 up until 2/1/2017.

starting on 5/1/1953 up until 2/1/2017. An Actuary s Guide to Financial Applications: Examples with EViews By William Bourgeois An actuary is a business professional who uses statistics to determine and analyze risks for companies. In this guide,

More information

Chapter 11: Inference for Distributions Inference for Means of a Population 11.2 Comparing Two Means

Chapter 11: Inference for Distributions Inference for Means of a Population 11.2 Comparing Two Means Chapter 11: Inference for Distributions 11.1 Inference for Means of a Population 11.2 Comparing Two Means 1 Population Standard Deviation In the previous chapter, we computed confidence intervals and performed

More information

Table of Contents. New to the Second Edition... Chapter 1: Introduction : Social Research...

Table of Contents. New to the Second Edition... Chapter 1: Introduction : Social Research... iii Table of Contents Preface... xiii Purpose... xiii Outline of Chapters... xiv New to the Second Edition... xvii Acknowledgements... xviii Chapter 1: Introduction... 1 1.1: Social Research... 1 Introduction...

More information

We will use an example which will result in a paired t test regarding the labor force participation rate for women in the 60 s and 70 s.

We will use an example which will result in a paired t test regarding the labor force participation rate for women in the 60 s and 70 s. Now let s review methods for one quantitative variable. We will use an example which will result in a paired t test regarding the labor force participation rate for women in the 60 s and 70 s. 17 The labor

More information

Data screening, transformations: MRC05

Data screening, transformations: MRC05 Dale Berger Data screening, transformations: MRC05 This is a demonstration of data screening and transformations for a regression analysis. Our interest is in predicting current salary from education level

More information

Two-Sample T-Test for Superiority by a Margin

Two-Sample T-Test for Superiority by a Margin Chapter 219 Two-Sample T-Test for Superiority by a Margin Introduction This procedure provides reports for making inference about the superiority of a treatment mean compared to a control mean from data

More information

Two-Sample T-Test for Non-Inferiority

Two-Sample T-Test for Non-Inferiority Chapter 198 Two-Sample T-Test for Non-Inferiority Introduction This procedure provides reports for making inference about the non-inferiority of a treatment mean compared to a control mean from data taken

More information

We will also use this topic to help you see how the standard deviation might be useful for distributions which are normally distributed.

We will also use this topic to help you see how the standard deviation might be useful for distributions which are normally distributed. We will discuss the normal distribution in greater detail in our unit on probability. However, as it is often of use to use exploratory data analysis to determine if the sample seems reasonably normally

More information

The Assumption(s) of Normality

The Assumption(s) of Normality The Assumption(s) of Normality Copyright 2000, 2011, 2016, J. Toby Mordkoff This is very complicated, so I ll provide two versions. At a minimum, you should know the short one. It would be great if you

More information

appstats5.notebook September 07, 2016 Chapter 5

appstats5.notebook September 07, 2016 Chapter 5 Chapter 5 Describing Distributions Numerically Chapter 5 Objective: Students will be able to use statistics appropriate to the shape of the data distribution to compare of two or more different data sets.

More information

Chapter 18: The Correlational Procedures

Chapter 18: The Correlational Procedures Introduction: In this chapter we are going to tackle about two kinds of relationship, positive relationship and negative relationship. Positive Relationship Let's say we have two values, votes and campaign

More information

Stat 139 Homework 2 Solutions, Fall 2016

Stat 139 Homework 2 Solutions, Fall 2016 Stat 139 Homework 2 Solutions, Fall 2016 Problem 1. The sum of squares of a sample of data is minimized when the sample mean, X = Xi /n, is used as the basis of the calculation. Define g(c) as a function

More information

Wk 2 Hrs 1 (Tue, Jan 10) Wk 2 - Hr 2 and 3 (Thur, Jan 12)

Wk 2 Hrs 1 (Tue, Jan 10) Wk 2 - Hr 2 and 3 (Thur, Jan 12) Wk 2 Hrs 1 (Tue, Jan 10) Wk 2 - Hr 2 and 3 (Thur, Jan 12) Descriptive statistics: - Measures of centrality (Mean, median, mode, trimmed mean) - Measures of spread (MAD, Standard deviation, variance) -

More information

Descriptive Statistics

Descriptive Statistics Chapter 3 Descriptive Statistics Chapter 2 presented graphical techniques for organizing and displaying data. Even though such graphical techniques allow the researcher to make some general observations

More information

2018 AAPM: Normal and non normal distributions: Why understanding distributions are important when designing experiments and analyzing data

2018 AAPM: Normal and non normal distributions: Why understanding distributions are important when designing experiments and analyzing data Statistical Failings that Keep Us All in the Dark Normal and non normal distributions: Why understanding distributions are important when designing experiments and Conflict of Interest Disclosure I have

More information

Chapter 6 Part 3 October 21, Bootstrapping

Chapter 6 Part 3 October 21, Bootstrapping Chapter 6 Part 3 October 21, 2008 Bootstrapping From the internet: The bootstrap involves repeated re-estimation of a parameter using random samples with replacement from the original data. Because the

More information

NCSS Statistical Software. Reference Intervals

NCSS Statistical Software. Reference Intervals Chapter 586 Introduction A reference interval contains the middle 95% of measurements of a substance from a healthy population. It is a type of prediction interval. This procedure calculates one-, and

More information

The FREQ Procedure. Table of Sex by Gym Sex(Sex) Gym(Gym) No Yes Total Male Female Total

The FREQ Procedure. Table of Sex by Gym Sex(Sex) Gym(Gym) No Yes Total Male Female Total Jenn Selensky gathered data from students in an introduction to psychology course. The data are weights, sex/gender, and whether or not the student worked-out in the gym. Here is the output from a 2 x

More information

Summary of Statistical Analysis Tools EDAD 5630

Summary of Statistical Analysis Tools EDAD 5630 Summary of Statistical Analysis Tools EDAD 5630 Test Name Program Used Purpose Steps Main Uses/Applications in Schools Principal Component Analysis SPSS Measure Underlying Constructs Reliability SPSS Measure

More information

8.2 The Standard Deviation as a Ruler Chapter 8 The Normal and Other Continuous Distributions 8-1

8.2 The Standard Deviation as a Ruler Chapter 8 The Normal and Other Continuous Distributions 8-1 8.2 The Standard Deviation as a Ruler Chapter 8 The Normal and Other Continuous Distributions For Example: On August 8, 2011, the Dow dropped 634.8 points, sending shock waves through the financial community.

More information

Establishing a framework for statistical analysis via the Generalized Linear Model

Establishing a framework for statistical analysis via the Generalized Linear Model PSY349: Lecture 1: INTRO & CORRELATION Establishing a framework for statistical analysis via the Generalized Linear Model GLM provides a unified framework that incorporates a number of statistical methods

More information

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example... Chapter 4 Point estimation Contents 4.1 Introduction................................... 2 4.2 Estimating a population mean......................... 2 4.2.1 The problem with estimating a population mean

More information

Descriptive Statistics in Analysis of Survey Data

Descriptive Statistics in Analysis of Survey Data Descriptive Statistics in Analysis of Survey Data March 2013 Kenneth M Coleman Mohammad Nizamuddiin Khan Survey: Definition A survey is a systematic method for gathering information from (a sample of)

More information

Data Distributions and Normality

Data Distributions and Normality Data Distributions and Normality Definition (Non)Parametric Parametric statistics assume that data come from a normal distribution, and make inferences about parameters of that distribution. These statistical

More information

An approximate sampling distribution for the t-ratio. Caution: comparing population means when σ 1 σ 2.

An approximate sampling distribution for the t-ratio. Caution: comparing population means when σ 1 σ 2. Stat 529 (Winter 2011) Non-pooled t procedures (The Welch test) Reading: Section 4.3.2 The sampling distribution of Y 1 Y 2. An approximate sampling distribution for the t-ratio. The Sri Lankan analysis.

More information

Stat 5303 (Oehlert): Power and Sample Size 1

Stat 5303 (Oehlert): Power and Sample Size 1 Stat 5303 (Oehlert): Power and Sample Size 1 Cmd> # The Stat5303 package includes two functions of use here: power.anova.test() and sample.size.anova(). You won t be surprised to learn that the first computes

More information

Washington University Fall Economics 487

Washington University Fall Economics 487 Washington University Fall 2009 Department of Economics James Morley Economics 487 Project Proposal due Tuesday 11/10 Final Project due Wednesday 12/9 (by 5:00pm) (20% penalty per day if the project is

More information

The Two-Sample Independent Sample t Test

The Two-Sample Independent Sample t Test Department of Psychology and Human Development Vanderbilt University 1 Introduction 2 3 The General Formula The Equal-n Formula 4 5 6 Independence Normality Homogeneity of Variances 7 Non-Normality Unequal

More information

Copyright 2011 Pearson Education, Inc. Publishing as Addison-Wesley.

Copyright 2011 Pearson Education, Inc. Publishing as Addison-Wesley. Appendix: Statistics in Action Part I Financial Time Series 1. These data show the effects of stock splits. If you investigate further, you ll find that most of these splits (such as in May 1970) are 3-for-1

More information

Chapter 6. y y. Standardizing with z-scores. Standardizing with z-scores (cont.)

Chapter 6. y y. Standardizing with z-scores. Standardizing with z-scores (cont.) Starter Ch. 6: A z-score Analysis Starter Ch. 6 Your Statistics teacher has announced that the lower of your two tests will be dropped. You got a 90 on test 1 and an 85 on test 2. You re all set to drop

More information

Analysis of 2x2 Cross-Over Designs using T-Tests for Non-Inferiority

Analysis of 2x2 Cross-Over Designs using T-Tests for Non-Inferiority Chapter 235 Analysis of 2x2 Cross-Over Designs using -ests for Non-Inferiority Introduction his procedure analyzes data from a two-treatment, two-period (2x2) cross-over design where the goal is to demonstrate

More information

Point-Biserial and Biserial Correlations

Point-Biserial and Biserial Correlations Chapter 302 Point-Biserial and Biserial Correlations Introduction This procedure calculates estimates, confidence intervals, and hypothesis tests for both the point-biserial and the biserial correlations.

More information

The Range, the Inter Quartile Range (or IQR), and the Standard Deviation (which we usually denote by a lower case s).

The Range, the Inter Quartile Range (or IQR), and the Standard Deviation (which we usually denote by a lower case s). We will look the three common and useful measures of spread. The Range, the Inter Quartile Range (or IQR), and the Standard Deviation (which we usually denote by a lower case s). 1 Ameasure of the center

More information

CS227-Scientific Computing. Lecture 6: Nonlinear Equations

CS227-Scientific Computing. Lecture 6: Nonlinear Equations CS227-Scientific Computing Lecture 6: Nonlinear Equations A Financial Problem You invest $100 a month in an interest-bearing account. You make 60 deposits, and one month after the last deposit (5 years

More information

Learning Objectives for Ch. 7

Learning Objectives for Ch. 7 Chapter 7: Point and Interval Estimation Hildebrand, Ott and Gray Basic Statistical Ideas for Managers Second Edition 1 Learning Objectives for Ch. 7 Obtaining a point estimate of a population parameter

More information

DATA ANALYSIS. ratio as a measurement of bank s growth. (further details can bee seen in appendix A) 1. Permata Bank (BNLI) Central Asia Bank (BCA)

DATA ANALYSIS. ratio as a measurement of bank s growth. (further details can bee seen in appendix A) 1. Permata Bank (BNLI) Central Asia Bank (BCA) Chapter 4 DATA ANALYSIS This chapter discusses the capital structure in each groups and the effect of that differences which reflect on their debt, equity, debt to equity ratio and capital adequacy ratio

More information

Loan and Bond Amortization

Loan and Bond Amortization Loan and Bond Amortization 5 chapter In this chapter you will learn: How to use the payment function to calculate payments to retire a loan How to create a loan amortization schedule How to use a what-if

More information

Binomial Random Variable - The count X of successes in a binomial setting

Binomial Random Variable - The count X of successes in a binomial setting 6.3.1 Binomial Settings and Binomial Random Variables What do the following scenarios have in common? Toss a coin 5 times. Count the number of heads. Spin a roulette wheel 8 times. Record how many times

More information

Chapter 3 Describing Data Numerically and Graphically

Chapter 3 Describing Data Numerically and Graphically 42 Chapter 3 Describing Data Numerically and Graphically 3.1 Obtaining Numerical Summaries Follow along with me in this section by importing the SPSS file LarsonHall.Forgotten.sav into R. Make sure that

More information

Putting Things Together Part 1

Putting Things Together Part 1 Putting Things Together Part 1 These exercise blend ideas from various graphs (histograms and boxplots), differing shapes of distributions, and values summarizing the data. Data for 1, 5, and 6 are in

More information

Lecture Week 4 Inspecting Data: Distributions

Lecture Week 4 Inspecting Data: Distributions Lecture Week 4 Inspecting Data: Distributions Introduction to Research Methods & Statistics 2013 2014 Hemmo Smit So next week No lecture & workgroups But Practice Test on-line (BB) Enter data for your

More information

STATISTICAL DISTRIBUTIONS AND THE CALCULATOR

STATISTICAL DISTRIBUTIONS AND THE CALCULATOR STATISTICAL DISTRIBUTIONS AND THE CALCULATOR 1. Basic data sets a. Measures of Center - Mean ( ): average of all values. Characteristic: non-resistant is affected by skew and outliers. - Median: Either

More information

Statistics and Probability

Statistics and Probability Statistics and Probability Continuous RVs (Normal); Confidence Intervals Outline Continuous random variables Normal distribution CLT Point estimation Confidence intervals http://www.isrec.isb-sib.ch/~darlene/geneve/

More information

Chapter 7. Inferences about Population Variances

Chapter 7. Inferences about Population Variances Chapter 7. Inferences about Population Variances Introduction () The variability of a population s values is as important as the population mean. Hypothetical distribution of E. coli concentrations from

More information

Key Objectives. Module 2: The Logic of Statistical Inference. Z-scores. SGSB Workshop: Using Statistical Data to Make Decisions

Key Objectives. Module 2: The Logic of Statistical Inference. Z-scores. SGSB Workshop: Using Statistical Data to Make Decisions SGSB Workshop: Using Statistical Data to Make Decisions Module 2: The Logic of Statistical Inference Dr. Tom Ilvento January 2006 Dr. Mugdim Pašić Key Objectives Understand the logic of statistical inference

More information

μ: ESTIMATES, CONFIDENCE INTERVALS, AND TESTS Business Statistics

μ: ESTIMATES, CONFIDENCE INTERVALS, AND TESTS Business Statistics μ: ESTIMATES, CONFIDENCE INTERVALS, AND TESTS Business Statistics CONTENTS Estimating parameters The sampling distribution Confidence intervals for μ Hypothesis tests for μ The t-distribution Comparison

More information

3: Balance Equations

3: Balance Equations 3.1 Balance Equations Accounts with Constant Interest Rates 15 3: Balance Equations Investments typically consist of giving up something today in the hope of greater benefits in the future, resulting in

More information

MA 1125 Lecture 12 - Mean and Standard Deviation for the Binomial Distribution. Objectives: Mean and standard deviation for the binomial distribution.

MA 1125 Lecture 12 - Mean and Standard Deviation for the Binomial Distribution. Objectives: Mean and standard deviation for the binomial distribution. MA 5 Lecture - Mean and Standard Deviation for the Binomial Distribution Friday, September 9, 07 Objectives: Mean and standard deviation for the binomial distribution.. Mean and Standard Deviation of the

More information

Conover Test of Variances (Simulation)

Conover Test of Variances (Simulation) Chapter 561 Conover Test of Variances (Simulation) Introduction This procedure analyzes the power and significance level of the Conover homogeneity test. This test is used to test whether two or more population

More information

Dr. Allen Back. Oct. 28, 2016

Dr. Allen Back. Oct. 28, 2016 Dr. Allen Back Oct. 28, 2016 A coffee vending machine dispenses coffee into a paper cup. You re supposed to get 10 ounces of coffee., but the amount varies slightly from cup to cup. The amounts measured

More information

XLSTAT TIP SHEET FOR BUSINESS STATISTICS CENGAGE LEARNING

XLSTAT TIP SHEET FOR BUSINESS STATISTICS CENGAGE LEARNING XLSTAT TIP SHEET FOR BUSINESS STATISTICS CENGAGE LEARNING INTRODUCTION XLSTAT makes accessible to anyone a powerful, complete and user-friendly data analysis and statistical solution. Accessibility to

More information

STAB22 section 1.3 and Chapter 1 exercises

STAB22 section 1.3 and Chapter 1 exercises STAB22 section 1.3 and Chapter 1 exercises 1.101 Go up and down two times the standard deviation from the mean. So 95% of scores will be between 572 (2)(51) = 470 and 572 + (2)(51) = 674. 1.102 Same idea

More information

Monte Carlo Simulation (General Simulation Models)

Monte Carlo Simulation (General Simulation Models) Monte Carlo Simulation (General Simulation Models) Revised: 10/11/2017 Summary... 1 Example #1... 1 Example #2... 10 Summary Monte Carlo simulation is used to estimate the distribution of variables when

More information

Describing Data: One Quantitative Variable

Describing Data: One Quantitative Variable STAT 250 Dr. Kari Lock Morgan The Big Picture Describing Data: One Quantitative Variable Population Sampling SECTIONS 2.2, 2.3 One quantitative variable (2.2, 2.3) Statistical Inference Sample Descriptive

More information

WEB APPENDIX 8A 7.1 ( 8.9)

WEB APPENDIX 8A 7.1 ( 8.9) WEB APPENDIX 8A CALCULATING BETA COEFFICIENTS The CAPM is an ex ante model, which means that all of the variables represent before-the-fact expected values. In particular, the beta coefficient used in

More information

The instructions on this page also work for the TI-83 Plus and the TI-83 Plus Silver Edition.

The instructions on this page also work for the TI-83 Plus and the TI-83 Plus Silver Edition. The instructions on this page also work for the TI-83 Plus and the TI-83 Plus Silver Edition. The position of the graphically represented keys can be found by moving your mouse on top of the graphic. Turn

More information

Sampling Distributions Homework Answers

Sampling Distributions Homework Answers Sampling Distributions Homework Answers John Fox Soc. 6Z03, Fall 2016 11.2 41% of registered voters are Democrats: parameter 36% of registered voters are Republicans: parameter 34% of voters contacted

More information

Valid Missing Total. N Percent N Percent N Percent , ,0% 0,0% 2 100,0% 1, ,0% 0,0% 2 100,0% 2, ,0% 0,0% 5 100,0%

Valid Missing Total. N Percent N Percent N Percent , ,0% 0,0% 2 100,0% 1, ,0% 0,0% 2 100,0% 2, ,0% 0,0% 5 100,0% dimension1 GET FILE= validacaonestscoremédico.sav' (só com os 59 doentes) /COMPRESSED. SORT CASES BY UMcpEVA (D). EXAMINE VARIABLES=UMcpEVA BY NoRespostasSignif /PLOT BOXPLOT HISTOGRAM NPPLOT /COMPARE

More information

PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS

PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS Melfi Alrasheedi School of Business, King Faisal University, Saudi

More information

STA2601. Tutorial letter 105/2/2018. Applied Statistics II. Semester 2. Department of Statistics STA2601/105/2/2018 TRIAL EXAMINATION PAPER

STA2601. Tutorial letter 105/2/2018. Applied Statistics II. Semester 2. Department of Statistics STA2601/105/2/2018 TRIAL EXAMINATION PAPER STA2601/105/2/2018 Tutorial letter 105/2/2018 Applied Statistics II STA2601 Semester 2 Department of Statistics TRIAL EXAMINATION PAPER Define tomorrow. university of south africa Dear Student Congratulations

More information

BINARY OPTIONS: A SMARTER WAY TO TRADE THE WORLD'S MARKETS NADEX.COM

BINARY OPTIONS: A SMARTER WAY TO TRADE THE WORLD'S MARKETS NADEX.COM BINARY OPTIONS: A SMARTER WAY TO TRADE THE WORLD'S MARKETS NADEX.COM CONTENTS To Be or Not To Be? That s a Binary Question Who Sets a Binary Option's Price? And How? Price Reflects Probability Actually,

More information

Solutions for practice questions: Chapter 15, Probability Distributions If you find any errors, please let me know at

Solutions for practice questions: Chapter 15, Probability Distributions If you find any errors, please let me know at Solutions for practice questions: Chapter 15, Probability Distributions If you find any errors, please let me know at mailto:msfrisbie@pfrisbie.com. 1. Let X represent the savings of a resident; X ~ N(3000,

More information

You can find out what your

You can find out what your One of the more involved transactions you ll have to do occasionally as a club treasurer is to process a member full or partial withdrawal. Withdrawals Withdrawals Not just a cash payout A distribution

More information

PAIRS TRADING (just an introduction)

PAIRS TRADING (just an introduction) PAIRS TRADING (just an introduction) By Rob Booker Trading involves substantial risk of loss. Past performance is not necessarily indicative of future results. You can share this ebook with anyone you

More information

R & R Study. Chapter 254. Introduction. Data Structure

R & R Study. Chapter 254. Introduction. Data Structure Chapter 54 Introduction A repeatability and reproducibility (R & R) study (sometimes called a gauge study) is conducted to determine if a particular measurement procedure is adequate. If the measurement

More information

Hedge Fund Returns: You Can Make Them Yourself!

Hedge Fund Returns: You Can Make Them Yourself! ALTERNATIVE INVESTMENT RESEARCH CENTRE WORKING PAPER SERIES Working Paper # 0023 Hedge Fund Returns: You Can Make Them Yourself! Harry M. Kat Professor of Risk Management, Cass Business School Helder P.

More information

The Binomial Distribution

The Binomial Distribution The Binomial Distribution January 31, 2018 Contents The Binomial Distribution The Normal Approximation to the Binomial The Binomial Hypothesis Test Computing Binomial Probabilities in R 30 Problems The

More information

Expectation Exercises.

Expectation Exercises. Expectation Exercises. Pages Problems 0 2,4,5,7 (you don t need to use trees, if you don t want to but they might help!), 9,-5 373 5 (you ll need to head to this page: http://phet.colorado.edu/sims/plinkoprobability/plinko-probability_en.html)

More information

7. For the table that follows, answer the following questions: x y 1-1/4 2-1/2 3-3/4 4

7. For the table that follows, answer the following questions: x y 1-1/4 2-1/2 3-3/4 4 7. For the table that follows, answer the following questions: x y 1-1/4 2-1/2 3-3/4 4 - Would the correlation between x and y in the table above be positive or negative? The correlation is negative. -

More information

Putting Things Together Part 2

Putting Things Together Part 2 Frequency Putting Things Together Part These exercise blend ideas from various graphs (histograms and boxplots), differing shapes of distributions, and values summarizing the data. Data for, and are in

More information

Problem Set 7 Part I Short answer questions on readings. Note, if I don t provide it, state which table, figure, or exhibit backs up your point

Problem Set 7 Part I Short answer questions on readings. Note, if I don t provide it, state which table, figure, or exhibit backs up your point Business 35150 John H. Cochrane Problem Set 7 Part I Short answer questions on readings. Note, if I don t provide it, state which table, figure, or exhibit backs up your point 1. Mitchell and Pulvino (a)

More information

Non-Inferiority Tests for Two Means in a 2x2 Cross-Over Design using Differences

Non-Inferiority Tests for Two Means in a 2x2 Cross-Over Design using Differences Chapter 510 Non-Inferiority Tests for Two Means in a 2x2 Cross-Over Design using Differences Introduction This procedure computes power and sample size for non-inferiority tests in 2x2 cross-over designs

More information

The Binomial Distribution

The Binomial Distribution The Binomial Distribution January 31, 2019 Contents The Binomial Distribution The Normal Approximation to the Binomial The Binomial Hypothesis Test Computing Binomial Probabilities in R 30 Problems The

More information

Math 120 Introduction to Statistics Mr. Toner s Lecture Notes. Standardizing normal distributions The Standard Normal Curve

Math 120 Introduction to Statistics Mr. Toner s Lecture Notes. Standardizing normal distributions The Standard Normal Curve 6.1 6.2 The Standard Normal Curve Standardizing normal distributions The "bell-shaped" curve, or normal curve, is a probability distribution that describes many reallife situations. Basic Properties 1.

More information

Chapter 6: Random Variables. Ch. 6-3: Binomial and Geometric Random Variables

Chapter 6: Random Variables. Ch. 6-3: Binomial and Geometric Random Variables Chapter : Random Variables Ch. -3: Binomial and Geometric Random Variables X 0 2 3 4 5 7 8 9 0 0 P(X) 3???????? 4 4 When the same chance process is repeated several times, we are often interested in whether

More information

MM and ML for a sample of n = 30 from Gamma(3,2) ===============================================

MM and ML for a sample of n = 30 from Gamma(3,2) =============================================== and for a sample of n = 30 from Gamma(3,2) =============================================== Generate the sample with shape parameter α = 3 and scale parameter λ = 2 > x=rgamma(30,3,2) > x [1] 0.7390502

More information

Robust Critical Values for the Jarque-bera Test for Normality

Robust Critical Values for the Jarque-bera Test for Normality Robust Critical Values for the Jarque-bera Test for Normality PANAGIOTIS MANTALOS Jönköping International Business School Jönköping University JIBS Working Papers No. 00-8 ROBUST CRITICAL VALUES FOR THE

More information

SFSU FIN822 Project 1

SFSU FIN822 Project 1 SFSU FIN822 Project 1 This project can be done in a team of up to 3 people. Your project report must be accompanied by printouts of programming outputs. You could use any software to solve the problems.

More information

Lecture 2 Describing Data

Lecture 2 Describing Data Lecture 2 Describing Data Thais Paiva STA 111 - Summer 2013 Term II July 2, 2013 Lecture Plan 1 Types of data 2 Describing the data with plots 3 Summary statistics for central tendency and spread 4 Histograms

More information

Statistical Models of Stocks and Bonds. Zachary D Easterling: Department of Economics. The University of Akron

Statistical Models of Stocks and Bonds. Zachary D Easterling: Department of Economics. The University of Akron Statistical Models of Stocks and Bonds Zachary D Easterling: Department of Economics The University of Akron Abstract One of the key ideas in monetary economics is that the prices of investments tend to

More information

Parametric Statistics: Exploring Assumptions.

Parametric Statistics: Exploring Assumptions. Parametric Statistics: Exploring Assumptions http://www.pelagicos.net/classes_biometry_fa17.htm Reading - Field: Chapter 5 R Packages Used in This Chapter For this chapter, you will use the following packages:

More information

Numerical Descriptive Measures. Measures of Center: Mean and Median

Numerical Descriptive Measures. Measures of Center: Mean and Median Steve Sawin Statistics Numerical Descriptive Measures Having seen the shape of a distribution by looking at the histogram, the two most obvious questions to ask about the specific distribution is where

More information

Overview/Outline. Moving beyond raw data. PSY 464 Advanced Experimental Design. Describing and Exploring Data The Normal Distribution

Overview/Outline. Moving beyond raw data. PSY 464 Advanced Experimental Design. Describing and Exploring Data The Normal Distribution PSY 464 Advanced Experimental Design Describing and Exploring Data The Normal Distribution 1 Overview/Outline Questions-problems? Exploring/Describing data Organizing/summarizing data Graphical presentations

More information

Probability. An intro for calculus students P= Figure 1: A normal integral

Probability. An intro for calculus students P= Figure 1: A normal integral Probability An intro for calculus students.8.6.4.2 P=.87 2 3 4 Figure : A normal integral Suppose we flip a coin 2 times; what is the probability that we get more than 2 heads? Suppose we roll a six-sided

More information

Basic Procedure for Histograms

Basic Procedure for Histograms Basic Procedure for Histograms 1. Compute the range of observations (min. & max. value) 2. Choose an initial # of classes (most likely based on the range of values, try and find a number of classes that

More information

Monthly Treasurers Tasks

Monthly Treasurers Tasks As a club treasurer, you ll have certain tasks you ll be performing each month to keep your clubs financial records. In tonights presentation, we ll cover the basics of how you should perform these. Monthly

More information

Chapter 3. Numerical Descriptive Measures. Copyright 2016 Pearson Education, Ltd. Chapter 3, Slide 1

Chapter 3. Numerical Descriptive Measures. Copyright 2016 Pearson Education, Ltd. Chapter 3, Slide 1 Chapter 3 Numerical Descriptive Measures Copyright 2016 Pearson Education, Ltd. Chapter 3, Slide 1 Objectives In this chapter, you learn to: Describe the properties of central tendency, variation, and

More information

9/17/2015. Basic Statistics for the Healthcare Professional. Relax.it won t be that bad! Purpose of Statistic. Objectives

9/17/2015. Basic Statistics for the Healthcare Professional. Relax.it won t be that bad! Purpose of Statistic. Objectives Basic Statistics for the Healthcare Professional 1 F R A N K C O H E N, M B B, M P A D I R E C T O R O F A N A L Y T I C S D O C T O R S M A N A G E M E N T, LLC Purpose of Statistic 2 Provide a numerical

More information

ECON 214 Elements of Statistics for Economists 2016/2017

ECON 214 Elements of Statistics for Economists 2016/2017 ECON 214 Elements of Statistics for Economists 2016/2017 Topic The Normal Distribution Lecturer: Dr. Bernardin Senadza, Dept. of Economics bsenadza@ug.edu.gh College of Education School of Continuing and

More information

Washington University Fall Economics 487. Project Proposal due Monday 10/22 Final Project due Monday 12/3

Washington University Fall Economics 487. Project Proposal due Monday 10/22 Final Project due Monday 12/3 Washington University Fall 2001 Department of Economics James Morley Economics 487 Project Proposal due Monday 10/22 Final Project due Monday 12/3 For this project, you will analyze the behaviour of 10

More information

A useful modeling tricks.

A useful modeling tricks. .7 Joint models for more than two outcomes We saw that we could write joint models for a pair of variables by specifying the joint probabilities over all pairs of outcomes. In principal, we could do this

More information

Non-Inferiority Tests for the Ratio of Two Means in a 2x2 Cross-Over Design

Non-Inferiority Tests for the Ratio of Two Means in a 2x2 Cross-Over Design Chapter 515 Non-Inferiority Tests for the Ratio of Two Means in a x Cross-Over Design Introduction This procedure calculates power and sample size of statistical tests for non-inferiority tests from a

More information

Intraday Open Pivot Setup

Intraday Open Pivot Setup Intraday Open Pivot Setup The logistics of this plan are relatively simple and take less than two minutes to process from collection of the previous session s history data to the order entrance. Once the

More information

On Some Test Statistics for Testing the Population Skewness and Kurtosis: An Empirical Study

On Some Test Statistics for Testing the Population Skewness and Kurtosis: An Empirical Study Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 8-26-2016 On Some Test Statistics for Testing the Population Skewness and Kurtosis:

More information

Assignment 3-Solutions

Assignment 3-Solutions Assignment 3-Solutions Question 1. - Joint Probability Mass Function Consider the function x y 1.0 1.0 1.5 2.0 1.5 3.0 2.5 4.0 3.0 4.0 Determine the following: (a) Show that If is a valid probability mass

More information

Percentiles, STATA, Box Plots, Standardizing, and Other Transformations

Percentiles, STATA, Box Plots, Standardizing, and Other Transformations Percentiles, STATA, Box Plots, Standardizing, and Other Transformations Lecture 3 Reading: Sections 5.7 54 Remember, when you finish a chapter make sure not to miss the last couple of boxes: What Can Go

More information

Empirical Distribution Testing of Economic Scenario Generators

Empirical Distribution Testing of Economic Scenario Generators 1/27 Empirical Distribution Testing of Economic Scenario Generators Gary Venter University of New South Wales 2/27 STATISTICAL CONCEPTUAL BACKGROUND "All models are wrong but some are useful"; George Box

More information

PROBABILITY AND STATISTICS CHAPTER 4 NOTES DISCRETE PROBABILITY DISTRIBUTIONS

PROBABILITY AND STATISTICS CHAPTER 4 NOTES DISCRETE PROBABILITY DISTRIBUTIONS PROBABILITY AND STATISTICS CHAPTER 4 NOTES DISCRETE PROBABILITY DISTRIBUTIONS I. INTRODUCTION TO RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS A. Random Variables 1. A random variable x represents a value

More information