1 / 23

BIBD and Adjusted Sums of Squares

BIBD and Adjusted Sums of Squares. Type I and Type III Sums of Squares. Confounding in Unbalanced Designs. When designs are “unbalanced”, typically with missing values, our estimates of Treatment Effects can be biased.

junius
Télécharger la présentation

BIBD and Adjusted Sums of Squares

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. BIBD and Adjusted Sums of Squares Type I and Type III Sums of Squares

  2. Confounding in Unbalanced Designs • When designs are “unbalanced”, typically with missing values, our estimates of Treatment Effects can be biased. • When designs are “unbalanced”, the usual computation formulas for Sums of Squares can give misleading results, since some of the variability in the data can be explained by two or more variables.

  3. Example BIBDfrom Hicks

  4. Type I vs. Type III in partitioning variation • If an experimental design is not a balanced and complete factorial design, it is not an orthogonal design. • If a two factor design is not orthogonal, then the SSModel will not partition into unique components, i.e., some components of variation may be explained by either factor individually (or simultaneously). • Type I SS are computing according to the order in which terms are entered in the model. • Type III SS are computed in an order independent fashion, i.e. each term gets the SS as though it were the last term entered for Type I SS.

  5. Notation for Hicks’ example There are only two possible factors, Block and Trt. There are only three possible simple additive models one could run. In SAS syntax they are: • Model 1: Model Y=Block; • Model 2: Model Y=Trt; • Model 3: Model Y=Block Trt;

  6. Adjusted SS notation Each model has its own “Model Sums of Squares”. These are used to derive the “Adjusted Sums of Squares”. • SS(Block)=Model Sums of Squares for Model 1 • SS(Trt)=Model Sums of Squares for Model 2 • SS(Block,Trt)=Model Sums of Squares for Model 3

  7. The Sums of Squares for Block and Treatment can be adjusted to remove any possible confounding. Adjusting Block Sums of Squares for the effect of Trt: SS(Block|Trt)= SSModel(Block,Trt)- SSModel(Trt) Adjusting TrtSums of Squares for the effect of Block: SS(Trt|Block)= SSModel(Block,Trt)- SSModel(Block)

  8. From Hicks’ Example • SS(Block)=100.667 • SS(Trt)=975.333 • SS(Block,Trt)=981.500

  9. For SAS model Y=Block Trt; Source df Type I SS Type III SS Block 3 SS(Block) SS(Block|Trt) =100.667 =981.50-975.333 Trt 3 SS(Trt|Block) SS(Trt|Block) =981.50-100.667 =981.50-100.667

  10. ANOVA Type III and Type I(Block first term in Model)

  11. For SAS model Y=Trt Block; Source df Type I SS Type III SS Trt 3 SS(Trt) SS(Trt|Block) =975.333 =981.50-100.667 Block 3 SS(Block|Trt) SS(Block|Trt) =981.50-975.333 =981.50-975.333

  12. ANOVA Type III and Type I(Trt. First term in Model)

  13. How does variation partition?

  14. How this can work-IHicks example

  15. When does case I happen? In Regression, when two Predictor variables are positively correlated, either one could explain the “same” part of the variation in the Response variable. The overlap in their ability to predict is what is adjusted “out” of their Sums of Squares.

  16. Example BIBDFrom Montgomery (things can go the other way)

  17. ANOVA with Adjusted and Unadjusted Sums of Squares

  18. Sequential Fit with Block first

  19. Sequential Fit with Treatment first

  20. LS Means Plot

  21. LS Means for Treatment, Tukey HSD

  22. How this can work- IIMontgomery example

  23. When does case II happen? Sometimes two Predictor variables can predict the Response better in combination than the total of they might predict by themselves. In Regression this can occur when Predictor variables are negatively correlated.

More Related