1 / 22

Intro to Statistics Part2

Intro to Statistics Part2. Arier Lee University of Auckland. Standard error. Standard error – the standard deviation of the sampling distribution of a statistic

walden
Télécharger la présentation

Intro to Statistics Part2

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Intro to Statistics Part2 Arier Lee University of Auckland

  2. Standard error • Standard error – the standard deviation of the sampling distribution of a statistic • The standard deviation of the sample means is called the standard error of the mean and it measures how precisely the population mean is estimated by the sample mean • The standard error is a measure of the precision of the estimated mean whereas the standard deviation summarises the variability or the spread of the observations • Standard error <= standard deviation • The larger the sample size the smaller the standard error

  3. Confidence intervals • A 95% confidence interval for a mean is calculated by (mean-1.96*SE, mean+1.96*SE) • An example: In a sample of 2000 pregnant women, serum cholesterol was measured and it was found that the sample mean is 5.62 and SE=0.15. 95% confidence interval: (5.33, 5.91)

  4. Confidence intervals • 95% CI does not mean that there is a 95% chance that the true mean lies between 5.33 and 5.91 • If we repeat the study over and over again, calculating a 95% confidence interval each time, about 95 of 100 such intervals would include the true mean • Whether the one that we have obtained from our study is one of them we will never know – but we have some confidence • It is a measure of precision of our estimate • Bigger confidence interval -> less precision

  5. Graphical presentation of the data • Exploratory data analysis • Presentation of results • Examples: Bar charts, Line graphs, Scatter plots, Box plots, Kaplan Meier Plots etc. • Graphs can only be as good as the data they display • No amount of creativity can produce a good graph from dubious data

  6. Bar chart 2005 maternity report

  7. Line graph

  8. Box plot Obs beyond end of whisker Q3 1.5 x (Q3-Q1) median Q1 Smallest obs marks end of whisker

  9. Data to chart ratio Mental health score by treatment groups Good Bad

  10. Inadequate chart type Effect of ethnicity on road traffic injury deaths and hospitalisations, 2000-8, Auckland region, by age group, adjusted for gender and deprivation (using National Minimum Data Set and Mortality Collection data) Graphs of risk or rate ratio should be presented with • Points with error bars • Log scale

  11. Odds ratio presented with logarithmic scale Outcome: Blindness

  12. Unnecessary 3D effects How often do you read to your child

  13. Inadequate labelling

  14. Graphical presentation of the data • Use appropriate graph types for the appropriate purpose, e.g. line chart for trend • All axes, tick marks, title, should be labelled • Appropriate scale used • Adequate data to chart ratio • Avoid unnecessary complexity such as • Irrelevant decoration • Too much colours • 3D effects • Keep it simple!

  15. Research process Research question Analyse data Primary and secondary endpoints Study design Interpret results Sampling and/or randomisation scheme Disseminate Power and sample size calculation Pre-define analyses methods

  16. Sample size and power of a study • One of the statistical, economical and ethical issues of the design of medical studies • Statistical: Ensure the study is large enough to detect an effect if it exists • Economical: Ensure not enlist more patients than are needed • Ethical: unethical to engage more people in a trial than are needed • Larger samples -> more precise estimates • How large?

  17. Sample size and power of a study • The power of a test is the probability of detecting a true difference • The size of the sample needed depends on • required power • detectable difference • variability in the population • level of significance (probability of falsely reject the NULL) • statistical test being used • Need information to calculated a meaningful sample size – literature search

  18. Sample size and power of a study- an example • A double blind randomised controlled study on treatment for chronic hypertension during pregnancy • Comparing two treatments: • Standard treatment • New treatment

  19. Sample size and power of a study- an example • Based on current evidence, assume • Detectable difference: 10mmHg • Standard deviation: 15 mmHg • 90% power • 5% significance level • Two-sided test • 1:1 ratio • Using PS (a power and sample size calculation software) – 48 subjects per group • After considering drop-out rate, say 10%, round to, say, 60 subjects per group

  20. Sample size and power of a study Chronic hypertension during pregnancy example • To detect a difference of 10mmHg • SD varies from 5 to 30mmHg

  21. Sample size and power of a study Sample size calculation is an evidence based best guess • Relies on assumptions • Not a precise number • No guarantee of significant effect at the end of a study

  22. Any Questions?

More Related