Statistical Core Didactic - PowerPoint PPT Presentation

About This Presentation
Title:

Statistical Core Didactic

Description:

Title: Oral Biology 201 Author: Donald E Mercante Last modified by: pcaba2 Created Date: 9/11/2006 1:35:29 AM Document presentation format: On-screen Show (4:3) – PowerPoint PPT presentation

Number of Views:69
Avg rating:3.0/5.0
Slides: 57
Provided by: Donal265
Category:

less

Transcript and Presenter's Notes

Title: Statistical Core Didactic


1
Statistical Core Didactic
  • Introduction to
  • Biostatistics
  • Donald E. Mercante, PhD

2
Randomized Experimental Designs
3
Randomized Experimental Designs
  • Three Design Principles
  • 1. Replication
  • 2. Randomization
  • 3. Blocking

4
Randomized Experimental Designs
  • 1. Replication
  • Allows estimation of experimental error, against
    which, differences in trts are judged.

5
Randomized Experimental Designs
  • Replication
  • Allows estimation of exptl error, against which,
    differences in trts are judged.
  • Experimental Error
  • Measure of random variability.
  • Inherent variability between subjects treated
    alike.

6
Randomized Experimental Designs
  • If you dont replicate . . .
  • . . . You cant estimate!

7
Randomized Experimental Designs
  • To ensure the validity of our estimates of
    exptl error and treatment effects we rely on ...

8
Randomized Experimental Designs
  • . . . Randomization

9
Randomized Experimental Designs
  • 2. Randomization
  • leads to unbiased estimates of
  • treatment effects

10
Randomized Experimental Designs
  • Randomization
  • leads to unbiased estimates of
  • treatment effects
  • i.e., estimates free from systematic differences
    due to uncontrolled variables

11
Randomized Experimental Designs
  • Without randomization, we may need to adjust
    analysis by
  • stratifying
  • covariate adjustment

12
Randomized Experimental Designs
  • 3. Blocking
  • Arranging subjects into similar groups to
  • account for systematic differences

13
Randomized Experimental Designs
  • Blocking
  • Arranging subjects into similar groups (i.e.,
    blocks) to account for systematic differences
  • - e.g., clinic site, gender, or age.

14
Randomized Experimental Designs
  • Blocking
  • leads to increased sensitivity of statistical
    tests by reducing exptl error.

15
Randomized Experimental Designs
  • Blocking
  • Result More powerful statistical test

16
Randomized Experimental Designs
  • Summary
  • Replication allows us to estimate Exptl Error
  • Randomization ensures unbiased estimates of
    treatment effects
  • Blocking increases power of statistical tests

17
Randomized Experimental Designs
  • Three Aspects of Any Statistical Design
  • Treatment Design
  • Sampling Design
  • Error Control Design

18
Randomized Experimental Designs
  • 1. Treatment Design
  • How many factors
  • How many levels per factor
  • Range of the levels
  • Qualitative vs quantitative factors

19
Randomized Experimental Designs
  • One Factor Design Examples
  • Comparison of multiple bonding agents
  • Comparison of dental implant techniques
  • Comparing various dose levels to achieve numbness

20
Randomized Experimental Designs
  • Multi-Factor Design Examples
  • Factorial or crossed effects
  • Bonding agent and restorative compound
  • Type of perio procedure and dose of antibiotic
  • Nested or hierarchical effects
  • Surface disinfection procedures within clinic type

21
Randomized Experimental Designs
  • 2. Sampling or Observation Design
  • Is observational unit experimental unit ?
  • or,
  • is there subsampling of EU ?

22
Randomized Experimental Designs
  • Sampling or Observation Design
  • For example,
  • Is one measurement taken per mouth, or are
    multiple sites measured?
  • Is one blood pressure reading obtained or are
    multiple blood pressure readings taken?

23
Randomized Experimental Designs
  • 3. Error Control Design
  • concerned with actual arrangement of the exptl
    units
  • How treatments are assigned to eus

24
Randomized Experimental Designs
  • 3. Error Control Design
  • Goal Decrease experimental error

25
Randomized Experimental Designs
  • 3. Error Control Design
  • Examples
  • CRD Completely Randomized Design
  • RCB Randomized Complete Block Design
  • Split-mouth designs (whole incomplete block)
  • Cross-Over Design

26
Inferential Statistics
  • Hypothesis Testing
  • Confidence Intervals

27
Hypothesis Testing
  • Start with a research question
  • Translate this into a testable hypothesis

28
Hypothesis Testing
  • Specifying hypotheses
  • H0 null or no effect hypothesis
  • H1 research or alternative hypothesis
  • Note Only the null is tested.

29
Errors in Hypothesis Testing
  • When testing hypotheses, the chance of making a
    mistake always exists.
  • Two kinds of errors can be made
  • Type I Error
  • Type II Error

30
Errors in Hypothesis Testing
Reality ? ? Decision H0 True H0 False
Fail to Reject H0 ? Type II (?)
Reject H0 Type I (?) ?
31
Errors in Hypothesis Testing
  • Type I Error
  • Rejecting a true null hypothesis
  • Type II Error
  • Failing to reject a false null hypothesis

32
Errors in Hypothesis Testing
  • Type I Error
  • Experimenter controls or explicitly sets this
    error rate - ?
  • Type II Error
  • We have no direct control over this error rate - ?

33
Randomized Experimental Designs
  • When constructing an hypothesis
  • Since you have direct control over Type I error
    rate, put what you think is likely to happen in
    the alternative.
  • Then, you are more likely to reject H0, since
    you know the risk level (?).

34
Errors in Hypothesis Testing
  • Goal of Hypothesis Testing
  • Simultaneously minimize chance of making either
    error

35
Errors in Hypothesis TestingIndirect Control of ß
  • Power
  • Ability to detect a false null hypothesis
  • POWER 1 - ?

36
Steps in Hypothesis Testing
  • General framework
  • Specify null alternative hypotheses
  • Specify test statistic and ?-level
  • State rejection rule (RR)
  • Compute test statistic and compare to RR
  • State conclusion

37
Steps in Hypothesis Testing
  • test statistic
  • Summary of sample evidence relevant to
    determining whether the null or the alternative
    hypothesis is more likely true.

38
Steps in Hypothesis Testing
  • test statistic
  • When testing hypotheses about means, test
    statistics usually take the form of a standardize
    difference between the sample and hypothesized
    means.

39
Steps in Hypothesis Testing
  • test statistic
  • For example, if our hypothesis is
  • Test statistic might be

40
Steps in Hypothesis Testing
  • Rejection Rule (RR)
  • Rule to base an Accept or Reject null
    hypothesis decision.
  • For example,
  • Reject H0 if t gt 95th percentile of
    t-distribution

41
Hypothesis Testing
  • P-values
  • Probability of obtaining a result (i.e., test
    statistic) at least as extreme as that observed,
    given the null is true.

42
Hypothesis Testing
  • P-values
  • Probability of obtaining a result at least as
    extreme given the null is true.
  • P-values are probabilities
  • 0 lt p lt 1 lt-- valid range
  • Computed from distribution of the test
    statistic

43
Hypothesis Testing
  • P-values
  • Generally, plt0.05 considered significant

44
Hypothesis Testing
45
Hypothesis Testing
  • Example
  • Suppose we wish to study the effect on blood
    pressure of an exercise regimen consisting of
    walking 30 minutes twice a day.
  • Let the outcome of interest be resting systolic
    BP.
  • Our research hypothesis is that following the
    exercise regimen will result in a reduction of
    systolic BP.

46
Hypothesis Testing
  • Study Design 1 Take baseline SBP (before
    treatment) and at the end of the therapy period.
  • Primary analysis variable difference in SBP
    between the baseline and final measurements.

47
Hypothesis Testing
  • Null Hypothesis
  • The mean change in SBP (pre post) is equal to
    zero.
  • Alternative Hypothesis
  • The mean change in SBP (pre post) is
    different from zero.

48
Hypothesis Testing
  • Test Statistic
  • The mean change in SBP (pre post) divided by
    the standard error of the differences.

49
Hypothesis Testing
  • Study Design 2 Randomly assign patients to
    control and experimental treatments. Take
    baseline SBP (before treatment) and at the end of
    the therapy period (post-treatment).
  • Primary analysis variable difference in SBP
    between the baseline and final measurements in
    each group.

50
Hypothesis Testing
  • Null Hypothesis
  • The mean change in SBP (pre post) is equal in
    both groups.
  • Alternative Hypothesis
  • The mean change in SBP (pre post) is
    different between the groups.

51
Hypothesis Testing
  • Test Statistic
  • The difference in mean change in SBP (pre
    post) between the two groups divided by the
    standard error of the differences.

52
Interval Estimation
  • Statistics such as the sample mean, median, and
    variance are called
  • point estimates
  • -vary from sample to sample
  • -do not incorporate precision

53
Interval Estimation
  • Take as an example the sample mean
  • X gt ?
  • (popn mean)
  • Or the sample variance
  • S2 gt ?2
  • (popn variance)

Estimates
Estimates
54
Interval Estimation
  • Recall, a one-sample t-test on the population
    mean. The test statistic was
  • This can be rewritten to yield

55
Interval Estimation
Confidence Interval for ?
The basic form of most CI Estimate
Multiple of Std Error of the Estimate
56
Interval Estimation
  • Example Standing SBP
  • Mean 140.8, S.D. 9.5, N 12
  • 95 CI for ?
  • 140.8 2.201 (9.5/sqrt(12))
  • 140.8 6.036
  • (134.8, 146.8)
Write a Comment
User Comments (0)
About PowerShow.com