- ↑
STA258 Lecture 12
STA258 Lecture 12 Raw
STA258 Lecture 12 Flashcards
-
Completed Notes Status
- Completed insertions: 52
- Ambiguities left unresolved: 1
-
Lecture Summary
- Central objective: Introduce the framework of Hypothesis Testing, evaluating decision rules through error probabilities, and establishing connections between test statistics, confidence intervals, and p-values.
- Key concepts:
- Hypothesis Testing: Contrasts a null hypothesis (
) against an alternative hypothesis ( ), differentiating between simple (singleton) and composite hypotheses. - Type I Error and Type II Error:
(Type I) is the probability of rejecting a true , while (Type II) is the probability of failing to reject a false , demonstrating a seesaw trade-off where researchers generally prioritize minimizing . - Critical Region: The subset of the sample space that triggers the rejection of
, mathematically driven by bounding below a given threshold. - P-value: The probability of observing a test statistic at least as extreme as the one computed, assuming
is true, acting as a continuous metric for evidence against the null.
- Hypothesis Testing: Contrasts a null hypothesis (
- Connections:
- Two-sided hypothesis tests inherently link to Confidence Intervals; failing to reject
at level is mathematically equivalent to the null parameter value falling inside the confidence interval for the mean.
- Two-sided hypothesis tests inherently link to Confidence Intervals; failing to reject
-
Practice Questions
- Remember/Understand:
- What is the fundamental difference between a simple hypothesis and a composite hypothesis?
- Define a Type I error and a Type II error in plain language.
- How is the significance level
related to the critical region ?
- Apply/Analyze:
- If a researcher decreases the significance level
from to , what is the expected impact on the probability of a Type II error ( ), assuming sample size remains constant? - Given
, construct a rejection region to test vs ensuring that . - Explain why a two-sided hypothesis test decision boundary aligns perfectly with the boundaries of a confidence interval.
- If a researcher decreases the significance level
- Evaluate/Create:
- Critically evaluate the standard practice of rigidly adhering to
as a rejection threshold; what issues arise when versus in applied fields like finance? - Propose a scenario where a Type II error is vastly more dangerous than a Type I error, and justify how you would adjust the testing framework to accommodate this priority.
- Critically evaluate the standard practice of rigidly adhering to
- Remember/Understand:
-
Challenging Concepts
- Type I Error vs Type II Error (The Seesaw Effect):
- Why it's challenging: Balancing the minimization of both errors mathematically requires understanding that constraining one tail probability directly widens the non-rejection region, inherently inflating the complementary error risk under the alternative distribution.
- Study strategy: Draw overlapping probability density curves (one for
, one for ). Shade the areas representing and to visually trace how moving the critical value line shifts the balance between the two regions.
- Translating binomial parameters into optimal critical regions:
- Why it's challenging: Unlike continuous
or tests, discrete distributions like the binomial do not allow for exact matching (e.g., exactly 0.05), requiring analysts to map out probability tables and manually optimize for while keeping below a strict cap. - Study strategy: Re-calculate the
example from the notes entirely by hand, calculating for every single possible valid critical region to solidify the mechanical process of optimization.
- Why it's challenging: Unlike continuous
- Type I Error vs Type II Error (The Seesaw Effect):
-
Action Plan
- Immediate review actions:
- Practice and application:
- Deep dive study:
- Verification and integration:
- Immediate review actions:
-
Footnotes