Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
All Business Statistics Foundation. Includes probability, Anova, single and two data sample means and etc.,
Typology: Cheat Sheet
1 / 18
COMM 291 – Final Formula Sheet
Chapter 5: Displaying and Describing Quantitative Data
Measures of Centre Excel Mean
Mode Highest point in the histogram =MODE.SING Measures of Spread Range (^) Maximum value – minimum value
2
Percentiles 75 th^ percentile = 3rd^ or upper quartile = Q 25 th^ percentile = 1st^ or lower quartile = Q
Interquartile range (IQR) IQR = Q 3 – Q 1 Outliers (^) Greater than: Q3 + 1.5 (IQR) Less than: Q1 – 1.5 (IQR) Extreme Outliers Lower outer fence = Q1 – 3(IQR) Upper outer fence = Q3 + 3(IQR)
Chapter 6: Correlation and Linear Progression
Measure Formula Excel Correlation Coefficient (^) r = 1 𝑛− 1 ∑^ (
𝑥𝑖−𝑥̅ 𝑆𝑥^ )^ (
𝑦𝑖−𝑦̅ 𝑆𝑦^ )^
Observed y – predicted y
Data analysis tool pack (‘Regression’) b 1
Data analysis tool pack (‘Regression’)
(‘Regression’)
(‘Regression’); (=CORREL)^2 Standard deviation of the residuals (^) 𝑠𝑒 = √ ∑^ 𝑒𝑖
2
𝑛 − 2
Chapter 8: Random Variables and Probability Models
COMM290 Review : If X is a random variable:
Measure Formula Excel Expected Value E(X)=∑X*P(X) =SUMPRODUCT E(X) = = np Variance Var(X)=∑(x-μ)^2 *P(X) ^2 = 𝑛𝑝𝑞, where 𝑞 = 1 − 𝑝 Standard Deviation (^) σ = √Var(X) =STDEV.S
Linear transformation: a + bX
Expected Value
Variance 𝑉𝑎𝑟(𝑎 + 𝑏𝑋) = 𝑏^2 𝑉𝑎𝑟(𝑋)
Standard Deviation
Sum of two INDEPENDENT random variables: X+Y
Expected Value
(Mean of a sum = sum of the means) Variance 𝑉𝑎𝑟(𝑋 + 𝑌) = 𝑉𝑎𝑟(𝑋)
Standard Deviation
Difference of two INDEPENDENT random variables: X–Y
Expected Value
Variance 𝑉𝑎𝑟(𝑋 − 𝑌) = 𝑉𝑎𝑟(𝑋)
Standard Deviation
Linear combination of two INDEPENDENT random variables: aX + bY
Expected Value
Variance 𝑉𝑎𝑟(𝑎𝑋 + 𝑏𝑌) = 𝑎^2 𝑉𝑎𝑟(𝑋)
Standard Deviation
Linear combination of two DEPENDENT random variables: aX + bY
Expected Value
Variance (^) 𝑉𝑎𝑟(𝑎𝑋 + 𝑏𝑌) (^) = 𝑎^2 𝑉𝑎𝑟(𝑋)
Standard Deviation
The Empirical Rule
68% of the area under the normal curve lies within μ ± σ
95% of the area under the normal curve lies within μ ± 2σ
99.7% of the area under the normal curve lies within μ ± 3σ
=NORM.DIST(X, μ, σ, TRUE): Given X, find probability.
=NORM.INV: Given probability, find X.
Empirical Rule of Thumb
s ≈ Range/6 (for bell- shaped distributions and large n)
s ≈ Range/4 (for bell- shaped distributions and small n, approx. 20)
Standardization
=NORM.S.DIST: the standardized version; given Z, find probability.
=NORM.S.INV: Given probability, find z.
In our applications, X and Y will always be independent. But, if they are not, the variances cannot always simply be added.
Inference and Sampling Distribution for Proportions
When to use: when a binary predictor is used to measure a binary outcome.
Where p = successes and q = failures,
One sample z test
Excel
Parameter p
Estimate (^) 𝑝̂ = 𝑋 𝑛 (X = number of successes, n = sample) Model One sample z test
SE (estimate of standard dev) SD(𝑝̂^ ) =^ √
𝑝 0 𝑞 0 𝑛
Alternative Hypothesis
(Can be one sided)
Test Statistic
𝑝̂ −𝑝 0 √𝑝^0 𝑛𝑞 0 = NORM.S.INV(prob)
P-Value
HA: p > p 0 p = Pr(z > z-stat) (^) = 1 – NORM.S.DIST(z-
stat, cumulative) HA: p < p 0 p = Pr(z < z-stat) (^) = NORM.S.DIST(z-stat,
cumulative) HA: p ≠ p 0 p = 2 x Pr(z < z- stat)
= 2 x NORM.S.DIST(z- stat, cumulative)
Critical Values (z) (if you cannot find p-value)*
Sig Level (%)
One sided
Two sided
5 1.645 1.96 = NORM.S.INV(prob) 1 2.33 2. 0.1 3.09 3. Confidence Interval (Always 2-sided) 𝑝̂ ± z*√
𝑝̂ ( 1 −𝑝̂ ) 𝑛 Sample Size for a certain confidence interval n =^
(𝑧∗)^2 𝑝̂ ( 1 −𝑝̂ ) (𝑀𝐸)^2 (approximation with 95% CI) n =
1 (𝑀𝐸)^2
Two Samples z test
When to use: When there are 2 binary predictors (2 categories) and 2 binary outcomes ( categories).
Excel
Parameter p 1 – p2, Mean of (𝑝̂ 1 – 𝑝̂ 2 ) Estimate 𝑝̂ 1 – 𝑝̂ 2 , where (𝑝̂ 1 = X 1 / n 1 and 𝑝̂ 2 = X 2 / n 2 ) Model Two sample z test
SE (estimate of standard dev) (^) √𝑝̂^1 𝑞̂^1 𝑛 1
Test Statistic Z^ =^
(𝑝̂ 1 – 𝑝̂ 2 )−∆ 0
√𝑝
̂ 1 𝑞̂ 1 𝑛 1 +
𝑝̂ 2 𝑞̂ (^2) 𝑛 2 (If Δ 0 =0, we use a pooled p) 𝑝̂ 1 =
𝑋 1 𝑛 1 and^ 𝑝̂^2 =^
𝑋 2 𝑛 2 so 𝑝̂ = 𝑋 𝑛^1 +𝑋^2 1 +𝑛 2 Z =
(𝑝̂ 1 − 𝑝̂ 2 ) √𝑝̂ 𝑞̂ ( (^) 𝑛^11 + (^) 𝑛^12 )
where 𝑞̂ = 1–𝑝̂
P-Value
Ha: p 1 – p 2 > Δ 0 p = Pr(z > z- stat)
= 1 – NORM.S.DIST(z- stat, cumulative) Ha: p 1 – p 2 < Δ 0 p = Pr(z < z- stat)
= NORM.S.DIST(z-stat, cumulative) Ha: p 1 – p 2 ≠ Δ 0 p = 2 x Pr(z < z-stat)
= 2 x NORM.S.DIST(z- stat, cumulative)
Critical Values (z) (if you cannot find p-value)*
Sig Level (%)
One sided
Two sided
5 1.645 1.96 = NORM.S.INV(prob) 1 2.33 2. 0.1 3.09 3. Confidence Interval (Always 2-sided) (𝑝̂^1 –^ 𝑝̂^2 )^ ±^ z
𝑝̂ 2 𝑞̂ 2 𝑛 2
Inference and Hypothesis Tests for Means
One sample t test of means
Excel
Parameter μ
Estimate (^) 𝑥̅
Model One sample t test
SE (estimate of standard dev)
Null Hypothesis H 0 : μ = 0
Alternative Hypothesis
HA: μ ≠ 0 HA: μ > 0 HA: μ < 0
Test Statistic t =^
𝑥̅ −𝜇 0 𝑠 √𝑛
,
with degrees of freedom n-
P-Value
HA: μ > 0 Pr(t > t-stat) (^) =T.DIST.RT(t-stat,df)
HA: μ < 0 Pr(t < t-stat) (^) =T.DIST.RT(t-stat,df)
HA: μ ≠ 0 2 x Pr(t > t-stat) (^) =T.DIST.2T((t-stat,df)
Critical Values (t) (for CI)*
=T.INV. 2T(sig level,df)
Confidence Interval (Always 2-sided)
𝑥̅ ± t*n- 1
𝑠 √𝑛 Sample Size n =
(𝑡∗)^2 𝑠^2 (𝑀𝐸)^2 s ≈
𝑟𝑎𝑛𝑔𝑒 6
Two-sample t-test of Independent Means
Characteristic Unequal (or separate) Variance Version
Equal (or pooled) Variance Version
Excel
Parameter μ 1 – μ 2 Estimate 𝑥̅ 1 – 𝑥̅ 2 SE
Alternative Hypothesis
When to use pooled version
= pooled variance
Regression
t-statistic
2
𝑠 22 𝑛 2
1
1 𝑛 2
Regression
Distribution (^) Approximate Exact Degrees of freedom
Both If n1=n2, the two test statistics are the same. The P-value, however, is not.
=T.DIST.2T(t-stat, degrees of freedom) =T.DIST.RT(t-stat, degrees of freedom) Critical Value =T.INV.2T(sig level,df), Regression
Confidence Interval
s 12 n 1
s 22 n 2
1
1 𝑛 2 Regression
Two-sample t-test of dependent means (paired t-test)
When to use: same sample, different situations. EX: same person’s midterm and final results.
Excel
Estimate Sample mean of the differences, 𝑑̅
Model Paired t test
sd (estimate of standard dev)
Compute the differences and then find their standard deviation on excel. SE 𝑠𝑑
Null Hypothesis H 0 : μd = 0
Alternative Hypothesis
HA: μd ≠ 0 HA: μd > 0 HA: μd < 0
Test Statistic t =^
𝑠𝑑 √𝑛
,
with degrees of freedom n-
P-Value
HA: μ > 0 Pr (t > t-stat) (^) =T.DIST.RT(t-stat,df)
HA: μ < 0 Pr(t < t-stat) (^) =T.DIST.RT(t-stat,df)
HA: μ ≠ 0 2 x Pr(t > t-stat) (^) =T.DIST.2T((t-stat,df)
Critical Values (z) (for CI)*
=T.INV. 2T(prob,df)
Confidence Interval
Types of Errors in Hypothesis Testing
Decision Accept Ho Reject Ho True State
Ho True Correct Type I error (False positive) Ho False Type II error (False negative)
Correct
Chapter 13: Inference for Counts—Chi Square Tests
‘Goodness of Fit’ Test
When to use: by assessing the discrepancy between observed frequencies for a categorical variable and the frequencies that would be expected under a certain theory.
Excel
Parameter NO PARAMETERS
Model Chi Square test
Null Hypothesis H 0 : p 1 = p 2 = p 3 = (observed proportions)
Alternative Hypothesis
HA: at least one p 1 is not observed value
Test Statistic (always one sided)
𝜒^2 =
𝑠𝑑 √𝑛
,
with degrees of freedom n-
P-Value (Always one sided) P-value = Pr (𝜒^2 > 𝜒^2 -stat)
=CHISQ.DIST.RT(test- statistic, degrees of freedom)
Critical Values (X) (for CI)*
Chi-square test of two way tables (See summary for steps)
When to use:
A 2x2 table can be tested using either a two-sample z-test for two proportions or a 𝜒^2 table.
Excel
Parameter NO PARAMETERS
Model (^) Chi Square test
Null Hypothesis Ho: Row and column classifications are independent (or unassociated, or unrelated)
Alternative Hypothesis
Ha: Row and column classifications are dependent (or associated, or related)
Expected Counts
Test Statistic 𝜒^2 = ∑
(𝑂𝑏𝑠−𝐸𝑥𝑝)^2 𝐸𝑥𝑝
, where the
sum is taken over all the cells in the table.
P-Value (Always one sided) (^) P-value = Pr (𝜒^2 > 𝜒^2 - stat); Degrees of freedom = (r–1)(c–1), (r = number of rows, c = number of columns)
=CHISQ.DIST.RT(test- statistic, degrees of freedom)
Critical Values (X) (for CI)*
Chapter 16: Introduction to Statistical Models
𝑌 = 𝛽 0 + 𝛽 1 𝑋 + 𝜀 Y is the unknown dependent variable (outcome) X is the known independent variable (predictor) ε is a random variable representing random error β 0 and β 1 are parameters
β 1
β 0
𝑆𝑥
Mean β 1 β 0
, or
√∑(𝑥𝑖−𝑥̅ )^2 𝑠𝑒√
1 𝑛
𝑥̅ 2 (𝑛− 1 )𝑠𝑥^2
, or
𝑠𝑒√
1 𝑛
𝑥̅ 2 ∑(𝑥𝑖−𝑥̅ )^2 S^2 e
Sampling distribution Normal Null Hypothesis HO: β 1 = 0 i.e.:
HO: β 1 = 0 (Not often tested)
Alternative Hypothesis (^) Ha: β 1 ≠ 0 (Always two sided for this case)
Ha: β 0 ≠ 0
Test statistics 𝑡 =
P value P-value = 2 x Pr (t > |t-stat|) P-value = 2 x Pr (t > |t-stat|)
1 ±^ t*n-^2
𝑠𝑒 𝑠𝑥√𝑛− 1
β 0 : 𝑏 0 ± t*n- 2 𝑠𝑒√
1 𝑛
𝑥̅ 2 (𝑛− 1 )𝑠𝑥^2
Confidence Interval for a Mean Response
When to use:
𝑠𝑒^2 𝑛
𝑠𝑒^2 𝑛
Confidence interval depends on:
Prediction Interval for a Future Observation
𝑠𝑒^2
2
100(1-α)% Prediction Interval
𝑠𝑒^2
2
EXCEPTION: When n is large, this simply becomes:
𝑦̂𝑥∗^ ± 𝑡𝑛−2∗^ 𝑠𝑒 where 𝑦̂𝑥∗^ = 𝑏 0 + 𝑏 1 𝑥∗
For a 95% prediction interval, this becomes simply:
𝑦̂𝑥∗^ ± 2𝑠𝑒
Extrapolation Penalty = 𝑆𝐸^2 (𝑏 1 ) × (𝑥∗^ − 𝑥̅)^2
First Hypothesis Test for Correlation
When to use:
Parameter Correlation coefficient of the population (ρ). Estimate r
Null Hypothesis Ho: ρ = 0
Alternative Hypothesis Ha: ρ ≠ 0
Test Statistic t =
P-Value (Two sided)
P-value = Pr (𝜒^2 > 𝜒^2 -stat) =T.DIST(t-stat,n-2)
Another Test for β1: Analysis of Variance for Regression (ANOVA TABLE)
When to use: to answer whether the X variable is a useful predictor of Y; if the model is worthwhile.
Sum of Squares Total (SST) = Sum of Squares Model (SSM) + Sum of Squares Error (SSE) (Regression) (Residual)
Source of Variation
Sum of Squares
Degrees of Freedom
Mean Square F-stat
Model SSM 1 MSM=SSM/df MSM/MSE Error SSE n-2 MSE=SSE/(n-2) Total SST n-
Analysis of Variance to Test the Worthiness of the Model
Excel
Parameter β 1
Se^2 𝑠𝑒^2 =
∑ 𝑒𝑖^2
𝑛 − 2
“Standard Error” in the Regression analysis of excel, (standard error of the residuals) Null Hypothesis Ho: β 1 = 0 (The model is not worthwhile; x has no predictive info about y)
Alternative Hypothesis
Ha: β 1 ≠ 0 (The model is worthwhile; x has some predictive information over y) Test Statistic (Overall F test)
Test statistic: F = MSM / MSE = F.DIST.RT(F- stat,df1(k),df2(n-k-1) = T.DIST.2T(SQRT(F- stat), n-k-1) F-stat and t-stat (for simple regression only)
t^2 n- 2 = F1, n- 2 t = (^) √𝐹
P-Value (Always one sided) P-value = Pr (F > F-stat) = F.DIST.RT(f-stat,k,n-
𝑆𝑆𝑀 𝑆𝑆𝑇
= proportion of total variation explained by the model
r = √
𝑆𝑆𝑀 𝑆𝑆𝑇 , (excel says R-multiple, but you have to find the sign from the slope sign)
Chapter 15: Multiple Regression
When to use: to answer the questions
Excel
Parameter 𝛽 0 , 𝛽 1 , 𝛽 2 , … , 𝛽𝑘
Estimates (^) b 0 , b 1 , b 2 , …, bk
𝒔𝒆^ 𝟐 𝑠𝑒^2 =
∑ (^) 𝑒𝑖^2
𝑛 − 𝑘 − 1
“Standard error” in the ANOVA table (the standard error of the residuals) Model ANOVA
Null Hypothesis Ho: β 1 = β 2 = … = βk = 0
of the X-variables has any explanatory power or value)
Alternative Hypothesis
Ha: at least one βj is not 0 (The model is worthwhile; there are some valuable X-variables somewhere in the model) Test Statistic (Overall F test)
Test statistic: F = MSM / MSE = F.DIST.RT(F- stat,df1(k),df2(n-k-1) = T.DIST.2T(SQRT(F- stat), n-k-1)
P-Value (Always one sided) P-value = Pr (F > F-stat) = F.DIST.RT(f-stat,k,n- k- 1 )
Note: If H0 is rejected, the next steps are to find the smallest set of predictors that adequately fit the data.
ANOVA Table (The same, with minor changes in the Degrees of Freedom column).
Source of Variation
Sum of Squares
Degrees of Freedom
Mean Square F-stat
Model/Regression SSM k MSM=SSM/k MSM/MSE Error/Residual SSE n-k-1 MSE=SSE/(n-k-1) Total SST n-
R-squared is now formally called the Coefficient of Multiple Determination, but is defined the same way:
R^2 =
= proportion of total variation explained by the model
F-test to Compare Full and Reduced Models
When to use: to see if all present variables are relevant, and to avoid multicollinearity. Full Model: Use all available predictor variables Reduced Model: Subset of Full Model: drop q of the predictor variables
Null Hypothesis H 0 : The reduced model is adequate (The extra variables in the full model that are not in the reduced model do not provide a significant improvement in the model's predictive ability)
Alternative Hypothesis
Ha: The reduced model is not adequate (At least one of the extra predictors is worthwhile, so the reduced model discards some worthwhile predictors; therefore, it is better to keep the full model rather than this reduced model.) Test Statistic (F test to compare full and reduced models)
k = number of X-variables in the Full Model q = number of variables dropped from the Full Model to get the Reduced Model
P-Value (Always one sided) P-value = Pr (F > F-stat) = F.DIST.RT(f- stat,k,n-k-1)
Regression Output
Other formulas:
(3): R^2 : 1 – (1–R^2 )[(n–1)/(n–2)]
(15): Pr[F1,n- 2 > (14)]
(18): 𝑠𝑒√
1 𝑛
𝑥̅ 2 ∑(𝑥𝑖−𝑥̅ )^2
𝑠𝑒 𝑠𝑥√𝑛−
(22): 2×Pr[ t n- 2 > |(20)|]
(23): 2×Pr[ t n- 2 > |( 21 )|]
(24): (16) – t *n- 2 ×(18)
(25): (17) – t *n- 2 ×(19)
(26): (16) + t *n- 2 ×(18)
(27): (17) + t *n- 2 ×(19)