Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
Tutorial on Probability and Estimation, Dhruv Batra, Probability, Continuous Random Variables, Bias, Probabilistic Model, Probability Distributions, Sequence Probability, Parameter Estimation Problem, Maximum Likelihood Estimator, Bernoulli, Bayes Rule, MAP Estimation, Discrete Vs. Continuous RV, Beta Prior for Bernoulli, Conjugate Prior, MAP Estimate, Bernoulli, Beta Distributions, Gaussian Distributions, Greg Shakhnarovich, Lecture Slides, Introduction to Machine Learning, Computer Science, To
Typology: Lecture notes
1 / 89
TTIC 31020: Introduction to Machine Learning
Instructor: Greg Shakhnarovich, Lecture by Dhruv Batra
TTI–Chicago
September 27, 2010
TTIC 31020, Introduction to Machine Learning
MWF 9:30-10:20am
Instructor: Greg Shakhnarovich, [email protected]
TA: Feng Zhao, [email protected]
Greg is traveling this week; all administrative details will be discussed next Monday.
TTIC 31020, Introduction to Machine Learning
MWF 9:30-10:20am
Instructor: Greg Shakhnarovich, [email protected]
TA: Feng Zhao, [email protected]
Greg is traveling this week; all administrative details will be discussed next Monday.
Plan for this week:
This class is mostly about statistical methods and models in machine learning
Probability is fundamental in dealing with uncertainty inherent in real world problems
Statistics leverages laws of probability to evaluate important properties of the world from data, and make intelligent predictions about future
Things you should have seen before
Things you should have seen before
Things you should have seen before
Things you should have seen before
Things you should have seen before
Things you should have seen before
Things you should have seen before
This refresher WILL revise these topics.
A single coin toss produces H or T.
A sequence of n coin tosses produces a sequence of values; n = 4
A single coin toss produces H or T.
A sequence of n coin tosses produces a sequence of values; n = 4 T ,H,T ,H
A single coin toss produces H or T.
A sequence of n coin tosses produces a sequence of values; n = 4 T ,H,T ,H H,H,T ,T
A single coin toss produces H or T.
A sequence of n coin tosses produces a sequence of values; n = 4 T ,H,T ,H H,H,T ,T T ,T ,T ,H
A single coin toss produces H or T.
A sequence of n coin tosses produces a sequence of values; n = 4 T ,H,T ,H H,H,T ,T T ,T ,T ,H
A probabilistic model allows us to model the uncertainly inherent in the process (randomness in tossing a coin), as well as our uncertainty about the properties of the source (fairness of the coin).
First, for convenience, convert H → 1, T → 0.
First, for convenience, convert H → 1, T → 0.
Bernoulli distribution with parameter μ:
Pr(X = 1; μ) = μ.
First, for convenience, convert H → 1, T → 0.
Bernoulli distribution with parameter μ:
Pr(X = 1; μ) = μ.
We will write for simplicity p(x) or p(x; μ) instead of Pr(X = x; μ)
First, for convenience, convert H → 1, T → 0.
Bernoulli distribution with parameter μ:
Pr(X = 1; μ) = μ.
We will write for simplicity p(x) or p(x; μ) instead of Pr(X = x; μ)
The parameter μ ∈ [0, 1] specifies the bias of the coin
Discrete random variable X taking values in set X = {x 1 , x 2 ,.. .}
Discrete random variable X taking values in set X = {x 1 , x 2 ,.. .}
Probability mass function p : X → [0, 1] satisfies the law of total probability: ∑
x∈X
p(X = x) = 1
Discrete random variable X taking values in set X = {x 1 , x 2 ,.. .}
Probability mass function p : X → [0, 1] satisfies the law of total probability: ∑
x∈X
p(X = x) = 1
Hence, for Bernoulli distribution we know
p(0) = 1 − p(1; μ) = 1 − μ.
Now consider two tosses of the same coin, 〈 X 1 , X 2 〉
Now consider two tosses of the same coin, 〈 X 1 , X 2 〉
We can consider a number of probability distributions:
Joint distribution p(X 1 , X 2 ) Conditional distributions p(X 1 | X 2 ), p(X 2 | X 1 ), Marginal distributions p(X 1 ), p(X 2 )