CS计算机代考程序代写 Bayesian flex case study PowerPoint Presentation – cscodehelp代写

PowerPoint Presentation

Lecturer: Ben Rubinstein

Lecture 3. Linear Regression

COMP90051 Statistical Machine Learning

Copyright: University of Melbourne

COMP90051 Statistical Machine Learning

This lecture

• Linear regression
∗ Simple model (convenient maths at expense of flexibility)
∗ Often needs less data, “interpretable”, lifts to non-linear
∗ Derivable under all Statistical Schools: Lect 2 case study

• This week: Frequentist + Decision theory derivations
Later in semester: Bayesian approach

∗ Convenient optimisation: Training by “analytic” (exact) solution

• Basis expansion: Data transform for more expressive models

2

COMP90051 Statistical Machine Learning

Linear Regression
via Decision Theory

A warm-up example

3

COMP90051 Statistical Machine Learning

Example: Predict humidity from temperature
Temperature Humidity

TRAINING DATA

85 85

80 90

83 86

70 96

68 80

65 70

64 65

72 95

69 70

75 80

TEST DATA

75 70

In regression, the task is to predict numeric
response (aka dependent variable) from features
(aka predictors or independent variables)
Assume a linear relation: 𝐻𝐻 = 𝑎𝑎 + 𝑏𝑏𝑏𝑏
(𝐻𝐻 – humidity; 𝑏𝑏 – temperature; 𝑎𝑎, 𝑏𝑏 – parameters)

4

? ?

COMP90051 Statistical Machine Learning

5

• The model is
𝐻𝐻 = 𝑎𝑎 + 𝑏𝑏𝑏𝑏

• Fitting the model =
finding “best” 𝑎𝑎, 𝑏𝑏
values for data at
hand

• Important criterion:
minimise the sum
of squared errors
(aka residual sum of
squares)

Example: Problem statement

? ?

COMP90051 Statistical Machine Learning

To find 𝑎𝑎, 𝑏𝑏 that minimise 𝐿𝐿 = ∑𝑖𝑖=1
10 𝐻𝐻𝑖𝑖 − 𝑎𝑎 + 𝑏𝑏 𝑏𝑏𝑖𝑖

2

set derivatives to zero:
𝜕𝜕𝐿𝐿
𝜕𝜕𝑎𝑎

= −2�
𝑖𝑖=1

10
𝐻𝐻𝑖𝑖 − 𝑎𝑎 − 𝑏𝑏 𝑏𝑏𝑖𝑖 = 0

if we know 𝑏𝑏, then �𝑎𝑎 = 1
10
∑𝑖𝑖=1
10 𝐻𝐻𝑖𝑖 − 𝑏𝑏 𝑏𝑏𝑖𝑖

𝜕𝜕𝐿𝐿
𝜕𝜕𝑏𝑏

= −2�
𝑖𝑖=1

10
𝑏𝑏𝑖𝑖 𝐻𝐻𝑖𝑖 − 𝑎𝑎 − 𝑏𝑏 𝑏𝑏𝑖𝑖 = 0

if we know 𝑎𝑎, then �𝑏𝑏 = 1
∑𝑖𝑖=1
10 𝑇𝑇𝑖𝑖

2 ∑𝑖𝑖=1
10 𝑏𝑏𝑖𝑖 𝐻𝐻𝑖𝑖 − 𝑎𝑎

Can we be more systematic?

6

High-school optimisation:
• Write derivative
• Set to zero
• Solve for model
• (Check 2nd derivatives)

Example: Minimise Sum Squared Errors

COMP90051 Statistical Machine Learning

• We have two equations and two unknowns 𝑎𝑎, 𝑏𝑏

• Rewrite as a system of linear equations

10 ∑𝑖𝑖=1
10 𝑏𝑏𝑖𝑖

∑𝑖𝑖=1
10 𝑏𝑏𝑖𝑖 ∑𝑖𝑖=1

10 𝑏𝑏𝑖𝑖
2

𝑎𝑎
𝑏𝑏 =

∑𝑖𝑖=1
10 𝐻𝐻𝑖𝑖

∑𝑖𝑖=1
10 𝑏𝑏𝑖𝑖𝐻𝐻𝑖𝑖

• Analytic solution: 𝑎𝑎 = 25.3, 𝑏𝑏 = 0.77

• (Solve using numpy.linalg.solve or sim.)

7

Example: Analytic solution

COMP90051 Statistical Machine Learning

• Adopt a linear relationship between response 𝑦𝑦 ∈ ℝ and
an instance with features 𝑥𝑥1, … , 𝑥𝑥𝑚𝑚 ∈ ℝ

�𝑦𝑦 = 𝑤𝑤0 + �
𝑖𝑖=1

𝑚𝑚

𝑥𝑥𝑖𝑖𝑤𝑤𝑖𝑖

Here 𝑤𝑤0, … ,𝑤𝑤𝑚𝑚 ∈ ℝ denote weights (model parameters)

• Trick: add a dummy feature 𝑥𝑥0 = 1 and use vector
notation

�𝑦𝑦 = �
𝑖𝑖=0

𝑚𝑚

𝑥𝑥𝑖𝑖𝑤𝑤𝑖𝑖 = 𝒙𝒙′𝒘𝒘

8

More general decision rule

A lowercase symbol in bold face indicates a vector; 𝒙𝒙′ denotes transpose

COMP90051 Statistical Machine Learning

Mini Summary

• Linear regression
∗ Simple, effective, “interpretable”, basis for many

approaches
∗ Decision-theoretic frequentist derivation

Next:
Frequentist derivation; Solution/training approach

9

COMP90051 Statistical Machine Learning

Linear Regression
via Frequentist Probabilistic Model

Max-Likelihood Estimation

10

COMP90051 Statistical Machine Learning

Data is noisy!

11

Example: predict mark for Statistical Machine Learning (SML)
from mark for Intro ML (IML aka KT)

IML mark

SM
L

m
ar

k

* synthetic data 🙂

Training
data*

IML mark
SM

L
m

ar
k

COMP90051 Statistical Machine Learning

• Assume a probabilistic
model: 𝑌𝑌 = 𝑿𝑿′𝒘𝒘 + 𝜀𝜀
∗ Here 𝑿𝑿,𝑌𝑌 and 𝜀𝜀 are r.v.’s
∗ Variable 𝜀𝜀 encodes noise

• Next, assume Gaussian
noise (indep. of X):
𝜀𝜀~𝒩𝒩 0,𝜎𝜎2

12

Regression as a probabilistic model

• Recall that 𝒩𝒩 𝑥𝑥;𝜇𝜇,𝜎𝜎2 ≡ 1
2𝜋𝜋𝜎𝜎2

exp − 𝑥𝑥−𝜇𝜇
2

2𝜎𝜎2

• Therefore

𝑝𝑝𝒘𝒘,𝜎𝜎2 𝑦𝑦|𝒙𝒙 =
1
2𝜋𝜋𝜎𝜎2

exp −
𝑦𝑦 − 𝒙𝒙′𝒘𝒘 2

2𝜎𝜎2

𝑥𝑥

𝑦𝑦

this is a
squared
error!

COMP90051 Statistical Machine Learning

13

Parametric probabilistic model

• Using simplified notation, discriminative
model is:

𝑝𝑝 𝑦𝑦|𝒙𝒙 =
1
2𝜋𝜋𝜎𝜎2

exp −
𝑦𝑦 − 𝒙𝒙′𝒘𝒘 2

2𝜎𝜎2

• Unknown parameters: 𝒘𝒘,𝜎𝜎2

𝑥𝑥

𝑦𝑦

• Given observed data { 𝑿𝑿1,𝑌𝑌1 , … , 𝑿𝑿𝑛𝑛,𝑌𝑌𝑛𝑛 }, we want to find
parameter values that “best” explain the data

• Maximum-likelihood estimation: choose parameter values that
maximise the probability of observed data

COMP90051 Statistical Machine Learning

Maximum likelihood estimation
• Assuming independence of data points, the probability of data is

𝑝𝑝 𝑦𝑦1, … ,𝑦𝑦𝑛𝑛|𝒙𝒙1, … ,𝒙𝒙𝑛𝑛 = �
𝑖𝑖=1

𝑛𝑛

𝑝𝑝 𝑦𝑦𝑖𝑖|𝒙𝒙𝒊𝒊

• For 𝑝𝑝 𝑦𝑦𝑖𝑖|𝒙𝒙𝑖𝑖 =
1
2𝜋𝜋𝜎𝜎2

exp − 𝑦𝑦𝑖𝑖−𝒙𝒙𝑖𝑖′𝒘𝒘
2

2𝜎𝜎2

• “Log trick”: Instead of maximising this quantity, we can maximise its
logarithm (Why? Explained soon)


𝑖𝑖=1

𝑛𝑛

log𝑝𝑝 𝑦𝑦𝑖𝑖|𝒙𝒙𝒊𝒊 = −
1

2𝜎𝜎2

𝑖𝑖=1

𝑛𝑛

𝑦𝑦𝑖𝑖 − 𝒙𝒙𝑖𝑖′𝒘𝒘 2 + 𝐶𝐶

here 𝐶𝐶 doesn’t depend on 𝒘𝒘 (it’s a constant)

• Under this model, maximising log-likelihood as a function of 𝒘𝒘 is
equivalent to minimising the sum of squared errors

14

the sum of
squared
errors!

COMP90051 Statistical Machine Learning

• Training data: { 𝒙𝒙1, 𝑦𝑦1 , … , 𝒙𝒙𝑛𝑛,𝑦𝑦𝑛𝑛 }. Note bold face in 𝒙𝒙𝑖𝑖
• For convenience, place instances in rows (so attributes go in columns),

representing training data as an 𝑛𝑛 × (𝑚𝑚 + 1) matrix 𝑿𝑿, and 𝑛𝑛 vector 𝒚𝒚

• Probabilistic model/decision rule assumes 𝒚𝒚 ≈ 𝑿𝑿𝒘𝒘

• To find 𝒘𝒘, minimise the sum of squared errors

𝐿𝐿 = �
𝑖𝑖=1

𝑛𝑛
𝑦𝑦𝑖𝑖 −�

𝑗𝑗=0

𝑚𝑚

𝑋𝑋𝑖𝑖𝑗𝑗𝑤𝑤𝑗𝑗

2

• Setting derivative to zero and solving for 𝒘𝒘 yields
�𝒘𝒘 = 𝑿𝑿′𝑿𝑿 −1𝑿𝑿′𝒚𝒚

∗ This system of equations called the normal equations
∗ System is well defined only if the inverse exists

15

Method of least squares Analytic solution:• Write derivative
• Set to zero
• Solve for model

In this slide: UPPERCASE symbol in bold face means matrix; 𝑿𝑿′ denotes transpose; 𝑿𝑿−1 denotes matrix inverse

COMP90051 Statistical Machine Learning

Wherefore art thou: Bayesian derivation?

• Later in the semester: return of linear regression

• Fully Bayesian, with a posterior:
∗ Bayesian linear regression

• Bayesian (MAP) point estimate of weight vector:
∗ Adds a penalty term to sum of squared losses
∗ Equivalent to L2 “regularisation” to be covered next week
∗ Called: ridge regression

16

COMP90051 Statistical Machine Learning

Mini Summary

• Linear regression
∗ Simple, effective, “interpretable”, basis for many

approaches
∗ Probabilistic frequentist derivation
∗ Solution by normal equations
Later in semester: Bayesian approaches

Next: Basis expansion for non-linear regression

17

COMP90051 Statistical Machine Learning

Basis Expansion

Extending the utility of models via
data transformation

18

COMP90051 Statistical Machine Learning

Basis expansion for linear regression
• Real data is likely to be non-linear

• What if we still wanted to use a
linear regression?
∗ Simple, easy to understand,

computationally efficient, etc.

• How to marry non-linear data to a
linear method?

19

𝑦𝑦

𝑥𝑥

If you can’t beat’em, join’em

COMP90051 Statistical Machine Learning

Transform the data
• The trick is to transform the data: Map data into another

features space, s.t. data is linear in that space

• Denote this transformation 𝜑𝜑:ℝ𝑚𝑚 → ℝ𝑘𝑘. If 𝒙𝒙 is the
original set of features, 𝜑𝜑 𝒙𝒙 denotes new feature set

• Example: suppose there is just one feature 𝑥𝑥, and the
data is scattered around a parabola rather than a straight
line

20

𝑦𝑦

𝑥𝑥

COMP90051 Statistical Machine Learning

Example: Polynomial regression
• No worries, mate: define

𝜑𝜑1 𝑥𝑥 = 𝑥𝑥
𝜑𝜑2 𝑥𝑥 = 𝑥𝑥2

• Next, apply linear regression to 𝜑𝜑1,𝜑𝜑2
𝑦𝑦 = 𝑤𝑤0 + 𝑤𝑤1𝜑𝜑1 𝑥𝑥 + 𝑤𝑤2𝜑𝜑2 𝑥𝑥 = 𝑤𝑤0 + 𝑤𝑤1𝑥𝑥 + 𝑤𝑤2𝑥𝑥2

and here you have quadratic regression

• More generally, obtain polynomial regression if the
new set of attributes are powers of 𝑥𝑥

• Similar idea basis of autoregression for time series
21

𝑦𝑦

𝑥𝑥

COMP90051 Statistical Machine Learning

Example: linear classification
• Example binary classification problem: Dataset not linearly separable

• Define transformation as

𝜑𝜑𝑖𝑖 𝒙𝒙 = 𝒙𝒙 − 𝒛𝒛𝑖𝑖 , where 𝒛𝒛𝑖𝑖 some pre-defined constants

• Choose 𝒛𝒛1 = 0,0 ′, 𝒛𝒛2 = 0,1 ′, 𝒛𝒛3 = 1,0 ′, 𝒛𝒛4 = 1,1 ′

22

𝑥𝑥1 𝑥𝑥2 𝑦𝑦

0 0 Class A

0 1 Class B

1 0 Class B

1 1 Class A

𝑥𝑥1

𝑥𝑥2

𝜑𝜑1 𝜑𝜑2 𝜑𝜑3 𝜑𝜑4
0 1 1 2

1 0 2 1

1 2 0 1

2 1 1 0

𝑤𝑤1 𝑤𝑤2 𝑤𝑤3 𝑤𝑤4
1 0 0 1

𝝋𝝋′𝒘𝒘 𝑦𝑦

2 Class A

2 Class B

2 Class B

2 Class A

The transformed
data is linearly

separable!

there exist weights that make
new data separable, e.g.:

COMP90051 Statistical Machine Learning

Radial basis functions
• Previous example: motivated by approximation

theory where sums of RBFs approx. functions

• A radial basis function is a function of the form
𝜑𝜑 𝒙𝒙 = 𝜓𝜓 𝒙𝒙 − 𝒛𝒛 , where 𝒛𝒛 is a constant

• Examples:

• 𝜑𝜑 𝒙𝒙 = 𝒙𝒙 − 𝒛𝒛

• 𝜑𝜑 𝒙𝒙 = exp − 1
𝜎𝜎
𝒙𝒙 − 𝒛𝒛 2

23

𝜑𝜑 𝑥𝑥

𝑥𝑥

COMP90051 Statistical Machine Learning

Challenges of basis expansion
• Basis expansion can significantly increase the utility of

methods, especially, linear methods

• In the above examples, one limitation is that the
transformation needs to be defined beforehand
∗ Need to choose the size of the new feature set
∗ If using RBFs, need to choose 𝒛𝒛𝑖𝑖

• Regarding 𝒛𝒛𝑖𝑖, one can choose uniformly spaced points, or
cluster training data and use cluster centroids

• Another popular idea is to use training data 𝒛𝒛𝑖𝑖 ≡ 𝒙𝒙𝑖𝑖
∗ E.g., 𝜑𝜑𝑖𝑖 𝒙𝒙 = 𝜓𝜓 𝒙𝒙 − 𝒙𝒙𝑖𝑖
∗ However, for large datasets, this results in a large number of

features  computational hurdle

24

COMP90051 Statistical Machine Learning

Further directions
• There are several avenues for taking the idea of basis

expansion to the next level
∗ Will be covered later in this subject

• One idea is to learn the transformation 𝜑𝜑 from data
∗ E.g., Artificial Neural Networks

• Another powerful extension is the use of the kernel trick
∗ “Kernelised” methods, e.g., kernelised perceptron

• Finally, in sparse kernel machines, training depends only
on a few data points
∗ E.g., SVM

25

COMP90051 Statistical Machine Learning

Mini Summary

• Basis expansion
∗ Extending model expressiveness via data transformation
∗ Examples for linear and logistic regression
∗ Theoretical notes

Next time:
First/second-order iteration optimisation;
Logistic regression – linear probabilistic model for
classification.

26

�Lecturer: Ben Rubinstein
This lecture
Linear Regression�via Decision Theory
Example: Predict humidity from temperature
Example: Problem statement
Example: Minimise Sum Squared Errors
Example: Analytic solution
More general decision rule
Mini Summary
Linear Regression�via Frequentist Probabilistic Model
Data is noisy!
Regression as a probabilistic model
Parametric probabilistic model
Maximum likelihood estimation
Method of least squares
Wherefore art thou: Bayesian derivation?
Mini Summary
Basis Expansion
Basis expansion for linear regression
Transform the data
Example: Polynomial regression
Example: linear classification
Radial basis functions
Challenges of basis expansion
Further directions
Mini Summary

Posted in Uncategorized

Leave a Reply

Your email address will not be published. Required fields are marked *