• 검색 결과가 없습니다.

- Section 5 -

N/A
N/A
Protected

Academic year: 2022

Share "- Section 5 -"

Copied!
73
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

Advanced Engineering Statistics - Section 5 -

Jay Liu

Dept. Chemical Engineering

PKNU

(2)

Least squares regression

• What we will cover

(3)

[FYI]Least squares vs. interpolation

Given the data, there are two choices when we want to know the value of y at x = (x

1

+ x

2

)/2

least squares? or interpolation?

Interpolation is recommended when data are subject to negligible experimental error (or noise)

Ex. In using steam tables

Otherwise, least squares is recommended.

x

y

x1 x2

x y

x1 y1

x2 y2

(4)

Least squares - usage examples

Quantify relationship between 2 variables (or 2 sets of variables):

Manager: How does yield from the lactic acid batch fermentation relate to the purity of sucrose?

Engineer: The yield can be predicted from sucrose purity with an error of plus/minus 8%

Manager: And how about the relationship between yield and glucose purity?

Engineer: Over the range of our historical data, there is no discernible relationship.

(5)

Least squares - usage examples

Two general applications

Predictive modeling – usually when an exact model form is unknown.

Modeling data trends in order to predict future y values

Simulation – usually when parameters in the model are unknown.

Getting parameter values in the known model form (e.g., calculate activation energy from reaction data)

Terminology

y : response variables, output variables, dependent variables, … x : input variables, regressor variables, independent variables, …

(6)

Review: covariance

Consider measurements from a gas cylinder: temperature (K) and pressure (kPa).

Ideal gas law applies under moderate condition: pV = nRT

Fixed volume, V = 20 × 10−3m3 = 20 L

Moles of gas, n = 14.1 mols of chlorine gas, (1 kg gas) Gas constant, R = 8.314 J/(mol.K)

Simplify the ideal gas law to: p = b

1

T, where

1

nR b  V

(7)

Review: covariance (Cont.)

(8)

Review: covariance (Cont.)

Formal definition:

1. Calculate deviation variables:

Subtracting off mean centers the vector at zero.

2. Multiply the centered values:

16740 10080 5400 1440 180 60 1620 5700 10920 15660 3. Calculate the expected value (mean): 6780

4. Covariance has units: [K∙kPa]

c.f) Covariance between temperature and humidity is 202 [K∙%]

※ Covariance with itself is the variance:

  

 

cov( , )x yE xx yy where ( )E zz and

TT pp

T T

 

p p

  

 

centeredT centered or T p

(9)

Review: correlation

Q: Which one (pressure or temperature) has stronger relationship with temperature?

Covariance depends on units: e.g. different covariance for grams vs kilograms

Correlation removes the scaling effect:

Divides by the units of x and y: dimensionless result

Gas cylinder example:

corr(temperature, pressure) = 0.997 corr(temperature, humidity) = 0.380

  

 

cov( , ) ( , )

x y x y

E x x y y

corr x y x y

   

 

 

1 corr x y( , ) xy 1

   

(10)

Review: correlation (cont.)

(11)

Review: correlation (cont.)

Remember!

?

(12)

Least squares? Least squares regression?

Regression is the act of choosing the “best” values for the unknown parameters in a model on the basis of a set of measured data.

Linear regression is the special case where the model is linear in the parameters. A straight line has the form:

There are many possible ways to define the “best” fit. However, the most commonly used measure for bestness is the sum of squared residuals.

Least sum of squares of errors  least squares in short.

Important: error is from y, not from x.

0 1

( )

yaa xe

(13)

[FYI] why minimize the sum of squares ?

The least squares model:

has the lowest possible variance for a0 and a1 when certain assumptions are met (more later)

computationally tractable by hand

easy to prove various mathematical properties intuitive: penalize deviations quadratically

Other forms: multiple solutions, unstable, high variance solutions,

mathematical proofs are difficult

(14)

Least squares (regression)

It is the basis for :

DOE (Design of Experiments) Latent variable methods

We consider only 2 (sets of) variables : x and y (or x’s and y)

Simple least squares Multiple least squares Generalized least squares

(15)

Simple least squares

Wind tunnel example

How can we find the best line that describe the following data?

Data from wind tunnel experiments:

Drag force (F) at various wind velocities

(16)

Wind tunnel example (cont.)

From the plot, a linear line seems adequate.

y = a0 + a1x

At a data point (x

i

, y

i

), error between the line and the point is: (see the figure on the right)

ei = yi – = yia0 – a1xi

Earlier, least squares means least sum of squares of errors. For all data points, sum of squares of errors is:

We need to find model parameters a

0

and a

1

that minimize S

r

.

“Least squares”

ˆi y

(17)

Wind tunnel example (cont.)

How to find model parameters?

Take a look at Sr.

Sr is a parabolic function w.r.t ao and a1 and sign of are plus.

Sr becomes minimum where

a0 a1

2 2

and 1

ao a

0 1

0 & 0.

r r

S S

a a

   

 

Rearranging and solving for a0 and a1

(18)

Wind tunnel example (cont.)

Calculations

(19)

Wind tunnel example (cont.)

Calculations

This is called simple least squares.

(20)

Wind tunnel example (cont.)

Results

Is this OK with you?

(21)

General modeling procedure

Define modeling objective

Variable selection

Identify the response variables (i.e., y variables), and the regressor variables (i.e., x variables) that are to be

considered

Design of experiment

Design an experiment and use it to generate the data that will be used to fit the model

Define the model

Choose an appropriate form for the model

Fit the model

Estimate values for the parameters in the model

Does the model fit?

N

Y Use the model

Statistical tools + prior knowledge

(22)

Simple least squares

Summary

Model form: y = a0 + a1x + e

becomes minimizes where

Rearranging and solving for a0 and a1 0 1

0 & 0.

r r

S S

a a

   

 

(23)

Simple least squares (cont.)

Properties

ˆ

. ., i i 0 i e

yy

a0 a1

(24)

Simple least squares (cont.)

Questions

what if our model we want to find is non-linear?

Ex. Activation energy in rate constant

 Linearize !

0

ERT

kk e a1

a1 a0

a1 a0 a1

(25)

Linearization

Want to model non-linear relationships between independent (x) and dependent (y) variables.

1. Make a simple linear model through a suitable transformation.

y = f(x) + e  y = a0 + a1x + e 2. Use previous results (simple least squares)

※ Caution: nonlinear transformation also changes P.D.F of variables (and errors)

We will discuss about this in model assessment.

(26)

Linearization (Cont.)

(27)

Polynomial regression

For quadratic form

Sum of squares

Again, Sr has a parabolic shape w.r.t a0, a1, and a2. with plus signs of a02, a12, and .a22

2

0 1 2

0

2

0 1 2

1

2 2

0 1 2

2

2 ( ) 0

2 ( ) 0

2 ( ) 0

r

i i i

r

i i i i

r

i i i i

S y a a x a x

a

S x y a a x a x

a

S x y a a x a x

a

      

      

      

(28)

Polynomial regression (Cont.)

Rearranging the previous equations gives

the above equations can be solved easily. (three unknowns and three equations.)

For general polynomials

From the results of two cases (y = a0 + a1x & y = a0 + a1x + a2x2)

0 1

r r r 0

m

S S S

a a a

      

  

2

0

2 3

1

2 3 4 2

2

i i i

i i i i i

i i i i i

n x x a y

x x x a x y

x x x a x y

    

     

    

   

    

 

  

   

   

(29)

Multiple least squares

Consider when there are more than two independent variables, x

1

, x

2

,

…, x

m

.  regression plane.

For 2-D case, y = a

0

+ a

1

x

1

+ a

2

x

2

.

Again, Sr has a parabolic shape w.r.t a0, a1. e x

a x

a x a a

y01 12 2

m m

(

i 0 1 1,i 2 2,i

)

2

r y a a x a x

S

0 1 1, 2 2,

0

1, 0 1 1, 2 2,

1

2, 0 1 1, 2 2,

2

2 ( ) 0

2 ( ) 0

2 ( ) 0

r

i i i

r

i i i i

r

i i i i

S y a a x a x

a

S x y a a x a x

a

S x y a a x a x

a

      

      

      

a1 a2

(30)

Multiple least squares (Cont.)

Rearranging and solve for a0, a1 and a2 gives

For an m-dimensional plane,

Same as in general polynomials,

we need to solve (m+1) linear algebraic equations for (m+1) parameters.

e x

a x

a x a a

y01 12 2

m m

0 1

r r r 0

m

S S S

a a a

  

   

  

(31)

General least squares

The following form includes all cases (simple least squares, polynomial regression, multiple regression)

Ex. Simple and multiple least squares

polynomial regression

Same as before,

we need to solve (m+1) linear algebraic equations for (m+1) parameters.

0 1

r r r 0

m

S S S

a a a

      

  

(32)

Quantification of errors

 

y y

2

S

t i

 

2 , ,

1 1 ,

0 0 2

i m m i

i i

i r

z a z

a z

a y

e S

Total sum of squares around the mean for the response variable, y

Sum of squares of residuals around the regression line

(33)

Quantification of errors (Cont.)

 

1 1

1

2

 

 

n

y y nS

Sy i t

) 1 (

 

m n

S Sr

x y

Standard error of predicted y (SE)

 quantify appropriateness of regression

Standard deviation of y

(34)

Quantification of errors (Cont.)

Coefficients of determination, R

2

t r t

S S

R S

2 The amount of variability in the data explained by the regression model.

R2 = 1 when Sr = 0 : perfect fit (a regression curve passes through data points) R2 = 0 when Sr = St : as bad as doing nothing

It is evident from the figures that a parabola is adequate.

(35)

Quantification of errors (Cont.)

Warning! : R

2

≈ 1 does not guarantee that the model is adequate, nor the model will predict new data well.

It is possible to force R2 to be one by adding as many terms as there are observations.

Sr can be big when variance of random error is large.

(Usual assumption on error is that error is random is unpredictable)

Practice using Excel

(1) Wind tunnel example with higher polynomials (2) Simple regression with increasing random noise

(36)

Confidence intervals - coefficients

Coefficients in the regression model have confidence interval.

Why? They are also statistics like & s. That is, they are numerical quantities calculated in a sample (not entire population). They are estimated values of parameters.

statistic

statistic   A

Statistic that we want to find its confidence interval

Standard error of the statistic Value that depends on P.D.F of the statistic & confidence level a

x

statistic A statistic

za/2 tn,a/2

x x n

x sx n

The standard error of a statistic is the standard deviation of the sampling distribution

(37)

Confidence intervals – coefficients (cont.)

Matrix representation of GLS

e Za y  

Z

T

y

a

T

e

T

m+1: number of coefficients n: number of data points

(38)

Confidence intervals – coefficients (Cont.)

Example

Fitting quadratic polynomials to five data points

x y

e x

a x a a

y012 2

e Za y  









































5 4 3 2 1

2 1 0

0 . 1 0

. 1 1

25 . 0 5

. 0 1

0 . 0 0

. 0 1

25 . 0 5 . 0 1

0 . 1 0 . 1 1

0 . 2

5 . 0

0 . 0

5 . 0

0 . 1

e e e e e

a a

a Three unknowns

Five equations

(39)

Confidence intervals – coefficients (Cont.)

Solutions

1. LU decomposition or other methods to solve L.A.E

2. Matrix inversion

computationally not efficient, but statistically useful

e Za y  

y Za

 

y Za

e

e   

i T T

r e

S 2

Sum of squares of errors

0

a

Sr

 

ZTZ a ZTy

Called “normal equations”

 

ZTZ a ZTy

"

Ax b

"

 

ZTZ a ZTy a

 

ZTZ 1ZTy

(40)

Confidence intervals – coefficients (Cont.)

Matrix inversion approach

Denote as the diagonal element of Confidence interval of estimated coefficients

 

Z Z Z y

aT 1 T

 

ZTZ 1

1

Zii

2 1

1 ( 1), /2

i n m y ii

x

a

t

  a

S Z

( 1), /2

n m

t

  a

) 1 (

 

m n

S Sr

x y

Student t statistics

Standard error of estimate

(41)

Confidence intervals – coefficients (Cont.)

For a linear model, C.I. for a1(slope)

C.I. for a0 (intercept)

What if confidence intervals contain zero?

 

2

0 ( 1), /2 / 2

1

n m y x

i

a t S x

n x x

a

 

 

 

1 ( 1), /2 / 2

1

n m y x

i

a t S

x x

a

 

 

(42)

Confidence intervals – prediction

C.I for predicted y,

 

 

0

2 0

( 1), / 2 / 2

ˆ

n m y x

1

x

i

x x

y t S

n x x

a

  

 

ˆ

i

y

(43)

Model assessment

When we do not know the model form, we have to assess the model before use it after we fit a regression model.

However, in order to assess the model and make inferences about the parameters and predictions from the model, we will have to employ

statistics and make some assumptions about the nature of the disturbance.

Tools for model assessment

Sy/x, R2 (quantitative) ( Do not use) Residual Plots

(qualitative)

Normal probability chart

(qualitative or quantitative)

Test for lack of fit

(quantitative)

This is used when the dataset includes replicates. It is based on analysis of variance (ANOVA).

(44)

Model assessment - assumptions

What is the most desirable errors in regression ?

Assumptions on error

Error is additive

The variance of the error is constant and is not related to values of the response or values of the regressor variables.

There is no error associated with the values of the regressor variables.

Error is a random variable with Gaussian distribution N(0,2) (2 usually unknown)

unpredictable

e x a a

y01 1y (a0a1x1)e

(45)

Model assessment – residual plots

Recall the assumptions on error

Error is not related to the values of response or regressor variables.

Then, assumptions will not be valid if the model is wrong.

Following residual plots will reveal this.

Residuals vs. regressor variables Residuals vs. fitted y values ( )

Residuals vs. “lurking” variables (i.e. time or order)

 These plots will show “some patterns” when a model is inadequate.

ˆi y

(46)

Model assessment – residual plots (con’t) Examples

x

e

x

e

(47)

Model assessment – residual plots (con’t)

Examples of residual plots

(48)

Model assessment – normal probability plot

Recall the assumptions on error

Error is a random variable with Gaussian distribution N(0, 2 ) (2 usually unknown)

Then, errors will fall onto a straight line (y = x) in a normal probability plot. (especially useful when the number of data points is large)

Normal probability plot

Alternatively, normality test

can be used.

(49)

Model assessment – ANOVA (Test for lack of fit) The variance breakdown

Sr St

 

2

t i

S   yy

ˆ

2

r i i

S   yy

 

2

Reg

ˆ

i

S   yy

Se

Really? Prove to yourself

Re

t g r

S S S

  

(50)

Model assessment – ANOVA (Test for lack of fit) The variance breakdown

Ratio of SReg/Sr follows F distribution when corrected with degree of freedom.

If regression is not meaningful, the ratio (Se/Sr) is small and St ≒ Sr.

Re

t g r

SSS

Sr St

 

2

t i

S   yy

ˆ

2

r i i

S   yy

 

2

Reg

ˆ

i

S   yy

SReg

(51)

Model assessment – ANOVA (Test for lack of fit)

S

Re g

S

r

S

t

a1 a0

(52)

Model assessment – ANOVA (Test for lack of fit)

ANOVA Table

Source of Var. Sum of Squares

Degrees of Freedom

Mean Square F0

Regression SReg p MSReg=SReg/p MSReg/ MSE

Residual error Sr n-p MSE=Sr/(n-p)

Total St n-1

Compare F0 to the critical value Fp,n-p;a

What we are doing is a test of hypothesis.

We are testing the hypothesis:

H0 : b0 bp 0

H1 : at least one parameter is not equal to zero.

(53)

[FYI]Meaning of a p-value in hypothesis test

A measure of how much evidence we have against the null hypothesis.

Null hypothesis (H0) represents the hypothesis of no change or no effect.

Much research involves making a hypothesis and then collecting data to test that hypothesis. Then researchers will collect data and measure the consistency of this data with the null hypothesis.

A small p-value is evidence against the null hypothesis while a large p-value means little or no evidence against the null hypothesis.

Traditionally, researchers will reject a null hypothesis if the p-value is less than 0.05 (a = 0.05).

 p-value can mean that the possibility that you can be wrong when rejecting the null hypothesis.

(54)

Integer variables in the model

Integer variables 0 and 1 can represent qualitative variables.

Example: raw material from Spain, India, or Vietnam

y = a0 + a1x1 + . . . + akxk + r1d1 + r2d2 + r3d3 d1 = 1 and d2 = 0 and d3 = 0 for Spain

d1 = 0 and d2 = 1 and d3 = 0 for India d1 = 0 and d2 = 0 and d3 = 1 for Vietnam

Often called indicator variables for this reason

(55)

Integer variables in the model

Example

Want to predict yield when two different impeller used. Yield = f(temperature, impeller type)

Build two different models (one for axial, one for radial)

Build one model using indicator variable. y = a0 + a1T + rd y = a0 + a1T + rdi

di = 0 for axial, di = 1 for radial

axial radial

(56)

Leverage effect

Unusual observations influence the model parameters and our interpretation

To avoid the leverage effect,

Remove outliers before regression (but do not delete without investigation) Use different Sr (no longer least squares)

x

y

x

y

Outliers have an over-proportional effect on resulting regression curves.

(57)

Causal relation and correlation

Causal relation

Cause and effect relation

Has physical/chemical/engineering meanings x and y are not interchangeable

Direction exists.

Correlation

(Linear) relationship between two variables No physical/chemical/engineering meanings.

Average height of 20’s men vs. year x and y are interchangeable

(58)

Advanced topics

Testing of least-squares models

(59)

Advanced topics - Testing of least-squares models

(60)

Advanced topics

Correlated x’s

MLR solution

When two or more x’s are correlated, becomes nearly singular, i.e., ill-conditioned.

 

y Za e

Sr

ei2e eT

yZa

 

T yZa

r

0

S

a 

 

Z Z aT Z yT

 

T 1 T

 a Z Z Z y

 

Z ZT 1

(61)

Advanced topics – correlated x’s High/no correlation between x

1

and x

2

What if very small (measurement ) noises added to x’s

What will happen to your MLR model?

2012-05-16 Adv. Eng. Stat., Jay Liu© 60

 

1

1.0000 0.9999 0.9999 1.0000

5000.25 4999.75 4999.75 5000.25

T

T

 

  

 

  

  

Z Z

Z Z

 

1

1.0000 0.0 0.0 1.0000 1.0 0.0

0.0 1.0

T

T

 

  

 

 

  

 

Z Z

Z Z

 

1

1.0001 0.9999 0.9999 1.0000

3333.34 3333.11 3333.11 3333.34

T

T

 

  

 

  

  

Z Z

Z Z

 

1

1.0000 0.0 0.0 1.0001 0.9999 0.0

0.0 0.9999

T

T

 

  

 

 

  

 

Z Z

Z Z

(62)

Advanced topics – correlated x’s

If high correlation among x’s:

unstable solutions for a predictions uncertain also

Geometrically speaking

(63)

Advanced topics – correlated x’s

Remedies?

Use selected x variables  stepwise regression Use ridge regression

Use multivariate methods (will not be covered in this lecture)

(64)

Advanced topics

What we want to know:

How do we select the form of the model? Which variables should be

included? Should we include transformations of the regressor variables?

….

What we want:

We would like to build the “best” regression model

We would like to include as many regressor variables as is necessary to adequately describe the behaviour of y. At the same time, we want to keep the model as simple as possible.

 Stepwise regression start off by choosing an equation having the single best x variables and the attempts to build up with subsequent additions of x’s one at a time as long as these additions are worthwhile.

(65)

Advanced topics – stepwise regression

Procedure

1. Add a x variable to the model (the variable that is most highly correlated with y).

2. Check to see whether or not this has significantly improved the model. One way is to see whether or not the confidence interval for the parameter

includes zero. (of course you can use hypothesis test) If the new term or terms are not significant, remove them from the model.

3. Find one of the remaining x variables that is highly correlated with the residuals and repeat the procedure.

(66)

Advanced topics

Ridge regression (Hoerl, 1962; Hoerl & Kennard, 1970a,b)

A modified regression method specifically for ill-conditioned datasets that allows all variables to be kept in the model.

This is possible by adding additional information to the problem to remove the ill-conditioning.

The objective function to minimize:

Least squares estimates have no bias but large variance, while ridge regression estimates have small bias and small variance.

y Xb

 

T y Xb

kb bT

(67)

Advanced topics – ridge regression

Procedure

1. Mean center and scale all x’s to unit variance

2. Rewrite the model as:

i

i i

sx

xx f

1

1

1 1

1

p

p

p

x p x

x x

p p

a s a s

s s

b b

 

     

      

   

 

1

x x x x

y y

y f f

Y Fb ε

(68)

Advanced topics – ridge regression

3. The objective function to use for the optimization is to minimize

Therefore,

4. Solve the optimization problem in Step 3 for several values of k between 0 and 1 and choose that value of k at which the estimates of b see to stabilize.

Otherwise, choose k by validation on new data.

  

T

JYFb Y Fb kb b

 

1

* Tk T

b F FZ I F Y

(69)

Advanced topics

Non-linear regression

General form of a non-linear regression model

In a linear model, . In a non-linear model, f() woulds have any form. E.g.,

Remember that nonlinear transformation also changes P.D.F of variables (and errors)? What does this mean?

( , ) yf x a 

( , ) T f x ax a

b 

b

  x 1

y e

2

1x

(70)

Advanced topics – non-linear regression

The approach is exactly the same as for linear models We use the same objective function:

All we need is to minimize S over a. but how?

 

2 1

2 1

ˆ

( )

( , )

r n

i i

i n

i i

i

S

y y

y f

 

 

 

ε ε

x a

(71)

Advanced topics – non-linear regression

The big difference between linear and nonlinear regression is that in general, the optimization problem for a nonlinear model does not have an exact analytical solution.

Therefore, we have to use a numerical optimization algorithm such as:

Gauss-Newton Steepest Descent

Conjugated Gradients

Any other optimization algorithm

(72)

Advanced topics – non-linear regression

When using an optimization algorithm to solve nonlinear regression problems, one needs to be able to specify:

1. an expectation function (i.e. the form of the model) 2. Data

3. starting guesses for a 4. stopping criteria

5. possibly other “tuning” parameters associated with the optimization algorithm

(73)

Advanced topics – non-linear regression

Problems with Numerical Optimization

Failure to converge

Finding only a local minimum and not the global minimum Requires good starting guesses for the parameters

Can be sensitive to the choice of convergence criteria and other “tuning parameters” of the algorithm

Sometimes requires specification of the derivatives of the model with respect the the parameters.

참조

관련 문서

Solving the PDEs using separating variables and Fourier series and interpreting the solution... Solution

Note that the number of adjacent cells is the same as the number of input variables since it is equal to the number of bits... If we don’t use DC terms, the logic function f

local stationary behavior in the case of nonholonomic subsidiary conditions become identical with those for holonomic constraints.. The cable is of uniform

First, Among the variations in the variables by socio-demographic characteristics, the leader mentoring showed statistically significant differences in

Fourth, self-development factor of participation motive had significant influences on all variables of leisure satisfaction including psychology,

- Basic research on a compact, high b ST (Spherical Torus) with elongated chamber in partial solenoid configuration - Study on advanced tokamak scenario

Purpose Model predictive controller for simulating closed-loop systems with hard bounds on manipulated variables and/or controlled variables using linear models in the MPC

• Given certain conditions, the arithmetic mean (G) of a sufficiently large number of iterates (g i (x), i=1, ..n) of independent random variables, each with a well-defined