1 / 25

Design and Analysis of Experiments (5) Fitting Regression Models

Design and Analysis of Experiments (5) Fitting Regression Models. Kyung-Ho Park. In many problems one or more variables are related, it is of interest to model and explore this relationship. The model can be used for prediction, process optimization, or process control. Modelling.

umay
Télécharger la présentation

Design and Analysis of Experiments (5) Fitting Regression Models

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Design and Analysis of Experiments (5) Fitting Regression Models Kyung-Ho Park

  2. In many problems one or more variables are related, it is of interest to model and explore this relationship. • The model can be used for prediction, process optimization, or process control.

  3. Modelling

  4. Interpolation Method • interpolation is a method of constructing new data points within the range of a discrete set of known data points. • In engineering and science one often has a number of data points, as obtained by sampling or experiment, and tries to construct a function which closely fits those data points. • This is called curve fitting or regression analysis. Interpolation is a specific case of curve fitting, in which the function must go exactly through the data points.

  5. Piecewise constant interpolation The simplest interpolation method is to locate the nearest data value, and assign the same value. In one dimension, there are seldom good reasons to choose this one over linear interpolation, which is almost as cheap, but in higher dimensions, in multivariate interpolation, this can be a favourable choice for its speed and simplicity

  6. Linear interpolation One of the simplest methods is linear interpolation (sometimes known as lerp). Consider the above example of determining f(2.5). Since 2.5 is midway between 2 and 3, it is reasonable to take f(2.5) midway between f(2) = 0.9093 and f(3) = 0.1411, which yields 0.5252. Generally, linear interpolation takes two data points, say (xa,ya) and (xb,yb). Linear interpolation is quick and easy, but it is not very precise. Another disadvantage is that the interpolant is not differentiable at the point xk.

  7. Polynomial interpolation • Polynomial interpolation is a generalization of linear interpolation. Note that the linear interpolant is a linear function. We now replace this interpolant by a polynomial of higher degree. • Consider again the problem given above. The following sixth degree polynomial goes through all the seven points: • f(x) = − 0.0001521x6 − 0.003130x5 + 0.07321x4 − 0.3577x3 + 0.2255x2 + 0.9038x. • Substituting x = 2.5, we find that f(2.5) = 0.5965.

  8. Spline interpolation Remember that linear interpolation uses a linear function for each of intervals [xk,xk+1]. Spline interpolation uses low-degree polynomials in each of the intervals, and chooses the polynomial pieces such that they fit smoothly together. The resulting function is called a spline.

  9. Extrapolation Method • A sound choice of which extrapolation method to apply relies on a prior knowledge of the process that created the existing data points. • Crucial questions are for example if the data can be assumed to be continuous, smooth, possibly periodic etc.

  10. Linear extrapolation • Polynomial extrapolation • Conic extrapolation • A conic section can be created using five points near the end of the known data. If the conic section created is an ellipse or circle, it will loop back and rejoin itself. A parabolic or hyperbolic curve will not rejoin itself, but may curve back relative to the X-axis. This type of extrapolation could be done with a conic sections template (on paper) or with a computer.

  11. Simple Linear Regression and Correlation (Empirical Models) Example 6-1 y : the purity of oxygen produced in a chemical distillation process x : the percentage of hydrocarbons that are present in the main condenser of the distillation unit

  12. Simple Linear Regression and Correlation Stat > Regression > Fitted Line Plot

  13. Simple Linear Regression and Correlation (Empirical Models) Example 6-1

  14. Simple Linear Regression and Correlation (Empirical Models) Example 6-1 Regression Analysis: Purity y(%) versus Hydrocarbon level x(%) The regression equation is Purity y(%) = 74.28 + 14.95 Hydrocarbon level x(%) S = 1.08653 R-Sq = 87.7% R-Sq(adj) = 87.1% Analysis of Variance Source DF SS MS F P Regression 1 152.127 152.127 128.86 0.000 Error 18 21.250 1.181 Total 19 173.377

  15. Simple Linear Regression and Correlation (Empirical Models) Example 6-1

  16. Simple Linear Regression and Correlation (Empirical Models) Example 6-1 Predicted Values for New Observations New Obs Fit SE Fit 95% CI 95% PI 1 89.081 0.364 (88.316, 89.846) (86.674, 91.489) 2 89.530 0.336 (88.824, 90.235) (87.141, 91.919) 3 91.473 0.250 (90.947, 91.999) (89.130, 93.815) 4 93.566 0.273 (92.993, 94.138) (91.212, 95.919) 5 96.107 0.424 (95.216, 96.998) (93.656, 98.557) 6 94.612 0.325 (93.929, 95.295) (92.229, 96.995) 7 87.288 0.493 (86.251, 88.324) (84.781, 89.795) 8 92.669 0.247 (92.150, 93.188) (90.328, 95.010) 9 97.452 0.526 (96.348, 98.556) (94.916, 99.988) 10 95.210 0.362 (94.449, 95.971) (92.804, 97.616) 11 92.071 0.243 (91.560, 92.582) (89.732, 94.410) 12 91.473 0.250 (90.947, 91.999) (89.130, 93.815) 13 88.932 0.374 (88.146, 89.718) (86.518, 91.346) 14 89.380 0.345 (88.655, 90.105) (86.985, 91.775) 15 90.875 0.268 (90.312, 91.438) (88.524, 93.226) 16 92.220 0.243 (91.710, 92.731) (89.881, 94.559) 17 93.117 0.257 (92.577, 93.657) (90.771, 95.463) 18 94.014 0.293 (93.399, 94.629) (91.650, 96.378) 19 95.658 0.392 (94.834, 96.483) (93.231, 98.085) 20 88.483 0.405 (87.633, 89.334) (86.047, 90.919)

  17. Simple Linear Regression and Correlation (Empirical Models) Example 6-1 Hydrocarbon Purity Obs level x(%) y(%) Fit SE Fit Residual St Resid 1 0.99 90.010 89.081 0.364 0.929 0.91 2 1.02 89.050 89.530 0.336 -0.480 -0.46 3 1.15 91.430 91.473 0.250 -0.043 -0.04 4 1.29 93.740 93.566 0.273 0.174 0.17 5 1.46 96.730 96.107 0.424 0.623 0.62 6 1.36 94.450 94.612 0.325 -0.162 -0.16 7 0.87 87.590 87.288 0.493 0.302 0.31 8 1.23 91.770 92.669 0.247 -0.899 -0.85 9 1.55 99.420 97.452 0.526 1.968 2.07R 10 1.40 93.650 95.210 0.362 -1.560 -1.52 11 1.19 93.540 92.071 0.243 1.469 1.39 12 1.15 92.520 91.473 0.250 1.047 0.99 13 0.98 90.560 88.932 0.374 1.628 1.60 14 1.01 89.540 89.380 0.345 0.160 0.16 15 1.11 89.850 90.875 0.268 -1.025 -0.97 16 1.20 90.390 92.220 0.243 -1.830 -1.73 17 1.26 93.250 93.117 0.257 0.133 0.13 18 1.32 93.410 94.014 0.293 -0.604 -0.58 19 1.43 94.980 95.658 0.392 -0.678 -0.67 20 0.95 87.330 88.483 0.405 -1.153 -1.14

  18. Simple Linear Regression and Correlation (Empirical Models) Example 6-1

  19. Simple Linear Regression and Correlation (Empirical Models) Example 6-1

  20. Multiple Linear Regression Model

  21. Multiple Linear Regression Model Stat > Regression > Regression

  22. Multiple Linear Regression Model Regression Analysis: Viscosity versus Temp, Rate The regression equation is Viscosity = 1566 + 7.62 Temp + 8.58 Rate Predictor Coef SE Coef T P Constant 1566.08 61.59 25.43 0.000 Temp 7.6213 0.6184 12.32 0.000 Rate 8.585 2.439 3.52 0.004 S = 16.3586 R-Sq = 92.7% R-Sq(adj) = 91.6% Analysis of Variance Source DF SS MS F P Regression 2 44157 22079 82.50 0.000 Residual Error 13 3479 268 Total 15 47636

  23. Multiple Linear Regression Model

  24. Multiple Linear Regression Model

More Related