SlideShare uma empresa Scribd logo
1 de 20
SIMPLE LINEAR REGRESSION Reporters:  	Atty. Gener R. Gayam, CPA 	Agapito “pete” M. Cagampang, PM 	Raymond B. Cabling, MD
SIMPLE LINEAR REGRESSION The Scatter Diagram 	In solving problems that concern estimation and forecasting, a scatter diagram can be used as a graphical approach. This technique consists of joining the points corresponding to the paired scores of dependent and independent variables which are commonly represented by X and Y on the X – y coordinate system. 	Below is an illustration of a scatter diagram using the data in Table 6.1. This table shows the data about the six years working experience and the income of eight employees in a big industrial corporation.
Table 6.1
Figure 6.1 – A Scatter Diagram for Table 6.1 Data X X X X X X X X
For you to roughly predict the value of a dependent variable, such as years of working experience, from the dependent variable, which is income, your next step is to draw a trend line. This is a line passing through the series of points such that the total vertical measurement of the points below this line is more or less equal to the total measurements of the points  above the line. If these requirements are satisfied, you draw a correct trend Y. The illustration is shown in figure 6.1
Figure 6.2 - A trend line drawn on the linear direction between working experience and income of eight employees Trend Line
	Using the trend line draw in Figure 6.1 above, the value estimated for Y when X is 16, is 18. You should not fail to remember that if a “straight line” appears to describe the relationship, the algebraic approach called the regression formula can be used as explained in the next topic.
a = Ῡ - bX B. The Least Square Linear Regression Equation The least square linear regression equation can be understood through this formula known from algebra. 			Y = a + bx 	For instance the Y = a+bx in figure 6.1 in that line that gives the smallest sum of the squares of the vertical measurements or distance of the points from the line. 	In solving the regression equations, you need to solve first,
ΣX = 62  ΣY = 90 X = 7.75 Y = 11.25 Example: Solve the least squares regression line for the data scores in Table 6.1.
Solution:
After solving the values of b and a, your regression equation obtained from Table 6.1 is.
Now, we are interested in the distance of the Y values from Y₁ the corresponding ordinate of the regression line. Here, we are going to base our measure of dispersion or variation around the regression line on the distance (Y₁ ‒ Y)². This can be well understood by this standard error of estimate formula given below. Se =     Σ(Yi ‒ Ŷ)² 	 	              n ‒ 2 √ C. The standard Error of Estimate
However, this formula entails a very tedious process of computing the standard error of estimate, so that the formula by Basil P. Korin (1977), which is easier to solve suggested as follows: 		Se =        ΣYi² ‒ a(Yi) ‒ b(Xi ‒ Yi) 		    n ‒ 2 	Note: 	The symbol a and b stand for the intercept and the slope of the regression line. √
√ Example: 	Solve the standard error of estimate for the regression line which was derived from the data in Table 6.1. 		Se =      Σ(Yi ‒ Ŷ)² 	 	                  n ‒ 2
Step 1 – Compute the value of Y at each of the X values. 	Example: 			Y = 6.68 + .59 (2) 			   = 6.68 + 1.18 			   = 7.68 Do the rest by following the same procedure. Step 2 – Get the difference between (Yi ‒ Ŷ). 	Example: 			8 – 7.86 = .14 Step 3 – Square all the difference Yi ‒ Ŷ. 	Example: 			(.14)² = .0196
√ √ √ √ Step 4 – Apply the formula. 	Se =      Σ(Yi ‒ Ŷ)² 	 	      n ‒ 2 	     =       16.692 		     8 – 2 	     = 	  16.692 		       6 	     =	  2.782 	     =  1.67
Solution 2: 		Se =        ΣYi² ‒ a(Yi) ‒ b(Xi ‒ Yi) 		       n ‒ 2
√ √ √ √ √ Step 1 – Square Y₁  	Example: 		(8²) = 64 Step 2 – Multiply XiYi 	Example: 		2 X 8 = 16 Step 3 – Get the sum of Yi² and XiYi Step 4 – Apply the formula 			=      1084 – 6.68 (90) – .59 (791) 				            n – 2 			=      1084 – 601.2 – 466.69 				          8 – 2 			=      1084 – 1067.89 			                   8 – 2 			=      16.11 			            6 			=      2.685 			=   1.64
The standard error of estimate is interpreted as the standard deviation. For example, if we measure vertically three standard errors from the regression line above and below, we will find that the same value of X will always fall between the upper and lower 3Se Limits. 	In the example above of the standard error of estimate which is 1.64 you will come up with 4.92 units (3) (1.64) above and below the regression line. This means that these “bounds” of 4.92 unit above and below the regression line pertain to all observations taken for that particular sample. If you draw two parallel lines, each of them lying one Se from the regression line, you will expect two thirds of the observations falling between these bounds. See Figure 6.1 for the illustration of the data in Table 6.1.
Y = 6.68 + .59 X Figure 6.3 – A regression Line with One Standard Error Distance

Mais conteúdo relacionado

Mais procurados

Simple lin regress_inference
Simple lin regress_inferenceSimple lin regress_inference
Simple lin regress_inference
Kemal İnciroğlu
 
Least square method
Least square methodLeast square method
Least square method
Somya Bagai
 
Chapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares RegressionChapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares Regression
nszakir
 

Mais procurados (20)

Linear regression
Linear regressionLinear regression
Linear regression
 
Lesson 6 coefficient of determination
Lesson 6   coefficient of determinationLesson 6   coefficient of determination
Lesson 6 coefficient of determination
 
An Overview of Simple Linear Regression
An Overview of Simple Linear RegressionAn Overview of Simple Linear Regression
An Overview of Simple Linear Regression
 
Curve Fitting
Curve FittingCurve Fitting
Curve Fitting
 
Curve fitting
Curve fitting Curve fitting
Curve fitting
 
Ordinary least squares linear regression
Ordinary least squares linear regressionOrdinary least squares linear regression
Ordinary least squares linear regression
 
Basics of Regression analysis
 Basics of Regression analysis Basics of Regression analysis
Basics of Regression analysis
 
Simple lin regress_inference
Simple lin regress_inferenceSimple lin regress_inference
Simple lin regress_inference
 
04 regression
04 regression04 regression
04 regression
 
Presentation On Regression
Presentation On RegressionPresentation On Regression
Presentation On Regression
 
Introduction to Regression Analysis
Introduction to Regression AnalysisIntroduction to Regression Analysis
Introduction to Regression Analysis
 
Non Linear Equation
Non Linear EquationNon Linear Equation
Non Linear Equation
 
Non linear curve fitting
Non linear curve fitting Non linear curve fitting
Non linear curve fitting
 
Least square method
Least square methodLeast square method
Least square method
 
Presentation on Regression Analysis
Presentation on Regression AnalysisPresentation on Regression Analysis
Presentation on Regression Analysis
 
Curve fitting
Curve fittingCurve fitting
Curve fitting
 
Curve fitting
Curve fittingCurve fitting
Curve fitting
 
Applied numerical methods lec8
Applied numerical methods lec8Applied numerical methods lec8
Applied numerical methods lec8
 
Chapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares RegressionChapter 2 part3-Least-Squares Regression
Chapter 2 part3-Least-Squares Regression
 
Data Approximation in Mathematical Modelling Regression Analysis and Curve Fi...
Data Approximation in Mathematical Modelling Regression Analysis and Curve Fi...Data Approximation in Mathematical Modelling Regression Analysis and Curve Fi...
Data Approximation in Mathematical Modelling Regression Analysis and Curve Fi...
 

Semelhante a Simple linear regression

Correlation by Neeraj Bhandari ( Surkhet.Nepal )
Correlation by Neeraj Bhandari ( Surkhet.Nepal )Correlation by Neeraj Bhandari ( Surkhet.Nepal )
Correlation by Neeraj Bhandari ( Surkhet.Nepal )
Neeraj Bhandari
 

Semelhante a Simple linear regression (20)

Regression
RegressionRegression
Regression
 
ML Module 3.pdf
ML Module 3.pdfML Module 3.pdf
ML Module 3.pdf
 
Corr-and-Regress (1).ppt
Corr-and-Regress (1).pptCorr-and-Regress (1).ppt
Corr-and-Regress (1).ppt
 
Corr-and-Regress.ppt
Corr-and-Regress.pptCorr-and-Regress.ppt
Corr-and-Regress.ppt
 
Cr-and-Regress.ppt
Cr-and-Regress.pptCr-and-Regress.ppt
Cr-and-Regress.ppt
 
Corr-and-Regress.ppt
Corr-and-Regress.pptCorr-and-Regress.ppt
Corr-and-Regress.ppt
 
Corr-and-Regress.ppt
Corr-and-Regress.pptCorr-and-Regress.ppt
Corr-and-Regress.ppt
 
Corr-and-Regress.ppt
Corr-and-Regress.pptCorr-and-Regress.ppt
Corr-and-Regress.ppt
 
Correlation & Regression for Statistics Social Science
Correlation & Regression for Statistics Social ScienceCorrelation & Regression for Statistics Social Science
Correlation & Regression for Statistics Social Science
 
Bba 3274 qm week 6 part 1 regression models
Bba 3274 qm week 6 part 1 regression modelsBba 3274 qm week 6 part 1 regression models
Bba 3274 qm week 6 part 1 regression models
 
Corr And Regress
Corr And RegressCorr And Regress
Corr And Regress
 
Regression
Regression  Regression
Regression
 
Regression Ayalysis (2).ppt
Regression Ayalysis (2).pptRegression Ayalysis (2).ppt
Regression Ayalysis (2).ppt
 
Correlation by Neeraj Bhandari ( Surkhet.Nepal )
Correlation by Neeraj Bhandari ( Surkhet.Nepal )Correlation by Neeraj Bhandari ( Surkhet.Nepal )
Correlation by Neeraj Bhandari ( Surkhet.Nepal )
 
Simple linear regression
Simple linear regressionSimple linear regression
Simple linear regression
 
Regression
RegressionRegression
Regression
 
Rsh qam11 ch04 ge
Rsh qam11 ch04 geRsh qam11 ch04 ge
Rsh qam11 ch04 ge
 
Lecture 07 graphing linear equations
Lecture 07 graphing linear equationsLecture 07 graphing linear equations
Lecture 07 graphing linear equations
 
Chapter5
Chapter5Chapter5
Chapter5
 
Regression.pptx
Regression.pptxRegression.pptx
Regression.pptx
 

Mais de Maria Theresa

Syllabus educ200 methods of research
Syllabus educ200 methods of researchSyllabus educ200 methods of research
Syllabus educ200 methods of research
Maria Theresa
 
ED197 episodes 1 12 (manual)
ED197 episodes 1 12 (manual)ED197 episodes 1 12 (manual)
ED197 episodes 1 12 (manual)
Maria Theresa
 
Module 5 a job induction & orientation
Module 5 a  job induction & orientationModule 5 a  job induction & orientation
Module 5 a job induction & orientation
Maria Theresa
 
Topic 1 ed105 b introduction
Topic 1 ed105 b introductionTopic 1 ed105 b introduction
Topic 1 ed105 b introduction
Maria Theresa
 

Mais de Maria Theresa (20)

Syllabus educ200 methods of research
Syllabus educ200 methods of researchSyllabus educ200 methods of research
Syllabus educ200 methods of research
 
ED197 episodes 1 12 (manual)
ED197 episodes 1 12 (manual)ED197 episodes 1 12 (manual)
ED197 episodes 1 12 (manual)
 
Module 5 a job induction & orientation
Module 5 a  job induction & orientationModule 5 a  job induction & orientation
Module 5 a job induction & orientation
 
Module2 human resource information system
Module2  human resource information systemModule2  human resource information system
Module2 human resource information system
 
Module 1 human resource management
Module 1  human resource managementModule 1  human resource management
Module 1 human resource management
 
Tarpaulin making contest
Tarpaulin making contestTarpaulin making contest
Tarpaulin making contest
 
Documentary film
Documentary filmDocumentary film
Documentary film
 
Fs 2 episode 1
Fs 2 episode 1Fs 2 episode 1
Fs 2 episode 1
 
Fs 2 episode 7
Fs 2 episode 7Fs 2 episode 7
Fs 2 episode 7
 
Fs 2 episode 6
Fs 2 episode 6Fs 2 episode 6
Fs 2 episode 6
 
Fs 2 episode 6
Fs 2 episode 6Fs 2 episode 6
Fs 2 episode 6
 
Fs 2 episode 5
Fs 2 episode 5Fs 2 episode 5
Fs 2 episode 5
 
Fs 2 episode 4
Fs 2 episode 4Fs 2 episode 4
Fs 2 episode 4
 
Fs 2 episode 3
Fs 2 episode 3Fs 2 episode 3
Fs 2 episode 3
 
Fs 2 episode 2
Fs 2 episode 2Fs 2 episode 2
Fs 2 episode 2
 
Fs 2 episode 1
Fs 2 episode 1Fs 2 episode 1
Fs 2 episode 1
 
Topic 1 ed105 b introduction
Topic 1 ed105 b introductionTopic 1 ed105 b introduction
Topic 1 ed105 b introduction
 
Exercise (t able and mail merge)
Exercise (t able and mail merge)Exercise (t able and mail merge)
Exercise (t able and mail merge)
 
Sample format for appendices & bibliography
Sample format for appendices & bibliographySample format for appendices & bibliography
Sample format for appendices & bibliography
 
Title page final format
Title page final formatTitle page final format
Title page final format
 

Simple linear regression

  • 1. SIMPLE LINEAR REGRESSION Reporters: Atty. Gener R. Gayam, CPA Agapito “pete” M. Cagampang, PM Raymond B. Cabling, MD
  • 2. SIMPLE LINEAR REGRESSION The Scatter Diagram In solving problems that concern estimation and forecasting, a scatter diagram can be used as a graphical approach. This technique consists of joining the points corresponding to the paired scores of dependent and independent variables which are commonly represented by X and Y on the X – y coordinate system. Below is an illustration of a scatter diagram using the data in Table 6.1. This table shows the data about the six years working experience and the income of eight employees in a big industrial corporation.
  • 4. Figure 6.1 – A Scatter Diagram for Table 6.1 Data X X X X X X X X
  • 5. For you to roughly predict the value of a dependent variable, such as years of working experience, from the dependent variable, which is income, your next step is to draw a trend line. This is a line passing through the series of points such that the total vertical measurement of the points below this line is more or less equal to the total measurements of the points above the line. If these requirements are satisfied, you draw a correct trend Y. The illustration is shown in figure 6.1
  • 6. Figure 6.2 - A trend line drawn on the linear direction between working experience and income of eight employees Trend Line
  • 7. Using the trend line draw in Figure 6.1 above, the value estimated for Y when X is 16, is 18. You should not fail to remember that if a “straight line” appears to describe the relationship, the algebraic approach called the regression formula can be used as explained in the next topic.
  • 8. a = Ῡ - bX B. The Least Square Linear Regression Equation The least square linear regression equation can be understood through this formula known from algebra. Y = a + bx For instance the Y = a+bx in figure 6.1 in that line that gives the smallest sum of the squares of the vertical measurements or distance of the points from the line. In solving the regression equations, you need to solve first,
  • 9. ΣX = 62 ΣY = 90 X = 7.75 Y = 11.25 Example: Solve the least squares regression line for the data scores in Table 6.1.
  • 11. After solving the values of b and a, your regression equation obtained from Table 6.1 is.
  • 12. Now, we are interested in the distance of the Y values from Y₁ the corresponding ordinate of the regression line. Here, we are going to base our measure of dispersion or variation around the regression line on the distance (Y₁ ‒ Y)². This can be well understood by this standard error of estimate formula given below. Se = Σ(Yi ‒ Ŷ)² n ‒ 2 √ C. The standard Error of Estimate
  • 13. However, this formula entails a very tedious process of computing the standard error of estimate, so that the formula by Basil P. Korin (1977), which is easier to solve suggested as follows: Se = ΣYi² ‒ a(Yi) ‒ b(Xi ‒ Yi) n ‒ 2 Note: The symbol a and b stand for the intercept and the slope of the regression line. √
  • 14. √ Example: Solve the standard error of estimate for the regression line which was derived from the data in Table 6.1. Se = Σ(Yi ‒ Ŷ)² n ‒ 2
  • 15. Step 1 – Compute the value of Y at each of the X values. Example: Y = 6.68 + .59 (2) = 6.68 + 1.18 = 7.68 Do the rest by following the same procedure. Step 2 – Get the difference between (Yi ‒ Ŷ). Example: 8 – 7.86 = .14 Step 3 – Square all the difference Yi ‒ Ŷ. Example: (.14)² = .0196
  • 16. √ √ √ √ Step 4 – Apply the formula. Se = Σ(Yi ‒ Ŷ)² n ‒ 2 = 16.692 8 – 2 = 16.692 6 = 2.782 = 1.67
  • 17. Solution 2: Se = ΣYi² ‒ a(Yi) ‒ b(Xi ‒ Yi) n ‒ 2
  • 18. √ √ √ √ √ Step 1 – Square Y₁ Example: (8²) = 64 Step 2 – Multiply XiYi Example: 2 X 8 = 16 Step 3 – Get the sum of Yi² and XiYi Step 4 – Apply the formula = 1084 – 6.68 (90) – .59 (791) n – 2 = 1084 – 601.2 – 466.69 8 – 2 = 1084 – 1067.89 8 – 2 = 16.11 6 = 2.685 = 1.64
  • 19. The standard error of estimate is interpreted as the standard deviation. For example, if we measure vertically three standard errors from the regression line above and below, we will find that the same value of X will always fall between the upper and lower 3Se Limits. In the example above of the standard error of estimate which is 1.64 you will come up with 4.92 units (3) (1.64) above and below the regression line. This means that these “bounds” of 4.92 unit above and below the regression line pertain to all observations taken for that particular sample. If you draw two parallel lines, each of them lying one Se from the regression line, you will expect two thirds of the observations falling between these bounds. See Figure 6.1 for the illustration of the data in Table 6.1.
  • 20. Y = 6.68 + .59 X Figure 6.3 – A regression Line with One Standard Error Distance