How does a company forecast how much it can spend on future advertising in order to increase sales? Similarly, how much should a company spend in future training programs in order to improve productivities?
In both cases, those companies rely on the historical relationship between variables like, in the former case, the relationship between the sale value and advertising spending to predict the future outcome.
This is a typical regression problem that can be best solved using machine learning1.
This article describes a common type of regression analysis called linear regression2 and how the Intel® Data Analytics Acceleration Library (Intel® DAAL)3 helps optimize this algorithm when running it on systems equipped with Intel® Xeon® processors.
What is Linear Regression?
Linear regression (LR) is the most basic type of regression that is used for predictive analysis. LR shows a linear relationship between variables and how one variable can be affected by one or more variables. The variable which is impacted by the others is called a dependent, response, or outcome variable and the others are called independent, explanatory, or predictor variables.
In order to use LR analysis, we need to examine whether LR is suitable for this set of data. To do that we need to observe how the data is distributed. Let's look at the following two graphs:
Figure 1:A straight line can be fitted through the data points.
Figure 2:A straight line cannot be fitted through these data points.
In figure 1, we can fit a straight line through the data points; however, there is no way to do that with the data points in figure 2. Only the curved (non-linear) line can be fitted through the data points in figure 2. Therefore, linear regression analysis can be done on the dataset in figure 1, but not on that in figure 2.
Depending on the number of independent variables, LR is divided into two types: simple linear regression (SLR) and multiple linear regression (MLR).
LR is called SLR when there exists only one independent variable, whereas MLR has more than one independent variable.
The simplest form of the equation with one dependent and one independent variable is defined as:
y = Ax + B (1)
y: Dependent variable
x: Independent variable
A: Regression coefficient or slope of the line
The question is how to find the best-fit line so that the difference between the observed and predicted values of the dependent variable (y) is minimum. In other words, find A and B in equation (1) such that |yobserved – ypredicted| is minimum.
The task to find the best-fit line can be done using the least squares method4. It calculates the best-fit line by minimizing the sum of the squares of the vertical differences (difference between the observed and predicted values of y) from each data point to the line. The vertical difference can also be called residual.
Figure 3:Linear regression graph.
From figure 3 the green dots represent the actual data points. The black dots show the vertical (not perpendicular) projection of the data points onto the regression line (red line). The black dots are also called the predicted data. The vertical difference between the actual data and the predicted data is called residual.
Applications of Linear Regression
Some of the applications that can make good use of linear regression:
Advantages and Disadvantages of Linear Regression
Some of the pros and cons of LR:
Intel® Data Analytics Acceleration Library
Intel DAAL is a library consisting of many basic building blocks that are optimized for data analytics and machine learning. These basic building blocks are highly optimized for the latest features of the latest Intel® processors. LR is one of the predictive algorithms that Intel DAAL provides. In this article, we use the Python* API of Intel DAAL to build a basic LR predictor. To install DAAL, follow the instructions in How to install the Python Version of Intel DAAL in Linux*5.
Using Linear Regression Algorithm in Intel Data Analytics Acceleration Library
This section shows how to invoke the linear regression method in Python6 using Intel DAAL.
The reference section provides a link7 to the free datasets that can be used to test the application.
Do the following steps to invoke the algorithm from Intel DAAL:
1. Import the necessary packages using the commands from and import.
1. Import Numpy by issuing the command:
import numpy as np
2. Import the Intel DAAL numeric table by issuing the following command:
from daal.data_management import HomogenNumericTable
3. Import necessary functions to numeric tables to store data:
from daal.data_management import ( DataSourceIface, FileDataSource, HomogenNumericTable, MergeNumericTable, NumericTableIface)
4. Import the LR algorithm using the following commands:
from daal.algorithms.linear_regression import training, prediction
2. Initialize the file data source if the data input is from the .csv file:
trainDataSet = FileDataSource(
3. Create numeric tables for training data and dependent variables:
trainInput = HomogenNumericTable(nIndependentVar, 0, NumericTableIface.notAllocate)
trainDependentVariables = HomogenNumericTable(nDependentVariables, 0, NumericTableIface.notAllocate)
mergedData = MergedNumericTable(trainData, trainDependentVariables)
4. Loading input data:
5. Create a function to train the model.
1. First create an algorithm object to train the model using the following command:
algorithm = training.Batch_Float64NormEqDense()
Note: This algorithm uses the normal equation to solve the linear least squares problem. DAAL also supports QR decomposition/factorization.
2. Pass the training dataset and dependent variables to the algorithm using the following commands:
algorithm: The algorithm object as defined in step a above. train
Input: Training data. trainDependent
Variables: Training dependent variables
3. Train the model using the following command:
trainResult = algorithm.compute()
algorithm: The algorithm object as defined in step a above.
6. Create a function to test the model.
1. Similar to steps 2, 3, and 4 above, we need to create the test dataset for testing:
testDataSet = FileDataSource(
testInput = HomogenNumericTable(nIndependentVar, 0, NumericTableIface.notAllocate)
testTruthValues = HomogenNumericTable(nDependentVariables, 0, NumericTableIface.notAllocate)
mergedData = MergedNumericTable(testDataSet, testTruthValues)
2. Create an algorithm object to test/predict the model using the following command:
algorithm = prediction.Batch()
3. Pass the testing data and the train model to the model using the following commands:
algorithm: The algorithm object as defined in step a above.
testInput: Testing data.
4. Test/predict the model using the following command:
Prediction = algorithm.compute()
Linear regression is a very common predictive algorithm. Intel DAAL optimized the linear regression algorithm. By using Intel DAAL, developers can take advantage of new features in future generations of Intel Xeon processors without having to modify their applications. They only need to link their applications to the latest version of Intel DAAL.
For more such intel IoT resources and tools from Intel, please visit the Intel® Developer Zone
Digit caters to the largest community of tech buyers, users and enthusiasts in India. The all new Digit in continues the legacy of Thinkdigit.com as one of the largest portals in India committed to technology users and buyers. Digit is also one of the most trusted names when it comes to technology reviews and buying advice and is home to the Digit Test Lab, India's most proficient center for testing and reviewing technology products.
We are about leadership-the 9.9 kind! Building a leading media company out of India.And,grooming new leaders for this promising industry.