Regression analysis is an analysis technique that calculates the estimated relationship between a dependent variable and one or more explanatory variables. With regression analysis, you can model the relationship between the chosen variables as well as predict values based on the model.

## Regression analysis overview

Regression analysis uses a chosen estimation method, a dependent variable, and one or more explanatory variables to create an equation that estimates values for the dependent variable.

The regression model includes outputs, such as R^{2} and p-values, to provide information on how well the model estimates the dependent variable.

Charts, such as scatter plot matrices, histograms, and point charts, can also be used in regression analysis to analyze relationships and test assumptions.

Regression analysis can be used to solve the following types of problems:

- Determine which explanatory variables are related to the dependent variable.
- Understand the relationship between the dependent and explanatory variables.
- Predict unknown values of the dependent variable.

## Examples

An analyst for a small retail chain is studying the performance of different store locations. The analyst wants to know why some stores are having an unexpectedly low sales volume. The analyst creates a regression model with explanatory variables like median age and income in the surrounding neighborhood, as well as distance to retail centers and public transit, to determine which variables are influencing sales.

An analyst for a department of education is studying the effects of school breakfast programs. The analyst creates a regression model of educational attainment outcomes, such as graduation rate, using explanatory variables like class size, household income, school budget per capita, and proportion of students eating breakfast daily. The equation of the model can be used to determine the relative effect of each variable on the educational attainment outcomes.

An analyst for a nongovernmental organization is studying global greenhouse gas emissions. The analyst creates a regression model for the latest emissions for each country using explanatory variables like gross domestic product (GDP), population, electricity production using fossil fuels, and vehicle usage. The model can then be used to predict future greenhouse gas emissions using forecasted GDP and population values.

## Ordinary Least Squares

Regression analysis in ArcGIS Insights is modeled using the Ordinary Least Squares (OLS) method.

The OLS method is a form of multiple linear regression, meaning the relationship between the dependent variables and the independent variables must be modeled by fitting a linear equation to the observed data.

An OLS model uses the following equation:

`y`_{i}=β_{0}+β_{1}x_{1}+β_{2}x_{2}+...+β_{n}x_{n}+ε

where:

- y
_{i}=the observed value of the dependent variable at point i - β
_{0}=the y-intercept (constant value) - β
_{n}=the regression coefficient or slope for explanatory variable N at point i - x
_{n}=the value of variable N at point i - ε=the error of the regression equation

### Assumptions

Each regression method has several assumptions that must be met for the equation to be considered reliable. The OLS assumptions should be validated when creating a regression model.

The following assumptions should be tested and met when using the OLS method:

- The model must be linear.
- The data must be randomly sampled.
- The explanatory variables must not be collinear.
- The explanatory variables must have negligible error in measurement.
- The residuals have an expected sum of zero.
- The residuals have homogeneous variance.
- The residuals are normally distributed.
- Adjacent residuals must not show autocorrelation.

#### The model must be linear

OLS regression can only be used to create a linear model. Linearity can be tested between the dependent variable and the explanatory variables using a scatter plot. A scatter plot matrix can test all the variables, provided there are no more than five variables in total.

#### The data must be randomly sampled

The data being used in regression analysis should be sampled in such a way that the samples themselves are not dependent on any external factor. Random sampling can be tested using the residuals from the regression model. The residuals, which are an output from the regression model, should have no correlation when plotted against the explanatory variables on a scatter plot or scatter plot matrix.

#### The explanatory variables must not be collinear

Collinearity refers to a linear relationship between explanatory variables, which creates redundancy in the model. In some cases, the model can be created with collinearity. However, if one of the collinear variables seems to be dependent on the other, you may want to consider dropping that variable from the model. Collinearity can be tested using a scatter plot or scatter plot matrix of the explanatory variables.

#### The explanatory variables must have negligible error in measurement

A regression model is only as accurate as its input data. If the explanatory variables have large margins of error, the model cannot be accepted as accurate. When performing regression analysis, it is important to only use datasets from known and trusted sources to ensure that the error is negligible.

#### The residuals have an expected sum of zero

Residuals are the difference between observed and estimated values in a regression analysis. Observed values that fall above the regression curve will have a positive residual value, and observed values that fall below the regression curve will have a negative residual value. The regression curve should lie along the center of the data points; therefore, the sum of residuals should be zero. The sum of a field can be calculated in a summary table.

#### The residuals have homogeneous variance

The variance should be the same for all residuals. This assumption can be tested using a scatter plot of the residuals (y-axis) and the estimated values (x-axis). The resulting scatter plot should appear as a horizontal band of randomly plotted points across the plot.

#### The residuals are normally distributed

A normal distribution, also called a bell curve, is a naturally occurring distribution, where the frequency of a phenomenon is high near the mean and tapers off as the distance from the mean increases. A normal distribution is often used as the null hypothesis in a statistical analysis. The residuals must be normally distributed to show that the line of best fit is optimized centrally within the observed data points, not skewed toward some and away from others. This assumption can be tested by creating a histogram with the residuals. The normal distribution curve can be overlaid and the skewness and kurtosis measures are reported on the back of the histogram card.

#### Adjacent residuals must not show autocorrelation

This assumption is based on time-ordered data. If the data is time ordered, each data point must be independent of the preceding or subsequent data point. Therefore, it is important to make sure that your time-ordered data is organized in the correct order when performing a regression analysis. This assumption can be calculated using a Durbin-Watson test.

The Durbin-Watson test is a measure of autocorrelation in residuals in a regression model. The Durbin-Watson test uses a scale of 0 to 4, with values 0 to 2 indicating positive autocorrelation, 2 indicating no autocorrelation, and 2 to 4 indicating negative autocorrelation. Therefore, values near 2 are required to meet the assumption of no autocorrelation in the residuals. In general, values between 1.5 and 2.5 are considered acceptable, whereas values less than 1.5 or greater than 2.5 indicate that the model does not fit the assumption of no autocorrelation.

## Model validity

The accuracy of a regression equation is an important part of regression analysis. All models will include an amount of error, but understanding the statistics will help you determine if the model can be used in your analysis, or if adjustments need to be made.

There are two techniques for determining the validity of a regression model: exploratory analysis and confirmatory analysis.

### Exploratory analysis

Exploratory analysis is a method of understanding your data using a variety of visual and statistical techniques. Throughout the course of your exploratory analysis, you will test the assumptions of OLS regression and compare the effectiveness of different explanatory variables. Exploratory analysis will allow you to compare the effectiveness and accuracy of different models, but it does not determine whether you should use or reject your model. Exploratory analysis should be performed before confirmatory analysis for each regression model and reiterated to make comparisons between models.

The following charts and statistics can be used as part of exploratory analysis:

- Scatter plot and scatter plot matrix
- Histogram and normal distribution
- Regression equation and predicting new observations
- Coefficient of determination, R
^{2}and Adjusted R^{2} - Residual standard error
- Point chart

Exploratory analysis should begin while you are choosing explanatory variables and before you create a regression model. Since OLS is a method of linear regression, one of the main assumptions is that the model must be linear. A scatter plot or scatter plot matrix can be used to assess linearity between the dependent variable and the explanatory variables. A scatter plot matrix can display up to four explanatory variables along with the dependent variable, making it an important tool for large-scale comparisons between all variables. A single scatter plot only displays two variables: one dependent and one independent or explanatory. Viewing a scatter plot of the dependent variable and a single explanatory variable allows you to make a more acute assessment of the relationship between the variables. Linearity can be tested before you create a regression model to help determine which explanatory variables will create an acceptable model.

Several statistical outputs are available after you create a regression model, including the regression equation, R^{2} value, and Durbin-Watson test. Once you've created a regression model, you should use the outputs and necessary charts and tables to test the remaining assumptions of OLS regression. If your model meets the assumptions, you can continue with the remaining exploratory analysis.

The regression equation gives valuable information about the influence of each explanatory variable on the predicted values, including the regression coefficient for each explanatory variable. The slope values can be compared to determine the relative influence of each explanatory variable on the dependent variable; the further the slope value is from zero (either positive or negative), the larger the influence. The regression equation can also be used to predict values for the dependent variable by entering values for each explanatory variable.

The coefficient of determination, symbolized as R^{2}, measures how well the regression equation models the actual data points. The R^{2} value is a number between 0 and 1, with values closer to 1 indicating more accurate models. An R^{2} value of 1 indicates a perfect model, which is highly unlikely in real-world situations given the complexity of interactions between different factors and unknown variables. Therefore, you should strive to create a regression model with the highest R^{2} value possible, while recognizing that the value may not be close to 1.

When performing regression analysis, there is a risk of creating a regression model that has an acceptable R^{2} value by adding explanatory variables that cause a better fit based on chance alone. The adjusted R^{2} value, which is also a value between 0 and 1, accounts for additional explanatory variables, reducing the role that chance plays in the calculation. Adjusted R^{2} should be used for models using many explanatory variables, or when comparing models with different numbers of explanatory variables.

The residual standard error measures the accuracy with which the regression model can predict values with new data. Smaller values indicate a more accurate model; therefore, when multiple models are compared, the model with the smallest value will be the model that minimizes residual standard error.

Point charts can be used to analyze your explanatory variables for patterns like clustering and outliers, which may affect the accuracy of the model.

### Confirmatory analysis

Confirmatory analysis is the process of testing your model against a null hypothesis. In regression analysis, the null hypothesis is that there is no relationship between the dependent variable and the explanatory variables. A model with no relationship would have slope values of 0. If the elements of your confirmatory analysis are statistically significant, you can reject the null hypothesis (in other words, statistical significance indicates that a relationship does exist between the dependent and explanatory variables).

The following statistical outputs are used to determine significance as part of confirmatory analysis:

- F statistic and its associated p-value
- t statistics and their associated p-values
- Confidence intervals

The F statistic is a global statistic returned from an F-test, which indicates the predictive capability of the regression model by determining if all the regression coefficients in the model are significantly different from 0. The F-test analyzes the combined influence of the explanatory variables, rather than testing the explanatory variables individually. The F statistic has an associated p-value, which indicates the probability that the relationships in your data are happening by chance. Since p-values are based on probabilities, the values are given on a scale from 0.0 to 1.0. A small p-value, usually 0.05 or less, is required to determine that the relationships in the model are real (in other words, not happening by chance) and to reject the null hypothesis. In that case, the probability that the relationships in the model are happening by chance is 0.05, or 1 in 20. Alternatively, the probability that the relationships are real is 0.95, or 19 in 20.

The t statistic is a local statistic returned from a t-test, which indicates the predictive capability of each explanatory variable individually. Like the F-test, the t-test analyzes if the regression coefficients in the model are significantly different from zero. However, since a t-test is performed on each explanatory variable, the model will return a t statistic value for each explanatory variable, rather than one per model. Each t statistic has an associated p-value, which indicates the significance of the explanatory variable. Like the p-values for the F-test, the p-value for each t-test should be 0.05 or less to reject the null hypothesis. If an explanatory variable has a p-value greater than 0.05, the variable should be discarded and a new model should be created, even if the global p-value was significant.

Confidence intervals show the regression coefficient for each explanatory variable and the associated 90, 95, and 99 percent confidence intervals. Therefore, the confidence intervals can be used alongside the p-values from the t-tests to assess the null hypothesis for individual explanatory variables. The regression coefficients must not be equal to 0 if you are to reject the null hypothesis and continue using the model. Therefore, for each explanatory variable, the regression coefficient and the associated confidence intervals should not overlap with 0. If a 99 or 95 percent confidence interval for a given explanatory variable overlaps with 0, the explanatory variable has failed to reject the null hypothesis. Including such a variable in your model may have an effect on the overall significance of your model. If only the 90 percent confidence interval overlaps with 0, the explanatory variable may be included in the model provided the other global statistics are significant. Ideally, the confidence intervals for all explanatory variables should be far from 0.

## Other outputs

Other outputs, such as estimated values and residuals, are important for testing the assumptions of OLS regression. In this section, you will learn more about how these values are calculated.

### Estimated values

The estimated values are calculated using the regression equation and the values for each explanatory variable. Ideally, the estimated values would be equal to the observed values (in other words, the actual values of the dependent variable).

Estimated values are used with the observed values to calculate residuals.

### Residuals

The residual values in a regression analysis are the differences between the observed values in the dataset and the estimated values calculated with the regression equation.

The residuals A and B for the relationship above would be calculated as follows:

`residuals`_{A} = observed_{A} - estimated_{A}
residuals_{A} = 595 - 487.62
residuals_{A} = 107.38

`residuals`_{B} = observed_{B} - estimated_{B}
residuals_{B} = 392 - 527.27
residuals_{B} = -135.27

Residuals can be used to calculate error in a regression equation as well as to test several assumptions.