Statistics

Residual Analysis

Residual Analysis: The Complete Guide for Students and Data Analysts | Ivy League Assignment Help
Statistics & Data Analysis Guide

Residual Analysis: The Complete Guide for Students and Data Analysts

Residual analysis is one of the most underused and most consequential skills in statistics. Students rush through it. Analysts skip it. Then their models fail in ways they never anticipated — because the residuals were trying to tell them something they ignored. This guide changes that. If you are in a statistics course, working on a regression assignment, or building predictive models, this is the guide that bridges what your textbook explains in half a page and what you actually need to know to use residual analysis with confidence.

You will find clear definitions, step-by-step diagnostic procedures, and practical interpretations of every major residual type and residual plot — from raw residuals and studentized residuals to Cook's Distance, the Durbin-Watson test, the Breusch-Pagan test, and the hat matrix. Each concept is explained in terms of what it means for your model, not just what the formula says.

This guide also addresses the real-world applications of residual analysis in fields like econometrics, machine learning, biostatistics, and social science research — where detecting heteroscedasticity, autocorrelation, and influential observations is not an academic exercise but a professional requirement for valid inference.

Whether you are completing your first regression homework or auditing a complex predictive model, every section maps directly to the statistical competencies your instructors and employers actually assess — grounded in current statistical theory and software practice.

What Is Residual Analysis?

Residual analysis is the systematic examination of the differences between the values a regression model predicts and the values actually observed in the data. That gap — called a residual — is where the real story of your model lives. Summary statistics like R² or RMSE give you a bird's-eye view of model performance. Residual analysis gives you the forensic detail: where your model fails, why it fails, and what you can do about it. Skipping it is like reading only the headline and missing the entire article. Regression analysis rests entirely on a set of assumptions that only residual analysis can verify.

The formal definition is precise: a residual eᵢ for the i-th observation is the observed value yᵢ minus the predicted value ŷᵢ from the fitted regression equation. Written out: eᵢ = yᵢ − ŷᵢ. Every single data point in your dataset has exactly one residual. Positive residuals mean the model underpredicted the actual value; negative residuals mean it overpredicted. If the model were perfect — which no model ever is — every residual would be zero. The goal of residual analysis is not to eliminate residuals but to ensure they behave in ways that are consistent with the model's underlying assumptions.

4
Core regression assumptions that residual analysis directly tests: linearity, independence, homoscedasticity, and normality
6+
Distinct types of residual plots used in comprehensive regression diagnostics
~50%
Of hospital admissions show medication discrepancies — a figure residual analysis helps identify analogously in model "error auditing"

Why does this matter beyond getting assignments right? In applied settings — finance, medicine, engineering, social science — the validity of your conclusions depends entirely on whether your regression assumptions hold. Regression model assumptions are not just theoretical niceties; they are the conditions under which ordinary least squares (OLS) estimates are unbiased, efficient, and produce valid confidence intervals and hypothesis tests. When those assumptions are violated and residual analysis goes unchecked, you can end up with statistically significant results that are artefacts of a misspecified model — not genuine relationships in the data.

What Is the Difference Between a Residual and an Error?

Students frequently conflate residuals and errors. The distinction is conceptually important. An error is the theoretical deviation between an observed value and the true population regression line — the line you would have if you knew the entire population. Since you never observe the entire population, errors are never directly observable. A residual is the difference between an observed value and the estimated regression line — the one you fit from your sample data. Residuals are what you compute; errors are what you assume follow a distribution. When statisticians say "the errors are normally distributed," they are making a theoretical assumption verified in practice by examining whether the residuals appear approximately normally distributed.

"Since all models are wrong, the scientist must be alert to what is importantly wrong." — George E.P. Box, statistician at the University of Wisconsin-Madison, in a statement that has guided residual analysis practice for decades.

Where Does Residual Analysis Sit in the Modeling Process?

Residual analysis is not a final step you add at the end of a regression homework. It is an iterative diagnostic loop embedded throughout model development. You fit a model, examine the residuals, identify violations, refine the model, and repeat. Simple linear regression introduces residuals as part of the ordinary least squares fitting procedure — the OLS estimator literally minimizes the sum of squared residuals, making the residual structure fundamental to the whole methodology. Understanding residuals conceptually before diving into diagnostic plots is not just academically correct — it makes every subsequent diagnostic technique immediately interpretable.

The "Leftover" Intuition: Think of residuals as what your model could not explain. After your regression model has extracted all the systematic variation in the outcome that your predictors can account for, what remains in the residuals should look like pure noise — random, structureless, and unpatterned. Any systematic structure left in the residuals is evidence that your model is missing something: a non-linear relationship, an interaction effect, a missing variable, or a violated assumption. Residual analysis is the process of looking at that "leftover" noise and checking whether it truly is random.

Types of Residuals in Regression Analysis

Not all residuals are created equal. The raw residual eᵢ = yᵢ − ŷᵢ is the starting point, but raw residuals are not always the most informative diagnostic tool. Different types of residuals serve different purposes — detecting outliers, assessing leverage, flagging influential observations, and comparing residuals across observations on a standardized scale. Understanding which residual type to use for which diagnostic question is a genuine statistical competency. Multiple linear regression models, where residual patterns become more complex than in simple regression, make this distinction especially important.

Raw Residuals

The raw residual is the most direct: eᵢ = yᵢ − ŷᵢ. It is easy to compute and intuitively interpretable. The problem is that raw residuals have different variances depending on the leverage of each observation — observations with high leverage tend to have smaller raw residuals, which can make them look well-fit when they are actually influential. This is why raw residuals alone are insufficient for a complete diagnostic analysis. They are a good starting point, but the adjusted forms below are necessary for robust outlier and influence detection. Raw residuals are what you see first in statistical reporting outputs, but they require context to interpret correctly.

Standardized Residuals

Standardized residuals divide each raw residual by an estimate of its standard deviation, producing a dimensionless quantity that can be compared across observations regardless of the scale of the outcome variable. The standardized residual for observation i is approximately eᵢ / s, where s is the estimated residual standard error from the model. Under the assumption that errors are normally distributed, standardized residuals should follow a standard normal distribution. Values beyond ±2 are worth examining; values beyond ±3 are strong outlier candidates. Most statistical software reports standardized residuals by default alongside regression output.

Studentized Residuals

Studentized residuals take the standardization further. Rather than dividing by a single standard error estimate from the whole model, internally studentized residuals divide by a standard error that accounts for each observation's leverage — specifically, the diagonal elements of the hat matrix hᵢᵢ. The internally studentized residual is: rᵢ = eᵢ / (s × √(1 − hᵢᵢ)). This adjustment makes studentized residuals more sensitive to genuine outliers because it removes the dampening effect that high-leverage points have on their own raw residuals. Residual analysis in statistical modeling consistently recommends studentized residuals over raw residuals for outlier detection precisely because of this correction.

Externally Studentized Residuals (Jackknife Residuals)

Externally studentized residuals — also called jackknife residuals or deleted studentized residuals — go one step further. They estimate the standard error using a model fitted without observation i. This means the residual for observation i is scaled by the variance from a model that does not include it. This approach is particularly powerful for detecting outliers that are large enough to inflate the model's overall variance estimate, which would otherwise mask their own extremeness. Externally studentized residuals follow a t-distribution with n − p − 2 degrees of freedom under the null hypothesis of no outlier, making formal outlier testing straightforward. The t-distribution underpins these formal tests, making familiarity with it essential for rigorous residual-based outlier detection.

Residual Type Formula Basis Primary Use Key Threshold Scale
Raw Residuals eᵢ = yᵢ − ŷᵢ Basic model fit assessment, residual plots None (scale-dependent) Same as outcome variable
Standardized Residuals eᵢ / s Initial outlier screening across observations |value| > 2 or 3 Dimensionless
Internally Studentized eᵢ / (s √(1−hᵢᵢ)) Outlier detection adjusting for leverage |value| > 2 or 3 Approximately standard normal
Externally Studentized (Jackknife) Deletes observation i before estimating s Formal outlier testing; most sensitive Follow t(n−p−2) distribution t-distribution
PRESS Residuals yᵢ − ŷ₋ᵢ (leave-one-out prediction) Cross-validated predictive accuracy Used to compute PRESS statistic Same as outcome variable

PRESS Residuals and Predictive Accuracy

PRESS residuals (Prediction Residual Error Sum of Squares) are the difference between each observed value and the prediction made by a model fitted on all data except that observation. They represent a leave-one-out cross-validation directly embedded in the regression framework. The sum of squared PRESS residuals — the PRESS statistic — measures a model's predictive accuracy more honestly than the standard R² on training data, because it is evaluated on observations the model never "saw." A model with a much higher PRESS statistic than its training-data R² suggests is overfitting. Cross-validation and bootstrapping extend this logic to more general predictive validation contexts in machine learning and statistical modeling.

The Four Regression Assumptions Residual Analysis Tests

Residual analysis is fundamentally about verifying assumptions. Ordinary least squares regression produces reliable estimates only when four core assumptions about the error term hold. None of these can be confirmed by looking at the data alone — they are verified by examining how the residuals behave. When students ask "why do we do residual analysis," this is the complete answer: because every inferential statement you make about your regression model — every p-value, confidence interval, and hypothesis test — depends on these four conditions being approximately true. Knowing what each violation looks like in a residual plot is the practical competency being tested in every statistics course that covers regression analysis.

Assumption 1: Linearity

Linearity means the relationship between each predictor variable and the outcome variable is linear — a straight-line relationship rather than a curve. When this assumption holds, a plot of residuals against fitted values should show a flat, structureless horizontal band around zero. When linearity fails, you see a systematic curve or arch in the residual plot. Residuals are positive at low fitted values, negative in the middle, then positive again — or the reverse. This pattern means your model is systematically over- or underpredicting in certain ranges, which is diagnostic of a non-linear relationship that OLS is not capturing.

The fix is usually adding a polynomial term (quadratic, cubic) for the relevant predictor, applying a transformation to the outcome or predictor (log, square root), or using a more flexible non-linear modeling approach. In polynomial regression, the addition of higher-order terms directly addresses this kind of residual curvature. The Ramsey RESET test provides a formal statistical test for non-linearity in regression models, though graphical inspection of the residual plot is usually more informative.

Assumption 2: Independence

Independence means the residuals from different observations are not correlated with each other. In other words, knowing the residual for one observation tells you nothing about the residual for any other observation. This assumption is most commonly violated in time series data, longitudinal data, clustered data (students nested within schools, patients within hospitals), and spatial data. When independence fails, the residuals show autocorrelation — a systematic pattern where consecutive residuals tend to be either all positive, all negative, or alternating in a regular pattern.

The primary diagnostic tool for autocorrelation is the Durbin-Watson test, which tests specifically for first-order serial correlation. Values close to 2 indicate no correlation; below 1.5 suggests positive autocorrelation; above 2.5 suggests negative autocorrelation. Time series analysis with ARIMA models is one of the most important contexts where residual independence must be checked rigorously — autocorrelated residuals in a time series regression indicate that the temporal structure of the data has not been adequately modeled.

Assumption 3: Homoscedasticity

Homoscedasticity means the residuals have constant variance at all levels of the fitted values and predictor variables. When this holds, a residual plot shows a consistent, even horizontal band of points with no widening or narrowing. When this fails — a condition called heteroscedasticity — the residual plot shows a fan or funnel shape: residuals spread out more as fitted values increase (or decrease). Heteroscedasticity is extremely common in real-world data, particularly in economic and financial data where larger values tend to have larger variability, and in biological data where variance scales with the mean.

The formal tests for heteroscedasticity are the Breusch-Pagan test (which regresses squared residuals on the predictors and tests for significant relationships) and the White test (a more general version that also tests for interaction effects between predictors). When heteroscedasticity is detected, standard remedies include log-transforming the outcome variable, using robust standard errors (Huber-White sandwich estimators), or weighted least squares regression. Logistic regression and other generalized linear models inherently model non-constant variance, making them appropriate when the outcome's distributional structure generates systematic heteroscedasticity.

Assumption 4: Normality of Errors

Normality means the residuals are approximately normally distributed. This assumption is least critical for large samples (by the Central Limit Theorem, coefficient estimates are asymptotically normal regardless) but matters considerably in small samples where t-tests and F-tests depend on it for exact validity. The primary diagnostic tools for normality are the Q-Q (quantile-quantile) plot — comparing the empirical quantiles of residuals to theoretical normal quantiles — and formal normality tests like the Shapiro-Wilk test or Jarque-Bera test.

A Q-Q plot with points falling along a straight diagonal line indicates normality. Heavy tails produce an S-shaped deviation; right skewness produces an upward arch; left skewness produces a downward arch. The Shapiro-Wilk test is the most powerful normality test for small to moderate samples and is the default in many statistical software packages. The Jarque-Bera test, common in econometrics, tests whether the skewness and kurtosis of the residuals match a normal distribution. Normal distributions, kurtosis, and skewness are the statistical concepts that connect directly to interpreting Q-Q plots and normality tests in residual analysis. The Jarque-Bera normality test is among the most widely cited residual diagnostics in econometrics.

Normality Testing Trap: In large samples, formal normality tests like Shapiro-Wilk will almost always reject normality, even when the deviation from normal is trivially small and practically meaningless for inference. This is because test power increases with sample size — the test detects minuscule departures from normality that have no real-world consequence. For large samples, rely primarily on the Q-Q plot for visual assessment and only flag normality as a real concern when the Q-Q plot shows extreme skewness, heavy tails, or multiple modes.

Struggling With Residual Analysis Assignments?

Our statistics experts help you master regression diagnostics, residual plots, and model validation — for any course level.

Get Statistics Help Now Log In

Residual Plots: How to Read and Interpret Them

If you understand nothing else about residual analysis, understand how to read a residual plot. It is the single most informative diagnostic graphic in regression, and it is the one most consistently misread or ignored. Regression experts at institutions like the University of Wisconsin-Madison, the London School of Economics, and the MIT Statistics Department consistently recommend plotting residuals as the first diagnostic step — before any formal statistical tests — because visual inspection often reveals patterns that summary statistics mask entirely. Creating professional charts for assignments applies directly here — residual plots are standard academic deliverables in regression homework and research papers.

Residuals vs. Fitted Values Plot

The residuals vs. fitted values plot is the foundational diagnostic. Residuals go on the y-axis; fitted values (predicted values) go on the x-axis. The horizontal reference line at zero represents perfect prediction. What you are looking for is a random, featureless scatter of points around this line — no curves, no funnels, no systematic drift. This pattern confirms that the linearity and homoscedasticity assumptions hold simultaneously.

What you should worry about: a curved pattern (parabolic, S-shaped, or otherwise non-linear) indicates the linearity assumption is violated and your model is missing a non-linear relationship. A funnel or fan shape indicates heteroscedasticity — variance is not constant. Both can appear simultaneously, requiring different remedies. In practice, most student assignments require producing this plot and interpreting it in one to two sentences — but interpreting it correctly requires knowing exactly what each pattern implies for the model's validity. Misuse of statistics often stems directly from failing to check the residuals vs. fitted values plot and proceeding with invalid OLS estimates.

The Normal Q-Q Plot

The Normal Q-Q plot (quantile-quantile plot) places the theoretical quantiles of a standard normal distribution on the x-axis and the sample quantiles of your standardized residuals on the y-axis. If the residuals are normally distributed, the points fall along a straight 45-degree diagonal reference line. Deviations from this line indicate specific departures from normality:

If the points form an S-curve — bowing below the line at the lower end and above it at the upper end — the residuals have lighter tails than normal (platykurtic). The reverse S-curve means heavier tails (leptokurtic), which signals more extreme outliers than a normal distribution would produce. If the points curve upward at the right, the distribution is right-skewed; curving downward at the right indicates left skewness. Understanding kurtosis and skewness in distributions translates directly into reading Q-Q plot deviations accurately.

Scale-Location Plot (Spread-Location Plot)

The Scale-Location plot shows the square root of the absolute standardized residuals on the y-axis and fitted values on the x-axis. It is specifically designed to detect heteroscedasticity, often more clearly than the basic residuals vs. fitted values plot. When homoscedasticity holds, the points scatter randomly around a horizontal red line (in R's plot() output) with approximately constant spread. A clear upward or downward trend in the smoothed line indicates that residual variance is increasing or decreasing systematically with the fitted values — a clear heteroscedasticity signal.

Residuals vs. Leverage Plot

The residuals vs. leverage plot is arguably the most complex diagnostic plot but one of the most important for identifying influential observations. Leverage (hat value) measures how much an observation's predictor values differ from the average predictor values — high-leverage points are unusual in the predictor space. The y-axis shows standardized residuals. Cook's Distance contour lines are typically overlaid. An observation that is both high-leverage AND has a large standardized residual is potentially influential — it is pulling the regression line toward itself. Points outside Cook's Distance contours of 0.5 or 1.0 deserve individual investigation. Regression model diagnostics are incomplete without this plot because it identifies cases where one or two observations may be fundamentally distorting the model that all other observations would otherwise produce.

Good Residual Plot Characteristics

  • Points randomly scattered around the zero horizontal line
  • No discernible curve, arch, or systematic trend
  • Consistent spread (width) of points at all fitted values
  • No single point dramatically isolated from the others
  • Q-Q plot points close to the 45-degree diagonal
  • Scale-Location plot shows a flat, horizontal smooth line

Red Flag Patterns in Residuals

  • Curved or arch-shaped pattern → non-linearity violation
  • Fan or funnel expanding left to right → heteroscedasticity
  • Clustered positive then clustered negative → autocorrelation
  • S-curve on Q-Q plot → heavy or light tails, potential outliers
  • Points beyond Cook's Distance contours → influential cases
  • Isolated extreme points → potential outliers requiring review

What Does a "Good" Residual Plot Actually Look Like?

Here is the practical truth that textbooks often undersell: even a "good" residual plot from real-world data will not look perfectly random. There will be some scatter, some points that seem a bit high or low, and some very minor trends in the smoothed line. The question is always proportionality and pattern strength, not perfection. A slight upward trend in the Scale-Location plot in a large sample is probably inconsequential. A clear, strong funnel that doubles the residual spread across the range of fitted values is a genuine problem requiring remedy. Statistical relationships and patterns are what you are assessing — the threshold for action is always a matter of degree, practical significance, and the inferential claims you intend to make. Research on residual plot diagnostics confirms that over-interpretation of minor residual patterns is as common an error as under-interpretation of major ones.

Outliers, Leverage, and Influential Observations

One of the most practically important uses of residual analysis is identifying observations that are not just unusual but actually distorting your entire regression model. These are influential observations — and they are not the same as outliers, though the terms are often conflated. An outlier is an observation with an unusually large residual (unusual in the outcome space). A high-leverage point is unusual in the predictor space. An influential observation is one whose removal would substantially change the estimated regression coefficients. All three are identified through different residual analysis measures, and all three have different implications for how you respond. Correlation vs. causation analyses are particularly sensitive to influential observations, which can create or destroy apparent correlations entirely.

Leverage and the Hat Matrix

The concept of leverage comes from the hat matrix H = X(XᵀX)⁻¹Xᵀ, which projects the observed response vector onto the fitted values. The diagonal elements of this matrix — denoted hᵢᵢ — are the leverage values for each observation. They range from 1/n (minimum possible leverage) to 1 (maximum leverage, meaning one observation perfectly determines one prediction). High leverage means the observation's predictor values are far from the average predictor values; it has the potential to pull the regression line strongly toward itself. The conventional high-leverage threshold is hᵢᵢ > 2p/n, where p is the number of predictors and n is the sample size.

High leverage is not automatically a problem — an observation can have high leverage but still fall close to where the regression line would be without it. The danger arises when a high-leverage observation also has a large residual. That combination produces high influence. Think of a leverage point as a long lever: if your hand (the fitted value) is far from the fulcrum, small movements have large effects on the other end. Principal component analysis is one technique that can reduce the influence of high-leverage predictor configurations by recoding the predictor space into uncorrelated components.

Cook's Distance: Measuring Influence Directly

Cook's Distance (Dᵢ) is the single most widely used measure of overall influence in regression analysis. It quantifies how much all fitted values change when observation i is deleted from the model. Computationally, Cook's Distance combines leverage and discrepancy: Dᵢ = rᵢ² × hᵢᵢ / (p × (1 − hᵢᵢ)), where rᵢ is the internally studentized residual and hᵢᵢ is the hat value. The formula shows that influence increases both with large standardized residuals (discrepancy) and large leverage values — both components contribute independently.

Common thresholds: Cook's Distance greater than 4/n suggests an observation warrants investigation; values greater than 1 are typically considered highly influential. But these are heuristics, not hard rules. In practice, create a Cook's Distance plot (observation index on x-axis, Cook's D on y-axis) and look for observations that clearly separate from the rest of the distribution — those are your candidates for investigation. When you find them, the next step is not automatic deletion but investigation: is the observation a data entry error? A genuinely unusual but valid case? A different subpopulation? The answer determines your response. Cook's Distance in regression diagnostics is discussed extensively in biostatistics literature, where influential observations can directly affect clinical predictions.

DFFITS and DFBETAS

Beyond Cook's Distance, two related measures provide more targeted influence diagnostics. DFFITS (Difference in Fits) measures how much the fitted value for observation i changes when that observation is deleted — standardized by the standard error. It is particularly useful for identifying observations that have outsized influence on their own predicted value. DFBETAS (Difference in Betas) is more granular: it measures how much each individual regression coefficient changes when observation i is deleted. In a multiple regression model with, say, four predictors, DFBETAS produces four influence scores per observation — one for each coefficient — telling you precisely which coefficient relationships are being driven by specific observations. This is invaluable in models where you want to understand whether a particular predictor's apparent effect is genuine or artifact.

When to Remove vs. Retain Influential Observations

This is one of the most judgment-intensive questions in applied regression. The general principle: never remove an observation purely because it is influential. First, investigate its origin. If the value is a data entry error, correct or remove it. If it represents a legitimate but rare case (an extreme but valid data point), consider whether your research question includes or excludes such cases. If the case is valid and relevant, report the model with and without it — and if results differ substantially, acknowledge this sensitivity. Only remove observations that are demonstrably erroneous or explicitly outside your target population. Document every removal decision in your methods section. Transparent handling of influential observations is a mark of rigorous statistical practice. Transparent results reporting in statistics papers requires exactly this kind of documented decision-making around outlier handling.

Heteroscedasticity: Detection and Remediation

Heteroscedasticity is one of the most frequently encountered violations in applied regression analysis, and one of the most frequently mishandled. It occurs when the variance of the residuals varies across observations — typically as a function of one or more predictor variables or of the fitted values themselves. The name comes from Greek: "hetero" (different) + "skedasis" (dispersion). Its opposite, homoscedasticity ("same dispersion"), is the assumption OLS regression requires. Residual analysis is the primary vehicle for detecting heteroscedasticity, and the consequences of ignoring it are substantive: while OLS coefficient estimates remain unbiased under heteroscedasticity, the standard errors are wrong, making all your hypothesis tests and confidence intervals unreliable. Hypothesis testing results become invalid when standard errors are computed under a false homoscedasticity assumption.

Visual Detection: The Fan Pattern

In a residuals vs. fitted values plot, heteroscedasticity typically appears as a fan or funnel shape: residuals spread out more at higher (or lower) fitted values. In cross-sectional economic data, for instance, household income often shows this pattern — wealthier households have more variable spending behavior, so residuals from an income-spending regression get larger at higher income levels. In biological data, cell count measurements often show variance proportional to the mean, creating a funnel in residual plots. The Scale-Location plot is even more sensitive to heteroscedasticity because it shows the square root of absolute standardized residuals — a strong upward or downward trend in the smoothed line is unambiguous evidence.

Formal Tests for Heteroscedasticity

Two formal statistical tests are most widely used. The Breusch-Pagan test (Breusch and Pagan, 1979) regresses the squared residuals on the predictor variables and tests the null hypothesis of constant variance using a chi-square test. The null hypothesis is homoscedasticity; a significant result (p < 0.05) indicates heteroscedasticity. The White test (1980) is a more general version that also includes squared predictors and cross-products, testing for heteroscedasticity that depends on nonlinear functions of the predictors. The White test is preferred when you have no prior reason to expect a particular form of heteroscedasticity. In R, both are implemented in the lmtest package (functions bptest() and white_test()). Chi-square tests underlie these heteroscedasticity tests, making familiarity with the chi-square distribution essential for interpreting their output. The Breusch-Pagan test paper remains a foundational reference in econometrics and regression diagnostics.

Remedies for Heteroscedasticity

When heteroscedasticity is confirmed, several remedies are available depending on its form and severity. Log transformation of the outcome variable is the most common first-line remedy when variance scales with the mean — taking the log of income, price, or biological count data often stabilizes variance substantially. Square-root transformation is appropriate when variance scales with the mean of a count variable (Poisson-type data). Weighted least squares (WLS) is appropriate when you can model the variance function directly — you assign each observation a weight inversely proportional to its estimated variance, so higher-variance observations have less influence. Robust standard errors (Huber-White heteroscedasticity-consistent standard errors, or HC errors) are increasingly the default remedy in econometrics and social science regression — they correct the standard errors without changing the coefficient estimates, making inference valid even when the residuals are heteroscedastic. Regularization methods like Ridge and Lasso address a different set of regression problems but share the practical goal of producing models that generalize reliably rather than fitting training data artifacts.

Need Help With Regression Diagnostics?

From residual plots to Cook's Distance, heteroscedasticity tests to model interpretation — our statistics experts cover every aspect of regression analysis.

Get Expert Help Log In

Autocorrelation in Residuals: The Durbin-Watson Test and Beyond

Autocorrelation — also called serial correlation — occurs when residuals from sequential observations are correlated with each other. It violates the independence assumption of OLS regression. The most common context is time series data, where the error at time t is correlated with the error at time t−1 (first-order autocorrelation). But autocorrelation also arises in spatial data (geographic neighbors have correlated residuals), panel data (repeated measurements on the same individuals), and even cross-sectional data where observations were collected in a systematic order. When autocorrelation is present, OLS standard errors are biased — usually too small — making hypothesis tests appear more significant than they actually are. Time series analysis is the field where detecting and correcting autocorrelation in residuals is most central to valid inference.

The Durbin-Watson Test

The Durbin-Watson (DW) statistic tests for first-order positive autocorrelation in regression residuals. It is computed as: DW = Σ(eᵢ − eᵢ₋₁)² / Σeᵢ². The statistic ranges from 0 to 4. Values close to 2 indicate no autocorrelation; values close to 0 indicate strong positive autocorrelation (consecutive residuals tend to be similar); values close to 4 indicate strong negative autocorrelation (consecutive residuals tend to alternate in sign). The standard interpretation thresholds: DW between 1.5 and 2.5 suggests no serious autocorrelation issue; below 1.5 is a warning for positive autocorrelation; above 2.5 is a warning for negative autocorrelation. Precise critical values depend on sample size and number of predictors and are obtained from Durbin-Watson tables or statistical tables in most statistics references.

Visualizing Autocorrelation: ACF Plots

When the Durbin-Watson test suggests autocorrelation (or when you are working with time series data and need more detailed autocorrelation analysis), the Autocorrelation Function (ACF) plot is the appropriate visual tool. An ACF plot shows the correlation between residuals and their own lagged values at multiple time lags — lag 1, lag 2, lag 3, and so on. Under no autocorrelation, all autocorrelations beyond lag 0 should fall within the confidence bands (typically ±1.96/√n). Significant spikes at specific lags identify the structure of any autocorrelation present: a single significant spike at lag 1 suggests first-order AR process; a slowly decaying pattern suggests a higher-order autoregressive structure. This analysis directly informs how to build a corrected model using ARIMA or GLS approaches.

Remedies for Autocorrelated Residuals

Several remedies exist depending on the source of autocorrelation. If autocorrelation exists because a relevant lagged predictor variable is missing from the model, simply add it — the most common fix in economic and social science data. If autocorrelation reflects genuine temporal dependency in the error structure, use Generalized Least Squares (GLS) with a first-order autoregressive error structure (AR(1)) to model the dependency directly. The Cochrane-Orcutt procedure and the Prais-Winsten estimator are classical transformation approaches for first-order autocorrelation in time series regression. For more complex autocorrelation structures, ARIMA models (which explicitly model the autoregressive and moving average components of temporal dependency) are the appropriate framework. Causal inference studies are particularly sensitive to autocorrelated residuals in panel data settings, where failing to account for serial correlation can dramatically understate standard errors and produce false positives.

Performing Residual Analysis in R, Python, SPSS, and Excel

Understanding residual analysis conceptually is one thing. Doing it in actual statistical software is another — and that is what your assignments, research projects, and professional work actually require. The good news is that every major statistical platform provides built-in tools for the diagnostic procedures described in this guide. The differences are in flexibility, visualization quality, and the depth of available diagnostics. R and Python offer the most comprehensive residual analysis capabilities; SPSS and Excel are more limited but accessible for introductory courses. Statistics homework support for assignments that require software-based residual analysis is available when the implementation itself becomes the challenge.

Residual Analysis in R

R is the most powerful and flexible platform for residual analysis. The base R plot(lm_model) function produces the four standard diagnostic plots (residuals vs. fitted, Q-Q, Scale-Location, and residuals vs. leverage) automatically. The car package adds outlierTest() for formal Bonferroni-corrected outlier testing using externally studentized residuals. The lmtest package provides bgtest() (Breusch-Godfrey test for higher-order autocorrelation), bptest() (Breusch-Pagan test for heteroscedasticity), and resettest() (Ramsey RESET test for non-linearity). The sandwich package computes robust (Huber-White) standard errors.

# Basic residual analysis workflow in R
model <- lm(y ~ x1 + x2, data = mydata)

# Four standard diagnostic plots
par(mfrow = c(2,2))
plot(model)

# Formal tests
library(lmtest)
bptest(model) # Breusch-Pagan test for heteroscedasticity
bgtest(model) # Breusch-Godfrey test for autocorrelation
resettest(model) # RESET test for non-linearity

# Cook's Distance
plot(cooks.distance(model), type = "h")
abline(h = 4/nrow(mydata), col = "red")

Residual Analysis in Python

Python's statsmodels library is the primary platform for regression and residual analysis. The OLSResults object provides direct access to residuals, influence statistics, and diagnostic tests. statsmodels.stats.diagnostic contains het_breuschpagan(), acorr_breuschgodfrey(), and linear_reset(). For visualization, combine matplotlib or seaborn for residual plots with statsmodels.graphics.gofplots.qqplot() for Q-Q plots. The yellowbrick library provides high-quality visualizations specifically designed for regression diagnostics including residual plots, Cook's Distance plots, and prediction error plots.

import statsmodels.api as sm
import matplotlib.pyplot as plt
from statsmodels.stats.diagnostic import het_breuschpagan

# Fit model
X = sm.add_constant(X_vars)
model = sm.OLS(y, X).fit()

# Residuals
residuals = model.resid
fitted = model.fittedvalues

# Residuals vs Fitted plot
plt.scatter(fitted, residuals)
plt.axhline(0, color='red')

# Breusch-Pagan test for heteroscedasticity
bp_test = het_breuschpagan(residuals, X)
print(f"BP p-value: {bp_test[1]:.4f}")

Residual Analysis in SPSS

SPSS handles residual analysis through the Analyze → Regression → Linear menu. The "Save" sub-menu allows you to save unstandardized, standardized, studentized, deleted, and externally studentized residuals directly to your dataset. The "Plots" sub-menu generates residuals vs. fitted value plots, partial regression plots, and normal probability (P-P) plots for normality assessment. The "Statistics" sub-menu provides the Durbin-Watson statistic, collinearity diagnostics, and influence statistics including Cook's Distance, leverage values, and DFBETAS. SPSS is widely used in social science courses at UK and US universities and is the software students most commonly encounter in undergraduate statistics courses. Choosing the right statistical test in SPSS-based assignments also requires understanding which residual diagnostics are available and how to request them.

Residual Analysis in Excel

Excel's Data Analysis Toolpak provides basic regression with residual output — raw residuals, standardized residuals, and predicted values are all available. The toolpak generates a residual plot (residuals vs. each predictor) and a line fit plot. Normal probability plots are available for approximate normality assessment. However, Excel does not provide formal heteroscedasticity tests, Cook's Distance, leverage values, or autocorrelation tests — making it unsuitable for rigorous regression diagnostics. For introductory courses where Excel is the required tool, supplement with manual calculations of Durbin-Watson and visual inspection of residual plots. Statistical calculations in Excel walk through the core statistical functions available within Microsoft Excel for student assignments at the introductory level. Regression diagnostics in clinical research consistently recommend moving beyond Excel for any research-grade analysis requiring valid inference.

How to Perform Residual Analysis: A Complete Step-by-Step Workflow

Knowing the concepts is necessary. Having a workflow you can follow step by step — every time you fit a regression model — is what converts that knowledge into practice. The following procedure covers the complete residual analysis workflow from model fitting through diagnostic interpretation and remediation. It applies to any regression assignment in R, Python, SPSS, SAS, Minitab, or any other statistical platform. The steps are ordered to move from basic visual checks to more advanced influence diagnostics, mirroring the way applied statisticians actually approach model assessment. Research and analysis methodology principles apply directly here — systematic procedure prevents the confirmation bias of only checking what you expect to be right.

1

Fit Your Regression Model and Compute Residuals

Fit your ordinary least squares regression model and save the raw residuals (eᵢ = yᵢ − ŷᵢ), standardized residuals, and fitted values. Most software does this automatically. Verify the number of residuals equals your sample size — one per observation. Check that the sum of residuals is approximately zero (a mathematical property of OLS), which confirms the model was fitted correctly.

2

Produce the Residuals vs. Fitted Values Plot

This is your primary linearity and homoscedasticity check. Plot residuals on the y-axis, fitted values on the x-axis, with a horizontal reference line at zero. Add a smoothed LOESS line to detect non-linear trends. Describe what you see: is the scatter random? Is there a curve? Is there a funnel shape? This description belongs in your assignment's methods or results section.

3

Construct and Interpret the Q-Q Plot

Generate a normal Q-Q plot of the standardized residuals. Assess how closely the points follow the 45-degree diagonal reference line. Describe deviations: S-curves, arches, scattered extreme points. For assignments, state explicitly whether residuals appear approximately normally distributed and what — if anything — the deviations suggest about the data's error distribution.

4

Test for Heteroscedasticity Formally

Run the Breusch-Pagan test (or White test if you suspect non-linear heteroscedasticity). State the null hypothesis (constant variance), the test statistic, degrees of freedom, and p-value. If p < 0.05, reject homoscedasticity and note which remedy is appropriate — typically log transformation or robust standard errors for most applied settings.

5

Test for Autocorrelation (If Applicable)

If your data have a temporal or spatial ordering, run the Durbin-Watson test. Report the DW statistic and interpret it: values between 1.5 and 2.5 suggest no first-order autocorrelation; outside that range, flag the violation and consider remediation (adding lagged predictors, GLS, or ARIMA). For non-time-series cross-sectional data, autocorrelation is less commonly the primary concern.

6

Identify Influential Observations

Compute and plot Cook's Distance for all observations. Flag cases exceeding 4/n. Examine the residuals vs. leverage plot for observations in the upper right quadrant (high leverage AND large residual). Investigate flagged cases individually — check whether they are data errors or valid but unusual cases. Report your findings and decisions transparently.

7

Remediate, Refit, and Re-Diagnose

Apply any indicated remedies: transform variables, add missing non-linear terms, apply robust standard errors, or remove confirmed data errors. Refit the model and repeat the full diagnostic procedure on the new residuals. Confirm that the violations identified in Step 2–6 have been resolved. Document each iteration and the rationale for every modeling decision — this transparency is what separates rigorous regression analysis from ad hoc model tweaking.

Decision Framework: Which Remedy for Which Problem?

Non-linearity in residuals → Add polynomial terms or transform predictors/outcome. Heteroscedasticity → Log-transform outcome (if variance scales with mean), or use robust standard errors (general remedy). Autocorrelated residuals → Add lagged predictors, use GLS, or switch to ARIMA. Non-normality (small sample) → Transform outcome to reduce skewness; consider robust regression. Influential observations → Investigate individually; correct data errors, report sensitivity analysis. If multiple violations occur simultaneously, address non-linearity first (transformations often resolve heteroscedasticity simultaneously), then re-diagnose for remaining issues.

Residual Analysis Across Fields: Economics, Biostatistics, Machine Learning, and Social Science

Residual analysis is not a purely academic exercise. It is a standard professional practice in every field that uses regression modeling — which, in the data-driven world of 2026, means nearly every quantitative field. Understanding how residual analysis is applied in your specific domain makes both your assignments and your professional work more targeted and credible. Below are the most prominent application domains for students in US and UK universities, with field-specific examples and context. Inferential statistics is the overarching framework within which residual analysis operates — every domain below is ultimately concerned with making valid inferences from sample data.

Economics and Econometrics

In econometrics — as taught at institutions like the London School of Economics, the University of Chicago, MIT, and the Federal Reserve — residual analysis is the cornerstone of model validation. Econometric models routinely face heteroscedasticity (economic data with variance scaling by income, firm size, or GDP), autocorrelation (time series of GDP growth, inflation, or unemployment rates), and influential observations (economic crises or structural breaks that are real but dramatically unusual). The Breusch-Pagan test, the Durbin-Watson test, and robust standard errors are standard outputs in any published econometric analysis. Students in UK A-Level Further Mathematics and US AP Statistics programs encounter residual analysis as core curriculum content, and econometrics courses at the master's level treat it as foundational methodology. The foundational Breusch-Pagan paper has been cited thousands of times in the econometrics literature.

Biostatistics and Clinical Research

In biostatistics — practiced at institutions like Johns Hopkins Bloomberg School of Public Health, the Harvard T.H. Chan School of Public Health, the UK Medical Research Council, and the NIHR — residual analysis is integral to validating linear mixed-effects models, survival models, and dose-response regressions. Clinical trials frequently produce heteroscedastic residuals (patient response variability increases with dose level), and longitudinal studies produce autocorrelated residuals within individual patients over time. Cook's Distance is used to identify patients whose outcomes are so extreme that they may disproportionately drive treatment effect estimates. The stakes here are high: a poorly validated regression model in clinical research can lead to incorrect treatment recommendations. Regression diagnostics in clinical studies underscore that residual analysis is not optional in biostatistics — it is a reporting standard.

Machine Learning and Predictive Modeling

In machine learning, the concept of residual analysis extends into model performance diagnostics more broadly — examining prediction errors for systematic patterns, by demographic group (algorithmic fairness), by time period (temporal distribution shift), by geography (spatial generalization), or by feature values (conditional bias). When a neural network or gradient-boosted model produces residuals that are systematically larger for one subgroup of the data, this is exactly the kind of residual pattern that would flag a linearity violation in regression — and the remedy is the same: the model is missing something structurally important about that subgroup. Ridge and Lasso regularization are machine learning techniques that indirectly address some residual analysis concerns by reducing model complexity and thus limiting overfitting, which shows up as systematically small residuals on training data and large residuals on new data.

Social Science Research

In sociology, political science, education research, and psychology — as practiced at institutions like the University of Michigan Institute for Social Research, Oxford's Department of Sociology, and the Educational Testing Service (ETS) — residual analysis validates regression models used to analyze survey data, educational outcomes, and behavioral measures. Social science data frequently shows heteroscedasticity (variation in test scores or survey responses that differs systematically by socioeconomic status or other grouping variables) and non-normality (bounded, skewed, or categorical outcomes that are inappropriately analyzed with OLS). Residual analysis helps researchers decide whether OLS is appropriate or whether ordered logistic regression, multilevel modeling, or structural equation modeling would be more appropriate. Factor analysis is closely related to structural equation modeling, where residual diagnostics take the form of fit indices and residual covariance matrices. Local fit assessment in structural equation models extends residual analysis principles directly into SEM contexts.

Field Most Common Violation Standard Test Used Typical Remedy Notable Institutions
Econometrics Heteroscedasticity, autocorrelation Breusch-Pagan, Durbin-Watson Robust SE, GLS, log transform LSE, Chicago, MIT, Federal Reserve
Biostatistics Non-normality, autocorrelation (longitudinal) Shapiro-Wilk, Cook's Distance Transformation, mixed models Johns Hopkins, Harvard Chan, MRC
Machine Learning Conditional bias, distribution shift Grouped residual analysis Subgroup retraining, regularization Google AI, DeepMind, OpenAI, MIT CSAIL
Social Science Non-normality, heteroscedasticity Breusch-Pagan, Q-Q plots OLS alternatives, robust SE Michigan ISR, Oxford Sociology, ETS
Environmental Science Spatial autocorrelation, non-linearity Moran's I, RESET test Spatial GLS, polynomial terms NOAA, EPA, USGS, UK Environment Agency

Advanced Residual Analysis: Non-Linear Models, GLMs, and Beyond

Residual analysis in its standard form is designed for linear regression with continuous, normally distributed outcomes. But most real-world modeling goes beyond that. Students in econometrics, biostatistics, and machine learning courses quickly encounter logistic regression, Poisson regression, multilevel models, survival models, and other generalized linear models — all of which require adapted forms of residual analysis. The core logic is the same: examine what the model could not explain and check whether it behaves randomly. But the specific residual types, diagnostic plots, and tests differ. Generalized linear models (GLMs) each have their own residual analysis protocols that differ meaningfully from standard OLS diagnostics.

Residuals in Logistic Regression

Logistic regression predicts a binary outcome — the residual is more complex than in OLS. Three residual types are most used. Pearson residuals measure the difference between observed and fitted probabilities standardized by the binomial variance. Deviance residuals are based on the contribution of each observation to the model deviance — they are most commonly used in formal tests. Studentized residuals have a similar interpretation to OLS but with a different formula. Cook's Distance and leverage are both adapted for logistic regression and interpreted similarly to their OLS counterparts. Visual tools include residuals vs. fitted probability plots, index plots of Cook's Distance, and ROC curves (though the last is a performance metric rather than a residual diagnostic). Logistic regression assignments at the graduate level typically require demonstrating these diagnostics explicitly. Influence diagnostics in logistic regression are covered in detail in biostatistics methodology literature.

Residuals in Multilevel Models

Multilevel models (also called hierarchical linear models or mixed-effects models) produce two types of residuals: level-1 residuals (individual-level random errors) and level-2 residuals (group-level random effects). Both should be examined. The level-1 residuals should be approximately normally distributed with constant variance — checked with the same Q-Q and residual plots used in OLS. The level-2 residuals (random effects for groups, schools, hospitals, etc.) should also be approximately normally distributed and unrelated to level-2 predictors. When groups have extreme level-2 residuals, they may be functionally different from the rest of the sample — influencing the model's estimates of between-group variation. In educational research at institutions like the Institute of Education Sciences (IES) in the US and the National Foundation for Educational Research (NFER) in the UK, multilevel residual analysis is a standard component of any school effectiveness or educational assessment study.

Residuals in Machine Learning Models

Modern machine learning models — gradient boosted trees (XGBoost, LightGBM), random forests, and neural networks — produce residuals just like regression models. Examining them is critical for detecting systematic bias. Partial dependence residuals show how prediction errors vary with specific features. SHAP value decomposition (SHapley Additive exPlanations) partitions the residual into feature-level contributions, enabling you to identify which features drive model errors for specific subgroups. In machine learning fairness analysis — increasingly standard at companies like Google, Microsoft, and Amazon — examining residuals stratified by demographic group is a core methodology for detecting algorithmic bias. The residual pattern for protected groups (defined by race, gender, disability status) that differs systematically from the overall residual pattern is exactly the signal that triggers algorithmic fairness interventions. Causal inference with counterfactuals is the statistical framework that connects residual-based bias detection to fairness interventions in machine learning applications.

The Common Thread Across All Models: Whether you are analyzing OLS residuals, logistic regression deviance residuals, multilevel random effects, or machine learning prediction errors — the fundamental question is always the same: Does the structure of the "leftover" variation tell you something important that your model is missing? Randomness in residuals means the model has captured the systematic patterns. Structure in residuals means there is more to learn. That core insight applies universally, regardless of how sophisticated the model or how complex the data.

Statistics Assignment Due Soon?

Our expert statisticians help with every aspect — residual analysis, regression diagnostics, GLMs, R and Python coding, and final writeups.

Order Now Log In

Frequently Asked Questions About Residual Analysis

What is residual analysis? +
Residual analysis is the systematic examination of the differences between observed data values and the values predicted by a statistical regression model. These differences (residuals) are analyzed using diagnostic plots (residuals vs. fitted, Q-Q plots, Scale-Location, residuals vs. leverage) and formal statistical tests (Breusch-Pagan, Durbin-Watson, Shapiro-Wilk) to verify whether the model's assumptions hold — linearity, independence, homoscedasticity, and normality. When assumptions are violated, residual analysis identifies the nature of the violation and points toward appropriate remedies.
What is a residual in regression? +
A residual in regression is the difference between an observed value (y) and the value predicted by the fitted regression model (ŷ): eᵢ = yᵢ − ŷᵢ. Every observation in your dataset has exactly one residual. A positive residual means the model underpredicted (the actual value was higher than predicted); a negative residual means the model overpredicted. The sum of all residuals in an OLS regression is always zero (a mathematical property of least squares estimation). Residuals are the observable estimates of the theoretical error terms in the regression model.
What does a good residual plot look like? +
A good residuals vs. fitted values plot shows points randomly scattered around the horizontal zero line with no discernible pattern. There should be no curve or arch (which would indicate non-linearity), no funnel or fan shape (which would indicate heteroscedasticity), and no clustering of positive or negative residuals (which would indicate autocorrelation). Points should be evenly distributed above and below zero across the entire range of fitted values. The smoothed LOESS line in the plot should be approximately flat and horizontal. Some random scatter is expected — the goal is absence of systematic patterns, not perfect randomness.
What are the types of residuals in regression? +
The main types of residuals are: (1) Raw residuals — basic observed minus predicted difference; (2) Standardized residuals — raw residuals divided by the model's standard error; (3) Internally studentized residuals — standardized by an estimate that accounts for each observation's leverage (hᵢᵢ); (4) Externally studentized (jackknife) residuals — standardized using a model that excludes the observation, making them the most sensitive for outlier detection; (5) PRESS residuals — leave-one-out prediction residuals used to assess cross-validated predictive accuracy. The choice of residual type depends on the diagnostic purpose: standardized for general outlier screening, externally studentized for formal outlier tests, PRESS for predictive validation.
What is heteroscedasticity and how is it detected? +
Heteroscedasticity occurs when the variance of the residuals is not constant across observations — it changes as a function of the predictor variables or fitted values. It is detected visually by a fan or funnel shape in the residuals vs. fitted values plot, and more formally by the Breusch-Pagan test (regresses squared residuals on predictors, uses a chi-square test) or the White test (a more general version including interactions). A significant Breusch-Pagan test (p < 0.05) rejects the null hypothesis of constant variance. Remedies include log-transforming the outcome variable, using robust (Huber-White) standard errors, or weighted least squares regression.
What is Cook's Distance and how do you interpret it? +
Cook's Distance measures how much all fitted values change when a single observation is removed from the model. It combines two components: the observation's leverage (how extreme its predictor values are) and its discrepancy (how large its residual is). An observation with both high leverage and a large residual has high Cook's Distance and is considered influential. Common interpretation thresholds: values greater than 4/n (where n is sample size) flag an observation for investigation; values greater than 1 are considered highly influential. Influential observations should be investigated — not automatically deleted — to determine whether they represent data errors, outliers, or genuinely unusual but valid cases.
What is the Durbin-Watson test? +
The Durbin-Watson test detects first-order autocorrelation (serial correlation) in regression residuals. The test statistic ranges from 0 to 4, with values near 2 indicating no autocorrelation, values near 0 indicating strong positive autocorrelation, and values near 4 indicating negative autocorrelation. Standard interpretation: values between 1.5 and 2.5 are generally considered acceptable; below 1.5 or above 2.5 warrant concern. The Durbin-Watson test is most relevant for time series data where consecutive observations are ordered in time. For higher-order autocorrelation, the Breusch-Godfrey test is a more general alternative.
What is the difference between residuals and errors? +
Errors are the theoretical deviations between observed values and the true population regression line — they are unobservable because the true model is unknown. Residuals are the observable estimates of errors: deviations between observed values and the estimated regression line fitted from sample data. Regression assumptions (normality, independence, homoscedasticity) are stated in terms of errors but verified in practice using residuals. The key practical consequence: residuals are not independent of each other (they satisfy the OLS constraint that they sum to zero), while errors in theory are. This distinction becomes important in small samples where the approximation between residuals and errors is less reliable.
How do you interpret a Q-Q plot? +
A Q-Q (quantile-quantile) plot compares the empirical quantiles of your standardized residuals to the theoretical quantiles of a standard normal distribution. If residuals are normally distributed, points fall approximately along a straight 45-degree diagonal reference line. Deviations indicate non-normality: an S-curve (below the line at low quantiles, above at high) means heavy tails (leptokurtic); the reverse S-curve means light tails; an upward arch suggests right skewness; a downward arch suggests left skewness. In large samples, minor deviations are expected and often inconsequential. Focus on strong, systematic departures that suggest genuine non-normality requiring attention.
What software is best for residual analysis? +
R is the most comprehensive and flexible platform for residual analysis — the base plot(lm_model) function produces four standard diagnostic plots automatically, and packages like car, lmtest, and sandwich add formal tests and robust standard errors. Python (statsmodels + matplotlib) is the preferred alternative, especially for data science workflows. SPSS is widely used in social science and psychology courses and provides standard residual outputs through its regression menus. SAS is standard in pharmaceutical and biostatistics settings. Excel provides basic residual output through the Data Analysis Toolpak but lacks Cook's Distance, leverage values, and formal diagnostic tests — making it unsuitable for rigorous regression analysis.

author-avatar

About Byron Otieno

Byron Otieno is a professional writer with expertise in both articles and academic writing. He holds a Bachelor of Library and Information Science degree from Kenyatta University.

Leave a Reply

Your email address will not be published. Required fields are marked *