Multiple Linear Regression Calculator

Run a Multiple Linear Regression on your dataset right in the browser. Enter your X1, X2, and Y values (one per line or comma-separated), set your significance level (α), and the calculator returns the regression equation, , adjusted R², F-statistic, p-value, and individual coefficient p-values. Supports up to two predictor variables with optional outlier exclusion and data transformations.

Enter X1 values separated by commas, spaces, or new lines.

Enter X2 values separated by commas, spaces, or new lines.

Enter Y (dependent variable) values. Must match the count of X1 and X2.

Typically 0.05. Used to determine statistical significance.

Results

R² (Coefficient of Determination)

--

Adjusted R²

--

F-Statistic

--

Model P-Value

--

Intercept (β₀)

--

Coefficient X1 (β₁)

--

Coefficient X2 (β₂)

--

Standard Error (S)

--

Number of Observations (n)

--

Actual vs Fitted Values

Results Table

Frequently Asked Questions

What is multiple linear regression?

Multiple linear regression is a statistical technique that models the relationship between one dependent variable (Y) and two or more independent predictor variables (X1, X2, …). It fits a linear equation of the form Y = β₀ + β₁X1 + β₂X2 + ε, where β values are the estimated coefficients and ε is the error term.

What does R² mean in multiple linear regression?

R² (the coefficient of determination) measures what proportion of the variance in Y is explained by the predictors. A value of 0.85 means 85% of the variability in Y is accounted for by X1 and X2. Higher values indicate a better fit, but adding more predictors will always increase R² even if they are not meaningful — use Adjusted R² to account for this.

What is Adjusted R² and why does it matter?

Adjusted R² penalizes the model for adding predictors that do not improve explanatory power. Unlike plain R², it can decrease if a new variable adds noise rather than signal. It is generally a more honest measure of model fit when comparing models with different numbers of predictors.

How do I interpret the F-statistic and its p-value?

The F-statistic tests whether the regression model as a whole is statistically significant — i.e., whether at least one predictor is meaningfully related to Y. If the model p-value is below your chosen significance level α (e.g., 0.05), you reject the null hypothesis that all coefficients are zero and conclude the model has explanatory value.

What do the individual coefficient p-values tell me?

Each coefficient's p-value tests whether that specific predictor has a statistically significant relationship with Y, holding all other predictors constant. A p-value below α suggests the predictor contributes meaningfully to the model. A high p-value suggests the predictor may not be necessary.

When should I use a data transformation like Log or Sqrt?

Transformations are useful when your data violates the linearity or normality assumptions. For example, if Y grows exponentially with X, a log transformation of Y can linearize the relationship. If data is right-skewed, a square root transformation can reduce skewness and stabilize variance.

What does 'Force intercept through zero' do?

When you force zero intercept, the regression line is constrained to pass through the origin (0, 0), meaning β₀ = 0. This is appropriate only when you have theoretical or physical justification that Y must be zero when all predictors are zero. In most practical cases, you should include the intercept.

How are outliers detected and handled?

When 'Exclude Outliers' is selected, data points whose standardized residuals exceed ±2.5 standard deviations from the mean are removed before fitting the model. This can improve model fit when extreme values are due to data entry errors or anomalies, but should be used carefully so legitimate data is not discarded.

More Statistics Tools