Standardized coefficient
{{Short description|Estimates from regression analysis on data with unit variance}}
{{distinguish|Beta (finance)}}
{{Multiple issues|
{{More citations needed|date=December 2010}}
{{more footnotes|date=December 2010}}
}}
In statistics, standardized (regression) coefficients, also called beta coefficients or beta weights, are the estimates resulting from a regression analysis where the underlying data have been standardized so that the variances of dependent and independent variables are equal to 1.{{Citation|last=Menard|first=S.|editor1-last=Lewis-Beck|editor1-first=M.S.|editor2-last=Bryman|editor2-first=A.|editor3-last=Liao|editor3-first=T.F.|title=The Sage Encyclopedia of Social Science Research Methods|chapter=Standardized regression coefficients|pages=1069–1070|publisher=Sage Publications|place=Thousand Oaks, CA, USA|year=2004|doi=10.4135/9781412950589.n959|isbn=9780761923633 }} Therefore, standardized coefficients are unitless and refer to how many standard deviations a dependent variable will change, per standard deviation increase in the predictor variable.
Usage
Standardization of the coefficient is usually done to answer the question of which of the independent variables have a greater effect on the dependent variable in a multiple regression analysis where the variables are measured in different units of measurement (for example, income measured in dollars and family size measured in number of individuals).
It may also be considered a general measure of effect size, quantifying the "magnitude" of the effect of one variable on another.
For simple linear regression with orthogonal predictors, the standardized regression coefficient equals the correlation between the independent and dependent variables.
Implementation
A regression carried out on original (unstandardized) variables produces unstandardized coefficients. A regression carried out on standardized variables produces standardized coefficients. Values for standardized and unstandardized coefficients can also be re-scaled to one another subsequent to either type of analysis.
Suppose that is the regression coefficient resulting from a linear regression (predicting by ). The standardized coefficient simply results as , where and are the (estimated) standard deviations of and , respectively.
Sometimes, standardization is done only without respect to the standard deviation of the regressor (the independent variable ).{{cite journal|last1=Greenland|first1=S.|last2=Schlesselman|first2=J. J.|last3=Criqui|first3=M. H.|title=The fallacy of employing standardized regression coefficients and correlations as measures of effect|journal=American Journal of Epidemiology|volume=123|issue=2|year=1986|pages=203–208|doi=10.1093/oxfordjournals.aje.a114229|pmid=3946370 |doi-access=free}}{{cite journal|last1=Newman|first1=T. B.|last2=Browner|first2=W. S.|title=In defense of standardized regression coefficients|journal=Epidemiology|volume=2|issue=5|year=1991|pages=383–386|doi=10.1097/00001648-199109000-00014|pmid=1742391 |doi-access=free}}
Advantages and disadvantages
Standardized coefficients' advocates note that the coefficients are independent of the involved variables' units of measurement (i.e., standardized coefficients are unitless), which makes comparisons easy.
Critics voice concerns that such a standardization can be very misleading.{{cite journal|last1=Greenland|first1=S.|last2=Maclure|first2=M.|last3=Schlesselman|first3=J. J.|last4=Poole|first4=C.|last5=Morgenstern|first5=H.|title=Standardized regression coefficients: A further critique and review of some alternatives|journal=Epidemiology|volume=2|issue=5|year=1991|pages=387–392|doi=10.1097/00001648-199109000-00016|pmid=1742393 |doi-access=free}}
Due to the re-scaling based on sample standard deviations, any effect apparent in the standardized coefficient may be due to confounding with the particularities (especially: variability) of the involved data sample(s).
Also, the interpretation or meaning of a "one standard deviation change" in the regressor may vary markedly between non-normal distributions (e.g., when skewed, asymmetric or multimodal).
Terminology
See also
References
{{Reflist}}
Further reading
- {{cite book |first1=Larry D. |last1=Schroeder |first2=David L. |last2=Sjoquist |first3=Paula E. |last3=Stephan |year=1986 |title=Understanding Regression Analysis |publisher=Sage Publications |isbn=0-8039-2758-4 |pages=[https://archive.org/details/understandingreg00larr/page/31 31–32] |url=https://archive.org/details/understandingreg00larr/page/31 }}
- {{cite book |first1=Eric |last1=Vittinghoff |first2=David V. |last2=Glidden |first3=Stephen C. |last3=Shiboski |first4=Charles E. |last4=McCulloch |year=2005 |title=Regression Methods in Biostatistics: Linear, Logistic, Survival, and Repeated Measures Models |publisher=Springer |pages=75–76 |isbn=0-387-20275-7 }}
- {{cite book |first1=J.|last1=Neter|first2=M. H.|last2=Kutner|first3=C.J.|last3=Nachtsheim|first4=W.|last4=Wasserman|year=1996|title=Applied Linear Statistical Models|publisher=McGraw-Hill|section=7.5 Standardized multiple regression model|pages=281–284|edition=4th|isbn=0-256-11736-5}}
External links
- [http://www.jerrydallal.com/LHSP/importnt.htm Which Predictors Are More Important?] - why standardized coefficients are used
{{DEFAULTSORT:Standardized Coefficient}}