1992 | OriginalPaper | Buchkapitel
Normal Theory Models and Some Extensions
verfasst von : James K. Lindsey
Erschienen in: The Analysis of Stochastic Processes using GLIM
Verlag: Springer New York
Enthalten in: Professional Book Archive
Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.
Wählen Sie Textabschnitte aus um mit Künstlicher Intelligenz passenden Patente zu finden. powered by
Markieren Sie Textabschnitte, um KI-gestützt weitere passende Inhalte zu finden. powered by
One of the most widely used tools in all of statistics is linear regression. This is often misnamed least squares regression, but a least squares estimation refers to a deterministic process, whereby the best straight line is fitted through a series of points. In statistical analysis, the interpretation is much different although the technical calculations remain the same. Normal theory linear regression carries the assumption that the response variable has a normal or Gaussian distribution:1.1$$ f(y;\mu ,{{\sigma }^{2}}) = \exp [{{(y - \mu )}^{2}}/(2{{\sigma }^{2}})]/\sqrt {{2\pi {{\sigma }^{2}}}} $$The mean of this distribution changes in some deterministic way with the values of the explanatory variable(s), e.g. 1.2$$ {{\mu }_{i}} = {{\beta }_{0}} + \sum\limits_{j} {{{\beta }_{j}}{{X}_{{ij}}}} $$ while the variance remains constant. Then, the regression equation specifies how the mean of the distribution changes for each value of the explanatory variable(s); individual observations will be dispersed about the mean with the given variance. This is illustrated in Figure 1.1.