1994 | OriginalPaper | Chapter
Principal components and model selection
Authors : Beat E. Neuenschwander, Bernard D. Flury
Published in: Selecting Models from Data
Publisher: Springer New York
Included in: Professional Book Archive
Activate our intelligent search to find suitable subject content or patents.
Select sections of text to find matching patents with Artificial Intelligence. powered by
Select sections of text to find additional relevant content using AI-assisted search. powered by
Let the kp-variate random vector X be partitioned into k subvectors X i of dimension p each, and let the covariance matrix Ψ of X be partitioned analogously into submatrices Ψ ij . Based on principal component analysis we suggest a hierarchy of models, where the lowest level assumes independence of all X i , with identical covariance matrices, and the highest level makes no assumptions about Ψ beyond positive definiteness. The intermediate levels are characterized by a common orthogonal matrix ß which diagonalizes Ψ ij for all pairs (i, j), i.e., Ψ ij = ßΛ ij ß′, where Λ ij is diagonal. The hierarchy is motivated by both a practical example and theoretical arguments. Model selection based on likelihood ratio tests and information criteria is discussed.