Документ взят из кэша поисковой машины. Адрес
оригинального документа
: http://www.sao.ru/precise/Midas_doc/doc/94NOV/vol2/node224.html
Дата изменения: Fri Feb 23 14:02:03 1996
Дата индексирования: Tue Oct 2 18:37:06 2012
Кодировка:
Поисковые слова: п п п п п п п п п п п п п п п п п п п п р п п р п п р п п р п п р п п р п
|
Next: Cluster Analysis
Up: Multivariate Analysis Methods
Previous: Introduction
Among the objectives of Principal Components Analysis are the
following.
- dimensionality reduction;
- the determining of linear combinations of variables;
- feature selection: the choosing of the most useful variables;
- visualisation of multidimensional data;
- identification of underlying variables;
- identification of groups of objects or of
outliers.
The tasks required of the analyst to carry these out are as
follows:
- In case of a table of dimensions , each of the
n rows or objects can be regarded as an m--dimensional vector.
Finding a set of principal axes allows the objects to
be adequately characterised on a smaller number of
(artificial) variables. This is advantageous as a prelude to
further analysis
as the dimensions may often be ignored as
constituting noise; and, secondly, for storage economy
(sufficient information from the initial table is now
represented in a table with columns).
Reduction of dimensionality is practicable if the first
new axes account for approximately 75 % or more of
the variance. There is no set threshold, --- the analyst
must judge. The cumulative percentage of variance
explained by the principal axes is consulted in order
to make this choice.
- If the eigenvalue is zero, the variance of projections on the
associated eigenvector
is zero. Hence the eigenvector is reduced to a point. If this point is
additionally the origin (i.e. the data is centred), then
this allows linear combinations between
the variables to be found. In fact, we can go a
good deal further: by analysing second--order variables,
defined from the given variables, quadratic dependencies
can be straightforwardly sought. This means, for example,
that in analysing three variables, , , and ,
we would also input the variables , , ,
, , and . If the linear combination
exists, then we would find
it. Similarly we could feed in the logarithms or other functions
of variables.
- In feature selection we want to simplify the task
of characterising each object by a set of attributes.
Linear combinations among attributes must be found; highly
correlated attributes (i.e. closely located attributes in the
new space) allow some attributes to be
removed from consideration; and the proximity of attributes
to the new axes indicate the more relevant and important
attributes.
- In order to provide a convenient representation of
multidimensional data, planar plots are necessary. An
important consideration is the adequacy of the planar
representation: the percentage variance explained by the
pair of axes defining the plane must be looked at here.
- PCA is often motivated by the search for latent
variables. Often it is relatively easy to label the
highest or second highest components, but it becomes
increasingly difficult as less relevant axes are
examined. The objects with the highest loadings or
projections on the axes (i.e. those which are placed
towards the extremities of the axes) are usually worth
examining: the axis may be characterisable as a spectrum
running from a small number of objects with high positive
loadings to those with high negative loadings.
- A visual inspection of a planar plot indicates which
objects are grouped together, thus indicating that they
belong to the same family or result from the same
process. Anomalous objects can also be detected, and
in some cases it might be of interest to redo the
analysis with these excluded because of the perturbation
they introduce.
Next: Cluster Analysis
Up: Multivariate Analysis Methods
Previous: Introduction
Pascal Ballester
Tue Mar 28 16:52:29 MET DST 1995