The ClusterFinder is a heuristic to find subpopulations of outliers essentially by looking for secondary modes in a density estimate.
ClusterFinder1(X,...)
# S3 method for acomp
ClusterFinder1(X,...,sigma=0.3,radius=1,asig=1,minGrp=3,
robust=TRUE)
A list
a factor representing the group assignments, when the small groups are ignored
a table giving the number of members in each of these groups
a factor representing the found group assignments
a logical vector indicating for each observation,whether it represent a local maximum in the density estimate.
the infered probability to belong to the different groups given as an acomp composition.
a tabel giving the number of members of each group
the density estimated in each observation location
The infered likelihood see this observation, for each of the groups
the dataset to be clustered
Further arguments to MahalanobisDist(X,...,robust=robust,pairwise=TRUE)
numeric: The Bandwidth of the density estimation kernel in a robustly Mahalanobis transformed space. (i.e. in the transform, where the main group has unit variance)
The minimum size of a cluster in a robustly Mahalanobis transformed space. (i.e. in the transform, where the main group has unit variance)
a scaling factor for the geometry of the robustly Mahalanobis transformed space when computing the likelihood of an observation to belong to group (under a Gaussian assumption). Higher values
the minimum size of group to be used. Smaller groups are treated as single outliers
A robustness description for estimating the variance of
the main group. FALSE is probably not a usefull value. However later
other robustness techniques than mcd might be usefull. TRUE
just picks the default method of the package.
K.Gerald v.d. Boogaart http://www.stat.boogaart.de
See outliersInCompositions for a comprehensive introduction
into the outlier
treatment in compositions.
The ClusterFinder is labeled with a number to make clear that this is
just an implementation of some heuristic and not based on some eternal
truth. Other might give better Clusterfinders.
Unlike other Clustering Algorithms the basic model of this
algorithm assumes that there is one dominating subpopulation and an
unkown number of smaller subpopulations with a similar covariance
structure but a different mean. The algorithm thus first estimates the
covariance structure of the main population by a robust location scale
estimator. Then it uses a simplified (Gaussian) kernel density
estimator to find
nonrandom secondary modes. The it tries to a assign the different
observations according to discrimination analysis model to the
different modes. Groups under a given size are considered as single
outliers forming a seperate group. In this way the number of clusters
is kept low even if there are many erratic measurements in the dataset.
The main use of the
clusters is descriptive plotting. The advantage of these cluster
against other cluster techniques like k-mean or hclust is that it does
not tear appart the central mass of the data, as these methods do to
make the clusters as compact as possible.
data(SimulatedAmounts)
cl <- ClusterFinder1(sa.outliers5,sigma=0.4,radius=1)
plot(sa.outliers5,col=as.numeric(cl$types),pch=as.numeric(cl$types))
legend(1,1,legend=levels(cl$types),xjust=1,col=1:length(levels(cl$types)),
pch=1:length(levels(cl$types)))
Run the code above in your browser using DataLab