k
clusters ``around
medoids'', a more robust version of K-means.pam(x, k, diss = inherits(x, "dist"),
metric = "euclidean", stand = FALSE, cluster.only = FALSE,
keep.diss = !diss && !cluster.only && n < 100,
keep.data = !diss && !cluster.only)
diss
argument.In case of a matrix or data frame, each row corresponds to an observation, and each column corresponds to a variable. All
dist
or
dissimilarity
objects), then x
will be considered as a
dissimilarity matrix. If FALSE, then x
will be considered as
a matrix of observations by varx
are
standardized before calculating the dissimilarities. Measurements
are standardized for each variable (column), by subtracting the
variable's mean value and dividing by the variable's mex
should be kept in the result. Setting
these to FALSE
can give much smaller results and hence even save
memory allocation time."pam"
representing the clustering. See
?pam.object
for details.pam
is fully described in chapter 2 of Kaufman and Rousseeuw
(1990). Compared to the k-means approach in kmeans
, the
function pam
has the following features: (a) it also accepts a
dissimilarity matrix; (b) it is more robust because it minimizes a sum
of dissimilarities instead of a sum of squared euclidean distances;
(c) it provides a novel graphical display, the silhouette plot (see
plot.partition
) (d) it allows to select the number of clusters
using mean(silhouette(pr))
on the result
pr <- pam(..)
, or directly its component
pr$silinfo$avg.width
, see also pam.object
. When cluster.only
is true, the result is simply a (possibly
named) integer vector specifying the clustering, i.e.,
pam(x,k, cluster.only=TRUE)
is the same as
pam(x,k)$clustering
but computed more efficiently.
The pam
-algorithm is based on the search for k
representative objects or medoids among the observations of the
dataset. These observations should represent the structure of the
data. After finding a set of k
medoids, k
clusters are
constructed by assigning each observation to the nearest medoid. The
goal is to find k
representative objects which minimize the sum
of the dissimilarities of the observations to their closest
representative object.
The algorithm first looks for a good initial set of medoids (this is
called the build phase). Then it finds a local minimum for the
objective function, that is, a solution such that there is no single
switch of an observation with a medoid that will decrease the
objective (this is called the swap phase).
agnes
for background and references;
pam.object
, clara
, daisy
,
partition.object
, plot.partition
,
dist
.## generate 25 objects, divided into 2 clusters.
x <- rbind(cbind(rnorm(10,0,0.5), rnorm(10,0,0.5)),
cbind(rnorm(15,5,0.5), rnorm(15,5,0.5)))
pamx <- pam(x, 2)
pamx
summary(pamx)
plot(pamx)
pam(daisy(x, metric = "manhattan"), 2, diss = TRUE)
data(ruspini)
## Plot similar to Figure 4 in Stryuf et al (1996)
plot(pam(ruspini, 4), ask = TRUE)
<testonly>plot(pam(ruspini, 4))</testonly>
Run the code above in your browser using DataLab