When the k-means clustering algorithm runs, it uses a randomly generated seed to determine the starting centroids of the clusters. wiki article

If the feature variables exhibit patterns that automatically group them into visible clusters, then the starting seed will not have an impact on the final cluster memberships. However, if the data is evenly distributed, then we might end up with different cluster members based on the initial random variable. An example for such a behavior is shown.

R is used for the experiment. The code to load the data and the contents of the data are as follows. We try to group the samples based on two feature variables - age and bmi.

`data <- read.csv("AgeBMIEven.csv")`

`str(data)`

## 'data.frame': 1338 obs. of 2 variables:

```
## $ age: int 19 18 28 33 32 31 46 37 37 60 ...
## $ bmi: num 27.9 33.8 33 22.7 28.9 ...
```

summary(data)

## age bmi

```
## Min. :18.0 Min. :16.0
## 1st Qu.:27.0 1st Qu.:26.3
## Median :39.0 Median :30.4
## Mean :39.2 Mean :30.7
## 3rd Qu.:51.0 3rd Qu.:34.7
## Max. :64.0 Max. :53.1
```

plot(data$age, data$bmi)

```
```

`As we an see from the above plot, the data points are distributed almost evenly all over the scatter plot. The initial cluster center position would affect the final cluster shapes and memberships. We run the clustering 4 times to group this data as 4 clusters and plot the clusters outputs here.`

```
```

```
par(mfrow=c(2,2))
for (i in 1:4 ) {
clusters<- kmeans(data,4)
plot(data$age, data$bmi, col=clusters$cluster)
}
```

Each time the clustering algorithm runs, it is going to pick a random seed and that seem to impact the shapes and memberships of the clusters. The first two runs generate the same groups, but the next 2 give different groupings of the data. Setting the seed explicitly to a specific value is required to generate the same results every time.

This is why I use Fuzzy C-Means. By calculating the associativity of each node with each centroid and controlling the error function we can guarantee the same result and an even distribution each time.

ReplyDeleteNice test. Could you please repeat the experiment with kmeans++ (http://en.wikipedia.org/wiki/K-means%2B%2B) for initial centroid positioning, to see whether the same variance of output occurs?

ReplyDeleteWell done! It is so well written and interactive. Keep writing such brilliant piece of work. Glad i came across this post. Last night even i saw similar wonderful Python tutorial on youtube so you can check that too for more detailed knowledge on Python.https://www.youtube.com/watch?v=HcsvDObzW2U

ReplyDelete