聚类(clustering) 属于非监督学习 (unsupervised learning)
无类别标记(class label)
3.1 Clustering 中的经典算法,数据挖掘十大经典算法之一
3.2 算法接受参数 k ;然后将事先输入的n个数据对象划分为 k个聚类以便使得所获得的聚类满足:同一
聚类中的对象相似度较高;而不同聚类中的对象相似度较小。
3.3 算法思想:
以空间中k个点为中心进行聚类,对最靠近他们的对象归类。通过迭代的方法,逐次更新各聚类中心
的值,直至得到最好的聚类结果
3.4 算法描述:
(1)适当选择c个类的初始中心;
(2)在第k次迭代中,对任意一个样本,求其到c各中心的距离,将该样本归到距离最短的中心所在
的类;
(3)利用均值等方法更新该类的中心值;
(4)对于所有的c个聚类中心,如果利用(2)(3)的迭代法更新后,值保持不变,则迭代结束,
否则继续迭代。
3.5 算法流程:
输入:k, data[n];
(1) 选择k个初始中心点,例如c[0]=data[0],…c[k-1]=data[k-1];
(2) 对于data[0]….data[n], 分别与c[0]…c[k-1]比较,假定与c[i]差值最少,就标记为i;
(3) 对于所有标记为i点,重新计算c[i]={ 所有标记为i的data[j]之和}/标记为i的个数;
(4) 重复(2)(3),直到所有c[i]值的变化小于给定阈值。
停止
优点:速度快,简单
缺点:最终结果跟初始点选择相关,容易陷入局部最优,需直到k值
Reference: http://croce.ggf.br/dados/K%20mean%20Clustering1.pdf
import numpy as np
# Function: K Means
# -------------
# K-Means is an algorithm that takes in a dataset and a constant
# k and returns k centroids (which define clusters of data in the
# dataset which are similar to one another).
# 数据集, 分为几类, 最大循环次数
def kmeans(X, k, maxIt):
numPoints, numDim = X.shape
# 给数据集多加一列
dataSet = np.zeros((numPoints, numDim + 1))
dataSet[:, :-1] = X
# Initialize centroids randomly
#随机选取两个中心点,一共几行,随机选K行
centroids = dataSet[np.random.randint(numPoints, size = k), :]
#centroids = dataSet[0:2, :] 自己选取中心点
#Randomly assign labels to initial centorid
# 打上类标签
centroids[:, -1] = range(1, k +1)
# Initialize book keeping vars.
iterations = 0 # 从0开始循环
oldCentroids = None # 旧的中心点
# Run the main k-means algorithm
# oldCentroids: 旧的中心点
# centroids:新的中心点,新旧相同,停止
# iterations: 记录循环多少次
# maxIt:循序循环的最大次数
# 只要不停止就做什么?
while not shouldStop(oldCentroids, centroids, iterations, maxIt):
print "iteration: \n", iterations
print "dataSet: \n", dataSet
print "centroids: \n", centroids
# Save old centroids for convergence test. Book keeping.
# 如果直接使用等号=, 两个会同时变化呢
oldCentroids = np.copy(centroids)
iterations += 1
# Assign labels to each datapoint based on centroids
# 把数据集重新归类
updateLabels(dataSet, centroids)
# Assign centroids based on datapoint labels
# 计算新的中心点
centroids = getCentroids(dataSet, k)
# We can get the labels too by calling getLabels(dataSet, centroids)
return dataSet
# Function: Should Stop
# -------------
# Returns True or False if k-means is done. K-means terminates either
# because it has run a maximum number of iterations OR the centroids
# stop changing.
def shouldStop(oldCentroids, centroids, iterations, maxIt):
if iterations > maxIt:
return True
# 值相等 - 对象相等
return np.array_equal(oldCentroids, centroids)
# Function: Get Labels
# -------------
# Update a label for each piece of data in the dataset.
# 中心点事数据中的,计算每一行距离中心点的距离并分类
def updateLabels(dataSet, centroids):
# For each element in the dataset, chose the closest centroid.
# Make that centroid the element's label.
numPoints, numDim = dataSet.shape
for i in range(0, numPoints):
# 对每一行的最后标签列重新计算赋值,把最近中心点的label赋值该行
dataSet[i, -1] = getLabelFromClosestCentroid(dataSet[i, :-1], centroids)
# dataSetRow:一行一个实例。中心点(K行,列数相同的矩阵)
def getLabelFromClosestCentroid(dataSetRow, centroids):
# 初始化label
label = centroids[0, -1];
minDist = np.linalg.norm(dataSetRow - centroids[0, :-1])
for i in range(1 , centroids.shape[0]):
# np.linalg.norm: 计算两个向量值,(x1 - n1)^2 +(x2 - n2)^2
# centroids[i, :-1] 除去最后一列
dist = np.linalg.norm(dataSetRow - centroids[i, :-1])
if dist < minDist:
minDist = dist
label = centroids[i, -1]
print "minDist:", minDist
return label
# Function: Get Centroids
# -------------
# Returns k random centroids, each of dimension n.
def getCentroids(dataSet, k):
# Each centroid is the geometric mean of the points that
# have that centroid's label. Important: If a centroid is empty (no points have
# that centroid's label) you should randomly re-initialize it.
result = np.zeros((k, dataSet.shape[1]))
for i in range(1, k + 1):
# 找出所有事标签相同的类
oneCluster = dataSet[dataSet[:, -1] == i, :-1]
# 对着同一类的数据求均值 - 中心的新坐标值
result[i - 1, :-1] = np.mean(oneCluster, axis = 0)
result[i - 1, -1] = i
return result
x1 = np.array([1, 1])
x2 = np.array([2, 1])
x3 = np.array([4, 3])
x4 = np.array([5, 4])
# 纵向堆叠 4x2
testX = np.vstack((x1, x2, x3, x4))
result = kmeans(testX, 2, 10)
print "final result:"
print result