K-means算法(Spark Demo)

清华大佬耗费三个月吐血整理的几百G的资源,免费分享!....>>>

import java.util.Random
import spark.SparkContext
import spark.SparkContext._
import spark.examples.Vector._
 
object SparkKMeans {
    /**
     * line -> vector
     */
def parseVector (line: String) : Vector = {
        return new Vector (line.split (' ').map (_.toDouble) )
    }
 
    /**
     * 计算该节点的最近中心节点
     */
def closestCenter (p: Vector, centers: Array[Vector]) : Int = {
        var bestIndex = 0
        var bestDist = p.squaredDist (centers (0) ) //差平方之和
        for (i < - 1 until centers.length) {
            val dist = p.squaredDist (centers (i) )
            if (dist < bestDist) {
                bestDist = dist
                bestIndex = i
            }
        }
        return bestIndex
    }
 
def main (args: Array[String]) {
        if (args.length < 3) {
            System.err.println ("Usage: SparkKMeans <master> <file> <dimensions> <k> <iters>")
            System.exit (1)
        }
        val sc = new SparkContext (args (0), "SparkKMeans")
        val lines = sc.textFile (args (1), args (5).toInt)
                    val points = lines.map (parseVector (_) ).cache() //文本中每行为一个节点,再将每个节点转换成Vector
                                 val dimensions = args (2).toInt //节点的维度
                                         val k = args (3).toInt //聚类个数
                                                 val iterations = args (4).toInt //迭代次数
 
                                                         // 随机初始化k个中心节点
                                                         val rand = new Random (42)
        var centers = new Array[Vector] (k)
        for (i < - 0 until k)
            centers (i) = Vector (dimensions, _ => 2 * rand.nextDouble - 1)
                          println ("Initial centers: " + centers.mkString (", ") )
                          val time1 = System.currentTimeMillis()
            for (i < - 1 to iterations) {
                println ("On iteration " + i)
 
                // Map each point to the index of its closest center and a (point, 1) pair
                // that we will use to compute an average later
                val mappedPoints = points.map { p => (closestCenter (p, centers), (p, 1) ) }
 
                val newCenters = mappedPoints.reduceByKey {
                case ( (sum1, count1), (sum2, count2) ) => (sum1 + sum2, count1 + count2) //(向量相加, 计数器相加)
                    } .map {
                case (id, (sum, count) ) => (id, sum / count) //根据前面的聚类,重新计算中心节点的位置
                    } .collect
 
                // 更新中心节点
                for ( (id, value) < - newCenters) {
                    centers (id) = value
                }
            }
                       val time2 = System.currentTimeMillis()
                                   println ("Final centers: " + centers.mkString (", ") + ", time: " + (time2 - time1) )
    }
}