spark-kmedoids (homepage)

Spark implementation of k-medoids clustering algorithm

@tdebatty / (1)

Spark implementation of k-medoids. There are actually multiple algorithms implemented, like CLARANS and sibblings.


Tags

  • 1|machine learning
  • 1|clustering

How to

Include this package in your Spark Applications using:

spark-shell, pyspark, or spark-submit

> $SPARK_HOME/bin/spark-shell --packages tdebatty:spark-kmedoids:0.1.2

sbt

If you use the sbt-spark-package plugin, in your sbt build file, add:

spDependencies += "tdebatty/spark-kmedoids:0.1.2"

Otherwise,

resolvers += "Spark Packages Repo" at "https://repos.spark-packages.org/"

libraryDependencies += "tdebatty" % "spark-kmedoids" % "0.1.2"

Maven

In your pom.xml, add:
<dependencies>
  <!-- list of dependencies -->
  <dependency>
    <groupId>tdebatty</groupId>
    <artifactId>spark-kmedoids</artifactId>
    <version>0.1.2</version>
  </dependency>
</dependencies>
<repositories>
  <!-- list of other repositories -->
  <repository>
    <id>SparkPackagesRepo</id>
    <url>https://repos.spark-packages.org/</url>
  </repository>
</repositories>

Releases

Version: 0.1.2 ( d40262 | zip | jar ) / Date: 2017-09-24 / License: MIT