How to apply K-Medoids in PySpark?

the pyspark ml library does not provide any clustering methods for K-Medoids. So my question is, how can one apply K-Medoids in a pyspark context?

Topic pyspark apache-spark python clustering

Category Data Science


There is a k-medoids clustering for PySpark at spark-packages.org/package/tdebatty/spark-kmedoids and the source code is github.com/tdebatty/spark-kmedoids.

It can be installed with:

> $SPARK_HOME/bin/spark-shell --packages tdebatty:spark-kmedoids:0.1.2

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.