Spark DBSCAN is an implementation of the DBSCAN clustering algorithm on top of Apache Spark . It also includes 2 simple tools which will help you choose parameters of the DBSCAN algorithm.
This software is EXPERIMENTAL , it supports only Euclidean and Manhattan distance measures ( why? ) and it is not well optimized yet. I tested it only on small datasets (millions of records with 2 features in each record).
You can use Spark DBSCAN as a standalone application which you can submit to a Spark cluster ( Learn how ). Alternatively, you can include it into your own app - its API is documented and easy to use ( Learn how ).
Learn more about:
I was glad to receive contributions from other people and I'd like to say thank you:
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。