Skip to content

huang06/spark_svdpp

Repository files navigation

PySpark implementation of SVD++ for Top-N Recommendation

pyspark-flow

Prerequisites

You need to install Apache Hadoop and Apache Spark on every nodes of the cluster.

Install Hadoop

tar zxvf hadoop-3.y.z.tgz
ln -s /your/hadoop/path/hadoop-3.x.z /your/hadoop/path/hadoop

Install Spark

tar zxvf spark-2.y.z-bin-hadoop2.7.tgz
ln -s /your/spark/path/spark-2.y.z /your/spark/path/spark

Getting Started

Create the Python environment

make python

Run tests

make test

Run example

make example

References

About

An implementation of SVD++ using PySpark

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published