A connector for Spark that allows reading and writing to/from Redis cluster
@RedisLabs / (1)
A library for reading and writing data from and to Redis with Apache Spark, for Spark SQL and DataFrames.
Spark-Redis provides access to all of Redis' data structures - String, Hash, List, Set and Sorted Set - from Spark as RDDs. The library can be used both with Redis stand-alone as well as clustered databases. When used with Redis cluster, Spark-Redis is aware of its partitioning scheme and adjusts in response to resharding and node failure events.
Include this package in your Spark Applications using:
spark-shell, pyspark, or spark-submit
> $SPARK_HOME/bin/spark-shell --packages RedisLabs:spark-redis:0.1.1
If you use the sbt-spark-package plugin, in your sbt build file, add:
spDependencies += "RedisLabs/spark-redis:0.1.1"
resolvers += "Spark Packages Repo" at "http://dl.bintray.com/spark-packages/maven" libraryDependencies += "RedisLabs" % "spark-redis" % "0.1.1"
MavenIn your pom.xml, add:
<dependencies> <!-- list of dependencies --> <dependency> <groupId>RedisLabs</groupId> <artifactId>spark-redis</artifactId> <version>0.1.1</version> </dependency> </dependencies> <repositories> <!-- list of other repositories --> <repository> <id>SparkPackagesRepo</id> <url>http://dl.bintray.com/spark-packages/maven</url> </repository> </repositories>