All Projects → neo4j-contrib → Neo4j Spark Connector

neo4j-contrib / Neo4j Spark Connector

Licence: apache-2.0
Neo4j Connector for Apache Spark, which provides bi-directional read/write access to Neo4j from Spark, using the Spark DataSource APIs

Programming Languages

scala
5932 projects

Projects that are alternatives of or similar to Neo4j Spark Connector

Javaorbigdata Interview
Java开发者或者大数据开发者面试知识点整理
Stars: ✭ 203 (-17.14%)
Mutual labels:  spark
Ruby Spark
Ruby wrapper for Apache Spark
Stars: ✭ 221 (-9.8%)
Mutual labels:  spark
Azure Event Hubs
☁️ Cloud-scale telemetry ingestion from any stream of data with Azure Event Hubs
Stars: ✭ 233 (-4.9%)
Mutual labels:  spark
Example Spark
Spark, Spark Streaming and Spark SQL unit testing strategies
Stars: ✭ 205 (-16.33%)
Mutual labels:  spark
Spark Excel
A Spark plugin for reading Excel files via Apache POI
Stars: ✭ 216 (-11.84%)
Mutual labels:  spark
Spark.fish
▁▂▄▆▇█▇▆▄▂▁
Stars: ✭ 229 (-6.53%)
Mutual labels:  spark
Spark Practice
Apache Spark (PySpark) Practice on Real Data
Stars: ✭ 200 (-18.37%)
Mutual labels:  spark
Recommendationsystem
Book recommender system using collaborative filtering based on Spark
Stars: ✭ 244 (-0.41%)
Mutual labels:  spark
Sagemaker Spark
A Spark library for Amazon SageMaker.
Stars: ✭ 219 (-10.61%)
Mutual labels:  spark
Mastering Spark Sql Book
The Internals of Spark SQL
Stars: ✭ 234 (-4.49%)
Mutual labels:  spark
Hydro Serving
MLOps Platform
Stars: ✭ 213 (-13.06%)
Mutual labels:  spark
Gimel
Big Data Processing Framework - Unified Data API or SQL on Any Storage
Stars: ✭ 216 (-11.84%)
Mutual labels:  spark
Mydatascienceportfolio
Applying Data Science and Machine Learning to Solve Real World Business Problems
Stars: ✭ 227 (-7.35%)
Mutual labels:  spark
Spark Knn
k-Nearest Neighbors algorithm on Spark
Stars: ✭ 205 (-16.33%)
Mutual labels:  spark
Hadoop Docker
基于Docker构建的Hadoop开发测试环境,包含Hadoop,Hive,HBase,Spark
Stars: ✭ 238 (-2.86%)
Mutual labels:  spark
Mmlspark
Simple and Distributed Machine Learning
Stars: ✭ 2,899 (+1083.27%)
Mutual labels:  spark
Spark Workshop
Apache Spark™ and Scala Workshops
Stars: ✭ 224 (-8.57%)
Mutual labels:  spark
Dpark
Python clone of Spark, a MapReduce alike framework in Python
Stars: ✭ 2,668 (+988.98%)
Mutual labels:  spark
Video Stream Analytics
Stars: ✭ 240 (-2.04%)
Mutual labels:  spark
Installations mac ubuntu windows
Installations for Data Science. Anaconda, RStudio, Spark, TensorFlow, AWS (Amazon Web Services).
Stars: ✭ 231 (-5.71%)
Mutual labels:  spark

Neo4j Connector for Apache Spark

This repository contains the Neo4j Connector for Apache Spark.

License

This neo4j-connector-apache-spark is Apache 2 Licensed

Generating Documentation from Source

cd doc
# Install NodeJS dependencies
npm install
# Generate HTML/CSS from asciidoc
./node_modules/.bin/antora docs.yml
# Start local server to browse docs
npm run start

This will open http://localhost:8000/ which will serve development docs.

Building

Building for Spark 2.4

You can build for Spark 2.4 with both Scala 2.11 and Scala 2.12

./mvnw clean package -P spark-2.4 -P scala-2.11
./mvnw clean package -P spark-2.4 -P scala-2.12

These commands will generate the corresponding targets

  • spark-2.4/target/neo4j-connector-apache-spark_2.11_2.4-4.0.0.jar
  • spark-2.4/target/neo4j-connector-apache-spark_2.12_2.4-4.0.0.jar

Building for Spark 3

You can build for Spark 3 by running

./mvnw clean package -P spark-3 -P scala-2.12

This will generate spark-3/target/neo4j-connector-apache-spark_2.12_3-4.0.0.jar

Integration with Apache Spark Applications

spark-shell, pyspark, or spark-submit

$SPARK_HOME/bin/spark-shell --jars neo4j-connector-apache-spark_2.12_3-4.0.0.jar

$SPARK_HOME/bin/spark-shell --packages neo4j-contrib:neo4j-connector-apache-spark_2.12_3:4.0.0

sbt

If you use the sbt-spark-package plugin, in your sbt build file, add:

scala spDependencies += "neo4j-contrib/neo4j-connector-apache-spark_2.11_3:4.0.0"

Otherwise,

resolvers += "Spark Packages Repo" at "http://dl.bintray.com/spark-packages/maven"
libraryDependencies += "neo4j-contrib" % "neo4j-connector-apache-spark_2.11_2.4" % "4.0.0"

Or, for Spark 3

resolvers += "Spark Packages Repo" at "http://dl.bintray.com/spark-packages/maven"
libraryDependencies += "neo4j-contrib" % "neo4j-connector-apache-spark_2.12_3" % "4.0.0"

maven
In your pom.xml, add:

<dependencies>
  <!-- list of dependencies -->
  <dependency>
    <groupId>neo4j-contrib</groupId>
    <artifactId>neo4j-connector-apache-spark_2.11_2.4</artifactId>
    <version>4.0.0</version>
  </dependency>
</dependencies>
<repositories>
  <!-- list of other repositories -->
  <repository>
    <id>SparkPackagesRepo</id>
    <url>http://dl.bintray.com/spark-packages/maven</url>
  </repository>
</repositories>

In case of Spark 3

<dependencies>
  <!-- list of dependencies -->
  <dependency>
    <groupId>neo4j-contrib</groupId>
    <artifactId>neo4j-connector-apache-spark_2.12_3</artifactId>
    <version>4.0.0</version>
  </dependency>
</dependencies>
<repositories>
  <!-- list of other repositories -->
  <repository>
    <id>SparkPackagesRepo</id>
    <url>http://dl.bintray.com/spark-packages/maven</url>
  </repository>
</repositories>

For more info about the available version visit https://neo4j.com/developer/spark/overview/#_compatibility

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].