This library lets you expose Cassandra tables as Spark RDDs, write Spark RDDs to Cassandra tables, and execute arbitrary CQL queries in your Spark applications.
saveToCassandra
callWHERE
clauseThis project has been published to the Maven Central Repository. For SBT to download the connector binaries, sources and javadoc, put this in your project SBT config:
libraryDependencies += "com.datastax.spark" %% "spark-cassandra-connector" % "1.0.6" withSources() withJavadoc()
If you want to access the functionality of Connector from Java, you may want to add also a Java API module:
libraryDependencies += "com.datastax.spark" %% "spark-cassandra-connector-java" % "1.0.6" withSources() withJavadoc()
In the root directory run:
sbt assembly
A fat jar will be generated to both of these directories:
spark-cassandra-connector/target/scala-2.10/
spark-cassandra-connector-java/target/scala-2.10/
Select the former for Scala apps, the later for Java.
In the root directory run:
sbt package
sbt doc
The library package jars will be placed in:
spark-cassandra-connector/target/scala-2.10/
spark-cassandra-connector-java/target/scala-2.10/
The documentation will be generated to:
spark-cassandra-connector/target/scala-2.10/api/
spark-cassandra-connector-java/target/scala-2.10/api/
New issues should be reported using JIRA. Please do not use the built-in GitHub issue tracker. It is left for archival purposes and it will be disabled soon.
Questions etc can be submitted to the user mailing list.
To develop this project, we recommend using IntelliJ IDEA. Make sure you have installed and enabled the Scala Plugin. Open the project with IntelliJ IDEA and it will automatically create the project structure from the provided SBT configuration.
Before contributing your changes to the project, please make sure that all unit tests and integration tests pass. Don't forget to add an appropriate entry at the top of CHANGES.txt. Finally open a pull-request on GitHub and await review.
If your pull-request is going to resolve some opened issue, please add Fixes #xx at the end of each commit message (where xx is the number of the issue).
To run unit and integration tests:
sbt test
sbt it:test
By default, integration tests start up a separate, single Cassandra instance and run Spark in local mode. It is possible to run integration tests with your own Cassandra and/or Spark cluster. First, prepare a jar with testing code:
sbt test:package
Then copy the generated test jar to your Spark nodes and run:
export IT_TEST_CASSANDRA_HOST=<IP of one of the Cassandra nodes>
export IT_TEST_SPARK_MASTER=<Spark Master URL>
./sbt/sbt it:test
Copyright 2014-2015, DataStax, Inc.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。