Skip to content

Releases: datastax/spark-cassandra-connector

Release 1.4.0 Milestone 1

12 Jun 19:28
Compare
Choose a tag to compare
Pre-release

This preview release brings compatibility with Spark 1.4.0.

Release 1.3.0 Milestone 1

01 Jun 16:44
Compare
Choose a tag to compare
Pre-release

1.3.0 M1

  • Removed use of Thrift describe_ring and replaced it with native Java Driver
    support for fetching TokenRanges (SPARKC-93)
  • Support for converting Cassandra UDT column values to Scala case-class objects (SPARKC-4)
    • Introduced a common interface for TableDef and UserDefinedType
    • Removed ClassTag from ColumnMapper
    • Removed by-index column references and replaced them with by-name ColumnRefs
    • Created a GettableDataToMappedTypeConverter that can handle UDTs
    • ClassBasedRowReader delegates object conversion instead of doing it by itself;
      this improves unit-testability of code
  • Decoupled PredicatePushDown logic from Spark (SPARKC-166)
    • added support for Filter and Expression predicates
    • improved code testability and added unit-tests
  • Basic Datasource API integration and keyspace/cluster level settings (SPARKC-112, SPARKC-162)

Release 1.2.2

01 Jun 16:07
Compare
Choose a tag to compare

1.2.2

  • Updated Spark version to 1.2.2, Scala to 2.10.5 / 2.11.6
  • Enabled Java API artifact generation for Scala 2.11.x (SPARKC-130)
  • Fixed a bug preventing a custom type converter from being used
    when saving data. RowWriter implementations must
    perform type conversions now. (SPARKC-157)

Release 1.1.2

01 Jun 13:59
Compare
Choose a tag to compare

1.1.2

  • Backport SPARKC-8, retrieval of TTL and write time
  • Upgraded to Spark 1.1.1
  • Synchronized ReflectionUtil findScalaObject and findSingletonClassInstance methods
    to avoid problems with Scala 2.10 lack thread safety in the reflection subsystem (SPARKC-107)
  • Fixed populating ReadConf with properties from SparkConf (SPARKC-121)
  • Adds both hostname and hostaddress as partition preferredLocations (SPARKC-141, backport of SPARKC-126)

Release 1.2.1

15 May 10:23
Compare
Choose a tag to compare

1.2.1

  • Fixed problems with mixed case keyspaces in ReplicaMapper (SPARKC-159)

Release 1.2.0

04 May 20:33
Compare
Choose a tag to compare

1.2.0

  • Removed conversion method rom WriteOption which accepted object of Duration type
    from Spark Streaming (SPARKC-106)
  • Fixed compilation warnings (SPARKC-76)
  • Fixed ScalaDoc warnings (SPARKC-119)
  • Synchronized TypeTag access in various places (SPARKC-123)
  • Adds both hostname and hostaddress as partition preferredLocations (SPARKC-126)

1.2.0 RC 3

  • Select aliases are no longer ignored in CassandraRow objects (SPARKC-109)
  • Fix picking up username and password from SparkConf (SPARKC-108)
  • Fix creating CassandraConnectorSource in the executor environment (SPARKC-111)

1.2.0 RC 2

  • Cross cluster table join and write for Spark SQL (SPARKC-73)
  • Enabling / disabling metrics in metrics configuration file and other metrics fixes (SPARKC-91)
  • Provided a way to set custom auth config and connection factory properties (SPARKC-105)
  • Fixed setting custom connection factory and other properties (SPAKRC-102)
  • Fixed Java API (SPARKC-95)

1.2.0 RC 1

  • More Spark SQL predicate push (SPARKC-72)
  • Fixed some Java API problems and refactored its internals (SPARKC-77)
  • Allowing specification of column to property map (aliases) for reading and writing objects
    (SPARKC-9)
  • Added interface for doing primary key joins between arbitrary RDDs and Cassandra (SPARKC-25)
  • Added method for repartitioning an RDD based upon the replication of a Cassandra Table (SPARKC-25)
  • Fixed setting batch.level and batch.buffer.size in SparkConf. (SPARKC-84)
    • Renamed output.batch.level to output.batch.grouping.key.
    • Renamed output.batch.buffer.size to output.batch.grouping.buffer.size.
    • Renamed batch grouping key option "all" to "none".
  • Error out on invalid config properties (SPARKC-90)
  • Set Java driver version to 2.1.5 and Cassandra to 2.1.3 (SPARKC-92)
  • Moved Spark streaming related methods from CassandraJavaUtil to CassandraStreamingJavaUtil
    (SPARKC-80)

1.2.0 alpha 3

  • Exposed spanBy and spanByKey in Java API (SPARKC-39)
  • Added automatic generation of Cassandra table schema from a Scala type and
    saving an RDD to a new Cassandra table by saveAsCassandraTable method (SPARKC-38)
  • Added support for write throughput limiting (SPARKC-57)
  • Added EmptyCassandraRDD (SPARKC-37)
  • Exposed authConf in CassandraConnector
  • Overridden count() implementation in CassandraRDD which uses native Cassandra count (SPARKC-52)
  • Removed custom Logging class (SPARKC-54)
  • Added support for passing the limit clause to CQL in order to fetch top n results (SPARKC-31)
  • Added support for pushing down order by clause for explicitly specifying an order of rows within
    Cassandra partition (SPARKC-32)
  • Fixed problems when rows are mapped to classes with inherited fields (SPARKC-70)
  • Support for compiling with Scala 2.10 and 2.11 (SPARKC-22)

1.2.0 alpha 2

  • All connection properties can be set on SparkConf / CassandraConnectorConf objects and
    the settings are automatically distributed to Spark Executors (SPARKC-28)
  • Report Connector metrics to Spark metrics system (SPARKC-27)
  • Upgraded to Spark 1.2.1 (SPARKC-30)
  • Add conversion from java.util.Date to java.sqlTimestamp for Spark SQL (#512)
  • Upgraded to Scala 2.11 and scala version cross build (SPARKC-22)

1.2.0 alpha 1

  • Added support for TTL and timestamp in the writer (#153)
  • Added support for UDT column types (SPARKC-1)
  • Upgraded Spark to version 1.2.0 (SPARKC-15)
  • For 1.2.0 release, table name with dot is not supported for Spark SQL,
    it will be fixed in the next release
  • Added fast spanBy and spanByKey methods to RDDs useful for grouping Cassandra
    data by partition key / clustering columns. Useful for e.g. time-series data. (SPARKC-2)
  • Refactored the write path so that the writes are now token-aware (S

Preview Release 1.2.0 RC 3

31 Mar 11:04
Compare
Choose a tag to compare
Pre-release

1.2.0 RC 3

  • Select aliases are no longer ignored in CassandraRow objects (SPARKC-109)
  • Fix picking up username and password from SparkConf (SPARKC-108)
  • Fix creating CassandraConnectorSource in the executor environment (SPARKC-111)

Preview Release 1.2.0 RC 2

27 Mar 12:08
Compare
Choose a tag to compare
Pre-release

1.2.0 RC 2

  • Cross cluster table join and write for Spark SQL (SPARKC-73)
  • Enabling / disabling metrics in metrics configuration file and other metrics fixes (SPARKC-91)
  • Provided a way to set custom auth config and connection factory properties (SPARKC-105)
  • Fixed setting custom connection factory and other properties (SPAKRC-102)
  • Fixed Java API (SPARKC-95)

1.1.x (unreleased)

  • Synchronized ReflectionUtil findScalaObject and findSingletonClassInstance methods
    to avoid problems with Scala 2.10 lack thread safety in the reflection subsystem (SPARKC-107)

Preview Release 1.2.0 RC 1

19 Mar 13:08
Compare
Choose a tag to compare
Pre-release

1.2.0 rc 1

  • More Spark SQL predicate push (SPARKC-72)
  • Fixed some Java API problems and refactored its internals (SPARKC-77)
  • Allowing specification of column to property map (aliases) for reading and writing objects
    (SPARKC-9)
  • Added interface for doing primary key joins between arbitrary RDDs and Cassandra (SPARKC-25)
  • Added method for repartitioning an RDD based upon the replication of a Cassandra Table (SPARKC-25)
  • Fixed setting batch.level and batch.buffer.size in SparkConf. (SPARKC-84)
    • Renamed output.batch.level to output.batch.grouping.key.
    • Renamed output.batch.buffer.size to output.batch.grouping.buffer.size.
    • Renamed batch grouping key option "all" to "none".
  • Error out on invalid config properties (SPARKC-90)
  • Set Java driver version to 2.1.5 and Cassandra to 2.1.3 (SPARKC-92)
  • Moved Spark streaming related methods from CassandraJavaUtil to CassandraStreamingJavaUtil
    (SPARKC-80)

Preview Release 1.2.0 alpha 3

06 Mar 10:34
Compare
Choose a tag to compare
Pre-release

1.2.0 alpha 3

  • Exposed spanBy and spanByKey in Java API (SPARKC-39)
  • Added automatic generation of Cassandra table schema from a Scala type and
    saving an RDD to a new Cassandra table by saveAsCassandraTable method (SPARKC-38)
  • Added support for write throughput limiting (SPARKC-57)
  • Added EmptyCassandraRDD (SPARKC-37)
  • Exposed authConf in CassandraConnector
  • Overridden count() implementation in CassandraRDD which uses native Cassandra count (SPARKC-52)
  • Removed custom Logging class (SPARKC-54)
  • Added support for passing the limit clause to CQL in order to fetch top n results (SPARKC-31)
  • Added support for pushing down order by clause for explicitly specifying an order of rows within
    Cassandra partition (SPARKC-32)
  • Fixed problems when rows are mapped to classes with inherited fields (SPARKC-70)
  • Support for compiling with Scala 2.10 and 2.11 (SPARKC-22)