Releases: datastax/spark-cassandra-connector
Releases · datastax/spark-cassandra-connector
Release 1.4.0 Milestone 1
This preview release brings compatibility with Spark 1.4.0.
Release 1.3.0 Milestone 1
1.3.0 M1
- Removed use of Thrift describe_ring and replaced it with native Java Driver
support for fetching TokenRanges (SPARKC-93) - Support for converting Cassandra UDT column values to Scala case-class objects (SPARKC-4)
- Introduced a common interface for TableDef and UserDefinedType
- Removed ClassTag from ColumnMapper
- Removed by-index column references and replaced them with by-name ColumnRefs
- Created a GettableDataToMappedTypeConverter that can handle UDTs
- ClassBasedRowReader delegates object conversion instead of doing it by itself;
this improves unit-testability of code
- Decoupled PredicatePushDown logic from Spark (SPARKC-166)
- added support for Filter and Expression predicates
- improved code testability and added unit-tests
- Basic Datasource API integration and keyspace/cluster level settings (SPARKC-112, SPARKC-162)
Release 1.2.2
1.2.2
- Updated Spark version to 1.2.2, Scala to 2.10.5 / 2.11.6
- Enabled Java API artifact generation for Scala 2.11.x (SPARKC-130)
- Fixed a bug preventing a custom type converter from being used
when saving data. RowWriter implementations must
perform type conversions now. (SPARKC-157)
Release 1.1.2
1.1.2
- Backport SPARKC-8, retrieval of TTL and write time
- Upgraded to Spark 1.1.1
- Synchronized ReflectionUtil findScalaObject and findSingletonClassInstance methods
to avoid problems with Scala 2.10 lack thread safety in the reflection subsystem (SPARKC-107) - Fixed populating ReadConf with properties from SparkConf (SPARKC-121)
- Adds both hostname and hostaddress as partition preferredLocations (SPARKC-141, backport of SPARKC-126)
Release 1.2.1
1.2.1
- Fixed problems with mixed case keyspaces in ReplicaMapper (SPARKC-159)
Release 1.2.0
1.2.0
- Removed conversion method rom WriteOption which accepted object of Duration type
from Spark Streaming (SPARKC-106) - Fixed compilation warnings (SPARKC-76)
- Fixed ScalaDoc warnings (SPARKC-119)
- Synchronized TypeTag access in various places (SPARKC-123)
- Adds both hostname and hostaddress as partition preferredLocations (SPARKC-126)
1.2.0 RC 3
- Select aliases are no longer ignored in CassandraRow objects (SPARKC-109)
- Fix picking up username and password from SparkConf (SPARKC-108)
- Fix creating CassandraConnectorSource in the executor environment (SPARKC-111)
1.2.0 RC 2
- Cross cluster table join and write for Spark SQL (SPARKC-73)
- Enabling / disabling metrics in metrics configuration file and other metrics fixes (SPARKC-91)
- Provided a way to set custom auth config and connection factory properties (SPARKC-105)
- Fixed setting custom connection factory and other properties (SPAKRC-102)
- Fixed Java API (SPARKC-95)
1.2.0 RC 1
- More Spark SQL predicate push (SPARKC-72)
- Fixed some Java API problems and refactored its internals (SPARKC-77)
- Allowing specification of column to property map (aliases) for reading and writing objects
(SPARKC-9) - Added interface for doing primary key joins between arbitrary RDDs and Cassandra (SPARKC-25)
- Added method for repartitioning an RDD based upon the replication of a Cassandra Table (SPARKC-25)
- Fixed setting batch.level and batch.buffer.size in SparkConf. (SPARKC-84)
- Renamed output.batch.level to output.batch.grouping.key.
- Renamed output.batch.buffer.size to output.batch.grouping.buffer.size.
- Renamed batch grouping key option "all" to "none".
- Error out on invalid config properties (SPARKC-90)
- Set Java driver version to 2.1.5 and Cassandra to 2.1.3 (SPARKC-92)
- Moved Spark streaming related methods from CassandraJavaUtil to CassandraStreamingJavaUtil
(SPARKC-80)
1.2.0 alpha 3
- Exposed spanBy and spanByKey in Java API (SPARKC-39)
- Added automatic generation of Cassandra table schema from a Scala type and
saving an RDD to a new Cassandra table by saveAsCassandraTable method (SPARKC-38) - Added support for write throughput limiting (SPARKC-57)
- Added EmptyCassandraRDD (SPARKC-37)
- Exposed authConf in CassandraConnector
- Overridden count() implementation in CassandraRDD which uses native Cassandra count (SPARKC-52)
- Removed custom Logging class (SPARKC-54)
- Added support for passing the limit clause to CQL in order to fetch top n results (SPARKC-31)
- Added support for pushing down order by clause for explicitly specifying an order of rows within
Cassandra partition (SPARKC-32) - Fixed problems when rows are mapped to classes with inherited fields (SPARKC-70)
- Support for compiling with Scala 2.10 and 2.11 (SPARKC-22)
1.2.0 alpha 2
- All connection properties can be set on SparkConf / CassandraConnectorConf objects and
the settings are automatically distributed to Spark Executors (SPARKC-28) - Report Connector metrics to Spark metrics system (SPARKC-27)
- Upgraded to Spark 1.2.1 (SPARKC-30)
- Add conversion from java.util.Date to java.sqlTimestamp for Spark SQL (#512)
- Upgraded to Scala 2.11 and scala version cross build (SPARKC-22)
1.2.0 alpha 1
- Added support for TTL and timestamp in the writer (#153)
- Added support for UDT column types (SPARKC-1)
- Upgraded Spark to version 1.2.0 (SPARKC-15)
- For 1.2.0 release, table name with dot is not supported for Spark SQL,
it will be fixed in the next release - Added fast spanBy and spanByKey methods to RDDs useful for grouping Cassandra
data by partition key / clustering columns. Useful for e.g. time-series data. (SPARKC-2) - Refactored the write path so that the writes are now token-aware (S
Preview Release 1.2.0 RC 3
1.2.0 RC 3
- Select aliases are no longer ignored in CassandraRow objects (SPARKC-109)
- Fix picking up username and password from SparkConf (SPARKC-108)
- Fix creating CassandraConnectorSource in the executor environment (SPARKC-111)
Preview Release 1.2.0 RC 2
1.2.0 RC 2
- Cross cluster table join and write for Spark SQL (SPARKC-73)
- Enabling / disabling metrics in metrics configuration file and other metrics fixes (SPARKC-91)
- Provided a way to set custom auth config and connection factory properties (SPARKC-105)
- Fixed setting custom connection factory and other properties (SPAKRC-102)
- Fixed Java API (SPARKC-95)
1.1.x (unreleased)
- Synchronized ReflectionUtil findScalaObject and findSingletonClassInstance methods
to avoid problems with Scala 2.10 lack thread safety in the reflection subsystem (SPARKC-107)
Preview Release 1.2.0 RC 1
1.2.0 rc 1
- More Spark SQL predicate push (SPARKC-72)
- Fixed some Java API problems and refactored its internals (SPARKC-77)
- Allowing specification of column to property map (aliases) for reading and writing objects
(SPARKC-9) - Added interface for doing primary key joins between arbitrary RDDs and Cassandra (SPARKC-25)
- Added method for repartitioning an RDD based upon the replication of a Cassandra Table (SPARKC-25)
- Fixed setting batch.level and batch.buffer.size in SparkConf. (SPARKC-84)
- Renamed output.batch.level to output.batch.grouping.key.
- Renamed output.batch.buffer.size to output.batch.grouping.buffer.size.
- Renamed batch grouping key option "all" to "none".
- Error out on invalid config properties (SPARKC-90)
- Set Java driver version to 2.1.5 and Cassandra to 2.1.3 (SPARKC-92)
- Moved Spark streaming related methods from CassandraJavaUtil to CassandraStreamingJavaUtil
(SPARKC-80)
Preview Release 1.2.0 alpha 3
1.2.0 alpha 3
- Exposed spanBy and spanByKey in Java API (SPARKC-39)
- Added automatic generation of Cassandra table schema from a Scala type and
saving an RDD to a new Cassandra table by saveAsCassandraTable method (SPARKC-38) - Added support for write throughput limiting (SPARKC-57)
- Added EmptyCassandraRDD (SPARKC-37)
- Exposed authConf in CassandraConnector
- Overridden count() implementation in CassandraRDD which uses native Cassandra count (SPARKC-52)
- Removed custom Logging class (SPARKC-54)
- Added support for passing the limit clause to CQL in order to fetch top n results (SPARKC-31)
- Added support for pushing down order by clause for explicitly specifying an order of rows within
Cassandra partition (SPARKC-32) - Fixed problems when rows are mapped to classes with inherited fields (SPARKC-70)
- Support for compiling with Scala 2.10 and 2.11 (SPARKC-22)