-
Notifications
You must be signed in to change notification settings - Fork 279
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Apache Spark - Docs refactoring #2789
Conversation
import TOCInline from '@theme/TOCInline'; | ||
|
||
# Spark JDBC | ||
One of the most used data sources supported by Spark is JDBC. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we need to provide any specific recommendations for the JDBC driver version?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do you have something specific version in mind?
@mzitnik is there a specific version we recommend on?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@mshustov did you meant what version of ClickHouse JDBC we should recommend?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
did you meant what version of ClickHouse JDBC we should recommend?
yes
docs/en/integrations/data-ingestion/apache-spark/spark-native-connector.md
Outdated
Show resolved
Hide resolved
docs/en/integrations/data-ingestion/apache-spark/spark-native-connector.md
Outdated
Show resolved
Hide resolved
docs/en/integrations/data-ingestion/apache-spark/spark-native-connector.md
Outdated
Show resolved
Hide resolved
The above examples demonstrate SparkSQL queries, which you can run within your application using any API—Java, Scala, PySpark, or shell. | ||
|
||
|
||
## Supported Data Types |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
could you extend the docs with the following sections:
- configuration options
- adjusting clickhouse settings (if possible, see Spark: Support read with settings spark-clickhouse-connector#367)
- logging
…connector.md Co-authored-by: Mikhail Shustov <[email protected]>
…connector.md Co-authored-by: Mikhail Shustov <[email protected]>
docs/en/integrations/data-ingestion/apache-spark/spark-native-connector.md
Show resolved
Hide resolved
docs/en/integrations/data-ingestion/apache-spark/spark-native-connector.md
Outdated
Show resolved
Hide resolved
docs/en/integrations/data-ingestion/apache-spark/spark-native-connector.md
Outdated
Show resolved
Hide resolved
:::important | ||
It's essential to include the [clickhouse-jdbc JAR](https://mvnrepository.com/artifact/com.clickhouse/clickhouse-jdbc) with the "all" classifier, | ||
as the connector relies on [clickhouse-http](https://mvnrepository.com/artifact/com.clickhouse/clickhouse-http-client) and [clickhouse-client](https://mvnrepository.com/artifact/com.clickhouse/clickhouse-client) —both of which are bundled in clickhouse-jdbc:all. | ||
Alternatively, you can add [clickhouse-client JAR](https://mvnrepository.com/artifact/com.clickhouse/clickhouse-client) and [clickhouse-http](https://mvnrepository.com/artifact/com.clickhouse/clickhouse-http-client) individually if you prefer not to use the full JDBC package. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
IMO, giving two many alternatives can cause confusion.
As part of our effort to improve Spark's documentation, this PR includes: