Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Add support for unity catalog in source config #1104

Closed
3 tasks done
data-blade opened this issue Sep 11, 2024 · 3 comments
Closed
3 tasks done

[Feature] Add support for unity catalog in source config #1104

data-blade opened this issue Sep 11, 2024 · 3 comments
Labels
enhancement New feature or request triage

Comments

@data-blade
Copy link

data-blade commented Sep 11, 2024

Is this your first time submitting a feature request?

  • I have read the expectations for open source contributors
  • I have searched the existing issues, and I could not find an existing issue for this feature
  • I am requesting a straightforward extension of existing dbt-spark functionality, rather than a Big Idea better suited to a discussion

Describe the feature

setting sources in dbt-databricks works like this:

sources:
  - name: <name>
    catalog|database: <catalog>
    schema: <schema>

dbt-spark should support unity in the same way

Describe alternatives you've considered

in dbt-spark, using...

database: <catalog>
schema: <schema>

will run into this error.

and setting...

catalog: <catalog>
schema: <schema>

will fail too, since catalog is unexpected.

the only working setting is

schema: <catalog>.<schema>

maybe only for us:
the above prevents the parallel use of dbt-databricks adapter (which we require), so we overcome this by

database: "{{ none if target.type == 'spark' else '<catalog>' }}"
schema: "{{ '<catalog>.<schema>' if target.type == 'spark' else '<schema>' }}"

Who will this benefit?

all clients using dbt-spark adapter on databricks' unity catalog (to not use hacks anymore)

Are you interested in contributing this feature?

if time allows

Anything else?

No response

@data-blade data-blade added enhancement New feature or request triage labels Sep 11, 2024
@amychen1776
Copy link
Contributor

Hi @data-blade ! We recommend that users on Databricks use dbt-databricks rather than dbt-spark. Would you be able to clarify why you need to use dbt-spark?

@data-blade
Copy link
Author

to run production on job clusters to save cost.

if dbt-databricks could connect to job clusters, i'd be ecstatic for a week!

@amychen1776
Copy link
Contributor

Ah I see - they have a PR open to operate on serverless job clusters that are getting released soon so I think you should take a look over there :)

That said - I am going to close this ticket for now because this is not something we are going to tackle in our near term roadmap, especially as our recommendation is for Databricks customers to use the dbt-databricks adapter.

@amychen1776 amychen1776 closed this as not planned Won't fix, can't repro, duplicate, stale Sep 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request triage
Projects
None yet
Development

No branches or pull requests

2 participants