Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Another attempt to improve metadata gathering performance #503

Merged
merged 3 commits into from
Nov 10, 2023

Conversation

benc-db
Copy link
Collaborator

@benc-db benc-db commented Nov 10, 2023

Resolves #496, I hope

Description

So we think when no models are selected, stuffing the where clause with every model is probably non-performant. In dbt core, they'll only call get_catalog if no models are selected, or there are more than 100 models selected. In this case, we can filter by schemas, but probably a bad idea to make our optimizer try to handle a where clause with 200+ components.

Checklist

  • I have run this code in development and it appears to resolve the stated issue
  • This PR includes tests, or tests are not required/relevant for this PR
  • I have updated the CHANGELOG.md and added information about my change to the "dbt-databricks next" section.

@benc-db benc-db merged commit c9d9797 into main Nov 10, 2023
21 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

dbt-databricks 1.6.X overeager introspection (describe extended ...) of all tables in schema -- REGRESSION
2 participants