diff --git a/website/docs/docs/cloud/configure-cloud-cli.md b/website/docs/docs/cloud/configure-cloud-cli.md index 854950f5d8c..df94f9354de 100644 --- a/website/docs/docs/cloud/configure-cloud-cli.md +++ b/website/docs/docs/cloud/configure-cloud-cli.md @@ -52,21 +52,29 @@ Once you install the dbt Cloud CLI, you need to configure it to connect to a dbt The config file looks like this: - ```yaml - version: "1" - context: - active-project: "" - active-host: "" - defer-env-id: "" - projects: - - project-id: "" - account-host: "" - api-key: "" - - - project-id: "" - account-host: "" - api-key: "" - ``` + ```yaml + version: "1" + context: + active-project: "" + active-host: "" + defer-env-id: "" + projects: + - project-name: "" + project-id: "" + account-name: "" + account-id: "" + account-host: "" # for example, "cloud.getdbt.com" + token-name: "" + token-value: "" + + - project-name: "" + project-id: "" + account-name: "" + account-id: "" + account-host: "" # for example, "cloud.getdbt.com" + token-name: "" + token-value: "" + ``` 3. After downloading the config file and creating your directory, navigate to a dbt project in your terminal: @@ -195,4 +203,4 @@ This command moves the `dbt_cloud.yml` from the `Downloads` folder to the `.dbt` By default, [all artifacts](/reference/artifacts/dbt-artifacts) are downloaded when you execute dbt commands from the dbt Cloud CLI. To skip these files from being downloaded, add `--download-artifacts=false` to the command you want to run. This can help improve run-time performance but might break workflows that depend on assets like the [manifest](/reference/artifacts/manifest-json). - \ No newline at end of file + diff --git a/website/docs/guides/databricks-qs.md b/website/docs/guides/databricks-qs.md index 700da7198f6..ba93ff74540 100644 --- a/website/docs/guides/databricks-qs.md +++ b/website/docs/guides/databricks-qs.md @@ -41,36 +41,33 @@ You can check out [dbt Fundamentals](https://learn.getdbt.com/courses/dbt-fundam ## Create a Databricks workspace -1. Use your existing account or [sign up for a Databricks account](https://databricks.com/). Complete the form with your user information. +1. Use your existing account or [sign up for a Databricks account](https://databricks.com/). Complete the form with your user information and click **Continue**.
-2. For the purpose of this tutorial, you will be selecting AWS as our cloud provider but if you use Azure or GCP internally, please choose one of them. The setup process will be similar. -3. Check your email to complete the verification process. -4. After setting up your password, you will be guided to choose a subscription plan. Select the `Premium` or `Enterprise` plan to access the SQL Compute functionality required for using the SQL warehouse for dbt. We have chosen `Premium` for this tutorial. Click **Continue** after selecting your plan. - -
- +2. On the next screen, select your cloud provider. This tutorial uses AWS as the cloud provider, but if you use Azure or GCP internally, please select your platform. The setup process will be similar. Do not select the **Get started with Community Edition** option, as this will not provide the required compute for this guide. + +
+
-5. Click **Get Started** when you come to this below page and then **Confirm** after you validate that you have everything needed. +3. Check your email and complete the verification process. +4. After completing the verification processes, you will be brought to the first setup screen. Databricks defaults to the `Premium` plan and you can change the trial to `Enterprise` on this page. +
- -
-
- +
-6. Now it's time to create your first workspace. A Databricks workspace is an environment for accessing all of your Databricks assets. The workspace organizes objects like notebooks, SQL warehouses, clusters, etc into one place. Provide the name of your workspace and choose the appropriate AWS region and click **Start Quickstart**. You might get the checkbox of **I have data in S3 that I want to query with Databricks**. You do not need to check this off for the purpose of this tutorial. +5. Now, it's time to create your first workspace. A Databricks workspace is an environment for accessing all of your Databricks assets. The workspace organizes objects like notebooks, SQL warehouses, clusters, and more so into one place. Provide the name of your workspace, choose the appropriate AWS region, and click **Start Quickstart**. You might get the checkbox of **I have data in S3 that I want to query with Databricks**. You do not need to check this off for this tutorial.
- +
-7. By clicking on `Start Quickstart`, you will be redirected to AWS and asked to log in if you haven’t already. After logging in, you should see a page similar to this. +6. By clicking on `Start Quickstart`, you will be redirected to AWS and asked to log in if you haven’t already. After logging in, you should see a page similar to this.
@@ -80,7 +77,7 @@ You can check out [dbt Fundamentals](https://learn.getdbt.com/courses/dbt-fundam If you get a session error and don’t get redirected to this page, you can go back to the Databricks UI and create a workspace from the interface. All you have to do is click **create workspaces**, choose the quickstart, fill out the form and click **Start Quickstart**. ::: -8. There is no need to change any of the pre-filled out fields in the Parameters. Just add in your Databricks password under **Databricks Account Credentials**. Check off the Acknowledgement and click **Create stack**. +7. There is no need to change any of the pre-filled out fields in the Parameters. Just add in your Databricks password under **Databricks Account Credentials**. Check off the Acknowledgement and click **Create stack**.
@@ -89,11 +86,11 @@ If you get a session error and don’t get redirected to this page, you can go b
-10. Go back to the Databricks tab. You should see that your workspace is ready to use. +8. Go back to the Databricks tab. You should see that your workspace is ready to use.
-11. Now let’s jump into the workspace. Click **Open** and log into the workspace using the same login as you used to log into the account. +9. Now let’s jump into the workspace. Click **Open** and log into the workspace using the same login as you used to log into the account. ## Load data diff --git a/website/docs/reference/global-configs/resource-type.md b/website/docs/reference/global-configs/resource-type.md index 25ef426a6a7..ad8897a745c 100644 --- a/website/docs/reference/global-configs/resource-type.md +++ b/website/docs/reference/global-configs/resource-type.md @@ -60,7 +60,7 @@ The available resource types are: ## Example -Instead of targeting specific resources, use the `--resource-flag` or `--exclude-resource-type` flags to target all resources of a certain type: `dbt build --resource-type RESOURCE_TYPE` replacing `RESOURCE_TYPE` with the resource type you want to exclude. +Instead of targeting specific resources, use the `--resource-flag` or `--exclude-resource-type` flags to target all resources of a certain type: `dbt build --resource-type RESOURCE_TYPE` replacing `RESOURCE_TYPE` with the resource type you want to include. - For example, use the following command to include _all_ snapshots from your dbt build process: diff --git a/website/docs/reference/resource-configs/grants.md b/website/docs/reference/resource-configs/grants.md index 6e960115ea1..99b61ef2413 100644 --- a/website/docs/reference/resource-configs/grants.md +++ b/website/docs/reference/resource-configs/grants.md @@ -249,13 +249,14 @@ models:
-* Granting to / revoking from is only fully supported for Redshift users (not groups or roles). +* Granting to / revoking from is only fully supported for Redshift users (not [groups](https://docs.aws.amazon.com/redshift/latest/dg/r_Groups.html) or [roles](https://docs.aws.amazon.com/redshift/latest/dg/r_roles-managing.html)).
* dbt accounts for the [`copy_grants` configuration](/reference/resource-configs/snowflake-configs#copying-grants) when calculating which grants need to be added or removed. +* Granting to / revoking from is only fully supported for Snowflake roles (not [database roles](https://docs.snowflake.com/user-guide/security-access-control-overview#types-of-roles)).
diff --git a/website/docs/reference/resource-configs/schema.md b/website/docs/reference/resource-configs/schema.md index 57a357767cb..1e2ff47729c 100644 --- a/website/docs/reference/resource-configs/schema.md +++ b/website/docs/reference/resource-configs/schema.md @@ -76,6 +76,15 @@ This results in the generated relation being located in the `snapshots` schema s + + + +```yml +saved-queries: + +schema: metrics +``` + + Customize the schema for storing test results in your `dbt_project.yml` file. diff --git a/website/static/img/databricks_tutorial/images/choose_plan.png b/website/static/img/databricks_tutorial/images/choose_plan.png index 055f232fda3..04565ab2d4f 100644 Binary files a/website/static/img/databricks_tutorial/images/choose_plan.png and b/website/static/img/databricks_tutorial/images/choose_plan.png differ diff --git a/website/static/img/databricks_tutorial/images/choose_provider.png b/website/static/img/databricks_tutorial/images/choose_provider.png new file mode 100644 index 00000000000..cf5d94d5fd7 Binary files /dev/null and b/website/static/img/databricks_tutorial/images/choose_provider.png differ diff --git a/website/static/img/databricks_tutorial/images/signup_form.png b/website/static/img/databricks_tutorial/images/signup_form.png index 612d847c8f5..5fa60ce37ef 100644 Binary files a/website/static/img/databricks_tutorial/images/signup_form.png and b/website/static/img/databricks_tutorial/images/signup_form.png differ diff --git a/website/static/img/databricks_tutorial/images/start_quickstart.png b/website/static/img/databricks_tutorial/images/start_quickstart.png new file mode 100644 index 00000000000..033250d4e5d Binary files /dev/null and b/website/static/img/databricks_tutorial/images/start_quickstart.png differ