Skip to content

Commit

Permalink
Merge branch 'current' into databricks-updates
Browse files Browse the repository at this point in the history
  • Loading branch information
matthewshaver authored Sep 24, 2024
2 parents 94ed865 + 9206e64 commit f4298bb
Show file tree
Hide file tree
Showing 7 changed files with 41 additions and 22 deletions.
40 changes: 24 additions & 16 deletions website/docs/docs/cloud/configure-cloud-cli.md
Original file line number Diff line number Diff line change
Expand Up @@ -52,21 +52,29 @@ Once you install the dbt Cloud CLI, you need to configure it to connect to a dbt

The config file looks like this:

```yaml
version: "1"
context:
active-project: "<project id from the list below>"
active-host: "<active host from the list>"
defer-env-id: "<optional defer environment id>"
projects:
- project-id: "<project-id>"
account-host: "<account-host>"
api-key: "<user-api-key>"

- project-id: "<project-id>"
account-host: "<account-host>"
api-key: "<user-api-key>"
```
```yaml
version: "1"
context:
active-project: "<project id from the list below>"
active-host: "<active host from the list>"
defer-env-id: "<optional defer environment id>"
projects:
- project-name: "<project-name>"
project-id: "<project-id>"
account-name: "<account-name>"
account-id: "<account-id>"
account-host: "<account-host>" # for example, "cloud.getdbt.com"
token-name: "<pat-or-service-token-name>"
token-value: "<pat-or-service-token-value>"

- project-name: "<project-name>"
project-id: "<project-id>"
account-name: "<account-name>"
account-id: "<account-id>"
account-host: "<account-host>" # for example, "cloud.getdbt.com"
token-name: "<pat-or-service-token-name>"
token-value: "<pat-or-service-token-value>"
```
3. After downloading the config file and creating your directory, navigate to a dbt project in your terminal:
Expand Down Expand Up @@ -195,4 +203,4 @@ This command moves the `dbt_cloud.yml` from the `Downloads` folder to the `.dbt`
By default, [all artifacts](/reference/artifacts/dbt-artifacts) are downloaded when you execute dbt commands from the dbt Cloud CLI. To skip these files from being downloaded, add `--download-artifacts=false` to the command you want to run. This can help improve run-time performance but might break workflows that depend on assets like the [manifest](/reference/artifacts/manifest-json).


</Expandable>
</Expandable>
2 changes: 1 addition & 1 deletion website/docs/reference/global-configs/resource-type.md
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ The available resource types are:

## Example

Instead of targeting specific resources, use the `--resource-flag` or `--exclude-resource-type` flags to target all resources of a certain type: `dbt build --resource-type RESOURCE_TYPE` replacing `RESOURCE_TYPE` with the resource type you want to exclude.
Instead of targeting specific resources, use the `--resource-flag` or `--exclude-resource-type` flags to target all resources of a certain type: `dbt build --resource-type RESOURCE_TYPE` replacing `RESOURCE_TYPE` with the resource type you want to include.

- For example, use the following command to include _all_ snapshots from your dbt build process:

Expand Down
3 changes: 2 additions & 1 deletion website/docs/reference/resource-configs/grants.md
Original file line number Diff line number Diff line change
Expand Up @@ -249,13 +249,14 @@ models:

<div warehouse="Redshift">

* Granting to / revoking from is only fully supported for Redshift users (not groups or roles).
* Granting to / revoking from is only fully supported for Redshift users (not [groups](https://docs.aws.amazon.com/redshift/latest/dg/r_Groups.html) or [roles](https://docs.aws.amazon.com/redshift/latest/dg/r_roles-managing.html)).

</div>

<div warehouse="Snowflake">

* dbt accounts for the [`copy_grants` configuration](/reference/resource-configs/snowflake-configs#copying-grants) when calculating which grants need to be added or removed.
* Granting to / revoking from is only fully supported for Snowflake roles (not [database roles](https://docs.snowflake.com/user-guide/security-access-control-overview#types-of-roles)).

</div>

Expand Down
9 changes: 9 additions & 0 deletions website/docs/reference/resource-configs/schema.md
Original file line number Diff line number Diff line change
Expand Up @@ -76,6 +76,15 @@ This results in the generated relation being located in the `snapshots` schema s

</TabItem>

<TabItem value="saved-queries" label="Saved queries">

<File name='dbt_project.yml'>
```yml
saved-queries:
+schema: metrics
```
</File>
</TabItem>
<TabItem value="tests" label="Test">

Customize the schema for storing test results in your `dbt_project.yml` file.
Expand Down
3 changes: 2 additions & 1 deletion website/docs/terms/data-catalog.md
Original file line number Diff line number Diff line change
Expand Up @@ -74,8 +74,9 @@ Does the data team fully understand the data and its lineage? If they don’t, i
Do you have sophisticated data models and sources that the business knows how to use? If not, it may be worth focusing on building out the right metrics and models to be used first.

Is the culture data-driven? If business users are caught up in their own processes like Google spreadsheets, they may not even utilize a catalog. You don’t want to pay for a tool that is too sophisticated for where the business and data teams currently stand. Don’t rush the process.

### Data catalog tools
Data teams may choose to use third-party tools with data cataloging capabilities such as [Atlan](https://ask.atlan.com/hc/en-us/articles/4433673207313-How-to-set-up-dbt-Cloud), Alation, [Collibra](https://marketplace.collibra.com/listings/dbt-lineage-to-collibra-integration/), [Metaphor](https://support.metaphor.io/hc/en-us/articles/9302185081627), [Select Star](https://docs.selectstar.com/integrations/dbt/dbt-cloud), [Stemma](https://docs.stemma.ai/docs/stemma/getting-started/what-we-need-from-you/dbt-integration/), [Castor](https://docs.castordoc.com/integrations/dbt), and others. These tools often integrate directly with your data pipelines and dbt workflows and offer zoomed-in data cataloging and lineage capabilities.
Data teams may choose to use third-party tools with data cataloging capabilities such as [Atlan](https://ask.atlan.com/hc/en-us/articles/4433673207313-How-to-set-up-dbt-Cloud), Alation, [Collibra](https://marketplace.collibra.com/listings/dbt-lineage-to-collibra-integration/), [Metaphor](https://support.metaphor.io/hc/en-us/articles/9302185081627), [Select Star](https://docs.selectstar.com/integrations/dbt/dbt-cloud), [Secoda](https://docs.secoda.co/integrations/data-transformation-tools/readme-1/dbt), [Castor](https://docs.castordoc.com/integrations/dbt), and others. These tools often integrate directly with your data pipelines and dbt workflows and offer zoomed-in data cataloging and lineage capabilities.

## Conclusion

Expand Down
2 changes: 1 addition & 1 deletion website/docs/terms/data-lineage.md
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ DAGs shouldn’t be dependent on manual updates. Instead, your DAG should be aut

### Third-party tooling

Data teams may also choose to use third-party tools with lineage capabilities such as [Atlan](https://ask.atlan.com/hc/en-us/articles/4433673207313-How-to-set-up-dbt-Cloud), Alation, [Collibra](https://marketplace.collibra.com/listings/dbt-lineage-to-collibra-integration/), [Datafold](https://www.datafold.com/column-level-lineage), Metaphor, [Monte Carlo](https://docs.getmontecarlo.com/docs/dbt-cloud), [Select Star](https://docs.selectstar.com/integrations/dbt/dbt-cloud), or [Stemma](https://docs.stemma.ai/docs/stemma/getting-started/what-we-need-from-you/dbt-integration/). These tools often integrate directly with your data pipelines and dbt workflows and offer zoomed-in data lineage capabilities such as column-level or business logic-level lineage.
Data teams may also choose to use third-party tools with lineage capabilities such as [Atlan](https://ask.atlan.com/hc/en-us/articles/4433673207313-How-to-set-up-dbt-Cloud), Alation, [Collibra](https://marketplace.collibra.com/listings/dbt-lineage-to-collibra-integration/), [Datafold](https://www.datafold.com/column-level-lineage), Metaphor, [Monte Carlo](https://docs.getmontecarlo.com/docs/dbt-cloud) or [Select Star](https://docs.selectstar.com/integrations/dbt/dbt-cloud). These tools often integrate directly with your data pipelines and dbt workflows and offer zoomed-in data lineage capabilities such as column-level or business logic-level lineage.

## Data lineage challenges

Expand Down
4 changes: 2 additions & 2 deletions website/docs/terms/json.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ Individual values from an array can be called by referencing the location of a v

Dictionaries and arrays can be nested in JSON objects as well as nested in each other. **Dictionaries and arrays can only be values. They can never be keys.**

Here’s an example of a JSON object describing a tweet from [Twitter’s developer platform](https://developer.twitter.com/en/docs/twitter-api/v1/data-dictionary/overview).
Here’s an example of a JSON object describing a tweet from [X's (formerly Twitter) developer platform](https://developer.x.com/en/docs/twitter-api/v1/data-dictionary/overview).

```json
tweet =
Expand Down Expand Up @@ -100,4 +100,4 @@ JSON is lightweight and often used to transfer data over a network connection. A

## Conclusion

The greatest strength of JSON also acts as its weakness—the data it contains informs the shape the object takes, rather than the other way around. Structured data is the bread and butter of analytics work, but a semi-structured format is an alternative option when a tabular format becomes too rigid to describe the relationships between different entities.
The greatest strength of JSON also acts as its weakness—the data it contains informs the shape the object takes, rather than the other way around. Structured data is the bread and butter of analytics work, but a semi-structured format is an alternative option when a tabular format becomes too rigid to describe the relationships between different entities.

0 comments on commit f4298bb

Please sign in to comment.