diff --git a/website/docs/best-practices/how-we-structure/4-marts.md b/website/docs/best-practices/how-we-structure/4-marts.md
index 21de31a9e0d..995dea7e96f 100644
--- a/website/docs/best-practices/how-we-structure/4-marts.md
+++ b/website/docs/best-practices/how-we-structure/4-marts.md
@@ -26,7 +26,8 @@ models/marts
✅ **Group by department or area of concern.** If you have fewer than 10 or so marts you may not have much need for subfolders, so as with the intermediate layer, don’t over-optimize too early. If you do find yourself needing to insert more structure and grouping though, use useful business concepts here. In our marts layer, we’re no longer worried about source-conformed data, so grouping by departments (marketing, finance, etc.) is the most common structure at this stage.
-✅ **Name by entity.** Use plain English to name the file based on the concept that forms the grain of the mart `customers`, `orders`. Note that for pure marts, there should not be a time dimension (`orders_per_day`) here, that is typically best captured via metrics.
+✅ **Name by entity.** Use plain English to name the file based on the concept that forms the grain of the mart’s `customers`, `orders`. Marts that don't include any time-based rollups (pure marts) should not have a time dimension (`orders_per_day`) here, typically best captured via metrics.
+
❌ **Build the same concept differently for different teams.** `finance_orders` and `marketing_orders` is typically considered an anti-pattern. There are, as always, exceptions — a common pattern we see is that, finance may have specific needs, for example reporting revenue to the government in a way that diverges from how the company as a whole measures revenue day-to-day. Just make sure that these are clearly designed and understandable as _separate_ concepts, not departmental views on the same concept: `tax_revenue` and `revenue` not `finance_revenue` and `marketing_revenue`.
diff --git a/website/docs/community/spotlight/bruno-de-lima.md b/website/docs/community/spotlight/bruno-de-lima.md
index f5ffaa6a970..3c373db06e8 100644
--- a/website/docs/community/spotlight/bruno-de-lima.md
+++ b/website/docs/community/spotlight/bruno-de-lima.md
@@ -2,42 +2,39 @@
id: bruno-de-lima
title: Bruno de Lima
description: |
- Hi all! I'm a Data Engineer, deeply fascinated by the awesomeness dbt. I love talking about dbt, creating content from daily tips to blogposts and engaging with this vibrant community!
-
- Started my career at the beginning of 2022 at Indicium as an Analytics Engineer, working with dbt from day 1. By 2023, my path took a global trajectory as I joined phData as a Data Engineer, expanding my experiences and forging connections beyond Brazil. While dbt is at the heart of my expertise, I've also delved into data warehouses such as Snowflake, Databricks, and BigQuery; visualization tools like Power BI and Tableau; and several minor modern data stack tools.
-
- I actively participate in the dbt community, having attended two dbt Meetups in Brazil organized by Indicium; writing about dbt-related topics in my Medium and LinkedIn profiles; contributing to the code; and frequently checking dbt Slack and Discourse, helping (and being helped by) other dbt practitioners. If you are a community member, you may have seen me around!
-image: /img/community/spotlight/bruno-de-lima.jpg
+ Hey all! I was born and raised in Florianopolis, Brazil, and I'm a Senior Data Engineer at phData. I live with my fiancée and I enjoy music, photography, and powerlifting.
+
+ I started my career in early 2022 at Indicium as an Analytics Engineer, working with dbt from day 1. By 2023, my path took a global trajectory as I joined phData as a Data Engineer, expanding my experiences and creating connections beyond Brazil. While dbt is my main expertise, because of my work in consultancy I have experience with a large range of tools, specially the ones related to Snowflake, Databricks, AWS and GCP; but I have already tried several other modern data stack tools too.
+
+ I actively participate in the dbt community, having organized dbt Meetups in Brazil (in Floripa and São Paulo); writing about dbt-related topics in my Medium and LinkedIn profiles; contributing to the dbt Core code and to the docs; and frequently checking dbt Slack and Discourse, helping (and being helped by) other dbt practitioners. If you are a community member, you may have seen me around!
+image: /img/community/spotlight/bruno-souza-de-lima-newimage.jpg
pronouns: he/him
location: Florianópolis, Brazil
-jobTitle: Data Engineer
+jobTitle: Senior Data Engineer
companyName: phData
-organization: ""
socialLinks:
- name: LinkedIn
link: https://www.linkedin.com/in/brunoszdl/
- name: Medium
link: https://medium.com/@bruno.szdl
-dateCreated: 2023-11-05
+dateCreated: 2024-11-03
hide_table_of_contents: true
communityAward: true
-communityAwardYear: 2023
+communityAwardYear: 2024
---
## When did you join the dbt community and in what way has it impacted your career?
-I was not truly happy with my academic life. My career took a new turn when I enrolled in the Analytics Engineer course by Indicium. That was my first contact with dbt, and I didn't realize how much it would transform my career. After that, I was hired at the company as an Analytics Engineer and worked extensively with dbt from day one.
+I was not truly happy with my academic life. My career took a new turn when I enrolled in the Analytics Engineer course by Indicium. That was my first contact with dbt, and I didn't realize how much it would transform my career. After that, I was hired at the company as an Analytics Engineer and worked extensively with dbt from day one.
It took me some time to become an active member of the dbt community. I started working with dbt at the beginning of 2022 and became more involved towards the end of that year, encouraged by Daniel Avancini. I regret not doing this earlier, because being an active community member has been a game-changer for me, as my knowledge of dbt has grown exponentially just by participating in daily discussions on Slack. I have found #advice-dbt-help and #advice-dbt-for-power-users channels particularly useful, as well as the various database-specific channels. Additionally, the #i-made-this and #i-read-this channels have allowed me to learn about the innovative things that community members are doing.
Inspired by other members, especially Josh Devlin and Owen Prough, I began answering questions on Slack and Discourse. For questions I couldn't answer, I would try engaging in discussions about possible solutions or provide useful links. I also started posting dbt tips on LinkedIn to help practitioners learn about new features or to refresh their memories about existing ones.
-By being more involved in the community, I felt more connected and supported. I received help from other members, and now, I could help others, too. I was happy with this arrangement, but more unexpected surprises came my way. My active participation in Slack, Discourse, and LinkedIn opened doors to new connections and career opportunities. I had the pleasure of meeting a lot of incredible people and receiving exciting job offers, including the one for working at phData.
+By being more involved in the community, I felt more connected and supported. I received help from other members, and now, I could help others, too. I was happy with this arrangement, but more unexpected surprises came my way. My active participation in Slack, Discourse, and LinkedIn opened doors to new connections and career opportunities. I had the pleasure of meeting a lot of incredible people and receiving exciting job offers, including the ones for working at phData and teaching at Zach Wilson's data engineering bootcamp.
Thanks to the dbt community, I went from feeling uncertain about my career prospects to having a solid career and being surrounded by incredible people.
-I would like to thank the Indicium folks for opening the first door for me for this career in data, and not just for me but for lots of people in Brazil trying to migrate from different fields who would not have this opportunity otherwise.
-
## What dbt community leader do you identify with? How are you looking to grow your leadership in the dbt community?
I identify with Gwen Windflower and Joel Labes, or at least they are the kind of leader I admire. Their strong presence and continuous interaction with all types of dbt enthusiasts make everyone feel welcomed in the community. They uplift those who contribute to the community, whether it's through a LinkedIn post or answering a question, and provide constructive feedback to help them improve. And of course they show a very strong knowledge about dbt and data in general, which is reflected in their contributions.
diff --git a/website/docs/community/spotlight/christophe-oudar.md b/website/docs/community/spotlight/christophe-oudar.md
new file mode 100644
index 00000000000..2381d88a381
--- /dev/null
+++ b/website/docs/community/spotlight/christophe-oudar.md
@@ -0,0 +1,35 @@
+---
+id: christophe-oudar
+title: Christophe Oudar
+description: |
+ I joined the dbt Community in November 2021 after exchanging some issues in Github. I currently work as a staff engineer at a scaleup in the ad tech industry called Teads, which I joined 11 years ago as a new grad. I've been using dbt Core on BigQuery since then. I write about data engineering both on Medium and Substack. I contribute on dbt-bigquery. I wrote an article that was then featured on the Developer Blog called BigQuery ingestion-time partitioning and partition copy with dbt.
+image: /img/community/spotlight/christophe-oudar.jpg
+pronouns: he/him
+location: Montpellier, France
+jobTitle: Staff Engineer
+companyName: Teads
+socialLinks:
+ - name: X
+ link: https://x.com/Kayrnt
+ - name: LinkedIn
+ link: https://www.linkedin.com/in/christopheoudar/
+ - name: Substack
+ link: https://smallbigdata.substack.com/
+dateCreated: 2024-11-08
+hide_table_of_contents: true
+communityAward: true
+communityAwardYear: 2024
+---
+
+## When did you join the dbt community and in what way has it impacted your career?
+
+I joined the community in November 2021 as a way to explore how to move our in-house data modeling layer to dbt. The transition took over a year while we ensured we could cover all our bases and add missing features to dbt-bigquery. That project was one of stepping stones that helped me to move from senior to staff level at my current job.
+
+## What dbt community leader do you identify with? How are you looking to grow your leadership in the dbt community?
+
+I identify with leaders that have strong convictions about how data engineering should move forward but remain open to innovation and ideas from everyone to bring the best to the field and make it as inclusive as possible to all cultures and profiles. I think that could mean people like Jordan Tigani or Mark Raasveldt. In the dbt community, my leadership has looked like helping people struggling and offering better ways to simplify one's day to day work when possible.
+
+## What have you learned from community members? What do you hope others can learn from you?
+
+I read a lot of articles about dbt, especially when I got started with it. It helped me a lot to build a proper Slim CI that could fit my company's ways of working. I also got to see how data pipelines were done in other companies and the pros and cons of my approaches. I hope I can share more of that knowledge for people to pick what's best for their needs.
+
diff --git a/website/docs/community/spotlight/fabiyi-opeyemi.md b/website/docs/community/spotlight/fabiyi-opeyemi.md
index 18a311fa437..9b78ff179c1 100644
--- a/website/docs/community/spotlight/fabiyi-opeyemi.md
+++ b/website/docs/community/spotlight/fabiyi-opeyemi.md
@@ -8,7 +8,7 @@ description: |
image: /img/community/spotlight/fabiyi-opeyemi.jpg
pronouns: he/him
location: Lagos, Nigeria
-jobTitle: Senior Analytics Engineer
+jobTitle: Analytics Manager
companyName: Data Culture
organization: Young Data Professionals (YDP)
socialLinks:
@@ -16,10 +16,10 @@ socialLinks:
link: https://twitter.com/Opiano_1
- name: LinkedIn
link: https://www.linkedin.com/in/opeyemifabiyi/
-dateCreated: 2023-11-06
+dateCreated: 2024-11-02
hide_table_of_contents: true
communityAward: true
-communityAwardYear: 2023
+communityAwardYear: 2024
---
## When did you join the dbt community and in what way has it impacted your career?
diff --git a/website/docs/community/spotlight/jenna-jordan.md b/website/docs/community/spotlight/jenna-jordan.md
new file mode 100644
index 00000000000..cb16ca0afdf
--- /dev/null
+++ b/website/docs/community/spotlight/jenna-jordan.md
@@ -0,0 +1,36 @@
+---
+id: jenna-jordan
+title: Jenna Jordan
+description: |
+ I am a Senior Data Management Consultant with Analytics8, where I advise clients on dbt best practices (especially regarding dbt Mesh and the various shifts in governance and strategy that come with it). My experiences working within a dbt Mesh architecture and all of the difficulties organizations could run into with such a major paradigm shift inspired my peer exchange (role-playing/simulation game) at Coalesce 2024: "Governance co-lab: We the people, in order to govern data, do establish processes." I also experimented with bringing role-playing scenarios to data problems at the September 2024 Chicago dbt Meetup, hosted by Analytics8. I occasionally write long blog posts on my website, if you're up for the read.
+image: /img/community/spotlight/jenna-jordan.jpg
+pronouns: she/her
+location: Asheville, USA
+jobTitle: Senior Data Management Consultant
+companyName: Analytics8
+socialLinks:
+ - name: LinkedIn
+ link: https://www.linkedin.com/in/jennajordan1/
+ - name: Personal website
+ link: https://jennajordan.me/
+dateCreated: 2024-11-01
+hide_table_of_contents: true
+communityAward: true
+communityAwardYear: 2024
+---
+
+## When did you join the dbt community and in what way has it impacted your career?
+
+My dbt learning journey kicked off with the CoRise (now Uplimit) course Analytics Engineering with dbt, with Emily Hawkins and Jake Hannan, in February 2022 – less than a month after starting as a data engineer with the City of Boston Analytics Team. About a year later, I spearheaded the adoption of dbt at the City and got to build the project and associated architecture from scratch – which is probably the best learning experience you could ask for! I saw the value dbt could bring to improving data management processes at the City, and I knew there were other cities and local governments that could benefit from dbt as well, which motivated me to find my fellow co-speakers Ian Rose and Laurie Merrell to give a talk at Coalesce 2023 called "From Coast to Coast: Implementing dbt in the public sector". As a part of our goal to identify and cultivate a community of dbt practitioners in the public (and adjacent) sectors, we also started the dbt Community Slack channel #industry-public-sector. That experience allowed me to continue to grow my career and find my current role - as well as connect with so many amazing data folks!
+
+## What dbt community leader do you identify with? How are you looking to grow your leadership in the dbt community?
+
+There are many leaders in the dbt community that I admire and identify with – I won’t list them all out because I will invariably miss someone (but… you probably know who you are). Technical prowess is always enviable, but I most admire those who bring the human element to data work: those who aren’t afraid to be their authentic selves, cultivate a practice of empathy and compassion, and are driven by curiosity and a desire to help others. I’ve never set out to be a leader, and I still don’t really consider myself to be a leader – I’m much more comfortable in the role of a librarian. I just want to help people by connecting them to the information and resources that they may need.
+
+## What have you learned from community members? What do you hope others can learn from you?
+
+Pretty much everything I’ve learned about dbt and working in a mature analytics ecosystem I’ve learned from dbt community members. The dbt Community Slack is full of useful information and advice, and has also helped me identify experts about certain topics that I can chat with to learn even more. When I find someone sharing useful information, I usually try to find and follow them on social media so I can see more of their content. If there is one piece of advice I want to share, it is this: don’t be afraid to engage. Ask for help when you need it, but also offer help freely. Engage with the community with the same respect and grace you would offer your friends and coworkers.
+
+## Anything else interesting you want to tell us?
+
+Library Science is so much more than the Dewey Decimal System (seriously, ask a librarian about Dewey for a juicy rant). RDF triples (for knowledge graphs) are queried using SPARQL (pronounced “sparkle”). An antelope can be a document. The correct way to write a date/time is ISO-8601. The oldest known table (of the spreadsheet variety) is from 5,000 years ago – record-keeping predates literature by a significant margin. Zip codes aren’t polygons – they don’t contain an area or have boundaries. Computers don’t always return 0.3 when asked to add 0.1 + 0.2. SQL was the sequel to SQUARE. Before computers, people programmed looms (weaving is binary). What? You asked!! On a more serious note – data teams: start hiring librarians. No, seriously. No degree could have prepared me better for what I do in the data field than my M.S. in Library & Information Science. I promise, you want the skillset & mindset that a librarian will bring to your team.
diff --git a/website/docs/community/spotlight/meagan-palmer.md b/website/docs/community/spotlight/meagan-palmer.md
index ff45a3d6b7d..697a35d80f0 100644
--- a/website/docs/community/spotlight/meagan-palmer.md
+++ b/website/docs/community/spotlight/meagan-palmer.md
@@ -14,9 +14,10 @@ companyName: Altis Consulting
socialLinks:
- name: LinkedIn
link: https://www.linkedin.com/in/meaganpalmer/
-dateCreated: 2024-07-29
+dateCreated: 2024-11-04
hide_table_of_contents: true
-communityAward: false
+communityAward: true
+communityAwardYear: 2024
---
## When did you join the dbt community and in what way has it impacted your career?
@@ -27,9 +28,9 @@ I was fortunate that Jon Bradley at Nearmap had the vision to engage the then Fi
Being in Australia, I often see replies from Jeremy Yeo to people in the dbt Slack. His clarity of communication is impressive.
-For growth, I'm hoping that others can benefit from the wide range of experience I have. My newsletter, Analytics Engineering Today on LinkedIn aims to upskill the dbt Community and shed some light on some useful features that might not be well known.
+For growth, I'm hoping that others can benefit from the wide range of experience I have. My LinkedIn Newsletter, Analytics Engineering Today aims to upskill the dbt Community and shed some light on some useful features that might not be well known.
-I'll be at Coalesce and am doing some webinars/events later in the year. Come say hi, I love talking dbt and analytics engineering with people.
+I was at Coalesce Onlineand am doing some webinars/events later in the year. Come say hi, I love talking dbt and analytics engineering with people.
## What have you learned from community members? What do you hope others can learn from you?
diff --git a/website/docs/community/spotlight/mike-stanley.md b/website/docs/community/spotlight/mike-stanley.md
new file mode 100644
index 00000000000..1d1f13f1a93
--- /dev/null
+++ b/website/docs/community/spotlight/mike-stanley.md
@@ -0,0 +1,18 @@
+---
+id: mike-stanley
+title: Mike Stanley
+description: |
+ Mike has been very helpful in dbt Slack, answering questions in channels like #advice-for-dbt-power-users, #advice-dbt-help, #db-bigquery and more. From October 2023 to October 2024, Mike responded to others’ posts 2,355 times. Mike is a truly humble community member who needs no credit or awards, and doesn't even have a profile picture in Slack.
+image: /img/community/spotlight/mike-stanley.jpg
+pronouns: he/him
+location: Gloucester, England, United Kingdom
+jobTitle: Manager, Data
+companyName: Freetrade
+socialLinks:
+ - name: LinkedIn
+ link: https://www.linkedin.com/in/mike-stanley-31616994/
+dateCreated: 2024-11-05
+hide_table_of_contents: true
+communityAward: true
+communityAwardYear: 2024
+---
diff --git a/website/docs/community/spotlight/original-dbt-athena-maintainers.md b/website/docs/community/spotlight/original-dbt-athena-maintainers.md
new file mode 100644
index 00000000000..158aa30997b
--- /dev/null
+++ b/website/docs/community/spotlight/original-dbt-athena-maintainers.md
@@ -0,0 +1,42 @@
+---
+id: original-dbt-athena-maintainers
+title: The Original dbt-athena Maintainers
+description: |
+ The original dbt-athena Maintainers is a group of 5 people—Jérémy Guiselin, Mattia, Jesse Dobbelaere, Serhii Dimchenko, and Nicola Corda—who met via dbt Slack in the #db-athena channel, with the aim to make make dbt-athena a production-ready adapter.
+ In the first periods, Winter 2022 and Spring 2023, we focused on contributing directly to the adapter, adding relevant features like Iceberg and Lake Formation support, and stabilizing some internal behaviour.
+ On a second iteration our role was triaging, providing community support and bug fixing. We encouraged community members to make their first contributions, and helped them to merge their PRs.
+image: /img/community/spotlight/dbt-athena-groupheadshot.png
+location: Europe
+jobTitle: A group of data-engineers
+companyName: Mix of companies
+organization: dbt-athena (since November 2022)
+socialLinks:
+ - name: Jérémy's LinkedIn
+ link: https://www.linkedin.com/in/jrmyy/
+ - name: Mattia's LinkedIn
+ link: https://www.linkedin.com/in/mattia-sappa/
+ - name: Jesse's LinkedIn
+ link: https://www.linkedin.com/in/dobbelaerejesse/
+ - name: Serhii's LinkedIn
+ link: https://www.linkedin.com/in/serhii-dimchenko-075b3061/
+ - name: Nicola's LinkedIn
+ link: https://www.linkedin.com/in/nicolacorda/
+dateCreated: 2024-11-06
+hide_table_of_contents: true
+communityAward: true
+communityAwardYear: 2024
+---
+
+## When did you join the dbt community and in what way has it impacted your career?
+
+The dbt community allowed the dbt-athena maintainers to meet each other, and share the common goal of making the dbt-athena adapter production-ready.
+
+## Which dbt Community leader do you identify with? How are you looking to grow your leadership in the dbt community?
+
+As we grow, we are looking to embody democratic leadership.
+
+## What have you learned from community members? What do you hope others can learn from you?
+
+We learned that the power of the community was endless. People started to share best practises, and some of the best practises were incorporated directly in dbt-athena, allowing people to run the adapter smoothly in their production environment.
+We reached a point where people started to ask advice for their AWS architecture, which we found pretty awesome.
+
diff --git a/website/docs/community/spotlight/ruth-onyekwe.md b/website/docs/community/spotlight/ruth-onyekwe.md
new file mode 100644
index 00000000000..cf07e98a4f7
--- /dev/null
+++ b/website/docs/community/spotlight/ruth-onyekwe.md
@@ -0,0 +1,31 @@
+---
+id: ruth-onyekwe
+title: Ruth Onyekwe
+description: |
+ I've been working in the world of Data Analytics for over 5 years and have been part of the dbt community for the last 4. With a background in International Business and Digital Marketing, I experienced first hand the need for reliable data to fuel business decisions. This inspired a career move into the technology space to be able to work with the tools and the people that were facilitating this process. Today I am leading teams to deliver data modernization projects, as well as helping grow the analytics arm of my company on a day to day basis. I also have the privilege of organising the dbt Meetups in Barcelona, Spain - and am excited to continue to grow the community across Europe.
+image: /img/community/spotlight/ruth-onyekwe.jpeg
+pronouns: she/her
+location: Madrid, Spain
+jobTitle: Data Analytics Manager
+companyName: Spaulding Ridge
+socialLinks:
+ - name: LinkedIn
+ link: https://www.linkedin.com/in/ruth-onyekwe/
+dateCreated: 2024-11-07
+hide_table_of_contents: true
+communityAward: true
+communityAwardYear: 2024
+---
+
+## When did you join the dbt community and in what way has it impacted your career?
+
+I joined the dbt community in 2021, after meeting dbt Labs reps at a conference. Through partnering with dbt Labs and learning the technology, we (Spaulding Ridge) were able to open a whole new offering in our service catalogue, and meet the growing needs of our customers.
+
+## Which dbt Community leader do you identify with? How are you looking to grow your leadership in the dbt community?
+
+I identify with the transparent leaders - those willing to share their learnings, knowledge, and experiences. I want to encourage other dbt enthusiasts to stretch themselves professionally and actively participate in the analytics community.
+
+## What have you learned from community members? What do you hope others can learn from you?
+
+I've learnt that most of us working in data have experienced the same struggles, be it searching for the best testing frameworks, or deciding how to build optimised and scalable models, or searching for the answers to non-technical questions like how to best organise teams or how to communicate with business stakeholders and translate their needs - we're all faced with the same dilemmas. And the great thing I've learned being in the dbt community, is that if you're brave enough to share your stories, you'll connect with someone who has already gone through those experiences, and can help you reach a solution a lot faster than if you tried to start from scratch.
+
diff --git a/website/docs/docs/cloud/connect-data-platform/about-connections.md b/website/docs/docs/cloud/connect-data-platform/about-connections.md
index 6f2f140b724..89dd13808ec 100644
--- a/website/docs/docs/cloud/connect-data-platform/about-connections.md
+++ b/website/docs/docs/cloud/connect-data-platform/about-connections.md
@@ -20,9 +20,12 @@ dbt Cloud can connect with a variety of data platform providers including:
- [Starburst or Trino](/docs/cloud/connect-data-platform/connect-starburst-trino)
- [Teradata](/docs/cloud/connect-data-platform/connect-teradata)
-You can connect to your database in dbt Cloud by clicking the gear in the top right and selecting **Account Settings**. From the Account Settings page, click **+ New Project**.
+To connect to your database in dbt Cloud:
-
+1. Click your account name at the bottom of the left-side menu and click **Account settings**
+2. Select **Projects** from the top left, and from there click **New Project**
+
+
These connection instructions provide the basic fields required for configuring a data platform connection in dbt Cloud. For more detailed guides, which include demo project data, read our [Quickstart guides](https://docs.getdbt.com/guides)
@@ -41,7 +44,7 @@ Connections created with APIs before this change cannot be accessed with the [la
Warehouse connections are an account-level resource. As such you can find them under **Accounts Settings** > **Connections**:
-
+
Warehouse connections can be re-used across projects. If multiple projects all connect to the same warehouse, you should re-use the same connection to streamline your management operations. Connections are assigned to a project via an [environment](/docs/dbt-cloud-environments).
diff --git a/website/docs/docs/cloud/connect-data-platform/connect-amazon-athena.md b/website/docs/docs/cloud/connect-data-platform/connect-amazon-athena.md
index 0b2f844ccac..f1009f61274 100644
--- a/website/docs/docs/cloud/connect-data-platform/connect-amazon-athena.md
+++ b/website/docs/docs/cloud/connect-data-platform/connect-amazon-athena.md
@@ -5,7 +5,7 @@ description: "Configure the Amazon Athena data platform connection in dbt Cloud.
sidebar_label: "Connect Amazon Athena"
---
-# Connect Amazon Athena
+# Connect Amazon Athena
Your environment(s) must be on ["Versionless"](/docs/dbt-versions/versionless-cloud) to use the Amazon Athena connection.
diff --git a/website/docs/docs/cloud/git/authenticate-azure.md b/website/docs/docs/cloud/git/authenticate-azure.md
index 42028bf993b..5278c134f72 100644
--- a/website/docs/docs/cloud/git/authenticate-azure.md
+++ b/website/docs/docs/cloud/git/authenticate-azure.md
@@ -13,9 +13,9 @@ If you use the dbt Cloud IDE or dbt Cloud CLI to collaborate on your team's Azur
Connect your dbt Cloud profile to Azure DevOps using OAuth:
-1. Click the gear icon at the top right and select **Profile settings**.
-2. Click **Linked Accounts**.
-3. Next to Azure DevOps, click **Link**.
+1. Click your account name at the bottom of the left-side menu and click **Account settings**
+2. Scroll down to **Your profile** and select **Personal profile**.
+3. Go to the **Linked accounts** section in the middle of the page.
4. Once you're redirected to Azure DevOps, sign into your account.
diff --git a/website/docs/docs/collaborate/project-recommendations.md b/website/docs/docs/collaborate/project-recommendations.md
index 12007c6b88b..c9499579e54 100644
--- a/website/docs/docs/collaborate/project-recommendations.md
+++ b/website/docs/docs/collaborate/project-recommendations.md
@@ -20,7 +20,7 @@ The Recommendations overview page includes two top-level metrics measuring the t
- **Model test coverage** — The percent of models in your project (models not from a package or imported via dbt Mesh) with at least one dbt test configured on them.
- **Model documentation coverage** — The percent of models in your project (models not from a package or imported via dbt Mesh) with a description.
-
+
## List of rules
The following table lists the rules currently defined in the `dbt_project_evaluator` [package](https://hub.getdbt.com/dbt-labs/dbt_project_evaluator/latest/).
diff --git a/website/docs/docs/dbt-versions/release-notes.md b/website/docs/docs/dbt-versions/release-notes.md
index ae6ed5f01a4..010042ea49f 100644
--- a/website/docs/docs/dbt-versions/release-notes.md
+++ b/website/docs/docs/dbt-versions/release-notes.md
@@ -19,18 +19,6 @@ Release notes are grouped by month for both multi-tenant and virtual private clo
\* The official release date for this new format of release notes is May 15th, 2024. Historical release notes for prior dates may not reflect all available features released earlier this year or their tenancy availability.
## October 2024
-
-- **Behavior change:** [Multi-factor authentication](/docs/cloud/manage-access/mfa) is now enforced on all users who log in with username and password credentials.
-- **Enhancement**: The dbt Semantic Layer JDBC now allows users to paginate `semantic_layer.metrics()` and `semantic_layer.dimensions()` for metrics and dimensions using `page_size` and `page_number` parameters. Refer to [Paginate metadata calls](/docs/dbt-cloud-apis/sl-jdbc#querying-the-api-for-metric-metadata) for more information.
-- **Enhancement**: The dbt Semantic Layer JDBC now allows you to filter your metrics to include only those that contain a specific substring, using the `search` parameter. If no substring is provided, the query returns all metrics. Refer to [Fetch metrics by substring search](/docs/dbt-cloud-apis/sl-jdbc#querying-the-api-for-metric-metadata) for more information.
-- **Fix**: The [dbt Semantic Layer Excel integration](/docs/cloud-integrations/semantic-layer/excel) now correctly surfaces errors when a query fails to execute. Previously, it was not clear why a query failed to run.
-- **Fix:** Previously, POST requests to the Jobs API with invalid `cron` strings would return HTTP response status code 500s but would update the underlying entity. Now, POST requests to the Jobs API with invalid `cron` strings will result in status code 400s, without the underlying entity being updated.
-- **Fix:** Fixed an issue where the `Source` view page in dbt Explorer did not correctly display source freshness status if older than 30 days.
-- **Fix:** The UI now indicates when the description of a model is inherited from a catalog comment.
-- **Behavior change:** User API tokens have been deprecated. Update to [personal access tokens](/docs/dbt-cloud-apis/user-tokens) if you have any still in use.
-- **New**: The dbt Cloud IDE supports signed commits for Git, available for Enterprise plans. You can sign your Git commits when pushing them to the repository to prevent impersonation and enhance security. Supported Git providers are GitHub and GitLab. Refer to [Git commit signing](/docs/cloud/dbt-cloud-ide/git-commit-signing.md) for more information.
-- **New:** With dbt Mesh, you can now enable bidirectional dependencies across your projects. Previously, dbt enforced dependencies to only go in one direction. dbt checks for cycles across projects and raises errors if any are detected. For details, refer to [Cycle detection](/docs/collaborate/govern/project-dependencies#cycle-detection). There's also the [Intro to dbt Mesh](/best-practices/how-we-mesh/mesh-1-intro) guide to help you learn more best practices.
-
Documentation for new features and functionality announced at Coalesce 2024:
@@ -54,8 +42,17 @@ Release notes are grouped by month for both multi-tenant and virtual private clo
- [Python SDK](https://docs.getdbt.com/docs/dbt-cloud-apis/sl-python) is now generally available
-
-
+
+- **Behavior change:** [Multi-factor authentication](/docs/cloud/manage-access/mfa) is now enforced on all users who log in with username and password credentials.
+- **Enhancement**: The dbt Semantic Layer JDBC now allows users to paginate `semantic_layer.metrics()` and `semantic_layer.dimensions()` for metrics and dimensions using `page_size` and `page_number` parameters. Refer to [Paginate metadata calls](/docs/dbt-cloud-apis/sl-jdbc#querying-the-api-for-metric-metadata) for more information.
+- **Enhancement**: The dbt Semantic Layer JDBC now allows you to filter your metrics to include only those that contain a specific substring, using the `search` parameter. If no substring is provided, the query returns all metrics. Refer to [Fetch metrics by substring search](/docs/dbt-cloud-apis/sl-jdbc#querying-the-api-for-metric-metadata) for more information.
+- **Fix**: The [dbt Semantic Layer Excel integration](/docs/cloud-integrations/semantic-layer/excel) now correctly surfaces errors when a query fails to execute. Previously, it was not clear why a query failed to run.
+- **Fix:** Previously, POST requests to the Jobs API with invalid `cron` strings would return HTTP response status code 500s but would update the underlying entity. Now, POST requests to the Jobs API with invalid `cron` strings will result in status code 400s, without the underlying entity being updated.
+- **Fix:** Fixed an issue where the `Source` view page in dbt Explorer did not correctly display source freshness status if older than 30 days.
+- **Fix:** The UI now indicates when the description of a model is inherited from a catalog comment.
+- **Behavior change:** User API tokens have been deprecated. Update to [personal access tokens](/docs/dbt-cloud-apis/user-tokens) if you have any still in use.
+- **New**: The dbt Cloud IDE supports signed commits for Git, available for Enterprise plans. You can sign your Git commits when pushing them to the repository to prevent impersonation and enhance security. Supported Git providers are GitHub and GitLab. Refer to [Git commit signing](/docs/cloud/dbt-cloud-ide/git-commit-signing.md) for more information.
+- **New:** With dbt Mesh, you can now enable bidirectional dependencies across your projects. Previously, dbt enforced dependencies to only go in one direction. dbt checks for cycles across projects and raises errors if any are detected. For details, refer to [Cycle detection](/docs/collaborate/govern/project-dependencies#cycle-detection). There's also the [Intro to dbt Mesh](/best-practices/how-we-mesh/mesh-1-intro) guide to help you learn more best practices.
- **New**: The [dbt Semantic Layer Python software development kit](/docs/dbt-cloud-apis/sl-python) is now [generally available](/docs/dbt-versions/product-lifecycles). It provides users with easy access to the dbt Semantic Layer with Python and enables developers to interact with the dbt Semantic Layer APIs to query metrics/dimensions in downstream tools.
- **Enhancement**: You can now add a description to a singular data test in dbt Cloud Versionless. Use the [`description` property](/reference/resource-properties/description) to document [singular data tests](/docs/build/data-tests#singular-data-tests). You can also use [docs block](/docs/build/documentation#using-docs-blocks) to capture your test description. The enhancement will be included in upcoming dbt Core 1.9 release.
- **New**: Introducing the [microbatch incremental model strategy](/docs/build/incremental-microbatch) (beta), available in dbt Cloud Versionless and will soon be supported in dbt Core 1.9. The microbatch strategy allows for efficient, batch-based processing of large time-series datasets for improved performance and resiliency, especially when you're working with data that changes over time (like new records being added daily). To enable this feature in dbt Cloud, set the `DBT_EXPERIMENTAL_MICROBATCH` environment variable to `true` in your project.
diff --git a/website/docs/docs/dbt-versions/versionless-cloud.md b/website/docs/docs/dbt-versions/versionless-cloud.md
index 17a97b0f319..4bb843898c2 100644
--- a/website/docs/docs/dbt-versions/versionless-cloud.md
+++ b/website/docs/docs/dbt-versions/versionless-cloud.md
@@ -4,15 +4,15 @@ sidebar_label: "Upgrade to \"Versionless\" "
description: "Learn how to go versionless in dbt Cloud. You never have to perform an upgrade again. Plus, you'll be able to access new features and enhancements as soon as they become available. "
---
-dbt Cloud is going versionless. Soon, your environments and jobs will always run the latest features and functionality.
+Since May 2024, new capabilities in dbt are delivered continuously to dbt Cloud. We call this "versionless dbt," because your projects and environments are upgraded automatically.
This will require you to make one final update to your current jobs and environments. When that's done, you'll never have to think about managing, coordinating, or upgrading dbt versions again.
-Move your environments and jobs to "Versionless" to get all the functionality in the latest features before they're in dbt Core — and more! — along with access to the new features and fixes as soon as they’re released.
+By moving your environments and jobs to "Versionless," you can get all the functionality in the latest features before they're in dbt Core — and more! — along with access to the new features and fixes as soon as they’re released.
## Tips for upgrading {#upgrade-tips}
-If you regularly develop your dbt project in dbt Cloud and this is your first time trying “Versionless,” dbt Labs recommends that you start in development because it will be the fastest for investigation and iteration. [Override your dbt version in development](/docs/dbt-versions/upgrade-dbt-version-in-cloud#override-dbt-version). Then, launch the IDE or Cloud CLI and do your development work as usual. Everything should work as you expect.
+If you regularly develop your dbt project in dbt Cloud and this is your first time trying “Versionless,” dbt Labs recommends that you try upgrading your project in a development environment. [Override your dbt version in development](/docs/dbt-versions/upgrade-dbt-version-in-cloud#override-dbt-version). Then, launch the IDE or Cloud CLI and do your development work as usual. Everything should work as you expect.
If you do see something unexpected or surprising, revert back to the previous version and record the differences you observed. [Contact dbt Cloud support](/docs/dbt-support#dbt-cloud-support) with your findings for a more detailed investigation.
@@ -22,13 +22,23 @@ If your organization has multiple dbt projects, we recommend starting your upgra
## Considerations
-The following is our guidance on some important considerations regarding dbt projects as part of the upgrade.
+The following is our guidance on some important considerations regarding dbt projects as part of the upgrade.
To learn more about how dbt Labs deploys stable dbt upgrades in a safe manner to dbt Cloud, we recommend that you read our blog post [How we're making sure you can confidently go "Versionless" in dbt Cloud](https://docs.getdbt.com/blog/latest-dbt-stability) for details.
-
+If you're running dbt version 1.6 or older, please know that your version of dbt Core has reached [end-of-life (EOL)](/docs/dbt-versions/core#eol-version-support) and is no longer supported. We strongly recommend that you update to a newer version as soon as reasonably possible.
-If you're running dbt version 1.5 or older, please know that your version of dbt Core has reached [end-of-life (EOL)](/docs/dbt-versions/core#eol-version-support) and is no longer supported. We strongly recommend that you update to a newer version as soon as reasonably possible. In the coming months, we're planning to automatically migrate jobs and environments on these older, unsupported versions.
+dbt Labs has extended the critical support period of dbt Core v1.7 for dbt Cloud Enterprise customers.
+
+
+
+If you're running dbt version v1.6 or older, please know that your version of dbt Core has reached [end-of-life (EOL)](/docs/dbt-versions/core#eol-version-support) and is no longer supported. We strongly recommend that you update to a newer version as soon as reasonably possible.
+
+dbt Labs has extended the "Critical Support" period of dbt Core v1.7 for dbt Cloud Enterprise customers while we work through the migration with those customers to automatic upgrades. In the meantime, this means that v1.7 will continue to be accessible in dbt Cloud for Enteprise customers, jobs and environments on v1.7 for those customers will not be automatically migrated to "Versionless," and dbt Labs will continue to fix critical bugs and security issues.
+
+dbt Cloud accounts on the Developer and Team plans will be migrated to "Versionless" dbt after November 1, 2024. If you know that your project will not be compatible with the upgrade, for one of the reasons described here, or a different reason in your own testing, you should [contact dbt Cloud support](https://docs.getdbt.com/docs/dbt-support#dbt-cloud-support) to request an extension.
+
+If your account has been migrated to "Versionless," and you are seeing net-new failures in your scheduled dbt jobs, you should also [contact dbt Cloud support](https://docs.getdbt.com/docs/dbt-support#dbt-cloud-support) to request an extension.
@@ -38,6 +48,26 @@ The legacy dbt Semantic Layer was deprecated in the second half of 2023. We reco
+
+
+If you are upgrading from a very old unsupported version of dbt Core, you may run into one of these edge cases after the upgrade to a newer version:
+- [v1.3] Customers with non-dbt `.py` files defined within their project directories, such as `models/`. Since v1.3, dbt expects these files be valid [Python models](/docs/build/python-models). The customer needs to move these files out of their `models/` directory, or ignore them via `.dbtignore`
+- [v1.5] Customers who have `--m` in their job definitions, instead of `-m` or `--models`. This autocompletion (`--m[odels]` for `--models`) has never been officially documented or supported. It was an implicit behavior of argparse (CLI library used in dbt-core v1.0-1.4) that is not supported by `click` (the CLI library used in dbt-core since v1.5+).
+- [v1.5] Empty invalid `tests` config start raising a validation error](https://docs.getdbt.com/docs/dbt-versions/core-upgrade/upgrading-to-v1.5). Replace empty `tests` config with `tests: []` or remove it altogether.
+- [v1.6] Performance optimization to `load_result` means you cannot call it on the same query result multiple times (https://github.com/dbt-labs/dbt-core/pull/7371)
+
+You should [contact dbt Cloud support](https://docs.getdbt.com/docs/dbt-support#dbt-cloud-support) to request an extension, during which you will need to make those updates.
+
+
+
+
+
+For the vast majority of customers, there is no further action needed.
+
+If you see new failures in your scheduled jobs now that they are running on a newer version of dbt, you may need to update your project code to account for one of the edge cases described on this page. You should [contact dbt Cloud support](https://docs.getdbt.com/docs/dbt-support#dbt-cloud-support) to request an extension, during which you will need to make those updates.
+
+
+
When we talk about _latest version_, we’re referring to the underlying runtime for dbt, not the versions of packages you’re installing. Our continuous release for dbt includes testing against several popular dbt packages. This ensures that updates we make to dbt-core, adapters, or anywhere else are compatible with the code in those packages.
@@ -54,13 +84,15 @@ If the package you’ve installed relies on _undocumented_ functionality of db
-No. Going forward, “Versionless” will be how all customers are going to access new functionality and ongoing support in dbt Cloud. We believe this is the best way for us to offer a reliable, stable, and secure runtime for dbt with continuous and consistent updates.
+No. Going forward, customers will access new functionality and ongoing support in dbt Cloud by receiving automatic updates. We believe this is the best way for us to offer a reliable, stable, and secure runtime for dbt, and for you as dbt users to be able to consistently take advantage of new features.
In 2023 (and earlier), customers were expected to manage their own upgrades by selecting dbt Core versions, up to and including dbt Core v1.7, which was released in October 2023. (Way back in 2021, dbt Cloud customers would pick specific _patch releases_ of dbt Core, such as upgrading from `v0.21.0` to `v0.21.1`. We’ve come a long way since then!)
-In 2024, we've changed the way that new dbt functionality is made available for dbt Cloud customers: continuously. Behavior or breaking changes are gated behind opt-in flags. Users don't need to spend valuable time managing their own upgrades. This is called "Versionless" dbt Cloud and it's required for accessing any new functionality that we've put out in 2024+.
+In 2024, we've changed the way that new dbt functionality is made available for dbt Cloud customers. Behavior or breaking changes are gated behind opt-in flags. Users don't need to spend valuable time managing their own upgrades. Currently, it is possible to receive continuous (daily) updates. We are adding other release cadence options for managed customers of dbt Cloud by the end of the year.
+
+Opting into a release cadence with automated upgrades is required for accessing any new functionality that we've released in 2024, and going forward.
-We will absolutely continue to release new minor versions of dbt Core (OSS), including v1.9 which will be available later this year. When we do, it will be a subset of the functionality that's already available to dbt Cloud customers and always after the functionality has been available in dbt Cloud.
+We continue to release new minor versions of dbt Core (OSS), including v1.9 which will be available later this year. When we do, it will be a subset of the functionality that's already available to dbt Cloud customers, and always after the functionality has been available in dbt Cloud.
diff --git a/website/docs/faqs/Accounts/transfer-account.md b/website/docs/faqs/Accounts/transfer-account.md
index 693061c55c6..e694636cf68 100644
--- a/website/docs/faqs/Accounts/transfer-account.md
+++ b/website/docs/faqs/Accounts/transfer-account.md
@@ -10,7 +10,7 @@ You can transfer your dbt Cloud [access control](/docs/cloud/manage-access/about
| Account plan| Steps |
| ------ | ---------- |
-| **Developer** | You can transfer ownership by changing the email directly on your dbt Cloud profile page, which you can access using this URL when you replace `YOUR_ACCESS_URL` with the [appropriate Access URL](/docs/cloud/about-cloud/access-regions-ip-addresses) for your region and plan: `https://YOUR_ACCESS_URL/settings/profile`. Before doing this, please ensure that you unlink your GitHub profile. |
+| **Developer** | You can transfer ownership by changing the email directly on your dbt Cloud profile page, which you can access using this URL when you replace `YOUR_ACCESS_URL` with the [appropriate Access URL](/docs/cloud/about-cloud/access-regions-ip-addresses) for your region and plan: `https://YOUR_ACCESS_URL/settings/profile`. Before doing this, please ensure that you unlink your GitHub profile. The email address of the new account owner cannot be associated with another dbt Cloud account.|
| **Team** | Existing account admins with account access can add users to, or remove users from the owner group. |
| **Enterprise** | Account admins can add users to, or remove users from a group with Account Admin permissions. |
| **If all account owners left the company** | If the account owner has left your organization, you will need to work with _your_ IT department to have incoming emails forwarded to the new account owner. Once your IT department has redirected the emails, you can request to reset the user password. Once you log in, you can change the email on the Profile page when you replace `YOUR_ACCESS_URL` with the [appropriate Access URL](/docs/cloud/about-cloud/access-regions-ip-addresses) for your region and plan: `https://YOUR_ACCESS_URL/settings/profile`. |
diff --git a/website/docs/reference/dbt-commands.md b/website/docs/reference/dbt-commands.md
index 8386cf61731..ca9a7725eb2 100644
--- a/website/docs/reference/dbt-commands.md
+++ b/website/docs/reference/dbt-commands.md
@@ -11,7 +11,7 @@ A key distinction with the tools mentioned, is that dbt Cloud CLI and IDE are de
## Parallel execution
-dbt Cloud allows for parallel execution of commands, enhancing efficiency without compromising data integrity. This enables you to run multiple commands at the same time, however it's important to understand which commands can be run in parallel and which can't.
+dbt Cloud allows for concurrent execution of commands, enhancing efficiency without compromising data integrity. This enables you to run multiple commands at the same time. However, it's important to understand which commands can be run in parallel and which can't.
In contrast, [`dbt-core` _doesn't_ support](/reference/programmatic-invocations#parallel-execution-not-supported) safe parallel execution for multiple invocations in the same process, and requires users to manage concurrency manually to ensure data integrity and system stability.
diff --git a/website/docs/reference/global-configs/adapter-behavior-changes.md b/website/docs/reference/global-configs/adapter-behavior-changes.md
index bd0ba9f7404..a755f8cfe50 100644
--- a/website/docs/reference/global-configs/adapter-behavior-changes.md
+++ b/website/docs/reference/global-configs/adapter-behavior-changes.md
@@ -14,10 +14,17 @@ Some adapters can display behavior changes when certain flags are enabled. The f
+
+
+
-
\ No newline at end of file
+
diff --git a/website/docs/reference/global-configs/databricks-changes.md b/website/docs/reference/global-configs/databricks-changes.md
new file mode 100644
index 00000000000..ca24b822ae5
--- /dev/null
+++ b/website/docs/reference/global-configs/databricks-changes.md
@@ -0,0 +1,26 @@
+---
+title: "Databricks adapter behavior changes"
+id: "databricks-changes"
+sidebar: "Databricks"
+---
+
+The following are the current [behavior change flags](/docs/reference/global-configs/behavior-changes.md#behavior-change-flags) that are specific to `dbt-databricks`:
+
+| Flag | `dbt-databricks`: Intro | `dbt-databricks`: Maturity |
+| ----------------------------- | ----------------------- | -------------------------- |
+| `use_info_schema_for_columns` | 1.9.0 | TBD |
+| `use_user_folder_for_python` | 1.9.0 | TBD |
+
+### Use information schema for columns
+
+The `use_info_schema_for_columns` flag is `False` by default.
+
+Setting this flag to `True` will use `information_schema` rather than `describe extended` to get column metadata for Unity Catalog tables. This setting helps you avoid issues where `describe extended` truncates information when the type is a complex struct. However, this setting is not yet the default behavior, as there are performance impacts due to a Databricks metadata limitation because of the need to run `REPAIR TABLE {{relation}} SYNC METADATA` before querying to ensure the `information_schema` is complete.
+
+This flag will become the default behavior when this additional query is no longer needed.
+
+### Use user's folder for Python model notebooks
+
+The `use_user_folder_for_python` flag is `False` by default and results in writing uploaded python model notebooks to `/Shared/dbt_python_models/{{schema}}/`. Setting this flag to `True` will write notebooks to `/Users/{{current user}}/{{catalog}}/{{schema}}/` Writing to the `Shared` folder is deprecated by Databricks as it does not align with governance best practices.
+
+We plan to promote this flag to maturity in v1.10.0.
diff --git a/website/docs/reference/programmatic-invocations.md b/website/docs/reference/programmatic-invocations.md
index 09e41b1789f..61250e6debb 100644
--- a/website/docs/reference/programmatic-invocations.md
+++ b/website/docs/reference/programmatic-invocations.md
@@ -25,9 +25,9 @@ for r in res.result:
## Parallel execution not supported
-[`dbt-core`](https://pypi.org/project/dbt-core/) doesn't support [safe parallel execution](/reference/dbt-commands#parallel-execution) for multiple invocations in the same process. This means it's not safe to run multiple dbt commands at the same time. It's officially discouraged and requires a wrapping process to handle sub-processes. This is because:
+[`dbt-core`](https://pypi.org/project/dbt-core/) doesn't support [safe parallel execution](/reference/dbt-commands#parallel-execution) for multiple invocations in the same process. This means it's not safe to run multiple dbt commands concurrently. It's officially discouraged and requires a wrapping process to handle sub-processes. This is because:
-- Running simultaneous commands can unexpectedly interact with the data platform. For example, running `dbt run` and `dbt build` for the same models simultaneously could lead to unpredictable results.
+- Running concurrent commands can unexpectedly interact with the data platform. For example, running `dbt run` and `dbt build` for the same models simultaneously could lead to unpredictable results.
- Each `dbt-core` command interacts with global Python variables. To ensure safe operation, commands need to be executed in separate processes, which can be achieved using methods like spawning processes or using tools like Celery.
To run [safe parallel execution](/reference/dbt-commands#available-commands), you can use the [dbt Cloud CLI](/docs/cloud/cloud-cli-installation) or [dbt Cloud IDE](/docs/cloud/dbt-cloud-ide/develop-in-the-cloud), both of which does that additional work to manage concurrency (multiple processes) on your behalf.
diff --git a/website/docusaurus.config.js b/website/docusaurus.config.js
index b68e2e8ec5c..9fe63267d61 100644
--- a/website/docusaurus.config.js
+++ b/website/docusaurus.config.js
@@ -209,7 +209,7 @@ var siteSettings = {
>