Skip to content

Commit 3d22701

Browse files
authored
Update databricks-setup.md
1 parent a4224ec commit 3d22701

File tree

1 file changed

+5
-13
lines changed

1 file changed

+5
-13
lines changed

src/unify/data-graph/setup-guides/databricks-setup.md

Lines changed: 5 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ Segment assumes that you already have a workspace that includes the datasets you
1616
## Step 1: Create a new Service Principal user
1717
Segment recommends setting up a new Service Principal user and only giving this user permissions to access the required catalogs and schemas.
1818

19-
If you already have a Service Principal user you'd like to use, grant it "Can use" permissions for your data warehouse and proceed to [step 2](#Create-a-catalog-for-Segment-to-store-checkpoint-tables).
19+
If you already have a Service Principal user you'd like to use, grant it "Can use" permissions for your data warehouse and proceed to [Step 2](#step-2-create-a-catalog-for-segment-to-store-checkpoint-tables).
2020

2121
### a) Create a new Service Principal user
2222
1. Log in to the Databricks UI as an Admin.
@@ -37,24 +37,16 @@ If you already have a Service Principal user you'd like to use, grant it "Can us
3737
4. Add the Service Principal user and grant them “Can use” access.
3838
5. Click **Add**.
3939

40-
### c) (Optional) Confirm Service Principal permissions
41-
Confirm that the Service Principal user that you're using to connect to Segment has "Can use" permissions for your warehouse.
42-
43-
To confirm that your Service Principal user has "Can use" permission:
44-
1. In the Databricks console, navigate to SQL Warehouses and select your warehouse.
45-
2. Navigate to Overview and click **Permissions**.
46-
3. Verify that the Service Principal user has "Can use" permission.
47-
4840
## Step 2: Create a catalog for Segment to store checkpoint tables
4941
**Segment requires write access to this catalog for internal bookkeeping and to store checkpoint tables for the queries that are executed. Therefore, Segment recommends creating a new catalog for this purpose.** This is also the catalog you'll be required to specify when connecting Databricks with the Segment app.
5042

5143
> info ""
5244
> Segment recommends creating a new database for the Data Graph.
53-
> If you choose to use an existing database that has also been used for [Segment Reverse ETL](/docs/connections/reverse-etl/), you must follow the [additional instructions to update user access for the Segment Reverse ETL catalog](#update-user-access-for-segment-reverse-etl-catalog).
45+
> If you choose to use an existing database that has also been used for [Segment Reverse ETL](/docs/connections/reverse-etl/), you must follow the [additional instructions](#update-user-access-for-segment-reverse-etl-catalog) to update user access for the Segment Reverse ETL catalog.
5446
5547
```SQL
5648
CREATE CATALOG IF NOT EXISTS `SEGMENT_LINKED_PROFILES_DB`;
57-
-- Copy the Client ID by clicking “Generate secret” for the Service Principal user
49+
-- Copy the saved Client ID from previously generated secret
5850
GRANT USAGE ON CATALOG `SEGMENT_LINKED_PROFILES_DB` TO `${client_id}`;
5951
GRANT CREATE ON CATALOG `SEGMENT_LINKED_PROFILES_DB` TO `${client_id}`;
6052
GRANT SELECT ON CATALOG `SEGMENT_LINKED_PROFILES_DB` TO `${client_id}`;
@@ -122,10 +114,10 @@ To connect your warehouse to the Data Graph:
122114
1. Navigate to **Unify > Data Graph**. This should be a Unify space with Profiles Sync already set up.
123115
2. Click Connect warehouse.
124116
3. Select Databricks as your warehouse type.
125-
4. Enter your warehouse credentials. Segment requires the following settings to connect to your Databricks warehouse. You can find these details in your Databricks workspace by navigating to **SQL Warehouse > Connection details**.
117+
4. Enter your warehouse credentials. You can find these details in your Databricks workspace by navigating to **SQL Warehouse > Connection details**. Segment requires the following settings to connect to your Databricks warehouse:
126118
- **Hostname**: The address of your Databricks server
127119
- **Http Path**: The address of your Databricks compute resources
128-
- **Port**: The port used to connect to your Databricks warehouse. The default port is 443, but your port might be different.
120+
- **Port**: The port used to connect to your Databricks warehouse. The default port is 443, but your port might be different
129121
- **Catalog**: The catalog you designated in [Step 2](#step-2-create-a-catalog-for-segment-to-store-checkpoint-tables)
130122
- **Service principal client ID**: The client ID used to access to your Databricks warehouse
131123
- **OAuth secret**: The OAuth secret used to connect to your Databricks warehouse

0 commit comments

Comments
 (0)