Quick Start Guide
This guide walks you through getting LakeSentry running — from creating your account to viewing your first cost breakdown. The whole process takes about 15 minutes of active setup, plus a short wait for initial data collection.
Prerequisites
Section titled “Prerequisites”Before you start, make sure you have:
- A Databricks account with Unity Catalog enabled
- A service principal with read access to system tables, or a personal access token for quick evaluation
- At least one SQL warehouse available in your Databricks workspace
Step 1: Create your LakeSentry account
Section titled “Step 1: Create your LakeSentry account”- Go to lakesentry.io and click Start Free.
- Enter your email address.
- Check your inbox for a magic link and click it to verify your account.
- You’ll land on the LakeSentry setup screen.
Step 2: Connect your Databricks workspace
Section titled “Step 2: Connect your Databricks workspace”This step creates the link between LakeSentry and your Databricks environment. LakeSentry uses a Direct Connection — it connects to your workspace and extracts system table data automatically.
- From the setup screen, choose Direct Connection as the connection method.
- Choose your authentication method:
- Service Principal (OAuth) — Recommended. Uses a dedicated app identity with no user dependency.
- Personal Access Token — Fastest to set up. Uses a token tied to a user account.
- Enter your Workspace URL (e.g.,
https://adb-xxx.azuredatabricks.net). LakeSentry auto-detects the cloud provider and region from this URL. - Provide your credentials:
- For Service Principal: enter the Client ID and Client Secret.
- For Personal Access Token: enter your Databricks token (starts with
dapi).
- Click Validate Credentials. LakeSentry verifies the connection, checks for SQL warehouses, and confirms system table access.
- Once validated, click Connect Workspace.
LakeSentry immediately begins extracting data from your Databricks system tables.
For detailed instructions on creating the service principal and configuring permissions, see Connecting Your Databricks Account.
Step 3: Add additional regions (if needed)
Section titled “Step 3: Add additional regions (if needed)”Databricks system tables are regional — a connector in one region can’t read data from another. If you run workspaces across multiple regions, add a connector for each region.
- Go to Settings > Connector and click Add Connector.
- Enter the Workspace URL for a workspace in the additional region.
- Provide credentials (the same service principal can be used across regions).
- LakeSentry auto-detects the region and creates the connector.
Step 4: Wait for initial data
Section titled “Step 4: Wait for initial data”After LakeSentry starts its first extraction, the onboarding progress screen shows four pipeline stages:
- Data Extraction — LakeSentry queries your Databricks system tables.
- Ledger Transform — Raw data is transformed into the cost ledger.
- Metrics — Cost metrics are computed and materialized.
- Insights — Automated cost insights are generated.
You can check the status of each connector in Settings > Connector. Each connector shows its sync status:
| Status | Meaning |
|---|---|
| Synced (green) | Data was received and the connector is active |
| Error (red) | Validation or connection errors |
| Awaiting data (gray) | No data received yet (setup in progress) |
Step 5: Explore the dashboard
Section titled “Step 5: Explore the dashboard”Once data starts flowing, head to the Overview page to see:
- Cost trends — Total spend over time, broken down by compute type
- Top spenders — The workspaces, jobs, and users driving the most cost
- Anomalies — Unusual cost spikes flagged automatically
- Insights — Actionable findings about waste and optimization opportunities
Use the time range selector in the header to adjust the analysis window (Last 24 hours, 7 days, 30 days, 90 days, 1 year, month/year to date, or a custom range). Apply workspace, organization, or tag filters to focus on specific areas.
For a detailed walkthrough of the dashboard, see Understanding the Dashboard.
Step 6: Investigate your first cost spike
Section titled “Step 6: Investigate your first cost spike”Here’s a typical first investigation:
- On the Overview page, spot a cost anomaly or top spender.
- Click through to see the cost breakdown — which workspaces, compute types, and users contributed.
- Drill into a specific work unit (job, pipeline, SQL query, notebook, or serving endpoint) to see per-run cost history.
- Check the Insights page for LakeSentry’s automated findings — it may have already flagged the issue.
What’s next
Section titled “What’s next”- Connecting Your Databricks Account — Detailed setup for service principals and permissions
- Understanding the Dashboard — Full walkthrough of the Overview page
- Cost Attribution — Learn how costs are assigned to teams
- Insights & Actions — How to act on optimization findings