# Getting Started with Databricks Spark

The how-to guides linked on this page illustrate how to integrate Databricks Spark with Immuta.

**Requirements**

* If **Databricks Unity Catalog is enabled** in a Databricks workspace, you must use an [Immuta cluster policy](/SaaS/configuration/integrations/databricks/databricks-spark/reference-guides/databricks/customizing-the-integration.md#cluster-policies) when you set up the Databricks Spark integration to create an Immuta-enabled cluster.
* If **Databricks Unity Catalog is not enabled** in your Databricks workspace, you must disable Unity Catalog in your Immuta tenant before proceeding with your configuration of Databricks Spark:
  1. Navigate to the <i class="fa-gear">:gear:</i> **App Settings** page and click **Integration Settings**.
  2. Uncheck the **Enable Unity Catalog** checkbox.
  3. Click **Save**.

{% stepper %}
{% step %}
**Connect your technology**

These guides provide instructions for getting your data set up in Immuta.

1. [Configure your Databricks Spark integration](/SaaS/configuration/integrations/databricks/databricks-spark/how-to-guides/simplified.md).
2. [Register Databricks securable objects in Immuta as data sources](/SaaS/configuration/integrations/data-and-integrations/registering-metadata/register-data-sources/databricks-tutorial.md).
3. [Organize your data sources into domains and assign domain permissions to accountable teams (recommended)](/SaaS/configuration/domains/configure-domains.md): Use domains to segment your data and assign responsibilities to the appropriate team members. These domains will then be used in policies, audit, and identification.
   {% endstep %}

{% step %}
**Register your users**

These guides provide instructions on setting up your users in Immuta.

1. [Integrate an IAM with Immuta](/SaaS/configuration/people/index.md): Connect the IAM your organization already uses and allow Immuta to register your users for you.
2. [Map external user IDs from Databricks to Immuta](/SaaS/configuration/people/users-index/how-to-guides/external-user-mapping.md): Ensure the user IDs in Immuta, Databricks, and your IAM are aligned so that the right policies impact the right users.
   {% endstep %}

{% step %}
**Add data metadata**

These guides provide instructions on getting your data metadata set up in Immuta for use in policies.

1. [Connect an external catalog](/SaaS/configuration/tags/catalogs/configure.md): Connect the external catalog your organization already uses and allow Immuta to continually sync your tags with your data sources for you.
2. [Run identification](/SaaS/configuration/tags/data-discovery/how-to-guides/enable-sdd.md): Identification allows you to automate data tagging using identifiers that detect certain data patterns.
   {% endstep %}

{% step %}
**Protect and monitor data access**

These guides provide instructions on authoring policies and auditing data access.

* [Author a global subscription policy](/SaaS/govern/secure-your-data/authoring-policies-in-secure/section-contents/how-to-guides/subscription-policy-tutorial.md): Once you add your data metadata to Immuta, you can immediately create policies that utilize your tags and apply to your tables. Subscription policies can be created to dictate access to data sources.
* [Author a global data policy](/SaaS/govern/secure-your-data/authoring-policies-in-secure/data-policies/how-to-guides/data-policy-tutorial.md): Data metadata can also be used to create data policies that apply to data sources as they are registered in Immuta. Data policies dictate what data a user can see once they are granted access to a data source. Using catalog and identification tags you can create proactive policies, knowing that they will apply to data sources as they are added to Immuta with the automated tagging.
* [Configure audit](/SaaS/govern/detect-your-data/audit/reference-guides/index.md): Once you have your data sources and users, and policies granting them access, you can set up audit export. This will export the audit logs from user queries, policy changes, and tagging updates.
  {% endstep %}
  {% endstepper %}


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://documentation.immuta.com/SaaS/configuration/integrations/databricks/databricks-spark/getting-started-with-databricks-spark.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
