Databricks Support

From ER/Studio Data Architect
Jump to: navigation, search

Go Up to Supported Database Platforms

ER/Studio supports Databricks as a Core Platform, which enables you to:

  • reverse engineer from the Unity Data Catalog
  • generate DDL Code
  • import DDL Code
  • generate ALTER script

ER/Studio allows you to design and document Delta tables within a Unity Data Catalog instance.

Connecting to the Unity Data Catalog

Before connecting to the Unity Data Catalog, you must download and install the Databricks ODBC driver from the Databricks website at https://www.databricks.com/spark/odbc-drivers-download.

For more information about configuring these connections, see Configuring ODBC Data Source and Target Connections.

Once you configure your ODBC driver, you can connect to your Unity Data Catalog, and then either reverse engineer or use the Compare and Merge Utility for your Databricks Catalogs and Schemas. Note that the ER/Studio user interface refers to Database and Owners, which equate to Databricks Catalog and Schema Objects, respectively. For more information about reverse engineering, see Reverse Engineering an Existing Database. For more information about using the compare/merge feature, see Using the Compare and Merge Utility.

What Databricks features are available?

With ER/Studio’s Databricks support, you can work with the following features available on Delta tables in the Unity Data Catalog:

  • Primary Key, Foreign Key, and Check Constraints
  • Materialized Views
  • Managed & Unmanaged Tables
  • Clustering and Partitioning
  • Bloom Filter Indexes
  • Functions and Mask Functions in Tables
  • Views

Databricks includes data types for creating denormalised structures to improve performance. ER/Studio has some powerful tools to allow you to generate these nested structures both to and from logical data models with containment relationships. For more information about this feature, see Containment Relationships.