LogoLogo
Go to Syntho.AI
English
English
  • Welcome to Syntho
  • Overview
    • Get started
      • Syntho bootcamp
        • 1. What is Syntho?
        • 2. Introduction data anonymization
        • 3. Connectors & workspace creation
        • 4. PII scan
        • 5. Generators
          • Mockers
          • Maskers
          • AI synthesize
          • Calculated columns
          • Free text de-identification
        • 6. Referential integrity & foreign keys
        • 7. Workspace synchronization & validation
        • 8. Workspace & user management
        • 9. Large workloads​
        • 10. Data pre-processing
        • 11. Continuous Success
      • Prerequisites
      • Sample datasets
      • Introduction to data generators
    • Frequently asked questions
  • Setup Workspaces
    • View workspaces
    • Create a workspace
      • Connect to a database
        • PostgreSQL
        • MySQL / MariaDB
        • Oracle
        • Microsoft SQL Server
        • DB2
        • Databricks
          • Importing Data into Databricks
        • Hive
        • SAP Sybase
        • Azure Data Lake Storage (ADLS)
        • Amazon Simple Storage Service (S3)
      • Workspace modes
    • Edit a workspace
    • Duplicate a workspace
    • Transfer workspace ownership
    • Share a workspace
    • Delete a workspace
    • Workspace default settings
  • Configure a Data Generation Job
    • Configure table settings
    • Configure column settings
      • AI synthesize
        • Sequence model
          • Prepare your sequence data
        • QA report
        • Additional privacy controls
        • Cross-table relationships limitations
      • Mockers
        • Text
          • Supported languages
        • Numeric (integer)
        • Numeric (decimal)
        • Datetime
        • Other
      • Mask
        • Text
        • Numeric (integer)
        • Numeric (decimal)
        • Datetime
        • UUID
      • Duplicate
      • Exclude
      • Consistent mapping
      • Calculated columns
      • Key generators
        • Differences between key generators
      • JSON de-identification
    • Manage personally identifiable information (PII)
      • Privacy dashboard
      • Discover and de-identify PII columns
        • Identify PII columns manually
        • Automatic PII discovery with PII scanner
      • Remove columns from PII list
      • Automatic PII discovery and de-identification in free text columns
      • Supported PII & PHI entities
    • Manage foreign keys
      • Foreign key inheritance
      • Add virtual foreign keys
        • Add virtual foreign keys
        • Use foreign key scanner
        • Import foreign keys via JSON
        • Export foreign keys via JSON
      • Delete foreign keys
    • Validate and Synchronize workspace
    • View and adjust generation settings
  • Deploy Syntho
    • Introduction
      • Syntho architecture
      • Requirements
        • Requirements for Docker deployments
        • Requirements for Kubernetes deployments
      • Access Docker images
        • Online
        • Offline
    • Deploy Syntho using Docker
      • Preparations
      • Deploy using Docker Compose
      • Run the application
      • Manually saving logs
      • Updating the application
    • Deploy Syntho using Kubernetes
      • Preparations
      • Deploy Ray using Helm
        • Upgrading Ray CRDs
        • Troubleshooting
      • Deploy Syntho using Helm
      • Validate the deployment
      • Troubleshooting
      • Saving logs
      • Upgrading the applications
    • Manage users and access
      • Single Sign-On (SSO) in Azure
      • Manage admin users
      • Manage non-admin users
    • Logs and monitoring
      • Does Syntho collect any data?
      • Temporary data storage by application
  • Syntho API
    • Syntho REST API
Powered by GitBook
On this page
  • Before you begin
  • Connect and set up the workspace
  • Supported Databrick versions
  • Supported data types
  • Supported data types

Was this helpful?

  1. Setup Workspaces
  2. Create a workspace
  3. Connect to a database

Databricks

PreviousDB2NextImporting Data into Databricks

Last updated 8 days ago

Was this helpful?

Important

This connector can only be used as a source database. The generated data can be written to Local , or as Parquet files.

Before you begin

Before you begin, gather this connection information:

  • Name of the server that hosts the database you want to connect to and port number

  • The name of the database that you want to connect to

  • HTTP path to the data source

  • Personal Access Token

Connect and set up the workspace

Launch Syntho and select Connect to a database, or Create workspace. Then select Databricks from Type under The connection details. For a complete list of data connections, click Type under The connection details. Then do the following:

  1. Enter the server hostname.

  2. Enter the catalog name.

  3. Enter the database name.

  4. Enter the HTTP Path to the data source.

  5. Select Create Workspace.

    If Syntho can't make the connection, verify that your credentials are correct. If you still can't connect, your computer is having trouble locating the server. Contact your network administrator or database administrator.

Supported Databrick versions

The table below provides an overview of the supported Databricks versions and their corresponding Apache Spark versions.

Databricks Version
Spark Version

16.2

3.5.0

15.4 LTS

3.5.0

14.3 LTS

3.5.0

Note: Version 13 is no longer supported.

Supported data types

The following table summarizes the current support limitations for various data types when using connectors with Databricks. It indicates what's supported per generator type.

Supported data types

The following table summarizes the current support limitations for various data types when using connectors with Databricks. It indicates what's supported per generator type.

Data Type
AI-powered Generation
Mockers
Mask
Calculated Columns

INT

SMALLINT

TINYINT

BIGINT

DECIMAL

FLOAT

DOUBLE

STRING

BINARY

False

True*

True*

True*

BOOLEAN

False

False

True*

DATE

False

TIMESTAMP

False

TIMESTAMP_NTZ

False

ARRAY

False

True*

True*

True*

STRUCT

False

False

False

MAP

False

True*

True*

True*

VARIANT

False

True*

True*

True*

OBJECT

False

True*

True*

True*

ENUM

False

False

False

False

*Some data types are not actively supported; however, certain generators such as AI synthesize, mask, mockers, or calculated columns may still show 'True' for these fields. This means the generators can be applied, even though the types are not actively supported. Duplication is fully supported for these data types.

In Databricks, find your cluster server hostname and HTTP path using the instructions in on the Databricks website.

Enter Personal Access Token. (See on the Databricks website for information on access tokens.)

Construct the JDBC URL
Personal Access Tokens
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
Filesystem
Azure Data Lake Storage (ADLS)
Amazon Simple Storage Service (S3)
True
Source and Destination Databases
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True
True