Databricks Connector for CData Sync

Build 25.0.9539
  • Databricks
    • Establishing a Connection
    • Advanced Features
      • SSL Configuration
      • Firewall and Proxy
    • Data Model
    • Connection String Options
      • Authentication
        • AuthScheme
        • Server
        • ProtocolVersion
        • Database
        • HTTPPath
        • Token
      • AWS Authentication
        • AWSAccessKey
        • AWSSecretKey
        • AWSRegion
        • AWSS3Bucket
      • Azure Authentication
        • AzureStorageAccount
        • AzureAccessKey
        • AzureTenant
        • AzureBlobContainer
      • AzureServicePrincipal Authentication
        • AzureTenantId
        • AzureClientId
        • AzureClientSecret
      • OAuth
        • OAuthClientId
        • OAuthClientSecret
        • Scope
        • OAuthLevel
        • DatabricksAccountId
      • SSL
        • SSLClientCert
        • SSLClientCertType
        • SSLClientCertPassword
        • SSLClientCertSubject
        • SSLServerCert
      • Firewall
        • FirewallType
        • FirewallServer
        • FirewallPort
        • FirewallUser
        • FirewallPassword
      • Proxy
        • ProxyAutoDetect
        • ProxyServer
        • ProxyPort
        • ProxyAuthScheme
        • ProxyUser
        • ProxyPassword
        • ProxySSLType
        • ProxyExceptions
      • Logging
        • LogModules
      • Schema
        • Location
        • BrowsableSchemas
        • Tables
        • Views
        • Catalog
        • PrimaryKeyIdentifiers
      • Databricks
        • CloudStorageType
        • StoreTableInCloud
        • QueryTableDetails
        • UseUploadApi
        • UseCloudFetch
        • SupportMultiCatalog
        • QueryAllMetadata
        • CheckSQLWarehouseAvailability
      • Miscellaneous
        • AllowPreparedStatement
        • ConnectRetryWaitTime
        • ApplicationName
        • AsyncQueryTimeout
        • DefaultColumnSize
        • DescribeCommand
        • DetectView
        • IncludeSystemSchemas
        • MaxRows
        • Other
        • PseudoColumns
        • QueryPassthrough
        • ServerConfigurations
        • ServerTimeZone
        • Timeout
        • UseDescTableQuery
        • UseInsertSelectSyntax
        • UserDefinedViews
    • Third Party Copyrights

Databricks Connector for CData Sync

Overview

The CData Sync App provides a straightforward way to continuously pipeline your Databricks data to any database, data lake, or data warehouse, making it easily available for Analytics, Reporting, AI, and Machine Learning.

The Databricks connector can be used from the CData Sync application to pull data from Databricks and move it to any of the supported destinations.

Databricks Version Support

The CData Sync App leverages Databricks Thrift to enable bidirectional SQL access to Databricks data. The Sync App supports databases that run Databricks Runtime Version 9.1 and later. It also supports the Pro and Classic Databricks SQL versions. The data model is fully dynamic and automatically reflects the data that is available in your Databricks environment. Any updates to tables or schemas are immediately accessible through SQL.

Databricks Connector for CData Sync

Establishing a Connection

Adding a Connection to Databricks

To add a connection to Databricks:

  1. In the application console, navigate to the Connections page.
  2. At the Add Connections panel, select the icon for the connection you want to add.
  3. If the Databricks icon is not available, click the Add More icon to download and install the Databricks connector from the CData site.

For required properties, see the Settings tab.

For connection properties that are not typically required, see the Advanced tab.

Connecting to Databricks

To connect to a Databricks cluster, set the following properties:

  • Database: The name of the Databricks database.
  • Server: The Server Hostname of your Databricks cluster.
  • HTTPPath: The HTTP Path of your Databricks cluster.
  • Token: Your personal access token. You can obtain this value by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab.

You can find the required values in your Databricks instance by navigating to Clusters and selecting the desired cluster, and selecting the JDBC/ODBC tab under Advanced Options.

Configuring Cloud Storage

The Sync App supports DBFS, Azure Blob Storage, and AWS S3 for uploading CSV files.

DBFS Cloud Storage

To use DBFS for cloud storage, set the CloudStorageType property to DBFS.

Azure Blob Storage

Set the following properties:

  • CloudStorageType: Azure Blob storage.
  • StoreTableInCloud: True to store tables in cloud storage when creating a new table.
  • AzureStorageAccount: The name of your Azure storage account.
  • AzureAccessKey: The storage key associated with your Databricks account. Find this via the azure portal (using the root account). Select your storage account and click Access Keys to find this value.
  • AzureBlobContainer: Set to the name of your Azure Blob storage container.

AWS S3 Storage

Set the following properties:

  • CloudStorageType: AWS S3.
  • StoreTableInCloud: True to store tables in cloud storage when creating a new table.
  • AWSAccessKey: The AWS account access key. You can acquire this value from your AWS security credentials page.
  • AWSSecretKey: Your AWS account secret key. You can acquire this value from your AWS security credentials page.
  • AWSS3Bucket: The name of your AWS S3 bucket.
  • AWSRegion: The hosting region for your Amazon Web Services. You can obtain the AWS Region value by navigating to the Buckets List page of your Amazon S3 service, for example, us-east-1.

Authenticating to Databricks

CData supports the following authentication schemes:

  • Personal Access Token
  • Microsoft Entra ID (Azure AD)
  • Azure Service Principal
  • OAuthU2M
  • OAuthM2M

Personal Access Token

To authenticate, set the following:

  • AuthScheme: PersonalAccessToken.
  • Token: The token used to access the Databricks server. It can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab.

Microsoft Entra ID (Azure AD)

Note: Microsoft has rebranded Azure AD as Entra ID. In topics that require the user to interact with the Entra ID Admin site, we use the same names Microsoft does. However, there are still CData connection properties whose names or values reference "Azure AD".

Before you can authenticate using Entra ID, you must first register an application with the Entra ID endpoint in the Azure portal, as described in Creating an Entra ID (Azure AD) Application. (See also Microsoft's own Configure an app in Azure portal.) Once the application has been completed, set these properties:

  • AuthScheme: AzureAD.
  • AzureTenant: The "Directory(tenant) ID" in the AzureAD application "Overview" page
  • OAuthClientId: The "Application(client) ID" in the AzureAD application "Overview" page.
  • CallbackURL: The "Redirect URIs" in AzureAD application "Authentication" page

When connecting, a web page opens that prompts you to authenticate. After successful authentication, the connection is established.

Example connection string:

"Server=https://adb-8439982502599436.16.azuredatabricks.net;HTTPPath=sql/protocolv1/o/8439982502599436/0810-011933-odsz4s3r;database=default;
AuthScheme=AzureAD;InitiateOAuth=GETANDREFRESH;AzureTenant=94be69e7-edb4-4fda-ab12-95bfc22b232f;OAuthClientId=f544a825-9b69-43d9-bec2-3e99727a1669;CallbackURL=http://localhost;"

Azure Service Principal

To authenticate, set the following properties:

  • AuthScheme: AzureServicePrincipal.
  • AzureTenantId: The tenant ID of your Microsoft Azure Active Directory.
  • AzureClientId: The application (client) ID of your Microsoft Azure Active Directory application.
  • AzureClientSecret: The application (client) secret of your Microsoft Azure Active Directory application.

OAuthU2M

OAuthU2M (User-to-Machine) authentication allows users to grant applications, such as CLI or SDK, access to their workspace. It uses a secure OAuth token, eliminating the need to share the user's password.

The following explains how OAuthU2M works:

After a user signs in and consents to the OAuthU2M authentication request, the tool or SDK receives an OAuth token. This token allows the tool or SDK to authenticate on the user's behalf.

By default, the Sync App uses an embedded OAuth application with a redirect URL of http://localhost:8020 which requires no setup. However, to customize the redirect URL or scopes used during authentication, you can register a custom OAuth application in the Databricks Account Console.

For instructions on registering a custom OAuth application, see Creating a Custom OAuth Application.

The required settings are:

  • AuthScheme: OAuthU2M
  • OAuthLevel: Set to the level at which you want to request the token.
  • OAuthClientId: Assigned when you register your application with an OAuth authorization server.
  • CallbackURL: The redirect URL registered with your OAuth application.
  • DatabricksAccountId: Required only when the OAuthLevel is set to AccountLevel.

OAuthM2M

OAuthM2M (Machine-to-Machine) authentication verifies the identiy of devices or applications communicating over a network. It ensures that only authorized machines can securely exchange data and access resources without human intervention.

The following explains how OAuthM2M works:

Register your application with the authorization server to obtain a client ID and secret. When accessing a protected resource, your machine sends a request with these credentials and desired scopes. The server verifies the provided information and, if valid, returns an access token. This token is included in the request header for API calls to access the resource.

The required settings are:

  • AuthScheme: OAuthM2M
  • OAuthLevel: Set to the level at which you want to request the token.
  • OAuthClientId: Assigned when you register your application with an OAuth authorization server.
  • OAuthClientSecret: Assigned when you register your application with an OAuth authorization server.
  • DatabricksAccountId: Required only when the OAuthLevel is set to AccountLevel.

Databricks Connector for CData Sync

Advanced Features

This section details a selection of advanced features of the Databricks Sync App.

User Defined Views

The Sync App supports the use of user defined views, virtual tables whose contents are decided by a pre-configured user defined query. These views are useful when you cannot directly control queries being issued to the drivers. For an overview of creating and configuring custom views, see User Defined Views .

SSL Configuration

Use SSL Configuration to adjust how Sync App handles TLS/SSL certificate negotiations. You can choose from various certificate formats;. For further information, see the SSLServerCert property under "Connection String Options" .

Firewall and Proxy

Configure the Sync App for compliance with Firewall and Proxy, including Windows proxies and HTTP proxies. You can also set up tunnel connections.

Query Processing

The Sync App offloads as much of the SELECT statement processing as possible to Databricks and then processes the rest of the query in memory (client-side).

For further information, see Query Processing.

Logging

For an overview of configuration settings that can be used to refine CData logging, see Logging. Only two connection properties are required for basic logging, but there are numerous features that support more refined logging, which enables you to use the LogModules connection property to specify subsets of information to be logged.

Databricks Connector for CData Sync

SSL Configuration

Customizing the SSL Configuration

By default, the Sync App attempts to negotiate TLS with the server. The server certificate is validated against the default system trusted certificate store. You can override how the certificate gets validated using the SSLServerCert connection property.

To specify another certificate, see the SSLServerCert connection property.

Client SSL Certificates

The Databricks Sync App also supports setting client certificates. Set the following to connect using a client certificate.

  • SSLClientCert: The name of the certificate store for the client certificate.
  • SSLClientCertType: The type of key store containing the TLS/SSL client certificate.
  • SSLClientCertPassword: The password for the TLS/SSL client certificate.
  • SSLClientCertSubject: The subject of the TLS/SSL client certificate.

Databricks Connector for CData Sync

Firewall and Proxy

Connecting Through a Firewall or Proxy

HTTP Proxies

To authenticate to an HTTP proxy, set the following:

  • ProxyServer: the hostname or IP address of the proxy server that you want to route HTTP traffic through.
  • ProxyPort: the TCP port that the proxy server is running on.
  • ProxyAuthScheme: the authentication method the Sync App uses when authenticating to the proxy server.
  • ProxyUser: the username of a user account registered with the proxy server.
  • ProxyPassword: the password associated with the ProxyUser.

Other Proxies

Set the following properties:

  • To use a proxy-based firewall, set FirewallType, FirewallServer, and FirewallPort.
  • To tunnel the connection, set FirewallType to TUNNEL.
  • To authenticate, specify FirewallUser and FirewallPassword.
  • To authenticate to a SOCKS proxy, additionally set FirewallType to SOCKS5.

Databricks Connector for CData Sync

Data Model

The CData Sync App leverages Databricks Thrift to enable bidirectional SQL access to Databricks data. The Sync App supports databases that run Databricks Runtime Version 9.1 and later. It also supports the Pro and Classic Databricks SQL versions. The data model is fully dynamic and automatically reflects the data that is available in your Databricks environment. Any updates to tables or schemas are immediately accessible through SQL.

Discovering Schemas

The CData Sync App dynamically retrieves schema information from Databricks. To ensure that any metadata changes (for example, added or removed columns or modified data types) are reflected in the driver, reconnect to your data source.

Databricks Connector for CData Sync

Connection String Options

The connection string properties are the various options that can be used to establish a connection. This section provides a complete list of the options you can configure in the connection string for this provider. Click the links for further details.

For more information on establishing a connection, see Establishing a Connection.

Authentication


PropertyDescription
AuthSchemeSpecifies the authentication scheme that the provider uses to connect to Databricks.
ServerSpecifies the host name of your Databricks workspace or SQL warehouse endpoint.
ProtocolVersionSpecifies the protocol version used by the provider when authenticating and exchanging data with Databricks.
DatabaseSpecifies the name of the database in Databricks that the provider uses for the connection.
HTTPPathSpecifies the HTTP path for the compute resource in Databricks that the provider connects to.
TokenSpecifies the personal access token used by the provider to authenticate to Databricks.

AWS Authentication


PropertyDescription
AWSAccessKeySpecifies your AWS account access key. This value is accessible from your AWS security credentials page.
AWSSecretKeyYour AWS account secret key. This value is accessible from your AWS security credentials page.
AWSRegionThe hosting region for your Amazon Web Services.
AWSS3BucketSpecifies the name of the AWS S3 bucket that the provider uses for staging or transferring data when connecting to Databricks.

Azure Authentication


PropertyDescription
AzureStorageAccountThe name of your Azure storage account.
AzureAccessKeyThe storage key associated with your Azure account.
AzureTenantIdentifies the Databricks tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com ) or its directory (tenant) ID.
AzureBlobContainerSpecifies the name of the Azure Blob Storage container that the provider uses for staging or transferring data when connecting to Databricks.

AzureServicePrincipal Authentication


PropertyDescription
AzureTenantIdSpecifies the directory (tenant) ID of your Microsoft Entra ID that the provider uses to authenticate to Databricks.
AzureClientIdSpecifies the application (client) ID of the Microsoft Entra ID application that the provider uses to authenticate to Databricks.
AzureClientSecretSpecifies the client secret for the Microsoft Entra ID application that the provider uses to authenticate to Databricks.

OAuth


PropertyDescription
OAuthClientIdSpecifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication.
OAuthClientSecretSpecifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. (Custom OAuth applications only.).
ScopeSpecifies the scope of the authenticating user's access to the application, to ensure they get appropriate access to data. If a custom OAuth application is needed, this is generally specified at the time the application is created.
OAuthLevelSpecifies whether the provider performs OAuth authentication at the workspace level or at the account level in Databricks.
DatabricksAccountIdSpecifies the unique account ID for your Databricks account.

SSL


PropertyDescription
SSLClientCertSpecifies the TLS/SSL client certificate store for SSL Client Authentication (2-way SSL). This property works in conjunction with other SSL-related properties to establish a secure connection.
SSLClientCertTypeSpecifies the type of key store containing the TLS/SSL client certificate for SSL Client Authentication. Choose from a variety of key store formats depending on your platform and certificate source.
SSLClientCertPasswordSpecifes the password required to access the TLS/SSL client certificate store. Use this property if the selected certificate store type requires a password for access.
SSLClientCertSubjectSpecifes the subject of the TLS/SSL client certificate to locate it in the certificate store. Use a comma-separated list of distinguished name fields, such as CN=www.server.com, C=US. The wildcard * selects the first certificate in the store.
SSLServerCertSpecifies the certificate to be accepted from the server when connecting using TLS/SSL.

Firewall


PropertyDescription
FirewallTypeSpecifies the protocol the provider uses to tunnel traffic through a proxy-based firewall.
FirewallServerIdentifies the IP address, DNS name, or host name of a proxy used to traverse a firewall and relay user queries to network resources.
FirewallPortSpecifies the TCP port to be used for a proxy-based firewall.
FirewallUserIdentifies the user ID of the account authenticating to a proxy-based firewall.
FirewallPasswordSpecifies the password of the user account authenticating to a proxy-based firewall.

Proxy


PropertyDescription
ProxyAutoDetectSpecifies whether the provider checks your system proxy settings for existing proxy server configurations, rather than using a manually specified proxy server.
ProxyServerIdentifies the hostname or IP address of the proxy server through which you want to route HTTP traffic.
ProxyPortIdentifies the TCP port on your specified proxy server that has been reserved for routing HTTP traffic to and from the client.
ProxyAuthSchemeSpecifies the authentication method the provider uses when authenticating to the proxy server specified in the ProxyServer connection property.
ProxyUserProvides the username of a user account registered with the proxy server specified in the ProxyServer connection property.
ProxyPasswordSpecifies the password of the user specified in the ProxyUser connection property.
ProxySSLTypeSpecifies the SSL type to use when connecting to the proxy server specified in the ProxyServer connection property.
ProxyExceptionsSpecifies a semicolon-separated list of destination hostnames or IPs that are exempt from connecting through the proxy server set in the ProxyServer connection property.

Logging


PropertyDescription
LogModulesSpecifies the core modules to include in the log file. Use a semicolon-separated list of module names. By default, all modules are logged.

Schema


PropertyDescription
LocationSpecifies the location of a directory containing schema files that define tables, views, and stored procedures. Depending on your service's requirements, this may be expressed as either an absolute path or a relative path.
BrowsableSchemasOptional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC .
TablesOptional setting that restricts the tables reported to a subset of all available tables. For example, Tables=TableA,TableB,TableC .
ViewsOptional setting that restricts the views reported to a subset of the available tables. For example, Views=ViewA,ViewB,ViewC .
CatalogSpecifies the default catalog that the provider uses when connecting to Databricks.
PrimaryKeyIdentifiersSpecifies primary keys for tables in the hive_metastore catalog.

Databricks


PropertyDescription
CloudStorageTypeSpecifies which cloud storage service the provider uses for staging or transferring data.
StoreTableInCloudSpecifies whether the Databricks server creates and stores new tables in external cloud storage instead of the Databricks File System (DBFS).
QueryTableDetailsSpecifies whether the provider uses the DESCRIBE FORMATTED command in Databricks to retrieve detailed table metadata.
UseUploadApiSpecifies whether the provider uses the Databricks Upload API to optimize Bulk INSERT operations.
UseCloudFetchSpecifies whether the provider uses CloudFetch to optimize data transfer for large query result sets in Databricks.
SupportMultiCatalogSpecifies whether the provider enables multi-catalog support for Unity Catalog in Databricks.
QueryAllMetadataSpecifies whether the provider retrieves metadata from all available catalogs and schemas in Databricks or only from those specified in the connection.
CheckSQLWarehouseAvailabilitySpecifies whether the provider checks the availability of the SQL Warehouse in Databricks before establishing a connection.

Miscellaneous


PropertyDescription
AllowPreparedStatementSpecifies whether the provider prepares SQL statements before executing them to improve performance on repeated queries.
ConnectRetryWaitTimeSpecifies the number of seconds the provider waits before retrying a connection request.
ApplicationNameSpecifies the application name that the provider includes in the HTTP User-Agent header when connecting to Databricks.
AsyncQueryTimeoutSpecifies the number of seconds the provider waits for asynchronous requests that retrieve large result sets before timing out.
DefaultColumnSizeSpecifies the default length, in characters, of string fields when the provider cannot determine a size from metadata.
DescribeCommandSpecifies which command the provider uses to retrieve metadata from Databricks.
DetectViewSpecifies whether the provider uses the DESCRIBE FORMATTED command to determine whether an object in Databricks is a table or a view.
IncludeSystemSchemasSpecifies whether to include the system catalog 'system' and system schema 'information_schema'.
MaxRowsSpecifies the maximum number of rows returned for queries that do not include either aggregation or GROUP BY.
OtherSpecifies advanced connection properties for specialized scenarios. Use this property only under the guidance of our Support team to address specific issues.
PseudoColumnsSpecifies the pseudocolumns to expose as table columns, expressed as a string in the format 'TableName=ColumnName;TableName=ColumnName'.
QueryPassthroughThis option passes the query to the Databricks server as is.
ServerConfigurationsSpecifies configuration variables to override the default settings on the Databricks server.
ServerTimeZoneSpecifies how the provider interprets datetime values returned from Databricks.
TimeoutSpecifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error.
UseDescTableQuerySpecifies whether the provider retrieves table column metadata using a DESC TABLE query instead of the Thrift API.
UseInsertSelectSyntaxDEPRECATED. This property is no longer supported, and should not be used. It will be removed in a future release.
UserDefinedViewsSpecifies a filepath to a JSON configuration file that defines custom views. The provider automatically detects and uses the views specified in this file.
Databricks Connector for CData Sync

Authentication

This section provides a complete list of the Authentication properties you can configure in the connection string for this provider.


PropertyDescription
AuthSchemeSpecifies the authentication scheme that the provider uses to connect to Databricks.
ServerSpecifies the host name of your Databricks workspace or SQL warehouse endpoint.
ProtocolVersionSpecifies the protocol version used by the provider when authenticating and exchanging data with Databricks.
DatabaseSpecifies the name of the database in Databricks that the provider uses for the connection.
HTTPPathSpecifies the HTTP path for the compute resource in Databricks that the provider connects to.
TokenSpecifies the personal access token used by the provider to authenticate to Databricks.
Databricks Connector for CData Sync

AuthScheme

Specifies the authentication scheme that the provider uses to connect to Databricks.

Remarks

The AuthScheme property determines which authentication flow the Sync App uses when connecting to Databricks. Each option requires a different set of supporting connection properties.

When set to PersonalAccessToken, the Sync App authenticates by using a personal access token from Databricks.

When set to OAuthU2M, the Sync App uses the OAuth user-to-machine (U2M) flow. Set OAuthLevel, DatabricksAccountId (optional), OAuthClientId, and CallbackURL.

When set to OAuthM2M, the Sync App uses the OAuth machine-to-machine (M2M) flow. Set OAuthLevel, DatabricksAccountId (optional), OAuthClientId, and OAuthClientSecret. The client ID and secret can be generated by creating a Databricks service principal.

When this property is set to AzureServicePrincipal, the Sync App authenticates with an Azure service principal. Set AzureTenantId, AzureClientId, and AzureClientSecret. Follow the instructions in Get Microsoft Entra ID tokens for service principals to register an Azure AD application, and in Assign Azure roles using the Azure portal to assign appropriate roles in Azure.

When set to AzureAD, the Sync App authenticates through Azure Active Directory OAuth. Set AzureTenantId, OAuthClientId, OAuthClientSecret (optional), and CallbackURL. Follow the instructions in Configure an app in Azure portal to register an Azure AD application. The client secret is required only if the platform type in your Azure app is "Web".

When set to AzureMSI, the Sync App automatically obtains Azure Managed Service Identity credentials when running on an Azure VM.

Databricks Connector for CData Sync

Server

Specifies the host name of your Databricks workspace or SQL warehouse endpoint.

Remarks

The Server property identifies the host name used by the Sync App to connect to Databricks. This is typically the domain portion of your workspace URL, for example: acme.cloud.databricks.com.

You can find this value in your browser’s address bar when signed in to your Databricks workspace.

This property is required for all connection types and determines which workspace or endpoint the Sync App connects to.

Databricks Connector for CData Sync

ProtocolVersion

Specifies the protocol version used by the provider when authenticating and exchanging data with Databricks.

Remarks

The ProtocolVersion property defines the protocol version used between the Sync App and Databricks. This setting determines how requests and responses are encoded during authentication and query execution.

The default value of 8 represents the latest supported protocol version in Databricks.

Changing this value is typically only necessary when connecting to legacy environments or troubleshooting protocol compatibility issues.

Databricks Connector for CData Sync

Database

Specifies the name of the database in Databricks that the provider uses for the connection.

Remarks

A database in Databricks is a logical container for schemas and tables within a catalog. In Unity Catalog environments, each database belongs to a specific catalog, such as catalog_name.database_name.

Set this property to the database that you want the Sync App to query by default. If no value is provided, the default database for the configured catalog is used.

This property is useful for targeting a specific database when working with multiple data environments or Unity Catalog configurations in Databricks.

Databricks Connector for CData Sync

HTTPPath

Specifies the HTTP path for the compute resource in Databricks that the provider connects to.

Remarks

The HTTPPath property identifies the path component of the JDBC or ODBC endpoint for your compute resource in Databricks. This value directs the Sync App to the correct SQL Warehouse or cluster.

You can find the HTTP path in the Databricks workspace under Compute > your SQL Warehouse or Cluster > Connection details (or Advanced options) > JDBC/ODBC > HTTP Path.

An example format is sql/protocolv1/o/123456789/1234-123456-1ab2cdef.

This property is required for all authenticated connections to Databricks and is useful for ensuring the Sync App targets the correct compute resource.

Databricks Connector for CData Sync

Token

Specifies the personal access token used by the provider to authenticate to Databricks.

Remarks

The Token property provides the authentication token that the Sync App uses to access your Databricks workspace or SQL warehouse when AuthScheme is set to PersonalAccessToken.

You can generate a new token in the Databricks UI by navigating to User Settings > Access Tokens.

This token grants access according to your user permissions in Databricks and should be stored securely.

This property is required when using the PersonalAccessToken authentication scheme.

Databricks Connector for CData Sync

AWS Authentication

This section provides a complete list of the AWS Authentication properties you can configure in the connection string for this provider.


PropertyDescription
AWSAccessKeySpecifies your AWS account access key. This value is accessible from your AWS security credentials page.
AWSSecretKeyYour AWS account secret key. This value is accessible from your AWS security credentials page.
AWSRegionThe hosting region for your Amazon Web Services.
AWSS3BucketSpecifies the name of the AWS S3 bucket that the provider uses for staging or transferring data when connecting to Databricks.
Databricks Connector for CData Sync

AWSAccessKey

Specifies your AWS account access key. This value is accessible from your AWS security credentials page.

Remarks

To find your AWS account access key:

  1. Sign into the AWS Management console with the credentials for your root account.
  2. Select your account name or number.
  3. Select My Security Credentials in the menu.
  4. Click Continue to Security Credentials.
  5. To view or manage root account access keys, expand the Access Keys section.

Databricks Connector for CData Sync

AWSSecretKey

Your AWS account secret key. This value is accessible from your AWS security credentials page.

Remarks

Your AWS account secret key. This value is accessible from your AWS security credentials page:

  1. Sign into the AWS Management console with the credentials for your root account.
  2. Select your account name or number and select My Security Credentials in the menu that is displayed.
  3. Click Continue to Security Credentials and expand the Access Keys section to manage or create root account access keys.

Databricks Connector for CData Sync

AWSRegion

The hosting region for your Amazon Web Services.

Remarks

The hosting region for your Amazon Web Services. Available values are OHIO, NORTHERNVIRGINIA, NORTHERNCALIFORNIA, OREGON, CAPETOWN, HONGKONG, TAIPEI, HYDERABAD, JAKARTA, MALAYSIA, MELBOURNE, MUMBAI, OSAKA, SEOUL, SINGAPORE, SYDNEY, THAILAND, TOKYO, CENTRAL, CALGARY, BEIJING, NINGXIA, FRANKFURT, IRELAND, LONDON, MILAN, PARIS, SPAIN, STOCKHOLM, ZURICH, TELAVIV, MEXICOCENTRAL, BAHRAIN, UAE, SAOPAULO, GOVCLOUDEAST, GOVCLOUDWEST, ISOLATEDUSEAST, ISOLATEDUSEASTB, ISOLATEDUSEASTF, ISOLATEDUSSOUTHF, ISOLATEDUSWEST and ISOLATEDEUWEST.

Databricks Connector for CData Sync

AWSS3Bucket

Specifies the name of the AWS S3 bucket that the provider uses for staging or transferring data when connecting to Databricks.

Remarks

An Amazon S3 bucket is a top-level container for storing objects in AWS. The AWSS3Bucket property identifies the specific bucket the Sync App accesses for uploading, downloading, or staging data.

Enter the bucket name exactly as it appears in the AWS Management Console.

This property is useful when your Databricks connection relies on AWS storage for temporary data operations or large result transfers.

Databricks Connector for CData Sync

Azure Authentication

This section provides a complete list of the Azure Authentication properties you can configure in the connection string for this provider.


PropertyDescription
AzureStorageAccountThe name of your Azure storage account.
AzureAccessKeyThe storage key associated with your Azure account.
AzureTenantIdentifies the Databricks tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com ) or its directory (tenant) ID.
AzureBlobContainerSpecifies the name of the Azure Blob Storage container that the provider uses for staging or transferring data when connecting to Databricks.
Databricks Connector for CData Sync

AzureStorageAccount

The name of your Azure storage account.

Remarks

The name of your Azure storage account.

Databricks Connector for CData Sync

AzureAccessKey

The storage key associated with your Azure account.

Remarks

The storage key associated with your Databricks account. You can retrieve it as follows:

  1. Sign into the azure portal with the credentials for your root account. (https://portal.azure.com/)
  2. Click on storage accounts and select the storage account you want to use.
  3. Under settings, click Access keys.
  4. Your storage account name and key will be displayed on that page.

Databricks Connector for CData Sync

AzureTenant

Identifies the Databricks tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com ) or its directory (tenant) ID.

Remarks

A tenant is a digital container for your organization's users and resources, managed through Microsoft Entra ID (formerly Azure AD). Each tenant is associated with a unique directory ID, and often with a custom domain (for example, microsoft.com or contoso.onmicrosoft.com).

To find the directory (tenant) ID in the Microsoft Entra Admin Center, navigate to Microsoft Entra ID > Properties and copy the value labeled "Directory (tenant) ID".

This property is required in the following cases:

  • When AuthScheme is set to AzureServicePrincipal or AzureServicePrincipalCert
  • When AuthScheme is AzureAD and the user account belongs to multiple tenants

You can provide the tenant value in one of two formats:

  • A domain name (for example, contoso.onmicrosoft.com)
  • A directory (tenant) ID in GUID format (for example, c9d7b8e4-1234-4f90-bc1a-2a28e0f9e9e0)

Specifying the tenant explicitly ensures that the authentication request is routed to the correct directory, which is especially important when a user belongs to multiple tenants or when using service principal–based authentication.

If this value is omitted when required, authentication may fail or connect to the wrong tenant. This can result in errors such as unauthorized or resource not found.

Databricks Connector for CData Sync

AzureBlobContainer

Specifies the name of the Azure Blob Storage container that the provider uses for staging or transferring data when connecting to Databricks.

Remarks

An Azure Blob Storage container is a logical grouping of blobs (files) within a storage account. The AzureBlobContainer property identifies the container that the Sync App accesses for uploading, downloading, or staging data.

Enter the container name as it appears in your Azure Storage account.

This property is useful when your Databricks connection relies on Azure storage for temporary data operations or large result transfers.

Databricks Connector for CData Sync

AzureServicePrincipal Authentication

This section provides a complete list of the AzureServicePrincipal Authentication properties you can configure in the connection string for this provider.


PropertyDescription
AzureTenantIdSpecifies the directory (tenant) ID of your Microsoft Entra ID that the provider uses to authenticate to Databricks.
AzureClientIdSpecifies the application (client) ID of the Microsoft Entra ID application that the provider uses to authenticate to Databricks.
AzureClientSecretSpecifies the client secret for the Microsoft Entra ID application that the provider uses to authenticate to Databricks.
Databricks Connector for CData Sync

AzureTenantId

Specifies the directory (tenant) ID of your Microsoft Entra ID that the provider uses to authenticate to Databricks.

Remarks

The AzureTenantId property identifies the Microsoft Entra ID (formerly Azure Active Directory) tenant that owns the registered application used for authentication.

You can find the tenant ID in the Azure portal under Microsoft Entra ID > Overview > Tenant ID.

This property is required when the AuthScheme is set to AzureServicePrincipal and may also be required for AzureAD authentication, depending on your OAuth configuration.

Databricks Connector for CData Sync

AzureClientId

Specifies the application (client) ID of the Microsoft Entra ID application that the provider uses to authenticate to Databricks.

Remarks

The AzureClientId property identifies the registered application in Microsoft Entra ID (formerly Azure Active Directory) that represents your service principal.

You can find the application (client) ID in the Azure portal under Microsoft Entra ID > App registrations > Your application.

This property is required when the AuthScheme is set to AzureServicePrincipal or AzureAD.

Databricks Connector for CData Sync

AzureClientSecret

Specifies the client secret for the Microsoft Entra ID application that the provider uses to authenticate to Databricks.

Remarks

The AzureClientSecret property stores the client secret associated with your Microsoft Entra ID (formerly Azure Active Directory) application. The secret acts as a password that the Sync App uses when authenticating through an Azure service principal.

You can create and view client secrets in the Azure portal under Microsoft Entra ID > App registrations > Your application > Certificates & secrets.

This property is required when the AuthScheme is set to AzureServicePrincipal and may also be required for AzureAD if the registered platform type in your Azure app is "Web".

Databricks Connector for CData Sync

OAuth

This section provides a complete list of the OAuth properties you can configure in the connection string for this provider.


PropertyDescription
OAuthClientIdSpecifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication.
OAuthClientSecretSpecifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. (Custom OAuth applications only.).
ScopeSpecifies the scope of the authenticating user's access to the application, to ensure they get appropriate access to data. If a custom OAuth application is needed, this is generally specified at the time the application is created.
OAuthLevelSpecifies whether the provider performs OAuth authentication at the workspace level or at the account level in Databricks.
DatabricksAccountIdSpecifies the unique account ID for your Databricks account.
Databricks Connector for CData Sync

OAuthClientId

Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication.

Remarks

This property is required in two cases:

  • When using a custom OAuth application, such as in web-based authentication flows, service-based authentication, or certificate-based flows that require application registration.
  • If the driver does not provide embedded OAuth credentials.

(When the driver provides embedded OAuth credentials, this value may already be provided by the Sync App and thus not require manual entry.)

OAuthClientId is generally used alongside other OAuth-related properties such as OAuthClientSecret and OAuthSettingsLocation when configuring an authenticated connection.

OAuthClientId is one of the key connection parameters that need to be set before users can authenticate via OAuth. You can usually find this value in your identity provider’s application registration settings. Look for a field labeled Client ID, Application ID, or Consumer Key.

While the client ID is not considered a confidential value like a client secret, it is still part of your application's identity and should be handled carefully. Avoid exposing it in public repositories or shared configuration files.

For more information on how this property is used when configuring a connection, see Establishing a Connection.

Databricks Connector for CData Sync

OAuthClientSecret

Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. (Custom OAuth applications only.).

Remarks

This property (sometimes called the application secret or consumer secret) is required when using a custom OAuth application in any flow that requires secure client authentication, such as web-based OAuth, service-based connections, or certificate-based authorization flows. It is not required when using an embedded OAuth application.

The client secret is used during the token exchange step of the OAuth flow, when the driver requests an access token from the authorization server. If this value is missing or incorrect, authentication fails with either an invalid_client or an unauthorized_client error.

OAuthClientSecret is one of the key connection parameters that need to be set before users can authenticate via OAuth. You can obtain this value from your identity provider when registering the OAuth application.

Notes:

  • This value should be stored securely and never exposed in public repositories, scripts, or unsecured environments.
  • Client secrets may also expire after a set period. Be sure to monitor expiration dates and rotate secrets as needed to maintain uninterrupted access.

For more information on how this property is used when configuring a connection, see Establishing a Connection

Databricks Connector for CData Sync

Scope

Specifies the scope of the authenticating user's access to the application, to ensure they get appropriate access to data. If a custom OAuth application is needed, this is generally specified at the time the application is created.

Remarks

Scopes are set to define what kind of access the authenticating user will have; for example, read, read and write, restricted access to sensitive information. System administrators can use scopes to selectively enable access by functionality or security clearance.

When InitiateOAuth is set to GETANDREFRESH, you must use this property if you want to change which scopes are requested.

When InitiateOAuth is set to either REFRESH or OFF, you can change which scopes are requested using either this property or the Scope input.

Databricks Connector for CData Sync

OAuthLevel

Specifies whether the provider performs OAuth authentication at the workspace level or at the account level in Databricks.

Remarks

The OAuthLevel property determines the scope of OAuth authentication when connecting to Databricks. When this property is set to WorkspaceLevel, the Sync App authenticates to a specific workspace and obtains an OAuth token that grants access only within that workspace.

When this property is set to AccountLevel, the Sync App authenticates through the account console and obtains a token that grants access to account-level resources. This level is required for features such as Unity Catalog or user and permission management across multiple workspaces.

This property is useful for choosing the OAuth token scope that matches your intended access level and deployment configuration in Databricks.

Databricks Connector for CData Sync

DatabricksAccountId

Specifies the unique account ID for your Databricks account.

Remarks

The DatabricksAccountId property identifies your Databricks account for account-level authentication flows. Each account ID is unique and distinct from workspace identifiers.

You can find the account ID in the Databricks Account Console by selecting your username in the upper-right corner and viewing the Account ID in the drop-down menu.

This value is only visible in the account console and does not appear within individual workspaces.

Databricks Connector for CData Sync

SSL

This section provides a complete list of the SSL properties you can configure in the connection string for this provider.


PropertyDescription
SSLClientCertSpecifies the TLS/SSL client certificate store for SSL Client Authentication (2-way SSL). This property works in conjunction with other SSL-related properties to establish a secure connection.
SSLClientCertTypeSpecifies the type of key store containing the TLS/SSL client certificate for SSL Client Authentication. Choose from a variety of key store formats depending on your platform and certificate source.
SSLClientCertPasswordSpecifes the password required to access the TLS/SSL client certificate store. Use this property if the selected certificate store type requires a password for access.
SSLClientCertSubjectSpecifes the subject of the TLS/SSL client certificate to locate it in the certificate store. Use a comma-separated list of distinguished name fields, such as CN=www.server.com, C=US. The wildcard * selects the first certificate in the store.
SSLServerCertSpecifies the certificate to be accepted from the server when connecting using TLS/SSL.
Databricks Connector for CData Sync

SSLClientCert

Specifies the TLS/SSL client certificate store for SSL Client Authentication (2-way SSL). This property works in conjunction with other SSL-related properties to establish a secure connection.

Remarks

This property specifies the client certificate store for SSL Client Authentication. Use this property alongside SSLClientCertType, which defines the type of the certificate store, and SSLClientCertPassword, which specifies the password for password-protected stores. When SSLClientCert is set and SSLClientCertSubject is configured, the driver searches for a certificate matching the specified subject.

Certificate store designations vary by platform. On Windows, certificate stores are identified by names such as MY (personal certificates), while in Java, the certificate store is typically a file containing certificates and optional private keys.

The following are designations of the most common User and Machine certificate stores in Windows:

MYA certificate store holding personal certificates with their associated private keys.
CACertifying authority certificates.
ROOTRoot certificates.
SPCSoftware publisher certificates.

For PFXFile types, set this property to the filename. For PFXBlob types, set this property to the binary contents of the file in PKCS12 format.

Databricks Connector for CData Sync

SSLClientCertType

Specifies the type of key store containing the TLS/SSL client certificate for SSL Client Authentication. Choose from a variety of key store formats depending on your platform and certificate source.

Remarks

This property determines the format and location of the key store used to provide the client certificate. Supported values include platform-specific and universal key store formats. The available values and their usage are:

USER - defaultFor Windows, this specifies that the certificate store is a certificate store owned by the current user. Note that this store type is not available in Java.
MACHINEFor Windows, this specifies that the certificate store is a machine store. Note that this store type is not available in Java.
PFXFILEThe certificate store is the name of a PFX (PKCS12) file containing certificates.
PFXBLOBThe certificate store is a string (base-64-encoded) representing a certificate store in PFX (PKCS12) format.
JKSFILEThe certificate store is the name of a Java key store (JKS) file containing certificates. Note that this store type is only available in Java.
JKSBLOBThe certificate store is a string (base-64-encoded) representing a certificate store in JKS format. Note that this store type is only available in Java.
PEMKEY_FILEThe certificate store is the name of a PEM-encoded file that contains a private key and an optional certificate.
PEMKEY_BLOBThe certificate store is a string (base64-encoded) that contains a private key and an optional certificate.
PUBLIC_KEY_FILEThe certificate store is the name of a file that contains a PEM- or DER-encoded public key certificate.
PUBLIC_KEY_BLOBThe certificate store is a string (base-64-encoded) that contains a PEM- or DER-encoded public key certificate.
SSHPUBLIC_KEY_FILEThe certificate store is the name of a file that contains an SSH-style public key.
SSHPUBLIC_KEY_BLOBThe certificate store is a string (base-64-encoded) that contains an SSH-style public key.
P7BFILEThe certificate store is the name of a PKCS7 file containing certificates.
PPKFILEThe certificate store is the name of a file that contains a PuTTY Private Key (PPK).
XMLFILEThe certificate store is the name of a file that contains a certificate in XML format.
XMLBLOBThe certificate store is a string that contains a certificate in XML format.
BCFKSFILEThe certificate store is the name of a file that contains an Bouncy Castle keystore.
BCFKSBLOBThe certificate store is a string (base-64-encoded) that contains a Bouncy Castle keystore.

Databricks Connector for CData Sync

SSLClientCertPassword

Specifes the password required to access the TLS/SSL client certificate store. Use this property if the selected certificate store type requires a password for access.

Remarks

This property provides the password needed to open a password-protected certificate store. This property is necessary when using certificate stores that require a password for decryption, as is often recommended for PFX or JKS type stores.

If the certificate store type does not require a password, for example USER or MACHINE on Windows, this property can be left blank. Ensure that the password matches the one associated with the specified certificate store to avoid authentication errors.

Databricks Connector for CData Sync

SSLClientCertSubject

Specifes the subject of the TLS/SSL client certificate to locate it in the certificate store. Use a comma-separated list of distinguished name fields, such as CN=www.server.com, C=US. The wildcard * selects the first certificate in the store.

Remarks

This property determines which client certificate to load based on its subject. The Sync App searches for a certificate that exactly matches the specified subject. If no exact match is found, the Sync App looks for certificates containing the value of the subject. If no match is found, no certificate is selected.

The subject should follow the standard format of a comma-separated list of distinguished name fields and values. For example, CN=www.server.com, OU=Test, C=US. Common fields include the following:

FieldMeaning
CNCommon Name. This is commonly a host name like www.server.com.
OOrganization
OUOrganizational Unit
LLocality
SState
CCountry
EEmail Address

Note: If any field contains special characters, such as commas, the value must be quoted. For example: CN="Example, Inc.", C=US.

Databricks Connector for CData Sync

SSLServerCert

Specifies the certificate to be accepted from the server when connecting using TLS/SSL.

Remarks

If you are using a TLS/SSL connection, use this property to specify the TLS/SSL certificate to be accepted from the server. If you specify a value for this property, all other certificates that are not trusted by the machine are rejected.

This property can take the following forms:

Description Example
A full PEM Certificate (example shortened for brevity) -----BEGIN CERTIFICATE-----
MIIChTCCAe4CAQAwDQYJKoZIhv......Qw==
-----END CERTIFICATE-----
A path to a local file containing the certificate C:\cert.cer
The public key (example shortened for brevity) -----BEGIN RSA PUBLIC KEY-----
MIGfMA0GCSq......AQAB
-----END RSA PUBLIC KEY-----
The MD5 Thumbprint (hex values can also be either space- or colon-separated) ecadbdda5a1529c58a1e9e09828d70e4
The SHA1 Thumbprint (hex values can also be either space- or colon-separated) 34a929226ae0819f2ec14b4a3d904f801cbb150d

Note: It is possible to use '*' to signify that all certificates should be accepted, but due to security concerns this is not recommended.

Databricks Connector for CData Sync

Firewall

This section provides a complete list of the Firewall properties you can configure in the connection string for this provider.


PropertyDescription
FirewallTypeSpecifies the protocol the provider uses to tunnel traffic through a proxy-based firewall.
FirewallServerIdentifies the IP address, DNS name, or host name of a proxy used to traverse a firewall and relay user queries to network resources.
FirewallPortSpecifies the TCP port to be used for a proxy-based firewall.
FirewallUserIdentifies the user ID of the account authenticating to a proxy-based firewall.
FirewallPasswordSpecifies the password of the user account authenticating to a proxy-based firewall.
Databricks Connector for CData Sync

FirewallType

Specifies the protocol the provider uses to tunnel traffic through a proxy-based firewall.

Remarks

A proxy-based firewall (or proxy firewall) is a network security device that acts as an intermediary between user requests and the resources they access. The proxy accepts the request of an authenticated user, tunnels through the firewall, and transmits the request to the appropriate server.

Because the proxy evaluates and transfers data backets on behalf of the requesting users, the users never connect directly with the servers, only with the proxy.

Note: By default, the Sync App connects to the system proxy. To disable this behavior and connect to one of the following proxy types, set ProxyAutoDetect to false.

The following table provides port number information for each of the supported protocols.

Protocol Default Port Description
TUNNEL 80 The port where the Sync App opens a connection to Databricks. Traffic flows back and forth via the proxy at this location.
SOCKS4 1080 The port where the Sync App opens a connection to Databricks. SOCKS 4 then passes theFirewallUser value to the proxy, which determines whether the connection request should be granted.
SOCKS5 1080 The port where the Sync App sends data to Databricks. If the SOCKS 5 proxy requires authentication, set FirewallUser and FirewallPassword to credentials the proxy recognizes.

To connect to HTTP proxies, use ProxyServer and ProxyPort. To authenticate to HTTP proxies, use ProxyAuthScheme, ProxyUser, and ProxyPassword.

Databricks Connector for CData Sync

FirewallServer

Identifies the IP address, DNS name, or host name of a proxy used to traverse a firewall and relay user queries to network resources.

Remarks

A proxy-based firewall (or proxy firewall) is a network security device that acts as an intermediary between user requests and the resources they access. The proxy accepts the request of an authenticated user, tunnels through the firewall, and transmits the request to the appropriate server.

Because the proxy evaluates and transfers data backets on behalf of the requesting users, the users never connect directly with the servers, only with the proxy.

Databricks Connector for CData Sync

FirewallPort

Specifies the TCP port to be used for a proxy-based firewall.

Remarks

A proxy-based firewall (or proxy firewall) is a network security device that acts as an intermediary between user requests and the resources they access. The proxy accepts the request of an authenticated user, tunnels through the firewall, and transmits the request to the appropriate server.

Because the proxy evaluates and transfers data backets on behalf of the requesting users, the users never connect directly with the servers, only with the proxy.

Databricks Connector for CData Sync

FirewallUser

Identifies the user ID of the account authenticating to a proxy-based firewall.

Remarks

A proxy-based firewall (or proxy firewall) is a network security device that acts as an intermediary between user requests and the resources they access. The proxy accepts the request of an authenticated user, tunnels through the firewall, and transmits the request to the appropriate server.

Because the proxy evaluates and transfers data backets on behalf of the requesting users, the users never connect directly with the servers, only with the proxy.

Databricks Connector for CData Sync

FirewallPassword

Specifies the password of the user account authenticating to a proxy-based firewall.

Remarks

A proxy-based firewall (or proxy firewall) is a network security device that acts as an intermediary between user requests and the resources they access. The proxy accepts the request of an authenticated user, tunnels through the firewall, and transmits the request to the appropriate server.

Because the proxy evaluates and transfers data backets on behalf of the requesting users, the users never connect directly with the servers, only with the proxy.

Databricks Connector for CData Sync

Proxy

This section provides a complete list of the Proxy properties you can configure in the connection string for this provider.


PropertyDescription
ProxyAutoDetectSpecifies whether the provider checks your system proxy settings for existing proxy server configurations, rather than using a manually specified proxy server.
ProxyServerIdentifies the hostname or IP address of the proxy server through which you want to route HTTP traffic.
ProxyPortIdentifies the TCP port on your specified proxy server that has been reserved for routing HTTP traffic to and from the client.
ProxyAuthSchemeSpecifies the authentication method the provider uses when authenticating to the proxy server specified in the ProxyServer connection property.
ProxyUserProvides the username of a user account registered with the proxy server specified in the ProxyServer connection property.
ProxyPasswordSpecifies the password of the user specified in the ProxyUser connection property.
ProxySSLTypeSpecifies the SSL type to use when connecting to the proxy server specified in the ProxyServer connection property.
ProxyExceptionsSpecifies a semicolon-separated list of destination hostnames or IPs that are exempt from connecting through the proxy server set in the ProxyServer connection property.
Databricks Connector for CData Sync

ProxyAutoDetect

Specifies whether the provider checks your system proxy settings for existing proxy server configurations, rather than using a manually specified proxy server.

Remarks

When this connection property is set to True, the Sync App checks your system proxy settings for existing proxy server configurations (no need to manually supply proxy server details).

This connection property takes precedence over other proxy settings. If you want to configure the Sync App to connect to a specific proxy server, set ProxyAutoDetect to False.

To connect to an HTTP proxy, see ProxyServer. For other proxies, such as SOCKS or tunneling, see FirewallType.

Databricks Connector for CData Sync

ProxyServer

Identifies the hostname or IP address of the proxy server through which you want to route HTTP traffic.

Remarks

The Sync App only routes HTTP traffic through the proxy server specified in this connection property when ProxyAutoDetect is set to False.

If ProxyAutoDetect is set to True (the default), the Sync App instead routes HTTP traffic through the proxy server specified in your system proxy settings.

Databricks Connector for CData Sync

ProxyPort

Identifies the TCP port on your specified proxy server that has been reserved for routing HTTP traffic to and from the client.

Remarks

The Sync App only routes HTTP traffic through the ProxyServer port specified in this connection property when ProxyAutoDetect is set to False.

If ProxyAutoDetect is set to True (the default), the Sync App instead routes HTTP traffic through the proxy server port specified in your system proxy settings.

For other proxy types, see FirewallType.

Databricks Connector for CData Sync

ProxyAuthScheme

Specifies the authentication method the provider uses when authenticating to the proxy server specified in the ProxyServer connection property.

Remarks

Supported authentication types :

  • BASIC: The Sync App performs HTTP basic authentication.
  • DIGEST: The Sync App performs HTTP digest authentication.
  • NTLM: The Sync App retrieves an NTLM token.
  • NEGOTIATE: The Sync App retrieves an NTLM or Kerberos token based on the applicable protocol for authentication.
  • NONE: Signifies that the ProxyServer does not require authentication.

For all values other than NONE, you must also set the ProxyUser and ProxyPassword connection properties.

If you need to use another authentication type, such as SOCKS 5 authentication, see FirewallType.

Databricks Connector for CData Sync

ProxyUser

Provides the username of a user account registered with the proxy server specified in the ProxyServer connection property.

Remarks

The ProxyUser and ProxyPassword connection properties are used to connect and authenticate against the HTTP proxy specified in ProxyServer.

After selecting one of the available authentication types in ProxyAuthScheme, set this property as follows:

ProxyAuthScheme Value Value to set for ProxyUser
BASIC The username of a user registered with the proxy server.
DIGEST The username of a user registered with the proxy server.
NEGOTIATE The username of a Windows user who is a valid user in the domain or trusted domain that the proxy server is part of, in the format user@domain or domain\user.
NTLM The username of a Windows user who is a valid user in the domain or trusted domain that the proxy server is part of, in the format user@domain or domain\user.
NONE Do not set the ProxyPassword connection property.

Note: The Sync App only uses this username if ProxyAutoDetect is set to False. If ProxyAutoDetect is set to True (the default), the Sync App instead uses the username specified in your system proxy settings.

Databricks Connector for CData Sync

ProxyPassword

Specifies the password of the user specified in the ProxyUser connection property.

Remarks

The ProxyUser and ProxyPassword connection properties are used to connect and authenticate against the HTTP proxy specified in ProxyServer.

After selecting one of the available authentication types in ProxyAuthScheme, set this property as follows:

ProxyAuthScheme Value Value to set for ProxyPassword
BASIC The password associated with the proxy server user specified in ProxyUser.
DIGEST The password associated with the proxy server user specified in ProxyUser.
NEGOTIATE The password associated with the Windows user account specified in ProxyUser.
NTLM The password associated with the Windows user account specified in ProxyUser.
NONE Do not set the ProxyPassword connection property.

For SOCKS 5 authentication or tunneling, see FirewallType.

Note: The Sync App only uses this password if ProxyAutoDetect is set to False. If ProxyAutoDetect is set to True (the default), the Sync App instead uses the password specified in your system proxy settings.

Databricks Connector for CData Sync

ProxySSLType

Specifies the SSL type to use when connecting to the proxy server specified in the ProxyServer connection property.

Remarks

This property determines when to use SSL for the connection to the HTTP proxy specified by ProxyServer. You can set this connection property to the following values :

AUTODefault setting. If ProxyServer is set to an HTTPS URL, the Sync App uses the TUNNEL option. If ProxyServer is set to an HTTP URL, the component uses the NEVER option.
ALWAYSThe connection is always SSL enabled.
NEVERThe connection is not SSL enabled.
TUNNELThe connection is made through a tunneling proxy. The proxy server opens a connection to the remote host and traffic flows back and forth through the proxy.

Databricks Connector for CData Sync

ProxyExceptions

Specifies a semicolon-separated list of destination hostnames or IPs that are exempt from connecting through the proxy server set in the ProxyServer connection property.

Remarks

The ProxyServer is used for all addresses, except for addresses defined in this property. Use semicolons to separate entries.

Note: The Sync App uses the system proxy settings by default, without further configuration needed. If you want to explicitly configure proxy exceptions for this connection, set ProxyAutoDetect to False.

Databricks Connector for CData Sync

Logging

This section provides a complete list of the Logging properties you can configure in the connection string for this provider.


PropertyDescription
LogModulesSpecifies the core modules to include in the log file. Use a semicolon-separated list of module names. By default, all modules are logged.
Databricks Connector for CData Sync

LogModules

Specifies the core modules to include in the log file. Use a semicolon-separated list of module names. By default, all modules are logged.

Remarks

The Sync App writes details about each operation it performs into the logfile specified by the Logfile connection property.

Each of these logged operations are assigned to a themed category called a module, and each module has a corresponding short code used to labels individual Sync App operations as belonging to that module.

When this connection property is set to a semicolon-separated list of module codes, only operations belonging to the specified modules are written to the logfile. Note that this only affects which operations are logged moving forward and doesn't retroactively alter the existing contents of the logfile. For example: INFO;EXEC;SSL;META;

By default, logged operations from all modules are included.

You can explicitly exclude a module by prefixing it with a "-". For example: -HTTP

To apply filters to submodules, identify them with the syntax <module name>.<submodule name>. For example, the following value causes the Sync App to only log actions belonging to the HTTP module, and further refines it to exclude actions belonging to the Res submodule of the HTTP module: HTTP;-HTTP.Res

Note that the logfile filtering triggered by the Verbosity connection property takes precedence over the filtering imposed by this connection property. This means that operations of a higher verbosity level than the level specified in the Verbosity connection property are not printed in the logfile, even if they belong to one of the modules specified in this connection property.

The available modules and submodules are:

Module Name Module Description Submodules
INFO General Information. Includes the connection string, product version (build number), and initial connection messages.
  • Connec – Information related to creating or destroying connections.
  • Messag – Generic label for messages pertaining to connections, the connection string, and product version. These messages are typically specific to the Sync App, rather than being received and passed along directly from the service.
EXEC Query Execution. Includes execution messages for user-written SQL queries, parsed SQL queries, and normalized SQL queries. Success/failure messages for queries and query pages appear here as well.
  • Messag – Messages pertaining to query execution. These messages are typically specific to the Sync App, rather than being received and passed along directly from the service.
  • Normlz – Query normalization steps. Query normalization is when the product takes the user-submitted query and rewrites the query to get the same results with optimal performance.
  • Origin – This label applies to any messages recording a user's original query (the exact, unaltered, non-normalized query executed by the user).
  • Page – Messages related to query paging.
  • Parsed – Query parsing steps. Parsing is the process of converting the user-submitted query into a standardized format for easier processing.
HTTP HTTP protocol messages. Includes HTTP requests/responses (including POST messages), as well as Kerberos related messages.
  • KERB – HTTP requests related to Kerberos.
  • Messag – Messages pertaining to HTTP protocols. These messages are typically specific to the Sync App, rather than being received and passed along directly from the service.
  • Unpack – This label applies to messages about zipped data being returned from the service API and unpacked by the product.
  • Res – Messages containing HTTP responses.
  • Req – Messages containing HTTP requests.
WSDL Messages pertaining to the generation of WSDL/XSD files. —
SSL SSL certificate messages.
  • Certif – Messages pertaining to SSL certificates.
AUTH Authentication related failure/success messages.
  • Messag – Messages pertaining to authentication. These messages are typically specific to the Sync App, rather than being received and passed along directly from the service.
  • OAuth – Messages related to OAuth authentication.
  • Krbros – Kerberos-related authentication messages.
SQL Includes SQL transactions, SQL bulk transfer messages, and SQL result set messages.
  • Bulk – Messages pertaining to bulk query execution.
  • Cache – Messages related to reading row data from and writing row data to the product's cache for better performance.
  • Messag – Messages pertaining to SQL transactions. These messages are typically specific to the Sync App, rather than being received and passed along directly from the service.
  • ResSet – Query resultsets.
  • Transc – Messages related to handling transactions, including information about the number of jobs executed and backup table handling.
META Metadata cache and schema messages.
  • Cache – Messages related to reading from and modifying column and table definitions in the product's cache for better performance.
  • Schema – Messages related to retrieving metadata from or modifying the service schema.
  • MemSto – Messages related to writing to or reading from in-memory metadata cache.
  • Storag – Messages relating to storing metadata on disk or in an external data store, rather than in memory.
FUNC Information related to executing SQL functions.
  • Errmsg – Error messages related to executing SQL functions.
TCP Incoming and outgoing raw bytes on TCP transport layer messages.
  • Send – Raw data sent via the TCP protocol.
  • Receiv – Raw data received via the TCP protocol.
FTP Messages pertaining to the File Transfer Protocol.
  • Info – Status messages related to communication in the FTP protocol.
  • Client – Messages related to actions taken by the FTP client (the product) during FTP communication.
  • Server – Messages related to actions taken by the FTP server during FTP communication.
SFTP Messages pertaining to the Secure File Transfer Protocol.
  • Info – Status messages related to communication in the SFTP protocol.
  • To_Server – Messages related to actions taken by the SFTP client (the product) during SFTP communication.
  • From_Server – Messages related to actions taken by the SFTP server during SFTP communication.
POP Messages pertaining to data transferred via the Post Office Protocol.
  • Client – Messages related to actions taken by the POP client (the product) during POP communication.
  • Server – Messages related to actions taken by the POP server during POP communication.
  • Status – Status messages related to communication in the POP protocol.
SMTP Messages pertaining to data transferred via the Simple Mail Transfer Protocol.
  • Client – Messages related to actions taken by the SMTP client (the product) during SMTP communication.
  • Server – Messages related to actions taken by the SMTP server during SMTP communication.
  • Status – Status messages related to communication in the SMTP protocol.
CORE Messages relating to various internal product operations not covered by other modules. —
DEMN Messages related to SQL remoting. —
CLJB Messages about bulk data uploads (cloud job).
  • Commit – Submissions for bulk data uploads.
SRCE Miscellaneous messages produced by the product that don't belong in any other module. —
TRANCE Advanced messages concerning low-level product operations. —
TRFT Applies to log messages generated from the Thrift protocol. —

Databricks Connector for CData Sync

Schema

This section provides a complete list of the Schema properties you can configure in the connection string for this provider.


PropertyDescription
LocationSpecifies the location of a directory containing schema files that define tables, views, and stored procedures. Depending on your service's requirements, this may be expressed as either an absolute path or a relative path.
BrowsableSchemasOptional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC .
TablesOptional setting that restricts the tables reported to a subset of all available tables. For example, Tables=TableA,TableB,TableC .
ViewsOptional setting that restricts the views reported to a subset of the available tables. For example, Views=ViewA,ViewB,ViewC .
CatalogSpecifies the default catalog that the provider uses when connecting to Databricks.
PrimaryKeyIdentifiersSpecifies primary keys for tables in the hive_metastore catalog.
Databricks Connector for CData Sync

Location

Specifies the location of a directory containing schema files that define tables, views, and stored procedures. Depending on your service's requirements, this may be expressed as either an absolute path or a relative path.

Remarks

The Location property is only needed if you want to either customize definitions (for example, change a column name, ignore a column, etc.) or extend the data model with new tables, views, or stored procedures.

If left unspecified, the default location is %APPDATA%\\CData\\Databricks Data Provider\\Schema, where %APPDATA% is set to the user's configuration directory:

Platform %APPDATA%
Windows The value of the APPDATA environment variable
Linux ~/.config

Databricks Connector for CData Sync

BrowsableSchemas

Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC .

Remarks

Listing all available database schemas can take extra time, thus degrading performance. Providing a list of schemas in the connection string saves time and improves performance.

Databricks Connector for CData Sync

Tables

Optional setting that restricts the tables reported to a subset of all available tables. For example, Tables=TableA,TableB,TableC .

Remarks

Listing all available tables from some databases can take extra time, thus degrading performance. Providing a list of tables in the connection string saves time and improves performance.

If there are lots of tables available and you already know which ones you want to work with, you can use this property to restrict your viewing to only those tables. To do this, specify the tables you want in a comma-separated list. Each table should be a valid SQL identifier with any special characters escaped using square brackets, double-quotes or backticks. For example, Tables=TableA,[TableB/WithSlash],WithCatalog.WithSchema.`TableC With Space`.

Note: If you are connecting to a data source with multiple schemas or catalogs, you must specify each table you want to view by its fully qualified name. This avoids ambiguity between tables that may exist in multiple catalogs or schemas.

Databricks Connector for CData Sync

Views

Optional setting that restricts the views reported to a subset of the available tables. For example, Views=ViewA,ViewB,ViewC .

Remarks

Listing all available views from some databases can take extra time, thus degrading performance. Providing a list of views in the connection string saves time and improves performance.

If there are lots of views available and you already know which ones you want to work with, you can use this property to restrict your viewing to only those views. To do this, specify the views you want in a comma-separated list. Each view should be a valid SQL identifier with any special characters escaped using square brackets, double-quotes or backticks. For example, Views=ViewA,[ViewB/WithSlash],WithCatalog.WithSchema.`ViewC With Space`.

Note: If you are connecting to a data source with multiple schemas or catalogs, you must specify each view you want to examine by its fully qualified name. This avoids ambiguity between views that may exist in multiple catalogs or schemas.

Databricks Connector for CData Sync

Catalog

Specifies the default catalog that the provider uses when connecting to Databricks.

Remarks

A catalog in Databricks is a top-level data namespace that organizes schemas and tables. Unity Catalog environments can contain multiple catalogs, while legacy workspaces use the default Hive metastore.

When the SupportMultiCatalog property is set to true, the Catalog property must also be set to specify which catalog to use by default.

In most cases, this value should remain hive_metastore unless your workspace uses Unity Catalog and you want to connect to a different catalog.

This property is useful for defining the default data namespace when working with multiple catalogs or Unity Catalog configurations in Databricks.

Databricks Connector for CData Sync

PrimaryKeyIdentifiers

Specifies primary keys for tables in the hive_metastore catalog.

Remarks

The PrimaryKeyIdentifiers property defines primary keys for tables that do not natively include them in Databricks. This property only applies to the hive_metastore catalog.

Since Databricks does not enforce primary keys, defining them manually can enable DML operations such as UPDATE or DELETE and improve compatibility with tools that expect primary key metadata.

Primary keys are defined as a list of rules, where each rule maps tables to one or more key columns. Multiple rules are separated by semicolons.

For example:

PrimaryKeyIdentifiers="*=my_key;my_table=my_key2,my_key3;my_nokeys_table=;"

This example defines three rules:

  1. *=my_key: All tables without a specific rule have a single primary key column named my_key. Multiple keys can be set by using a comma-separated list (for example, *=my_key,my_key2).
  2. my_table=my_key2,my_key3: The my_table table has two primary key columns, my_key2 and my_key3. Missing columns are ignored.
  3. my_nokeys_table=: The my_nokeys_table table has no primary keys, overriding the default rule.

Table identifiers in each rule can include the table name only, the schema and table, or the catalog, schema, and table. Any name can be quoted using SQL-style delimiters:

/* Rules with just table names use the default connection Catalog and Schema.
   All these rules refer to the same table when Catalog=someCatalog and Schema=someSchema */

someTable=a,b,c
someSchema.someTable=a,b,c
someCatalog.someSchema.someTable=a,b,c

/* Any table or column name may be quoted */
`someCatalog`."someSchema".[someTable]=`a`,[b],"c"

This property is useful for defining primary key metadata in Databricks when working with tools or operations that rely on key-based updates or synchronization logic.

Databricks Connector for CData Sync

Databricks

This section provides a complete list of the Databricks properties you can configure in the connection string for this provider.


PropertyDescription
CloudStorageTypeSpecifies which cloud storage service the provider uses for staging or transferring data.
StoreTableInCloudSpecifies whether the Databricks server creates and stores new tables in external cloud storage instead of the Databricks File System (DBFS).
QueryTableDetailsSpecifies whether the provider uses the DESCRIBE FORMATTED command in Databricks to retrieve detailed table metadata.
UseUploadApiSpecifies whether the provider uses the Databricks Upload API to optimize Bulk INSERT operations.
UseCloudFetchSpecifies whether the provider uses CloudFetch to optimize data transfer for large query result sets in Databricks.
SupportMultiCatalogSpecifies whether the provider enables multi-catalog support for Unity Catalog in Databricks.
QueryAllMetadataSpecifies whether the provider retrieves metadata from all available catalogs and schemas in Databricks or only from those specified in the connection.
CheckSQLWarehouseAvailabilitySpecifies whether the provider checks the availability of the SQL Warehouse in Databricks before establishing a connection.
Databricks Connector for CData Sync

CloudStorageType

Specifies which cloud storage service the provider uses for staging or transferring data.

Remarks

The CloudStorageType property determines the storage system used by the Sync App when reading or writing large result sets or temporary files.

When this property is set to DBFS, the Sync App uses the internal Databricks File System provided by Databricks.

When this property is set to Azure Blob storage, the following properties are required: AzureStorageAccount, AzureAccessKey, and AzureBlobContainer.

When this property is set to AWS S3, the following properties are required: AWSAccessKey, AWSSecretKey, AWSS3Bucket, and AWSRegion.

When this property is set to Unity Catalog volume, the Sync App uses the internal Unity Catalog volume provided by Databricks.

Databricks Connector for CData Sync

StoreTableInCloud

Specifies whether the Databricks server creates and stores new tables in external cloud storage instead of the Databricks File System (DBFS).

Remarks

When this property is set to true, the Databricks server stores created tables in an external cloud storage location based on the value of CloudStorageType. In this case, CloudStorageType cannot be set to DBFS.

When this property is set to false, new tables are stored in the default Databricks File System.

This property is useful for directing table storage to managed cloud environments such as AWS S3 or Azure Blob Storage for scalability, durability, or data governance purposes.

Databricks Connector for CData Sync

QueryTableDetails

Specifies whether the provider uses the DESCRIBE FORMATTED command in Databricks to retrieve detailed table metadata.

Remarks

When this property is set to true, the Sync App runs a DESCRIBE FORMATTED query for each table to collect detailed metadata such as file format, table location, and partitioning information.

This can provide a more complete view of table structure, but may significantly increase metadata query times, especially in large environments.

When this property is set to false, only standard table metadata is queried which improves performance.

This property is useful when applications or workflows require extended table metadata beyond standard schema information.

Databricks Connector for CData Sync

UseUploadApi

Specifies whether the provider uses the Databricks Upload API to optimize Bulk INSERT operations.

Remarks

When this property is set to true, the Sync App uses the Databricks Upload API to stage and load data in bulk, improving performance for large insert operations by reducing network overhead.

When this property is set to false, data is inserted directly over the active connection, which can be slower for large datasets but may be preferable for smaller or transactional workloads.

This property is useful for accelerating high-volume data ingestion scenarios in Databricks.

Databricks Connector for CData Sync

UseCloudFetch

Specifies whether the provider uses CloudFetch to optimize data transfer for large query result sets in Databricks.

Remarks

When this property is set to true, the Sync App uses CloudFetch, a mechanism in Databricks that improves performance for large query results by temporarily storing and transferring data through cloud storage rather than the active connection.

When this property is set to false, the Sync App streams query results directly over the connection.

CloudFetch is most beneficial for queries that return very large result sets. For example, more than one million rows. Enabling it can reduce memory usage and improve data retrieval speed for large-scale operations.

Databricks Connector for CData Sync

SupportMultiCatalog

Specifies whether the provider enables multi-catalog support for Unity Catalog in Databricks.

Remarks

When this property is set to true, the Sync App supports multiple catalogs through Unity Catalog, allowing access to both Unity-managed and legacy catalogs such as hive_metastore.

When this property is set to false, multi-catalog support is disabled, and only a single catalog named CData is available for all operations.

This property is useful for connecting to environments that use Unity Catalog for cross-workspace data governance or for simplifying metadata discovery in single-catalog configurations.

Databricks Connector for CData Sync

QueryAllMetadata

Specifies whether the provider retrieves metadata from all available catalogs and schemas in Databricks or only from those specified in the connection.

Remarks

When this property is set to true, the Sync App queries metadata from all catalogs and schemas that the connection can access in Databricks.

When this property is set to false, the scope of metadata discovery is limited based on the Catalog and Database properties:

  • If only Catalog is set, metadata is queried from all schemas within that catalog.
  • If both Catalog and Database are set, metadata is queried only from the specified catalog and schema.
  • If neither property is set, metadata is queried from the default catalog and schema.

This property is useful for optimizing connection startup time and reducing the amount of metadata retrieved from Databricks when working with large or complex environments.

Databricks Connector for CData Sync

CheckSQLWarehouseAvailability

Specifies whether the provider checks the availability of the SQL Warehouse in Databricks before establishing a connection.

Remarks

When this property is set to true, the Sync App performs a preliminary check to verify that the target SQL Warehouse in Databricks is active before completing the connection. This helps prevent query failures caused by inactive or paused warehouses.

When this property is set to false, the Sync App skips the availability check and attempts to connect immediately. This can shorten connection time but may result in an error if the warehouse is not running.

This property is useful for ensuring connection stability in environments where SQL Warehouses may pause due to idle time or cost-saving configurations.

Databricks Connector for CData Sync

Miscellaneous

This section provides a complete list of the Miscellaneous properties you can configure in the connection string for this provider.


PropertyDescription
AllowPreparedStatementSpecifies whether the provider prepares SQL statements before executing them to improve performance on repeated queries.
ConnectRetryWaitTimeSpecifies the number of seconds the provider waits before retrying a connection request.
ApplicationNameSpecifies the application name that the provider includes in the HTTP User-Agent header when connecting to Databricks.
AsyncQueryTimeoutSpecifies the number of seconds the provider waits for asynchronous requests that retrieve large result sets before timing out.
DefaultColumnSizeSpecifies the default length, in characters, of string fields when the provider cannot determine a size from metadata.
DescribeCommandSpecifies which command the provider uses to retrieve metadata from Databricks.
DetectViewSpecifies whether the provider uses the DESCRIBE FORMATTED command to determine whether an object in Databricks is a table or a view.
IncludeSystemSchemasSpecifies whether to include the system catalog 'system' and system schema 'information_schema'.
MaxRowsSpecifies the maximum number of rows returned for queries that do not include either aggregation or GROUP BY.
OtherSpecifies advanced connection properties for specialized scenarios. Use this property only under the guidance of our Support team to address specific issues.
PseudoColumnsSpecifies the pseudocolumns to expose as table columns, expressed as a string in the format 'TableName=ColumnName;TableName=ColumnName'.
QueryPassthroughThis option passes the query to the Databricks server as is.
ServerConfigurationsSpecifies configuration variables to override the default settings on the Databricks server.
ServerTimeZoneSpecifies how the provider interprets datetime values returned from Databricks.
TimeoutSpecifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error.
UseDescTableQuerySpecifies whether the provider retrieves table column metadata using a DESC TABLE query instead of the Thrift API.
UseInsertSelectSyntaxDEPRECATED. This property is no longer supported, and should not be used. It will be removed in a future release.
UserDefinedViewsSpecifies a filepath to a JSON configuration file that defines custom views. The provider automatically detects and uses the views specified in this file.
Databricks Connector for CData Sync

AllowPreparedStatement

Specifies whether the provider prepares SQL statements before executing them to improve performance on repeated queries.

Remarks

Prepared statements precompile SQL queries so they can be executed multiple times without being re-parsed or recompiled.

When this property is set to true, the Sync App prepares and caches SQL statements before execution. This reduces parsing and compilation overhead for queries that run multiple times.

When this property is set to false, statements are parsed and executed each time without precompilation. This can improve performance when running many unique or one-time queries.

Prepared statements require the Sync App to maintain an active connection while the statement is prepared.

This property is useful for tuning performance based on whether your workloads reuse queries or generate new SQL dynamically.

Databricks Connector for CData Sync

ConnectRetryWaitTime

Specifies the number of seconds the provider waits before retrying a connection request.

Remarks

The ConnectRetryWaitTime property controls how long the Sync App waits before retrying when Databricks responds that a cluster is starting or temporarily unavailable.

When this property is set to a positive integer, the Sync App waits that number of seconds before each retry attempt. Typical values range from 30 to 60 seconds.

When this property is set to -1, the retry feature is disabled.

This property is useful for improving connection reliability when clusters may take time to start or resume in Databricks.

Databricks Connector for CData Sync

ApplicationName

Specifies the application name that the provider includes in the HTTP User-Agent header when connecting to Databricks.

Remarks

This property identifies the client application in HTTP requests sent by the Sync App. The value appears in the User-Agent header and and can help Databricks identify your client in request logs.

For Databricks Runtime (formerly known as Databricks Cluster), use the following format: [isv-name+product-name]/[product-version] [comment]

For Databricks SQL Warehouse, only the optional comment field can be set, using the format: [comment]

Each segment must avoid spaces, parentheses, commas, or new lines. Nested comments are not supported.

This property is useful for identifying your application in Databricks logs and API request tracking.

Databricks Connector for CData Sync

AsyncQueryTimeout

Specifies the number of seconds the provider waits for asynchronous requests that retrieve large result sets before timing out.

Remarks

The AsyncQueryTimeout property defines how long the Sync App allows asynchronous operations to run before cancelling them. It applies to the total execution time of the operation, rather than individual requests.

When this property is set to 0, asynchronous operations do not time out. They continue until they complete successfully or encounter an error condition.

When this property is set to any other value, the Sync App raises an error if the asynchronous request has not finished within the specified number of seconds.

This property is distinct from Timeout, which applies to synchronous operations. It is useful for controlling how long large or complex queries can run before timing out.

Databricks Connector for CData Sync

DefaultColumnSize

Specifies the default length, in characters, of string fields when the provider cannot determine a size from metadata.

Remarks

Sets the default length of a string field for a provider. If not set by the provider, the value will be 2000.

The DefaultColumnSize property defines the length assigned to string-type columns when the Sync App cannot obtain this information from Databricks metadata.

When this property is not set, the default value of 1048576 is used.

This property is useful when working with systems or queries that do not return explicit string length information, ensuring consistent column definitions across environments.

Databricks Connector for CData Sync

DescribeCommand

Specifies which command the provider uses to retrieve metadata from Databricks.

Remarks

The DescribeCommand property determines whether the Sync App uses the full DESCRIBE keyword or its shorthand DESC when issuing metadata queries to the Hive-compatible endpoint in Databricks.

When this property is set to DESCRIBE, the Sync App uses the full standard Hive SQL syntax.

When this property is set to DESC, the Sync App uses the shorthand form, which may be required for compatibility with certain engines or query parsers.

This property is useful for adjusting SQL compatibility when retrieving schema details from Databricks or Hive-based systems.

Databricks Connector for CData Sync

DetectView

Specifies whether the provider uses the DESCRIBE FORMATTED command to determine whether an object in Databricks is a table or a view.

Remarks

When this property is set to true, the Sync App issues a DESCRIBE FORMATTED command to retrieve detailed metadata and identify whether a specified object in Databricks is a table or a view.

When this property is set to false, the Sync App skips this additional check, which can improve performance but may prevent accurate detection of views.

This property is useful when working with mixed environments that contain both tables and views, ensuring that object types are correctly identified during metadata discovery.

Databricks Connector for CData Sync

IncludeSystemSchemas

Specifies whether to include the system catalog 'system' and system schema 'information_schema'.

Remarks

When this property is set to true, the Sync App queries and exposes system-defined schemas alongside user-defined schemas. These schemas typically contain metadata tables, built-in views, and internal objects managed by Databricks.

When this property is set to false, the Sync App omits these internal schemas, simplifying the schema view for most users.

This property is useful for advanced users who need access to system metadata or diagnostic information in Databricks.

Databricks Connector for CData Sync

MaxRows

Specifies the maximum number of rows returned for queries that do not include either aggregation or GROUP BY.

Remarks

The default value for this property, -1, means that no row limit is enforced unless the query explicitly includes a LIMIT clause. (When a query includes a LIMIT clause, the value specified in the query takes precedence over the MaxRows setting.)

Setting MaxRows to a whole number greater than 0 ensures that queries do not return excessively large result sets by default.

This property is useful for optimizing performance and preventing excessive resource consumption when executing queries that could otherwise return very large datasets.

Databricks Connector for CData Sync

Other

Specifies advanced connection properties for specialized scenarios. Use this property only under the guidance of our Support team to address specific issues.

Remarks

This property allows advanced users to configure hidden properties for specialized situations, with the advice of our Support team. These settings are not required for normal use cases but can address unique requirements or provide additional functionality. To define multiple properties, use a semicolon-separated list.

Note: It is strongly recommended to set these properties only when advised by the Support team to address specific scenarios or issues.

Integration and Formatting

PropertyDescription
DefaultColumnSizeSets the default length of string fields when the data source does not provide column length in the metadata. The default value is 2000.
ConvertDateTimeToGMT=TrueConverts date-time values to GMT, instead of the local time of the machine. The default value is False (use local time).
RecordToFile=filenameRecords the underlying socket data transfer to the specified file.

Databricks Connector for CData Sync

PseudoColumns

Specifies the pseudocolumns to expose as table columns, expressed as a string in the format 'TableName=ColumnName;TableName=ColumnName'.

Remarks

This property allows you to define which pseudocolumns the Sync App exposes as table columns.

To specify individual pseudocolumns, use the following format:

Table1=Column1;Table1=Column2;Table2=Column3

To include all pseudocolumns for all tables use:

*=*

Databricks Connector for CData Sync

QueryPassthrough

This option passes the query to the Databricks server as is.

Remarks

When this is set, queries are passed through directly to Databricks.

Databricks Connector for CData Sync

ServerConfigurations

Specifies configuration variables to override the default settings on the Databricks server.

Remarks

The ServerConfigurations property accepts a comma-separated list of configuration variables defined as name-value pairs. Each pair is sent to the Databricks server to override its default values for the current session.

For example: hive.enforce.bucketing=true,hive.enforce.sorting=true

These settings can be used to adjust query behavior, enforce specific runtime constraints, or fine-tune optimization parameters for your Databricks session.

Databricks Connector for CData Sync

ServerTimeZone

Specifies how the provider interprets datetime values returned from Databricks.

Remarks

The ServerTimeZone property controls how datetime values are interpreted and converted between the Databricks server and the local system.

When this property is set to UTC, the Sync App assumes the server stores datetime values in Coordinated Universal Time and converts them to the local time zone.

When this property is set to LOCAL, datetime values are interpreted as local time with no conversion applied.

This property is useful for ensuring consistent timestamp handling across environments with different regional or daylight-saving settings.

Databricks Connector for CData Sync

Timeout

Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error.

Remarks

The timeout applies to each individual communication with the server rather than the entire query or operation. For example, a query could continue running beyond 60 seconds if each paging call completes within the timeout limit.

Timeout is set to 60 seconds by default. To disable timeouts, set this property to 0.

Disabling the timeout allows operations to run indefinitely until they succeed or fail due to other conditions such as server-side timeouts, network interruptions, or resource limits on the server.

Note: Use this property cautiously to avoid long-running operations that could degrade performance or result in unresponsive behavior.

Databricks Connector for CData Sync

UseDescTableQuery

Specifies whether the provider retrieves table column metadata using a DESC TABLE query instead of the Thrift API.

Remarks

When this property is set to true, the Sync App issues a DESC TABLE query in Databricks to obtain column information for each table.

When this property is set to false, the Sync App retrieves column metadata using the Thrift API call GetColumns, which is supported in Apache Spark 3.0.0 and later.

Using the Thrift API can improve metadata query performance, but requires a compatible Spark version. The DESC TABLE query provides broader compatibility across Databricks environments.

Databricks Connector for CData Sync

UseInsertSelectSyntax

DEPRECATED. This property is no longer supported, and should not be used. It will be removed in a future release.

Remarks

When set to true, an INSERT INTO SELECT statement will be used when executing insert statements. When set to false, an INSERT INTO VALUES statement will be used.

Unless explicitly specified, this option will be configured accordingly based on the Databricks version.

Databricks Connector for CData Sync

UserDefinedViews

Specifies a filepath to a JSON configuration file that defines custom views. The provider automatically detects and uses the views specified in this file.

Remarks

UserDefinedViews allows you to define and manage custom views through a JSON-formatted configuration file called UserDefinedViews.json. These views are automatically recognized by the Sync App and enable you to execute custom SQL queries as if they were standard database views. The JSON file defines each view as a root element with a child element called "query", which contains the SQL query for the view.

For example:

{
	"MyView": {
		"query": "SELECT * FROM [CData].[Sample].Customers WHERE MyColumn = 'value'"
	},
	"MyView2": {
		"query": "SELECT * FROM MyTable WHERE Id IN (1,2,3)"
	}
}

You can use this property to define multiple views in a single file and specify the filepath. For example:

UserDefinedViews=C:\Path\To\UserDefinedViews.json
When you specify a view in UserDefinedViews, the Sync App only sees that view.

For further information, see User Defined Views.

Databricks Connector for CData Sync

Third Party Copyrights

LZMA from 7Zip LZMA SDK

LZMA SDK is placed in the public domain.

Anyone is free to copy, modify, publish, use, compile, sell, or distribute the original LZMA SDK code, either in source code form or as a compiled binary, for any purpose, commercial or non-commercial, and by any means.

LZMA2 from XZ SDK

Version 1.9 and older are in the public domain.

Xamarin.Forms

Xamarin SDK

The MIT License (MIT)

Copyright (c) .NET Foundation Contributors

All rights reserved.

Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

NSIS 3.10

Copyright (C) 1999-2025 Contributors THE ACCOMPANYING PROGRAM IS PROVIDED UNDER THE TERMS OF THIS COMMON PUBLIC LICENSE ("AGREEMENT"). ANY USE, REPRODUCTION OR DISTRIBUTION OF THE PROGRAM CONSTITUTES RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT.

1. DEFINITIONS

"Contribution" means:

a) in the case of the initial Contributor, the initial code and documentation distributed under this Agreement, and b) in the case of each subsequent Contributor:

i) changes to the Program, and

ii) additions to the Program;

where such changes and/or additions to the Program originate from and are distributed by that particular Contributor. A Contribution 'originates' from a Contributor if it was added to the Program by such Contributor itself or anyone acting on such Contributor's behalf. Contributions do not include additions to the Program which: (i) are separate modules of software distributed in conjunction with the Program under their own license agreement, and (ii) are not derivative works of the Program.

"Contributor" means any person or entity that distributes the Program.

"Licensed Patents " mean patent claims licensable by a Contributor which are necessarily infringed by the use or sale of its Contribution alone or when combined with the Program.

"Program" means the Contributions distributed in accordance with this Agreement.

"Recipient" means anyone who receives the Program under this Agreement, including all Contributors.

2. GRANT OF RIGHTS

a) Subject to the terms of this Agreement, each Contributor hereby grants Recipient a non-exclusive, worldwide, royalty-free copyright license to reproduce, prepare derivative works of, publicly display, publicly perform, distribute and sublicense the Contribution of such Contributor, if any, and such derivative works, in source code and object code form.

b) Subject to the terms of this Agreement, each Contributor hereby grants Recipient a non-exclusive, worldwide, royalty-free patent license under Licensed Patents to make, use, sell, offer to sell, import and otherwise transfer the Contribution of such Contributor, if any, in source code and object code form. This patent license shall apply to the combination of the Contribution and the Program if, at the time the Contribution is added by the Contributor, such addition of the Contribution causes such combination to be covered by the Licensed Patents. The patent license shall not apply to any other combinations which include the Contribution. No hardware per se is licensed hereunder.

c) Recipient understands that although each Contributor grants the licenses to its Contributions set forth herein, no assurances are provided by any Contributor that the Program does not infringe the patent or other intellectual property rights of any other entity. Each Contributor disclaims any liability to Recipient for claims brought by any other entity based on infringement of intellectual property rights or otherwise. As a condition to exercising the rights and licenses granted hereunder, each Recipient hereby assumes sole responsibility to secure any other intellectual property rights needed, if any. For example, if a third party patent license is required to allow Recipient to distribute the Program, it is Recipient's responsibility to acquire that license before distributing the Program.

d) Each Contributor represents that to its knowledge it has sufficient copyright rights in its Contribution, if any, to grant the copyright license set forth in this Agreement.

3. REQUIREMENTS

A Contributor may choose to distribute the Program in object code form under its own license agreement, provided that:

a) it complies with the terms and conditions of this Agreement; and

b) its license agreement:

i) effectively disclaims on behalf of all Contributors all warranties and conditions, express and implied, including warranties or conditions of title and non-infringement, and implied warranties or conditions of merchantability and fitness for a particular purpose;

ii) effectively excludes on behalf of all Contributors all liability for damages, including direct, indirect, special, incidental and consequential damages, such as lost profits;

iii) states that any provisions which differ from this Agreement are offered by that Contributor alone and not by any other party; and

iv) states that source code for the Program is available from such Contributor, and informs licensees how to obtain it in a reasonable manner on or through a medium customarily used for software exchange.

When the Program is made available in source code form:

a) it must be made available under this Agreement; and

b) a copy of this Agreement must be included with each copy of the Program.

Contributors may not remove or alter any copyright notices contained within the Program.

Each Contributor must identify itself as the originator of its Contribution, if any, in a manner that reasonably allows subsequent Recipients to identify the originator of the Contribution.

4. COMMERCIAL DISTRIBUTION

Commercial distributors of software may accept certain responsibilities with respect to end users, business partners and the like. While this license is intended to facilitate the commercial use of the Program, the Contributor who includes the Program in a commercial product offering should do so in a manner which does not create potential liability for other Contributors. Therefore, if a Contributor includes the Program in a commercial product offering, such Contributor ("Commercial Contributor") hereby agrees to defend and indemnify every other Contributor ("Indemnified Contributor") against any losses, damages and costs (collectively "Losses") arising from claims, lawsuits and other legal actions brought by a third party against the Indemnified Contributor to the extent caused by the acts or omissions of such Commercial Contributor in connection with its distribution of the Program in a commercial product offering. The obligations in this section do not apply to any claims or Losses relating to any actual or alleged intellectual property infringement. In order to qualify, an Indemnified Contributor must: a) promptly notify the Commercial Contributor in writing of such claim, and b) allow the Commercial Contributor to control, and cooperate with the Commercial Contributor in, the defense and any related settlement negotiations. The Indemnified Contributor may participate in any such claim at its own expense.

For example, a Contributor might include the Program in a commercial product offering, Product X. That Contributor is then a Commercial Contributor. If that Commercial Contributor then makes performance claims, or offers warranties related to Product X, those performance claims and warranties are such Commercial Contributor's responsibility alone. Under this section, the Commercial Contributor would have to defend claims against the other Contributors related to those performance claims and warranties, and if a court requires any other Contributor to pay any damages as a result, the Commercial Contributor must pay those damages.

5. NO WARRANTY

EXCEPT AS EXPRESSLY SET FORTH IN THIS AGREEMENT, THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is solely responsible for determining the appropriateness of using and distributing the Program and assumes all risks associated with its exercise of rights under this Agreement, including but not limited to the risks and costs of program errors, compliance with applicable laws, damage to or loss of data, programs or equipment, and unavailability or interruption of operations.

6. DISCLAIMER OF LIABILITY

EXCEPT AS EXPRESSLY SET FORTH IN THIS AGREEMENT, NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.

7. GENERAL

If any provision of this Agreement is invalid or unenforceable under applicable law, it shall not affect the validity or enforceability of the remainder of the terms of this Agreement, and without further action by the parties hereto, such provision shall be reformed to the minimum extent necessary to make such provision valid and enforceable.

If Recipient institutes patent litigation against a Contributor with respect to a patent applicable to software (including a cross-claim or counterclaim in a lawsuit), then any patent licenses granted by that Contributor to such Recipient under this Agreement shall terminate as of the date such litigation is filed. In addition, if Recipient institutes patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Program itself (excluding combinations of the Program with other software or hardware) infringes such Recipient's patent(s), then such Recipient's rights granted under Section 2(b) shall terminate as of the date such litigation is filed.

All Recipient's rights under this Agreement shall terminate if it fails to comply with any of the material terms or conditions of this Agreement and does not cure such failure in a reasonable period of time after becoming aware of such noncompliance. If all Recipient's rights under this Agreement terminate, Recipient agrees to cease use and distribution of the Program as soon as reasonably practicable. However, Recipient's obligations under this Agreement and any licenses granted by Recipient relating to the Program shall continue and survive.

Everyone is permitted to copy and distribute copies of this Agreement, but in order to avoid inconsistency the Agreement is copyrighted and may only be modified in the following manner. The Agreement Steward reserves the right to publish new versions (including revisions) of this Agreement from time to time. No one other than the Agreement Steward has the right to modify this Agreement. IBM is the initial Agreement Steward. IBM may assign the responsibility to serve as the Agreement Steward to a suitable separate entity. Each new version of the Agreement will be given a distinguishing version number. The Program (including Contributions) may always be distributed subject to the version of the Agreement under which it was received. In addition, after a new version of the Agreement is published, Contributor may elect to distribute the Program (including its Contributions) under the new version. Except as expressly stated in Sections 2(a) and 2(b) above, Recipient receives no rights or licenses to the intellectual property of any Contributor under this Agreement, whether expressly, by implication, estoppel or otherwise. All rights in the Program not expressly granted under this Agreement are reserved.

This Agreement is governed by the laws of the State of New York and the intellectual property laws of the United States of America. No party to this Agreement will bring a legal action under this Agreement more than one year after the cause of action arose. Each party waives its rights to a jury trial in any resulting litigation.

Apache Thrift Client v. 0.10.0

Copyright (c) 2006 - 2019, The Apache Software Foundation

The Apache License

Version 2.0, January 2004

TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION

1. Definitions.

"License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document.

"Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License.

"Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity.

"You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License.

"Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files.

"Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types.

"Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work (an example is provided in the Appendix below).

"Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work and Derivative Works thereof.

"Contribution" shall mean any work of authorship, including the original version of the Work and any modifications or additions to that Work or Derivative Works thereof, that is intentionally submitted to Licensor for inclusion in the Work by the copyright owner or by an individual or Legal Entity authorized to submit on behalf of the copyright owner. For the purposes of this definition, "submitted" means any form of electronic, verbal, or written communication sent to the Licensor or its representatives, including but not limited to communication on electronic mailing lists, source code control systems, and issue tracking systems that are managed by, or on behalf of, the Licensor for the purpose of discussing and improving the Work, but excluding communication that is conspicuously marked or otherwise designated in writing by the copyright owner as "Not a Contribution."

"Contributor" shall mean Licensor and any individual or Legal Entity on behalf of whom a Contribution has been received by Licensor and subsequently incorporated within the Work.

2. Grant of Copyright License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare Derivative Works of, publicly display, publicly perform, sublicense, and distribute the Work and such Derivative Works in Source or Object form.

3. Grant of Patent License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Work, where such license applies only to those patent claims licensable by such Contributor that are necessarily infringed by their Contribution(s) alone or by combination of their Contribution(s) with the Work to which such Contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Work or a Contribution incorporated within the Work constitutes direct or contributory patent infringement, then any patent licenses granted to You under this License for that Work shall terminate as of the date such litigation is filed.

4. Redistribution. You may reproduce and distribute copies of the Work or Derivative Works thereof in any medium, with or without modifications, and in Source or Object form, provided that You meet the following conditions:

(a) You must give any other recipients of the Work or Derivative Works a copy of this License; and

(b) You must cause any modified files to carry prominent notices stating that You changed the files; and

(c) You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and

(d) If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works, in at least one of the following places: within a NOTICE text file distributed as part of the Derivative Works; within the Source form or documentation, if provided along with the Derivative Works; or, within a display generated by the Derivative Works, if and wherever such third-party notices normally appear. The contents of the NOTICE file are for informational purposes only and do not modify the License. You may add Your own attribution notices within Derivative Works that You distribute, alongside or as an addendum to the NOTICE text from the Work, provided that such additional attribution notices cannot be construed as modifying the License.

You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Works as a whole, provided Your use, reproduction, and distribution of the Work otherwise complies with the conditions stated in this License.

5. Submission of Contributions. Unless You explicitly state otherwise, any Contribution intentionally submitted for inclusion in the Work by You to the Licensor shall be under the terms and conditions of this License, without any additional terms or conditions. Notwithstanding the above, nothing herein shall supersede or modify the terms of any separate license agreement you may have executed with Licensor regarding such Contributions.

6. Trademarks. This License does not grant permission to use the trade names, trademarks, service marks, or product names of the Licensor, except as required for reasonable and customary use in describing the origin of the Work and reproducing the content of the NOTICE file.

7. Disclaimer of Warranty. Unless required by applicable law or agreed to in writing, Licensor provides the Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Work and assume any risks associated with Your exercise of permissions under this License.

8. Limitation of Liability. In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this License or out of the use or inability to use the Work (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if such Contributor has been advised of the possibility of such damages.

9. Accepting Warranty or Additional Liability. While redistributing the Work or Derivative Works thereof, You may choose to offer, and charge a fee for, acceptance of support, warranty, indemnity, or other liability obligations and/or rights consistent with this License. However, in accepting such obligations, You may act only on Your own behalf and on Your sole responsibility, not on behalf of any other Contributor, and only if You agree to indemnify, defend, and hold each Contributor harmless for any liability incurred by, or claims asserted against, such Contributor by reason of your accepting any such warranty or additional liability.

END OF TERMS AND CONDITIONS

APPENDIX: How to apply the Apache License to your work.

To apply the Apache License to your work, attach the following boilerplate notice, with the fields enclosed by brackets "[]" replaced with your own identifying information. (Don't include the brackets!) The text should be enclosed in the appropriate comment syntax for the file format. We also recommend that a file or class name and description of purpose be included on the same "printed page" as the copyright notice for easier identification within third-party archives.

Copyright [yyyy] [name of copyright owner]

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

-------------------------------------------------- SOFTWARE DISTRIBUTED WITH THRIFT:

The Apache Thrift software includes a number of subcomponents with separate copyright notices and license terms. Your use of the source code for the these subcomponents is subject to the terms and conditions of the following licenses.

-------------------------------------------------- Portions of the following files are licensed under the MIT License:

lib/erl/src/Makefile.am

Please see doc/otp-base-license.txt for the full terms of this license.

-------------------------------------------------- For the aclocal/ax_boost_base.m4 and contrib/fb303/aclocal/ax_boost_base.m4 components:

# Copyright (c) 2007 Thomas Porschberg <[email protected]> # # Copying and distribution of this file, with or without # modification, are permitted in any medium without royalty provided # the copyright notice and this notice are preserved.

-------------------------------------------------- For the lib/nodejs/lib/thrift/json_parse.js:

/* json_parse.js 2015-05-02 Public Domain. NO WARRANTY EXPRESSED OR IMPLIED. USE AT YOUR OWN RISK.

*/ (By Douglas Crockford <[email protected]>) --------------------------------------------------

Copyright (c) 2026 CData Software, Inc. - All rights reserved.
Build 25.0.9539