Connection String Options
The connection string properties are the various options that can be used to establish a connection. This section provides a complete list of the options you can configure in the connection string for this provider. Click the links for further details.
For more information on establishing a connection, see Establishing a Connection.
Property | Description |
AuthScheme | The authentication scheme used. Accepted entries are PersonalAccessToken, Basic, OAuthU2M, AzureServicePrincipal and AzureAD. |
Server | The host name or IP address of the server hosting the Databricks database. |
User | The username used to authenticate with Databricks. |
ProtocolVersion | The Protocol Version used to authenticate with Databricks. |
Database | The name of the Databricks database. |
HTTPPath | The path component of the URL endpoint. |
Token | The token used to access the Databricks server. |
Property | Description |
AWSAccessKey | Your AWS account access key. This value is accessible from your AWS security credentials page. |
AWSSecretKey | Your AWS account secret key. This value is accessible from your AWS security credentials page. |
AWSRegion | The hosting region for your Amazon Web Services. |
AWSS3Bucket | The name of your AWS S3 bucket. |
Property | Description |
AzureStorageAccount | The name of your Azure storage account. |
AzureAccessKey | The storage key associated with your Azure account. |
AzureTenant | The Microsoft Online tenant being used to access data. If not specified, your default tenant is used. |
AzureBlobContainer | The name of your Azure Blob storage container. |
AzureServicePrincipal Authentication
Property | Description |
AzureTenantId | The Tenant id of your Microsoft Azure Active Directory. |
AzureClientId | The application(client) id of your Microsoft Azure Active Directory application. |
AzureClientSecret | The application(client) secret of your Microsoft Azure Active Directory application. |
Property | Description |
InitiateOAuth | Set this property to initiate the process to obtain or refresh the OAuth access token when you connect. |
OAuthClientId | The client Id assigned when you register your application with an OAuth authorization server. |
OAuthClientSecret | The client secret assigned when you register your application with an OAuth authorization server. |
OAuthAccessToken | The access token for connecting using OAuth. |
OAuthSettingsLocation | The location of the settings file where OAuth values are saved when InitiateOAuth is set to GETANDREFRESH or REFRESH . Alternatively, you can hold this location in memory by specifying a value starting with 'memory://'. |
CallbackURL | The OAuth callback URL to return to when authenticating. This value must match the callback URL you specify in your app settings. |
OAuthVerifier | The verifier code returned from the OAuth authorization URL. |
OAuthRefreshToken | The OAuth refresh token for the corresponding OAuth access token. |
OAuthExpiresIn | The lifetime in seconds of the OAuth AccessToken. |
OAuthTokenTimestamp | The Unix epoch timestamp in milliseconds when the current Access Token was created. |
OAuthLevel | You can generate an access token at either the Databricks account level or workspace level. |
DatabricksAccountId | The Databricks account ID. |
Property | Description |
SSLClientCert | The TLS/SSL client certificate store for SSL Client Authentication (2-way SSL). |
SSLClientCertType | The type of key store containing the TLS/SSL client certificate. |
SSLClientCertPassword | The password for the TLS/SSL client certificate. |
SSLClientCertSubject | The subject of the TLS/SSL client certificate. |
SSLServerCert | The certificate to be accepted from the server when connecting using TLS/SSL. |
Property | Description |
FirewallType | The protocol used by a proxy-based firewall. |
FirewallServer | The name or IP address of a proxy-based firewall. |
FirewallPort | The TCP port for a proxy-based firewall. |
FirewallUser | The user name to use to authenticate with a proxy-based firewall. |
FirewallPassword | A password used to authenticate to a proxy-based firewall. |
Property | Description |
ProxyAutoDetect | When this connection property is set to True, the provider checks your system proxy settings for existing proxy server configurations (no need to manually supply proxy server details). Set to False if you want to manually configure the provider to connect to a specific proxy server. |
ProxyServer | The hostname or IP address of the proxy server that you want to route HTTP traffic through. |
ProxyPort | The TCP port that the proxy server (specified in the ProxyServer connection property) is running on. |
ProxyAuthScheme | The authentication method the provider uses when authenticating to the proxy server specified in the ProxyServer connection property. |
ProxyUser | The username of a user account registered with the proxy server specified in the ProxyServer connnection property. |
ProxyPassword | The password associated with the user specified in the ProxyUser connection property. |
ProxySSLType | The SSL type to use when connecting to the ProxyServer proxy. |
ProxyExceptions | A semicolon separated list of destination hostnames or IPs that are exempt from connecting through the ProxyServer . |
Property | Description |
Logfile | A filepath which designates the name and location of the log file. |
Verbosity | The verbosity level that determines the amount of detail included in the log file. |
LogModules | Core modules to be included in the log file. |
MaxLogFileSize | A string specifying the maximum size in bytes for a log file (for example, 10 MB). |
MaxLogFileCount | A string specifying the maximum file count of log files. |
Property | Description |
Location | A path to the directory that contains the schema files defining tables, views, and stored procedures. |
BrowsableSchemas | This property restricts the schemas reported to a subset of the available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC. |
Tables | This property restricts the tables reported to a subset of the available tables. For example, Tables=TableA,TableB,TableC. |
Views | Restricts the views reported to a subset of the available tables. For example, Views=ViewA,ViewB,ViewC. |
Catalog | The default catalog name. |
PrimaryKeyIdentifiers | Set this property to define primary keys. |
Property | Description |
AutoCache | Automatically caches the results of SELECT queries into a cache database specified by either CacheLocation or both of CacheConnection and CacheProvider . |
CacheDriver | The database driver used to cache data. |
CacheConnection | The connection string for the cache database. This property is always used in conjunction with CacheProvider . Setting both properties will override the value set for CacheLocation for caching data. |
CacheLocation | Specifies the path to the cache when caching to a file. |
CacheTolerance | The tolerance for stale data in the cache specified in seconds when using AutoCache . |
Offline | Use offline mode to get the data from the cache instead of the live source. |
CacheMetadata | This property determines whether or not to cache the table metadata to a file store. |
Property | Description |
CloudStorageType | Determine which cloud storage service will be used. |
StoreTableInCloud | This option specifies whether Databricks server will create and save tables in cloud storage. |
QueryTableDetails | Specifies whether to use DESCRIBE FORMATTED ... to query detailed table information. If set to True, the query runs for a long time. |
UseUploadApi | This option specifies whether the Databricks Upload API will be used when executing Bulk INSERT operations. |
UseCloudFetch | This option specifies whether to use CloudFetch to improve query efficiency when the data volume of the table is large. |
UseLegacyDataModel | This option specifies whether to support Unity Catalog. |
QueryAllMetadata | This option controls whether to query all catalogs and schemas/databases or only specified ones. The default catalog is specified by the property Catalog . The default schema/database is specified by the property Database . |
Property | Description |
AllowPreparedStatement | Prepare a query statement before its execution. |
ConnectRetryWaitTime | This property specifies the number of seconds to wait prior to retrying a connection request. |
ApplicationName | The application name connection string property expresses the HTTP User-Agent. |
AsyncQueryTimeout | The timeout for asynchronous requests issued by the provider to download large result sets. |
BatchSize | The maximum size of each batch operation to submit. |
ConnectionLifeTime | The maximum lifetime of a connection in seconds. Once the time has elapsed, the connection object is disposed. |
ConnectOnOpen | This property specifies whether to connect to the Databricks when the connection is opened. |
DescribeCommand | The describe command used to communicate with the Hive server. Accepted entries are DESCRIBE and DESC. |
DetectView | Specifies whether to use DESCRIBE FORMATTED ... to detect the specified table is view or not. |
MaxRows | Limits the number of rows returned when no aggregation or GROUP BY is used in the query. This takes precedence over LIMIT clauses. |
Other | These hidden properties are used only in specific use cases. |
PoolIdleTimeout | The allowed idle time for a connection before it is closed. |
PoolMaxSize | The maximum connections in the pool. |
PoolMinSize | The minimum number of connections in the pool. |
PoolWaitTime | The max seconds to wait for an available connection. |
PseudoColumns | Specify a set of pseudocolumns to expose as columns. |
QueryPassthrough | This option passes the query to the Databricks server as is. |
Readonly | You can use this property to enforce read-only access to Databricks from the provider. |
RTK | The runtime key used for licensing. |
ServerConfigurations | A name-value list of server configuration variables to override the server defaults. |
ServerTimeZone | Determine how to interpret datetime values from the server. |
Timeout | The value in seconds until the timeout error is thrown, canceling the operation. |
UseConnectionPooling | This property enables connection pooling. |
UseDescTableQuery | This option specifies whether the columns will be retrieved using a DESC TABLE query or the GetColumns Thrift API.The GetColumns Thrift API works for the Apache Spark 3.0.0 or later. |
UseInsertSelectSyntax | Specifies whether to use an INSERT INTO SELECT statement. |
UserDefinedViews | A filepath pointing to the JSON configuration file containing your custom views. |