JDBC Driver for Spark SQL

Build 23.0.8839


This section provides a complete list of the Miscellaneous properties you can configure in the connection string for this provider.

AsyncQueryTimeoutThe timeout for asynchronous requests issued by the provider to download large result sets.
BatchSizeThe maximum size of each batch operation to submit.
ConnectionLifeTimeThe maximum lifetime of a connection in seconds. Once the time has elapsed, the connection object is disposed.
ConnectOnOpenThis property specifies whether to connect to the Spark SQL when the connection is opened.
DescribeCommandThe describe command to determine which describe command will use to communicate with the Hive server. Accepted entries are DESCRIBE and DESC.
DetectViewSpecifies whether to use DECRIBE FORMATTED ... to detect the specified table is view or not.
HTTPPathThe path component of the URL endpoint when using HTTP TransportMode.
MaxRowsLimits the number of rows returned when no aggregation or GROUP BY is used in the query. This takes precedence over LIMIT clauses.
OtherThese hidden properties are used only in specific use cases.
PoolIdleTimeoutThe allowed idle time for a connection before it is closed.
PoolMaxSizeThe maximum connections in the pool.
PoolMinSizeThe minimum number of connections in the pool.
PoolWaitTimeThe max seconds to wait for an available connection.
PseudoColumnsThis property indicates whether or not to include pseudo columns as columns to the table.
QueryPassthroughThis option passes the query to the Spark SQL server as is.
ReadonlyYou can use this property to enforce read-only access to Spark SQL from the provider.
RTKThe runtime key used for licensing.
ServerConfigurationsA name-value list of server configuration variables to override the server defaults.
TimeoutThe value in seconds until the timeout error is thrown, canceling the operation.
UseConnectionPoolingThis property enables connection pooling.
UseDatabricksUploadApiThis option specifies whether the Databricks Upload API will be use when executing batch insert.
UseDescTableQueryThis option specifies whether the columns will be retrieved using a DESC TABLE query or the GetColumns Thrift API.The GetColumns Thrift API works for the SparkSQL 3.0.0 or later.
UseInsertSelectSyntaxSpecifies whether to use an INSERT INTO SELECT statement.
UserDefinedViewsA filepath pointing to the JSON configuration file containing your custom views.
UseSSLSpecifies whether to use SSL Encryption when connecting to Hive.

Copyright (c) 2024 CData Software, Inc. - All rights reserved.
Build 23.0.8839