CData Cloud offers access to Snowflake across several standard services and protocols, in a cloud-hosted solution. Any application that can connect to a SQL Server database can connect to Snowflake through CData Cloud.
CData Cloud allows you to standardize and configure connections to Snowflake as though it were any other OData endpoint or standard SQL Server.
This page provides a guide to Establishing a Connection to Snowflake in CData Cloud, as well as information on the available resources, and a reference to the available connection properties.
Establishing a Connection shows how to authenticate to Snowflake and configure any necessary connection properties to create a database in CData Cloud
Accessing data from Snowflake through the available standard services and CData Cloud administration is documented in further details in the CData Cloud Documentation.
Connect to Snowflake by selecting the corresponding icon in the Database tab. Required properties are listed under Settings. The Advanced tab lists connection properties that are not typically required.
Before authenticating, set the following properties to define your Snowflake connection:
The Cloud supports OAuth authentication by default and also provides several options for federated identity providers and key-based login.
To specify the authentication method you want to use, set the AuthScheme property.
To authenticate with OAuth, set the AuthScheme to OAuth.
To complete OAuth authentication, see Creating a Custom OAuth App. Then configure the following connection properties:
The Cloud also supports additional OAuth-based authentication schemes for specific Identity Providers. These schemes extend the standard OAuth configuration and support either browser-based login or headless authentication. Depending on your identity provider, the following additional OAuth-based authentication schemes are supported. Set the AuthScheme to one of the following values and configure the required properties.
Get an OAuth Access Token
After setting the following, you are ready to connect:
Manually Get an OAuth Access Token
Set the following connection properties to obtain the OAuthAccessToken:
You can then call stored procedures to complete the OAuth exchange:
Option 1: Obtain and Exchange a Verifier Code
To obtain a verifier code, you must authenticate at the OAuth authorization URL.
See Creating a Custom OAuth App for a procedure. This section describes the procedure to authenticate and connect to data.
To obtain the verifier code, set the following properties on the headless machine:
Next, authenticate from another machine and obtain the OAuthVerifier connection property:
Finally, on the headless machine, set the following connection properties to obtain the OAuth authentication values:
Connect to Data
After the OAuth settings file is generated, set the following properties to connect to data:
Option 2: Transfer OAuth Settings
To install the Cloud on another machine, authenticate, and then transfer the resulting OAuth values:
Set User and Password to a Snowflake user and set AuthScheme to PASSWORD.
Note: Starting with accounts created using Snowflake's bundle 2024_08 (October 2024), password-based authentication is no longer supported due to security concerns. Instead, use alternative authentication methods such as OAuth or Private Key authentication.
The Cloud allows you to authenticate using key pair authentication by creating a secure token with the private key defined for your user account. To connect with this method, set AuthScheme to PrivateKey and set the following values:
Set the AuthScheme to OKTA. The following connection properties are used to connect to Okta:
The following is an example connection string:
AuthScheme=OKTA;User=username;Password=password;Url='https://myaccount.region.snowflakecomputing.com';Warehouse=My_warehouse;SSO Properties='Domain=https://cdata-okta.okta.com';
The following is an example connection string for OKTA MFA:
AuthScheme=OKTA;User=username;Password=password;MFAPasscode=8111461;Url='https://myaccount.region.snowflakecomputing.com';Warehouse=My_warehouse;SSO Properties='Domain=https://cdata-okta.okta.com;MFAType=OktaVerify;';
AuthScheme=AzureAD;Url=https://myaccount.region.snowflakecomputing.com;[email protected];
AuthScheme=PingFederate;Url=https://myaccount.region.snowflakecomputing.com;User=myuser@mydomain;Account=myaccount;Warehouse=mywarehouse;
If you choose to connect by means other than via a browser, you must specify values for ProofKey and ExternalToken. In this case, follow the steps below to authenticate:
The Cloud has generic support for SAML-based identity providers, such as OneLogin.
Set the AuthScheme to ExternalBrowser.
Set User to the Snowflake user you want to authenticate.
When you attempt a connection, the SAML provider launches a login prompt in your default web browser.
Provide the credentials associated with your SAML provider to authenticate to Snowflake.
If the authenticating user maps to a system-defined role, specify it in the RoleName property.
By default, the Cloud attempts to negotiate TLS with the server. The server certificate is validated against the default system trusted certificate store. You can override how the certificate gets validated using the SSLServerCert connection property.
To specify another certificate, see the SSLServerCert connection property.
To authenticate to an HTTP proxy, set the following:
Set the following properties:
The Cloud uses the Snowflake API to enable bidirectional SQL access.
The CData Cloud automatically retrieves the structure of your data from Snowflake, including warehouses, databases, and schemas. Specifying a database or schema is optional, but doing so limits the available tables and views to only those within the selected database or schema.
Stored Procedures are function-like interfaces to Snowflake. They are used for OAuth authentication.
Stored procedures are function-like interfaces that extend the functionality of the Cloud beyond simple SELECT/INSERT/UPDATE/DELETE operations with Snowflake.
Stored procedures accept a list of parameters, perform their intended function, and then return any relevant response data from Snowflake, along with an indication of whether the procedure succeeded or failed.
| Name | Description |
| GetSSOAuthorizationURL | Retrieves the Single Sign-On (SSO) authorization URL for browser-based authentication in Snowflake. The returned URL must be accessed in a web browser to request an access token for API connections. |
Retrieves the Single Sign-On (SSO) authorization URL for browser-based authentication in Snowflake. The returned URL must be accessed in a web browser to request an access token for API connections.
| Name | Type | Required | Description |
| Port | String | False | Specifies the listening port on the local machine for the callback URL where the authorization response will be received.
The default value is 80. |
| Name | Type | Description |
| ProofKey | String | A cryptographic key used for verifying the integrity of the authentication request in the SSO process. |
| SSOURL | String | The SSO authorization URL that users must visit to initiate authentication with Snowflake. |
| TokenURL | String | The URL used to request an access token after successful SSO authentication. |
The connection string properties are the various options that can be used to establish a connection. This section provides a complete list of the options you can configure in the connection string for this provider. Click the links for further details.
For more information on establishing a connection, see Establishing a Connection.
| Property | Description |
| AuthScheme | Specifies the authentication scheme used to establish a connection to Snowflake. It determines the authentication mechanism required for validating the user's identity and allows integration with various SSO and OAuth providers. Accepted entries are Password, OKTA, AzureAD, AzureMSI, PingFederate, PrivateKey, OAuth, OAuthClient, OAuthAzureAD, OAuthClientAzureAD, OAuthOKTA, OAuthClientOKTA, OAuthOther, OAuthClientOther, OAuthJWT or ExternalBrowser. |
| Account | The Snowflake account identifier used to connect to a Snowflake instance. The account identifier is usually derived from the URL automatically. |
| Warehouse | Specifies the name of the Snowflake warehouse to be used for query execution. A warehouse in Snowflake is a virtual compute resource that provides the necessary processing power to perform operations such as data loading, querying, and analytics. |
| User | Specifies the username used for authenticating with the Snowflake database. The username is required for establishing a successful connection, regardless of the authentication method being used (Password, OKTA, PrivateKey, AzureAD, or OAuth). |
| Password | Specifies the user's password used for authenticating with Snowflake. This property is required when using password-based authentication methods, including Password, PingFederate, and OKTA. |
| URL | Set this property to the URL of your Snowflake database instance. This URL is required for establishing a connection to your Snowflake environment and should follow the standard format provided by Snowflake. The URL is typically structured as follows: https://orgname-myaccount.snowflakecomputing.com. |
| MFAPasscode | Provides a Multi-Factor Authentication (MFA) code when connecting to Snowflake. This code is required when MFA is enabled for your Snowflake account and you are using time-based one-time passwords (TOTP) or push-based authentication methods. |
| RoleName | Specifies the role assigned to the Snowflake user during authentication. This property determines the access privileges and permissions the user has when executing queries and managing resources in Snowflake. |
| Property | Description |
| UseVirtualHosting | If true (default), buckets will be referenced in the request using the hosted-style request: http://yourbucket.s3.amazonaws.com/yourobject. If set to false, the bean will use the path-style request: http://s3.amazonaws.com/yourbucket/yourobject. Note that this property will be set to false, in case of an S3 based custom service when the CustomURL is specified. |
| Property | Description |
| AzureTenant | Identifies the Snowflake tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com ) or its directory (tenant) ID. |
| Property | Description |
| ProofKey | Specifies the Proof Key used for Single Sign-On (SSO) authentication with Snowflake when integrating with external Identity Providers (IdPs) such as Azure AD and PingFederate. This value is typically generated during the OAuth 2.0 Authorization Code Flow. |
| ExternalToken | Specifies the OAuth or SSO token used for authentication with Snowflake. This property is required when performing Single Sign-On (SSO) or OAuth-based authentication with external identity providers like Azure AD or PingFederate. |
| SSOProperties | Specifies additional properties required to connect to the identity provider (IdP). These properties are provided as a semicolon-separated list of key-value pairs, and are commonly used for Single Sign-On (SSO) authentication with Okta or other providers. |
| Property | Description |
| PrivateKey | Specifies the private key used for Key Pair Authentication when connecting to Snowflake. This property allows authentication via a public-private key pair, providing a highly secure alternative to standard password-based authentication. The PrivateKeyType field specifies the type of the certificate store specified by PrivateKey . |
| PrivateKeyPassword | Specifies the password used to decrypt the private key when using Key Pair Authentication with Snowflake. This property is required when the private key file is encrypted with a password for added security. |
| PrivateKeyType | Specifies the type of key store or format containing the private key used for Key Pair Authentication with Snowflake. This property allows the user to choose from various formats and storage mechanisms for supplying the private key. |
| PrivateKeySubject | Specifies the subject name of the certificate containing the private key used for Key Pair Authentication with Snowflake. This property is necessary when the PrivateKeyType is set to USER or MACHINE, indicating that the private key is stored in a certificate store rather than a file or PEM blob. |
| Property | Description |
| OAuthClientId | Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication. |
| OAuthClientSecret | Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. (Custom OAuth applications only.). |
| State | Specifies an optional state parameter used during the OAuth authorization process. This value serves as a mechanism to preserve the application's state between the authorization request and the redirect response. It is commonly used to prevent Cross-Site Request Forgery (CSRF) attacks by validating the authenticity of the request upon receiving the callback. |
| Scope | Determines the OAuth scopes that the application requests from Snowflake. Setting appropriate scopes defines the permissions that the application will be granted during the OAuth authorization process. |
| OAuthAuthorizationURL | The authorization URL for the OAuth service. |
| OAuthAccessTokenURL | The URL from which the OAuth access token is retrieved. |
| PKCEVerifier | Specifies the Proof Key for Code Exchange (PKCE) Verifier used during the OAuth 2.0 PKCE authentication flow. It is used as input when calling GetOAuthAccessToken . This random value is required when obtaining an access token from the authorization server after the initial authorization code is provided. |
| Property | Description |
| SSLServerCert | Specifies the certificate to be accepted from the server when connecting using TLS/SSL. |
| Property | Description |
| Verbosity | Specifies the verbosity level of the log file, which controls the amount of detail logged. Supported values range from 1 to 5. |
| Property | Description |
| BrowsableSchemas | Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC . |
| Database | The default database to use for the session when connecting to Snowflake. All SQL queries executed during the session target this database by default. |
| Schema | Specifies the schema within the Snowflake database to which the connection will be made. Providing a schema name helps narrow down the scope of database objects being accessed, improving performance when retrieving metadata and executing queries. |
| Property | Description |
| AllowPreparedStatement | Determines whether prepared statements are allowed when executing queries in Snowflake. |
| AllowUserVariables | Determines whether user-defined variables (prefixed by an $) can be used in SQL queries executed through the connection. |
| ApplicationName | Specifies the name of the application making the connection to Snowflake. This property sets the HTTP User-Agent header for the connection, allowing Snowflake to identify and log queries by application name. |
| AsyncQueryTimeout | Specifies the maximum duration, in seconds, that the provider will wait for an asynchronous request to complete when downloading large result sets. Asynchronous requests are typically used for executing complex queries or retrieving large datasets where immediate results are not expected. This property ensures that long-running operations do not hang indefinitely. |
| BatchMode | Specifies the batch operation mode for executing bulk data operations in Snowflake. This property determines whether the connection should use Binding API or Upload API based on the size of the data being processed. |
| BindingType | Specifies the binding type used for handling Date, Time, and Timestamp_* data types during data insertion operations. This property ensures consistent handling of temporal data when using different APIs for data ingestion. |
| ClientTimestampNTZTimezone | Controls how to handle the Timestamp_NTZ value since the Timestamp_NTZ type is a value without time zone. |
| CustomStage | Specifies the name of a custom stage to be used during bulk write operations in Snowflake. This can be either an internal stage (stored within Snowflake) or an external stage (hosted on cloud services like AWS S3 or Azure Blob Storage). |
| ExternalStageAWSAccessKey | Specifies the AWS Access Key ID used to authenticate with an AWS S3 external stage during bulk write operations in Snowflake. This property is necessary when using a CustomStage that references an AWS S3 bucket. |
| ExternalStageAWSSecretKey | Specifies the AWS Secret Access Key associated with your AWS Access Key ID ( ExternalStageAWSAccessKey ). This property is required for authenticating access to AWS S3 external stages during bulk write operations in Snowflake. |
| ExternalStageAzureSASToken | Specifies the Azure Blob Storage Shared Access Signature (SAS) token used to authenticate access to an Azure Blob Storage external stage. This property allows Snowflake to interact with Azure Blob Storage containers or objects for bulk data operations. This property is necessary when using a CustomStage that references an Azure location. |
| IgnoreCase | Controls whether case sensitivity is ignored for object names (for example, databases, schemas, tables, and columns) when executing SQL queries. |
| MaxRows | Specifies the maximum number of rows returned for queries that do not include either aggregation or GROUP BY. |
| MaxThreads | Specifies the number of concurrent requests. |
| MergeDelete | A Boolean property that determines whether batch DELETE statements are automatically converted to MERGE statements. This applies only when the DELETE statement’s WHERE clause exclusively contains the table’s primary key fields combined using the AND logical operator. When enabled, Snowflake attempts to handle deletions more efficiently by using the MERGE mechanism instead of standard batch deletion. |
| MergeInsert | A Boolean property that determines whether INSERT statements are automatically converted to MERGE statements when executed. This property is applicable only when the INSERT operation includes a table’s primary key field. When enabled, Snowflake attempts to upsert records by merging incoming data with existing rows if a primary key conflict occurs, rather than performing a simple insert operation. |
| MergeUpdate | A Boolean property that determines whether batch UPDATE statements are automatically converted to MERGE statements. This applies only when the UPDATE statement’s WHERE clause exclusively includes the table’s primary key fields combined using the AND logical operator. When enabled, Snowflake optimizes updates by leveraging the MERGE mechanism instead of standard batch updates. |
| RetryOnChunkTimeout | Specifies whether the connection should retry downloading data chunks from Snowflake when a network issue or timeout occurs. |
| S3Domain | Specifies the URI of the Amazon S3 bucket used as the Snowflake S3 stage. This property is required when defining external stages for data loading and unloading between Snowflake and S3. |
| SessionIdleTimeout | Specifies the timeout duration, in minutes, for idle sessions. This setting corresponds to Snowflake's session policy parameter 'SESSION_IDLE_TIMEOUT_MINS'. The default value is 240 minutes, meaning a session is terminated if it remains inactive for more than four hours. |
| SessionParameters | Specifies additional session parameters to customize the behavior of the Snowflake connection. These parameters are passed as a semicolon-separated list of key-value pairs. Common use cases include setting query tags or modifying identifier handling settings. |
| Timeout | Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error. |
This section provides a complete list of the Authentication properties you can configure in the connection string for this provider.
| Property | Description |
| AuthScheme | Specifies the authentication scheme used to establish a connection to Snowflake. It determines the authentication mechanism required for validating the user's identity and allows integration with various SSO and OAuth providers. Accepted entries are Password, OKTA, AzureAD, AzureMSI, PingFederate, PrivateKey, OAuth, OAuthClient, OAuthAzureAD, OAuthClientAzureAD, OAuthOKTA, OAuthClientOKTA, OAuthOther, OAuthClientOther, OAuthJWT or ExternalBrowser. |
| Account | The Snowflake account identifier used to connect to a Snowflake instance. The account identifier is usually derived from the URL automatically. |
| Warehouse | Specifies the name of the Snowflake warehouse to be used for query execution. A warehouse in Snowflake is a virtual compute resource that provides the necessary processing power to perform operations such as data loading, querying, and analytics. |
| User | Specifies the username used for authenticating with the Snowflake database. The username is required for establishing a successful connection, regardless of the authentication method being used (Password, OKTA, PrivateKey, AzureAD, or OAuth). |
| Password | Specifies the user's password used for authenticating with Snowflake. This property is required when using password-based authentication methods, including Password, PingFederate, and OKTA. |
| URL | Set this property to the URL of your Snowflake database instance. This URL is required for establishing a connection to your Snowflake environment and should follow the standard format provided by Snowflake. The URL is typically structured as follows: https://orgname-myaccount.snowflakecomputing.com. |
| MFAPasscode | Provides a Multi-Factor Authentication (MFA) code when connecting to Snowflake. This code is required when MFA is enabled for your Snowflake account and you are using time-based one-time passwords (TOTP) or push-based authentication methods. |
| RoleName | Specifies the role assigned to the Snowflake user during authentication. This property determines the access privileges and permissions the user has when executing queries and managing resources in Snowflake. |
Specifies the authentication scheme used to establish a connection to Snowflake. It determines the authentication mechanism required for validating the user's identity and allows integration with various SSO and OAuth providers. Accepted entries are Password, OKTA, AzureAD, AzureMSI, PingFederate, PrivateKey, OAuth, OAuthClient, OAuthAzureAD, OAuthClientAzureAD, OAuthOKTA, OAuthClientOKTA, OAuthOther, OAuthClientOther, OAuthJWT or ExternalBrowser.
string
"OAuth"
The Cloud supports the following authentication mechanisms. See the Getting Started chapter for authentication guides.
The Snowflake account identifier used to connect to a Snowflake instance. The account identifier is usually derived from the URL automatically.
string
""
The account identifier is usually derived from the Snowflake login URL. For example, if the login URL is https://myorganization-myaccount.snowflakecomputing.com, the account is myorganization-myaccount. If your account includes an AWS cloud region, it would have the format myorganization-myaccount.aws_region. Set this if your Snowflake database URL doesn't include the account name. Use the command SELECT CURRENT_ACCOUNT(); to return the exact account value.
Specifies the name of the Snowflake warehouse to be used for query execution. A warehouse in Snowflake is a virtual compute resource that provides the necessary processing power to perform operations such as data loading, querying, and analytics.
string
""
Provide the name of the Snowflake warehouse you wish to connect to. This warehouse name is required to enable query processing and data manipulation.
For example:
Warehouse='MYACCOUNT_WH'Ensure the specified warehouse is active and accessible for your user account.
Specifies the username used for authenticating with the Snowflake database. The username is required for establishing a successful connection, regardless of the authentication method being used (Password, OKTA, PrivateKey, AzureAD, or OAuth).
string
""
Provide the username associated with your Snowflake account to enable authentication. This username is required when connecting via various authentication schemes, including Password, OAuth, SSO (OKTA, AzureAD, PingFederate), and External Browser.
Specifies the user's password used for authenticating with Snowflake. This property is required when using password-based authentication methods, including Password, PingFederate, and OKTA.
string
""
The Password property is supplied during the connection process along with the User property.
The authentication mechanism (AuthScheme) determines how the password is processed:
Set this property to the URL of your Snowflake database instance. This URL is required for establishing a connection to your Snowflake environment and should follow the standard format provided by Snowflake. The URL is typically structured as follows: https://orgname-myaccount.snowflakecomputing.com.
string
""
To find your URL:
Provides a Multi-Factor Authentication (MFA) code when connecting to Snowflake. This code is required when MFA is enabled for your Snowflake account and you are using time-based one-time passwords (TOTP) or push-based authentication methods.
string
""
Allows connecting to Snowflake using a code generated by an MFA application such as Google Authenticator, Duo, or Microsoft Authenticator. The user must supply a one-time code during connection.
Specifies the role assigned to the Snowflake user during authentication. This property determines the access privileges and permissions the user has when executing queries and managing resources in Snowflake.
string
""
The default roles in Snowflake are: PUBLIC, SYSADMIN, or ACCOUNTADMIN.
This section provides a complete list of the Connection properties you can configure in the connection string for this provider.
| Property | Description |
| UseVirtualHosting | If true (default), buckets will be referenced in the request using the hosted-style request: http://yourbucket.s3.amazonaws.com/yourobject. If set to false, the bean will use the path-style request: http://s3.amazonaws.com/yourbucket/yourobject. Note that this property will be set to false, in case of an S3 based custom service when the CustomURL is specified. |
If true (default), buckets will be referenced in the request using the hosted-style request: http://yourbucket.s3.amazonaws.com/yourobject. If set to false, the bean will use the path-style request: http://s3.amazonaws.com/yourbucket/yourobject. Note that this property will be set to false, in case of an S3 based custom service when the CustomURL is specified.
bool
true
If true (default), buckets will be referenced in the request using the hosted-style request: http://yourbucket.s3.amazonaws.com/yourobject. If set to false, the bean will use the path-style request: http://s3.amazonaws.com/yourbucket/yourobject. Note that this property will be set to false, in case of an S3 based custom service when the CustomURL is specified.
This section provides a complete list of the Azure Authentication properties you can configure in the connection string for this provider.
| Property | Description |
| AzureTenant | Identifies the Snowflake tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com ) or its directory (tenant) ID. |
Identifies the Snowflake tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com ) or its directory (tenant) ID.
string
""
A tenant is a digital container for your organization's users and resources, managed through Microsoft Entra ID (formerly Azure AD). Each tenant is associated with a unique directory ID, and often with a custom domain (for example, microsoft.com or contoso.onmicrosoft.com).
To find the directory (tenant) ID in the Microsoft Entra Admin Center, navigate to Microsoft Entra ID > Properties and copy the value labeled "Directory (tenant) ID".
This property is required in the following cases:
You can provide the tenant value in one of two formats:
Specifying the tenant explicitly ensures that the authentication request is routed to the correct directory, which is especially important when a user belongs to multiple tenants or when using service principal–based authentication.
If this value is omitted when required, authentication may fail or connect to the wrong tenant. This can result in errors such as unauthorized or resource not found.
This section provides a complete list of the SSO properties you can configure in the connection string for this provider.
| Property | Description |
| ProofKey | Specifies the Proof Key used for Single Sign-On (SSO) authentication with Snowflake when integrating with external Identity Providers (IdPs) such as Azure AD and PingFederate. This value is typically generated during the OAuth 2.0 Authorization Code Flow. |
| ExternalToken | Specifies the OAuth or SSO token used for authentication with Snowflake. This property is required when performing Single Sign-On (SSO) or OAuth-based authentication with external identity providers like Azure AD or PingFederate. |
| SSOProperties | Specifies additional properties required to connect to the identity provider (IdP). These properties are provided as a semicolon-separated list of key-value pairs, and are commonly used for Single Sign-On (SSO) authentication with Okta or other providers. |
Specifies the Proof Key used for Single Sign-On (SSO) authentication with Snowflake when integrating with external Identity Providers (IdPs) such as Azure AD and PingFederate. This value is typically generated during the OAuth 2.0 Authorization Code Flow.
string
""
To obtain the ProofKey:
Specifies the OAuth or SSO token used for authentication with Snowflake. This property is required when performing Single Sign-On (SSO) or OAuth-based authentication with external identity providers like Azure AD or PingFederate.
string
""
To obtain an external token:
Specifies additional properties required to connect to the identity provider (IdP). These properties are provided as a semicolon-separated list of key-value pairs, and are commonly used for Single Sign-On (SSO) authentication with Okta or other providers.
string
""
This property allows you to define specific connection parameters needed to authenticate with the identity provider. The entries should be formatted as a semicolon-separated list of key-value pairs.
For example, when using Okta as the identity provider:
SSOProperties='Domain=myorg.okta.com;APIToken=YourOktaToken;'
This section provides a complete list of the KeyPairAuth properties you can configure in the connection string for this provider.
| Property | Description |
| PrivateKey | Specifies the private key used for Key Pair Authentication when connecting to Snowflake. This property allows authentication via a public-private key pair, providing a highly secure alternative to standard password-based authentication. The PrivateKeyType field specifies the type of the certificate store specified by PrivateKey . |
| PrivateKeyPassword | Specifies the password used to decrypt the private key when using Key Pair Authentication with Snowflake. This property is required when the private key file is encrypted with a password for added security. |
| PrivateKeyType | Specifies the type of key store or format containing the private key used for Key Pair Authentication with Snowflake. This property allows the user to choose from various formats and storage mechanisms for supplying the private key. |
| PrivateKeySubject | Specifies the subject name of the certificate containing the private key used for Key Pair Authentication with Snowflake. This property is necessary when the PrivateKeyType is set to USER or MACHINE, indicating that the private key is stored in a certificate store rather than a file or PEM blob. |
Specifies the private key used for Key Pair Authentication when connecting to Snowflake. This property allows authentication via a public-private key pair, providing a highly secure alternative to standard password-based authentication. The PrivateKeyType field specifies the type of the certificate store specified by PrivateKey .
string
""
The contents of this property is dependent on the PrivateKeyType. This property's content can take one of three forms.
Designations of certificate stores are platform-dependent.
The following are designations of the most common User and Machine certificate stores in Windows:
| MY | A certificate store holding personal certificates with their associated private keys. |
| CA | Certifying authority certificates. |
| ROOT | Root certificates. |
| SPC | Software publisher certificates. |
In Java, the certificate store normally is a file containing certificates and optional private keys.
Specifies the password used to decrypt the private key when using Key Pair Authentication with Snowflake. This property is required when the private key file is encrypted with a password for added security.
string
""
The password for the private key specified in the PrivateKey property, if required.
Specifies the type of key store or format containing the private key used for Key Pair Authentication with Snowflake. This property allows the user to choose from various formats and storage mechanisms for supplying the private key.
string
"PEMKEY_BLOB"
This property can take one of the following values. The value chosen affects the format of the PrivateKey connection property contents:
| USER | For Windows, this specifies that the certificate store is a certificate store owned by the current user. Note that this store type is not available in Java. |
| MACHINE | For Windows, this specifies that the certificate store is a machine store. Note that this store type is not available in Java. |
| PFXFILE | The certificate store is the name of a PFX (PKCS12) file containing certificates. |
| PFXBLOB | The certificate store is a string (base-64-encoded) representing a certificate store in PFX (PKCS12) format. |
| JKSFILE | The certificate store is the name of a Java key store (JKS) file containing certificates. Note that this store type is only available in Java. |
| JKSBLOB | The certificate store is a string (base-64-encoded) representing a certificate store in JKS format. Note that this store type is only available in Java. |
| PEMKEY_FILE | The certificate store is the name of a PEM-encoded file that contains a private key and an optional certificate. |
| PEMKEY_BLOB | The certificate store is a string (base64-encoded) that contains a private key and an optional certificate. |
| PUBLIC_KEY_FILE | The certificate store is the name of a file that contains a PEM- or DER-encoded public key certificate. |
| PUBLIC_KEY_BLOB | The certificate store is a string (base-64-encoded) that contains a PEM- or DER-encoded public key certificate. |
| SSHPUBLIC_KEY_FILE | The certificate store is the name of a file that contains an SSH-style public key. |
| SSHPUBLIC_KEY_BLOB | The certificate store is a string (base-64-encoded) that contains an SSH-style public key. |
| P7BFILE | The certificate store is the name of a PKCS7 file containing certificates. |
| PPKFILE | The certificate store is the name of a file that contains a PuTTY Private Key (PPK). |
| XMLFILE | The certificate store is the name of a file that contains a certificate in XML format. |
| XMLBLOB | The certificate store is a string that contains a certificate in XML format. |
Specifies the subject name of the certificate containing the private key used for Key Pair Authentication with Snowflake. This property is necessary when the PrivateKeyType is set to USER or MACHINE, indicating that the private key is stored in a certificate store rather than a file or PEM blob.
string
"*"
To obtain the PrivateKeySubject:
This section provides a complete list of the OAuth properties you can configure in the connection string for this provider.
| Property | Description |
| OAuthClientId | Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication. |
| OAuthClientSecret | Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. (Custom OAuth applications only.). |
| State | Specifies an optional state parameter used during the OAuth authorization process. This value serves as a mechanism to preserve the application's state between the authorization request and the redirect response. It is commonly used to prevent Cross-Site Request Forgery (CSRF) attacks by validating the authenticity of the request upon receiving the callback. |
| Scope | Determines the OAuth scopes that the application requests from Snowflake. Setting appropriate scopes defines the permissions that the application will be granted during the OAuth authorization process. |
| OAuthAuthorizationURL | The authorization URL for the OAuth service. |
| OAuthAccessTokenURL | The URL from which the OAuth access token is retrieved. |
| PKCEVerifier | Specifies the Proof Key for Code Exchange (PKCE) Verifier used during the OAuth 2.0 PKCE authentication flow. It is used as input when calling GetOAuthAccessToken . This random value is required when obtaining an access token from the authorization server after the initial authorization code is provided. |
Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication.
string
""
This property is required in two cases:
(When the driver provides embedded OAuth credentials, this value may already be provided by the Cloud and thus not require manual entry.)
OAuthClientId is generally used alongside other OAuth-related properties such as OAuthClientSecret and OAuthSettingsLocation when configuring an authenticated connection.
OAuthClientId is one of the key connection parameters that need to be set before users can authenticate via OAuth. You can usually find this value in your identity provider’s application registration settings. Look for a field labeled Client ID, Application ID, or Consumer Key.
While the client ID is not considered a confidential value like a client secret, it is still part of your application's identity and should be handled carefully. Avoid exposing it in public repositories or shared configuration files.
For more information on how this property is used when configuring a connection, see Establishing a Connection.
Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. (Custom OAuth applications only.).
string
""
This property (sometimes called the application secret or consumer secret) is required when using a custom OAuth application in any flow that requires secure client authentication, such as web-based OAuth, service-based connections, or certificate-based authorization flows. It is not required when using an embedded OAuth application.
The client secret is used during the token exchange step of the OAuth flow, when the driver requests an access token from the authorization server. If this value is missing or incorrect, authentication fails with either an invalid_client or an unauthorized_client error.
OAuthClientSecret is one of the key connection parameters that need to be set before users can authenticate via OAuth. You can obtain this value from your identity provider when registering the OAuth application.
Notes:
For more information on how this property is used when configuring a connection, see Establishing a Connection
Specifies an optional state parameter used during the OAuth authorization process. This value serves as a mechanism to preserve the application's state between the authorization request and the redirect response. It is commonly used to prevent Cross-Site Request Forgery (CSRF) attacks by validating the authenticity of the request upon receiving the callback.
string
""
This property is typically a randomly generated string that your application sends to the authorization server and expects to receive back unchanged. This helps to ensure that the response is coming from the intended request and not from a malicious source.
Example usage: State='xyz123RandomStringForValidation'
This property is for organizations with a particularly high security requirement.
Determines the OAuth scopes that the application requests from Snowflake. Setting appropriate scopes defines the permissions that the application will be granted during the OAuth authorization process.
string
""
By default, the Cloud will request authorization for all available scopes. To restrict permissions, you can specify this property as a space-separated list of desired OAuth scopes, allowing for more granular control over access permissions.
For example, to request only read and write access to user data, you might set the value of this property as:
"read_data write_data"If you only need to read user profile information, you might use:
"read_profile"
The authorization URL for the OAuth service.
string
""
The authorization URL for the OAuth service. At this URL, the user logs into the server and grants permissions to the application. In OAuth 1.0, if permissions are granted, the request token is authorized.
The URL from which the OAuth access token is retrieved.
string
""
In OAuth 1.0, the authorized request token is exchanged for the access token at this URL.
Specifies the Proof Key for Code Exchange (PKCE) Verifier used during the OAuth 2.0 PKCE authentication flow. It is used as input when calling GetOAuthAccessToken . This random value is required when obtaining an access token from the authorization server after the initial authorization code is provided.
string
""
Use PKCEVerifier when integrating with external Identity Providers (IdPs) that support PKCE, such as Azure AD, Okta, or Google Identity.
The PKCEVerifier is usually derived from the GetOAuthAuthorizationUrl call.
This section provides a complete list of the SSL properties you can configure in the connection string for this provider.
| Property | Description |
| SSLServerCert | Specifies the certificate to be accepted from the server when connecting using TLS/SSL. |
Specifies the certificate to be accepted from the server when connecting using TLS/SSL.
string
""
If you are using a TLS/SSL connection, use this property to specify the TLS/SSL certificate to be accepted from the server. If you specify a value for this property, all other certificates that are not trusted by the machine are rejected.
This property can take the following forms:
| Description | Example |
| A full PEM Certificate (example shortened for brevity) | -----BEGIN CERTIFICATE----- MIIChTCCAe4CAQAwDQYJKoZIhv......Qw== -----END CERTIFICATE----- |
| A path to a local file containing the certificate | C:\cert.cer |
| The public key (example shortened for brevity) | -----BEGIN RSA PUBLIC KEY----- MIGfMA0GCSq......AQAB -----END RSA PUBLIC KEY----- |
| The MD5 Thumbprint (hex values can also be either space- or colon-separated) | ecadbdda5a1529c58a1e9e09828d70e4 |
| The SHA1 Thumbprint (hex values can also be either space- or colon-separated) | 34a929226ae0819f2ec14b4a3d904f801cbb150d |
Note: It is possible to use '*' to signify that all certificates should be accepted, but due to security concerns this is not recommended.
This section provides a complete list of the Logging properties you can configure in the connection string for this provider.
| Property | Description |
| Verbosity | Specifies the verbosity level of the log file, which controls the amount of detail logged. Supported values range from 1 to 5. |
Specifies the verbosity level of the log file, which controls the amount of detail logged. Supported values range from 1 to 5.
string
"1"
This property defines the level of detail the Cloud includes in the log file. Higher verbosity levels increase the detail of the logged information, but may also result in larger log files and slower performance due to the additional data being captured.
The default verbosity level is 1, which is recommended for regular operation. Higher verbosity levels are primarily intended for debugging purposes. For more information on each level, refer to Logging.
When combined with the LogModules property, Verbosity can refine logging to specific categories of information.
This section provides a complete list of the Schema properties you can configure in the connection string for this provider.
| Property | Description |
| BrowsableSchemas | Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC . |
| Database | The default database to use for the session when connecting to Snowflake. All SQL queries executed during the session target this database by default. |
| Schema | Specifies the schema within the Snowflake database to which the connection will be made. Providing a schema name helps narrow down the scope of database objects being accessed, improving performance when retrieving metadata and executing queries. |
Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC .
string
""
Listing all available database schemas can take extra time, thus degrading performance. Providing a list of schemas in the connection string saves time and improves performance.
The default database to use for the session when connecting to Snowflake. All SQL queries executed during the session target this database by default.
string
""
The Database property simplifies your queries if you normally work with one database during a session. If you work with multiple databases in the same session, do not set this property. If you do not set this property, use the fully qualified path to tables and schemas in SQL queries.
Specifies the schema within the Snowflake database to which the connection will be made. Providing a schema name helps narrow down the scope of database objects being accessed, improving performance when retrieving metadata and executing queries.
string
""
By specifying a schema, you can reduce the scope of metadata retrieval, which is particularly useful when working with large databases. If not specified, the connection may default to the schema defined by the user's default role or configuration.
This section provides a complete list of the Miscellaneous properties you can configure in the connection string for this provider.
| Property | Description |
| AllowPreparedStatement | Determines whether prepared statements are allowed when executing queries in Snowflake. |
| AllowUserVariables | Determines whether user-defined variables (prefixed by an $) can be used in SQL queries executed through the connection. |
| ApplicationName | Specifies the name of the application making the connection to Snowflake. This property sets the HTTP User-Agent header for the connection, allowing Snowflake to identify and log queries by application name. |
| AsyncQueryTimeout | Specifies the maximum duration, in seconds, that the provider will wait for an asynchronous request to complete when downloading large result sets. Asynchronous requests are typically used for executing complex queries or retrieving large datasets where immediate results are not expected. This property ensures that long-running operations do not hang indefinitely. |
| BatchMode | Specifies the batch operation mode for executing bulk data operations in Snowflake. This property determines whether the connection should use Binding API or Upload API based on the size of the data being processed. |
| BindingType | Specifies the binding type used for handling Date, Time, and Timestamp_* data types during data insertion operations. This property ensures consistent handling of temporal data when using different APIs for data ingestion. |
| ClientTimestampNTZTimezone | Controls how to handle the Timestamp_NTZ value since the Timestamp_NTZ type is a value without time zone. |
| CustomStage | Specifies the name of a custom stage to be used during bulk write operations in Snowflake. This can be either an internal stage (stored within Snowflake) or an external stage (hosted on cloud services like AWS S3 or Azure Blob Storage). |
| ExternalStageAWSAccessKey | Specifies the AWS Access Key ID used to authenticate with an AWS S3 external stage during bulk write operations in Snowflake. This property is necessary when using a CustomStage that references an AWS S3 bucket. |
| ExternalStageAWSSecretKey | Specifies the AWS Secret Access Key associated with your AWS Access Key ID ( ExternalStageAWSAccessKey ). This property is required for authenticating access to AWS S3 external stages during bulk write operations in Snowflake. |
| ExternalStageAzureSASToken | Specifies the Azure Blob Storage Shared Access Signature (SAS) token used to authenticate access to an Azure Blob Storage external stage. This property allows Snowflake to interact with Azure Blob Storage containers or objects for bulk data operations. This property is necessary when using a CustomStage that references an Azure location. |
| IgnoreCase | Controls whether case sensitivity is ignored for object names (for example, databases, schemas, tables, and columns) when executing SQL queries. |
| MaxRows | Specifies the maximum number of rows returned for queries that do not include either aggregation or GROUP BY. |
| MaxThreads | Specifies the number of concurrent requests. |
| MergeDelete | A Boolean property that determines whether batch DELETE statements are automatically converted to MERGE statements. This applies only when the DELETE statement’s WHERE clause exclusively contains the table’s primary key fields combined using the AND logical operator. When enabled, Snowflake attempts to handle deletions more efficiently by using the MERGE mechanism instead of standard batch deletion. |
| MergeInsert | A Boolean property that determines whether INSERT statements are automatically converted to MERGE statements when executed. This property is applicable only when the INSERT operation includes a table’s primary key field. When enabled, Snowflake attempts to upsert records by merging incoming data with existing rows if a primary key conflict occurs, rather than performing a simple insert operation. |
| MergeUpdate | A Boolean property that determines whether batch UPDATE statements are automatically converted to MERGE statements. This applies only when the UPDATE statement’s WHERE clause exclusively includes the table’s primary key fields combined using the AND logical operator. When enabled, Snowflake optimizes updates by leveraging the MERGE mechanism instead of standard batch updates. |
| RetryOnChunkTimeout | Specifies whether the connection should retry downloading data chunks from Snowflake when a network issue or timeout occurs. |
| S3Domain | Specifies the URI of the Amazon S3 bucket used as the Snowflake S3 stage. This property is required when defining external stages for data loading and unloading between Snowflake and S3. |
| SessionIdleTimeout | Specifies the timeout duration, in minutes, for idle sessions. This setting corresponds to Snowflake's session policy parameter 'SESSION_IDLE_TIMEOUT_MINS'. The default value is 240 minutes, meaning a session is terminated if it remains inactive for more than four hours. |
| SessionParameters | Specifies additional session parameters to customize the behavior of the Snowflake connection. These parameters are passed as a semicolon-separated list of key-value pairs. Common use cases include setting query tags or modifying identifier handling settings. |
| Timeout | Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error. |
Determines whether prepared statements are allowed when executing queries in Snowflake.
bool
false
A prepared statement is a SQL query that is precompiled and can be executed multiple times with different parameters.
The default value for this property is 'false', meaning that statements are parsed each time they are executed. Leave the default setting if you are executing many different queries only once, or if you have complex queries.
If you are executing the same query repeatedly, you will generally see better performance by setting this property to 'true'. Preparing the query avoids recompiling the same query over and over. However, prepared statements also require the Cloud to keep the connection active and open while the statement is prepared.
Determines whether user-defined variables (prefixed by an $) can be used in SQL queries executed through the connection.
bool
false
When set to 'true', user variables (prefixed by an $) can be used in SQL queries. By default, identifiers prefixed with $ are treated as command parameters rather than user-defined variables.
Specifies the name of the application making the connection to Snowflake. This property sets the HTTP User-Agent header for the connection, allowing Snowflake to identify and log queries by application name.
string
""
Snowflake logs the application name in query history and session information, making it useful for tracking and monitoring queries from different applications connecting to the same Snowflake environment. It can identify which applications generate heavy workloads. If not specified, Snowflake assigns a default value (often the client driver name).
You can view the ApplicationName in Snowflake's QUERY_HISTORY table.
Specifies the maximum duration, in seconds, that the provider will wait for an asynchronous request to complete when downloading large result sets. Asynchronous requests are typically used for executing complex queries or retrieving large datasets where immediate results are not expected. This property ensures that long-running operations do not hang indefinitely.
int
300
If this property is set to 0, asynchronous operations will run until completion, regardless of how long they take, unless an error condition occurs. This property is different from Timeout, which governs the timeout for individual HTTP operations.
When dealing with large datasets, setting a reasonable timeout value ensures that the operation will terminate if it exceeds the specified duration, helping to prevent resource exhaustion or prolonged processing.
If the AsyncQueryTimeout limit is reached and the asynchronous request has not completed, the Cloud will raise an error condition indicating that the request was unsuccessful due to a timeout.
Specifies the batch operation mode for executing bulk data operations in Snowflake. This property determines whether the connection should use Binding API or Upload API based on the size of the data being processed.
string
"Auto"
There are three accepted values: Auto (default), Binding, and Upload.
Forcing Binding or Upload can improve performance if you know the data size will always be small or large, respectively.
Specifies the binding type used for handling Date, Time, and Timestamp_* data types during data insertion operations. This property ensures consistent handling of temporal data when using different APIs for data ingestion.
string
"DEFAULT"
There are two kinds of binding types: DEFAULT and TEXT.
The behavior for TIMESTAMP_* types is determined by the session parameter CLIENT_TIMESTAMP_TYPE_MAPPING. However, when using the Upload API, this session parameter does not apply, potentially causing inconsistent timestamp values when inserting different quantities of records.
If you need consistent timestamp handling across Binding API and Upload API, use BindingType=TEXT.
Controls how to handle the Timestamp_NTZ value since the Timestamp_NTZ type is a value without time zone.
string
"Local"
Specifies the name of a custom stage to be used during bulk write operations in Snowflake. This can be either an internal stage (stored within Snowflake) or an external stage (hosted on cloud services like AWS S3 or Azure Blob Storage).
string
""
If the stage is external, the AWS or Azure credentials must be provided as well via the ExternalStageAWSAccessKey/ExternalStageAWSSecretKey or ExternalStageAzureSASToken properties.
When the CustomStage property is left unspecified, the Cloud generates a temporary stage automatically during the upload process and delete it after the upload is complete.
To avoid parsing errors with the generated CSV, you should include the FIELD_OPTIONALLY_ENCLOSED_BY parameter on the stage definition and set it to the double quote character. Otherwise, you may face parsing issues if you have string values that contain special characters in CSV (such as commas or double quotes).
For example:
CREATE STAGE "TEST_STAGE_CDATA" FILE_FORMAT = (FIELD_OPTIONALLY_ENCLOSED_BY='\"' )
Use internal stages for small to medium-sized data or external stages for integrating with cloud storage systems like AWS S3 or Azure Blob Storage.
Specifies the AWS Access Key ID used to authenticate with an AWS S3 external stage during bulk write operations in Snowflake. This property is necessary when using a CustomStage that references an AWS S3 bucket.
string
""
Your AWS account access key. This value is accessible from your AWS security credentials page:
Specifies the AWS Secret Access Key associated with your AWS Access Key ID ( ExternalStageAWSAccessKey ). This property is required for authenticating access to AWS S3 external stages during bulk write operations in Snowflake.
string
""
Your AWS account secret key. This value is accessible from your AWS security credentials page:
Specifies the Azure Blob Storage Shared Access Signature (SAS) token used to authenticate access to an Azure Blob Storage external stage. This property allows Snowflake to interact with Azure Blob Storage containers or objects for bulk data operations. This property is necessary when using a CustomStage that references an Azure location.
string
""
The string value of the Azure Blob shared access signature.
In the Azure Portal, go to Shared access signature in the Settings section for your Azure Blob container. Click Generate SAS token and URL and copy the value from the Blob SAS token textbox. Please be careful to select the proper permissions (Create, Write, Delete) in the Permissions dropdown list and the validity of the Start and Expiration time before you generate the SAS token.
Controls whether case sensitivity is ignored for object names (for example, databases, schemas, tables, and columns) when executing SQL queries.
bool
false
The default value is 'false' (object names are case sensitive). This property is useful for legacy systems that do not enforce case sensitivity.
Specifies the maximum number of rows returned for queries that do not include either aggregation or GROUP BY.
int
-1
The default value for this property, -1, means that no row limit is enforced unless the query explicitly includes a LIMIT clause. (When a query includes a LIMIT clause, the value specified in the query takes precedence over the MaxRows setting.)
Setting MaxRows to a whole number greater than 0 ensures that queries do not return excessively large result sets by default.
This property is useful for optimizing performance and preventing excessive resource consumption when executing queries that could otherwise return very large datasets.
Specifies the number of concurrent requests.
string
"5"
This property allows you to issue multiple requests simultaneously, thereby improving performance.
A Boolean property that determines whether batch DELETE statements are automatically converted to MERGE statements. This applies only when the DELETE statement’s WHERE clause exclusively contains the table’s primary key fields combined using the AND logical operator. When enabled, Snowflake attempts to handle deletions more efficiently by using the MERGE mechanism instead of standard batch deletion.
bool
false
MergeDelete allows for upsert functionality. This property is primarily intended for use with tools where you have no direct control over the queries being executed. Otherwise, as long as Query Passthrough is 'true', you could execute the MERGE command directly.
When this property is 'false', DELETE bulk statements are executed against the server. When it is set to 'true' and the DELETE query contains the primary key field, Snowflake sends a MERGE query that executes a DELETE if match is found in Snowflake.
For example this query:
DELETE FROM "Table" WHERE "ID" = 1 AND "NAME" = 'Jerry'
is sent to Snowflake as the following MERGE request:
MERGE INTO "Table" AS "Target" USING "RTABLE1_TMP_20eca05b-c050-47dd-89bc-81c7f617f877" AS "Source" ON ("Target"."ID" = "Source"."ID" AND "Target"."NAME" = "Source"."NAME")
WHEN MATCHED THEN DELETE
A Boolean property that determines whether INSERT statements are automatically converted to MERGE statements when executed. This property is applicable only when the INSERT operation includes a table’s primary key field. When enabled, Snowflake attempts to upsert records by merging incoming data with existing rows if a primary key conflict occurs, rather than performing a simple insert operation.
bool
false
MergeInsert allows for upsert functionality. This property is primarily intended for use with tools where you have no direct control over the queries being executed. Otherwise, as long as Query Passthrough is 'true', you could execute the MERGE command directly.
When this property is 'false', INSERT statements are executed directly against the server. When it is set to 'true' and the INSERT query contains the primary key field, the Snowflake sends a MERGE query that executes an INSERT if no match is found in Snowflake or an UPDATE if it is found. For example this query:
INSERT INTO "Table" ("ID", "NAME", "AGE") VALUES (1, 'NewName', 10)
is sent to Snowflake as the following MERGE request:
MERGE INTO "Table" AS "Target" USING (SELECT 1 AS "ID") AS [Source] ON ("Target"."ID" = "Source"."ID")
WHEN NOT MATCHED THEN INSERT ("ID", "NAME", "AGE") VALUES (1, 'NewName', 10)
WHEN MATCHED THEN UPDATE SET "NAME" = 'NewName', "AGE" = 10
A Boolean property that determines whether batch UPDATE statements are automatically converted to MERGE statements. This applies only when the UPDATE statement’s WHERE clause exclusively includes the table’s primary key fields combined using the AND logical operator. When enabled, Snowflake optimizes updates by leveraging the MERGE mechanism instead of standard batch updates.
bool
false
MergeUpdate allows for upsert functionality. This property is primarily intended for use with tools where you have no direct control over the queries being executed. Otherwise, as long as Query Passthrough is 'true', you could execute the MERGE command directly.
When this property is 'false', UPDATE statements are executed directly against the server. When it is set to 'true' and the UPDATE query contains the primary key field, Snowflake sends a MERGE query that executes an UPDATE if a match is found in Snowflake. For example this query:
UPDATE "Table" SET "NAME" = 'NewName', "AGE" = 10 WHERE "ID" = 1is sent to Snowflake as the following MERGE request:
MERGE INTO "Table" AS "Target" USING "RTABLE1_TMP_20eca05b-c050-47dd-89bc-81c7f617f877" AS "Source" ON ("Target"."ID" = "Source"."ID")
WHEN MATCHED THEN UPDATE SET "Target"."NAME" = "Source"."NAME", "Target"."AGE" = "Source"."AGE"
Specifies whether the connection should retry downloading data chunks from Snowflake when a network issue or timeout occurs.
bool
false
Typically if a network issue such as a timeout occurs during chunk downloading of data, the CData Cloud throws an exception. Set this property to 'true' to cause the CData Cloud to attempt retrying the request before failing.
Specifies the URI of the Amazon S3 bucket used as the Snowflake S3 stage. This property is required when defining external stages for data loading and unloading between Snowflake and S3.
string
""
Generally, the default domain is "s3.amazonaws.com". However, for Chinese S3 regions, the default domain is "s3.{region}.amazonaws.com.cn". Ensure the specified domain matches the region of your S3 bucket for proper connectivity.
Specifies the timeout duration, in minutes, for idle sessions. This setting corresponds to Snowflake's session policy parameter 'SESSION_IDLE_TIMEOUT_MINS'. The default value is 240 minutes, meaning a session is terminated if it remains inactive for more than four hours.
int
240
For example, changing the setting of this property to "60" terminates an idle session after one hour of inactivity.
Specifies additional session parameters to customize the behavior of the Snowflake connection. These parameters are passed as a semicolon-separated list of key-value pairs. Common use cases include setting query tags or modifying identifier handling settings.
string
""
This property allows the user to define session-specific configurations for Snowflake.
For example, the following session parameters sets a query tag for tracking purposes and allows case-insensitive handling of quoted identifiers. They are separated by a semicolon.
This property is used in rare cases only.
SessionParameters='QUERY_TAG=MyTag;QUOTED_IDENTIFIERS_IGNORE_CASE=True;'
Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error.
int
120
The timeout applies to each individual communication with the server rather than the entire query or operation. For example, a query could continue running beyond 60 seconds if each paging call completes within the timeout limit.
Timeout is set to 60 seconds by default. To disable timeouts, set this property to 0.
Disabling the timeout allows operations to run indefinitely until they succeed or fail due to other conditions such as server-side timeouts, network interruptions, or resource limits on the server.
Note: Use this property cautiously to avoid long-running operations that could degrade performance or result in unresponsive behavior.
LZMA from 7Zip LZMA SDK
LZMA SDK is placed in the public domain.
Anyone is free to copy, modify, publish, use, compile, sell, or distribute the original LZMA SDK code, either in source code form or as a compiled binary, for any purpose, commercial or non-commercial, and by any means.
LZMA2 from XZ SDK
Version 1.9 and older are in the public domain.
Xamarin.Forms
Xamarin SDK
The MIT License (MIT)
Copyright (c) .NET Foundation Contributors
All rights reserved.
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
NSIS 3.10
Copyright (C) 1999-2025 Contributors THE ACCOMPANYING PROGRAM IS PROVIDED UNDER THE TERMS OF THIS COMMON PUBLIC LICENSE ("AGREEMENT"). ANY USE, REPRODUCTION OR DISTRIBUTION OF THE PROGRAM CONSTITUTES RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT.
1. DEFINITIONS
"Contribution" means:
a) in the case of the initial Contributor, the initial code and documentation distributed under this Agreement, and b) in the case of each subsequent Contributor:
i) changes to the Program, and
ii) additions to the Program;
where such changes and/or additions to the Program originate from and are distributed by that particular Contributor. A Contribution 'originates' from a Contributor if it was added to the Program by such Contributor itself or anyone acting on such Contributor's behalf. Contributions do not include additions to the Program which: (i) are separate modules of software distributed in conjunction with the Program under their own license agreement, and (ii) are not derivative works of the Program.
"Contributor" means any person or entity that distributes the Program.
"Licensed Patents " mean patent claims licensable by a Contributor which are necessarily infringed by the use or sale of its Contribution alone or when combined with the Program.
"Program" means the Contributions distributed in accordance with this Agreement.
"Recipient" means anyone who receives the Program under this Agreement, including all Contributors.
2. GRANT OF RIGHTS
a) Subject to the terms of this Agreement, each Contributor hereby grants Recipient a non-exclusive, worldwide, royalty-free copyright license to reproduce, prepare derivative works of, publicly display, publicly perform, distribute and sublicense the Contribution of such Contributor, if any, and such derivative works, in source code and object code form.
b) Subject to the terms of this Agreement, each Contributor hereby grants Recipient a non-exclusive, worldwide, royalty-free patent license under Licensed Patents to make, use, sell, offer to sell, import and otherwise transfer the Contribution of such Contributor, if any, in source code and object code form. This patent license shall apply to the combination of the Contribution and the Program if, at the time the Contribution is added by the Contributor, such addition of the Contribution causes such combination to be covered by the Licensed Patents. The patent license shall not apply to any other combinations which include the Contribution. No hardware per se is licensed hereunder.
c) Recipient understands that although each Contributor grants the licenses to its Contributions set forth herein, no assurances are provided by any Contributor that the Program does not infringe the patent or other intellectual property rights of any other entity. Each Contributor disclaims any liability to Recipient for claims brought by any other entity based on infringement of intellectual property rights or otherwise. As a condition to exercising the rights and licenses granted hereunder, each Recipient hereby assumes sole responsibility to secure any other intellectual property rights needed, if any. For example, if a third party patent license is required to allow Recipient to distribute the Program, it is Recipient's responsibility to acquire that license before distributing the Program.
d) Each Contributor represents that to its knowledge it has sufficient copyright rights in its Contribution, if any, to grant the copyright license set forth in this Agreement.
3. REQUIREMENTS
A Contributor may choose to distribute the Program in object code form under its own license agreement, provided that:
a) it complies with the terms and conditions of this Agreement; and
b) its license agreement:
i) effectively disclaims on behalf of all Contributors all warranties and conditions, express and implied, including warranties or conditions of title and non-infringement, and implied warranties or conditions of merchantability and fitness for a particular purpose;
ii) effectively excludes on behalf of all Contributors all liability for damages, including direct, indirect, special, incidental and consequential damages, such as lost profits;
iii) states that any provisions which differ from this Agreement are offered by that Contributor alone and not by any other party; and
iv) states that source code for the Program is available from such Contributor, and informs licensees how to obtain it in a reasonable manner on or through a medium customarily used for software exchange.
When the Program is made available in source code form:
a) it must be made available under this Agreement; and
b) a copy of this Agreement must be included with each copy of the Program.
Contributors may not remove or alter any copyright notices contained within the Program.
Each Contributor must identify itself as the originator of its Contribution, if any, in a manner that reasonably allows subsequent Recipients to identify the originator of the Contribution.
4. COMMERCIAL DISTRIBUTION
Commercial distributors of software may accept certain responsibilities with respect to end users, business partners and the like. While this license is intended to facilitate the commercial use of the Program, the Contributor who includes the Program in a commercial product offering should do so in a manner which does not create potential liability for other Contributors. Therefore, if a Contributor includes the Program in a commercial product offering, such Contributor ("Commercial Contributor") hereby agrees to defend and indemnify every other Contributor ("Indemnified Contributor") against any losses, damages and costs (collectively "Losses") arising from claims, lawsuits and other legal actions brought by a third party against the Indemnified Contributor to the extent caused by the acts or omissions of such Commercial Contributor in connection with its distribution of the Program in a commercial product offering. The obligations in this section do not apply to any claims or Losses relating to any actual or alleged intellectual property infringement. In order to qualify, an Indemnified Contributor must: a) promptly notify the Commercial Contributor in writing of such claim, and b) allow the Commercial Contributor to control, and cooperate with the Commercial Contributor in, the defense and any related settlement negotiations. The Indemnified Contributor may participate in any such claim at its own expense.
For example, a Contributor might include the Program in a commercial product offering, Product X. That Contributor is then a Commercial Contributor. If that Commercial Contributor then makes performance claims, or offers warranties related to Product X, those performance claims and warranties are such Commercial Contributor's responsibility alone. Under this section, the Commercial Contributor would have to defend claims against the other Contributors related to those performance claims and warranties, and if a court requires any other Contributor to pay any damages as a result, the Commercial Contributor must pay those damages.
5. NO WARRANTY
EXCEPT AS EXPRESSLY SET FORTH IN THIS AGREEMENT, THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is solely responsible for determining the appropriateness of using and distributing the Program and assumes all risks associated with its exercise of rights under this Agreement, including but not limited to the risks and costs of program errors, compliance with applicable laws, damage to or loss of data, programs or equipment, and unavailability or interruption of operations.
6. DISCLAIMER OF LIABILITY
EXCEPT AS EXPRESSLY SET FORTH IN THIS AGREEMENT, NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
7. GENERAL
If any provision of this Agreement is invalid or unenforceable under applicable law, it shall not affect the validity or enforceability of the remainder of the terms of this Agreement, and without further action by the parties hereto, such provision shall be reformed to the minimum extent necessary to make such provision valid and enforceable.
If Recipient institutes patent litigation against a Contributor with respect to a patent applicable to software (including a cross-claim or counterclaim in a lawsuit), then any patent licenses granted by that Contributor to such Recipient under this Agreement shall terminate as of the date such litigation is filed. In addition, if Recipient institutes patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Program itself (excluding combinations of the Program with other software or hardware) infringes such Recipient's patent(s), then such Recipient's rights granted under Section 2(b) shall terminate as of the date such litigation is filed.
All Recipient's rights under this Agreement shall terminate if it fails to comply with any of the material terms or conditions of this Agreement and does not cure such failure in a reasonable period of time after becoming aware of such noncompliance. If all Recipient's rights under this Agreement terminate, Recipient agrees to cease use and distribution of the Program as soon as reasonably practicable. However, Recipient's obligations under this Agreement and any licenses granted by Recipient relating to the Program shall continue and survive.
Everyone is permitted to copy and distribute copies of this Agreement, but in order to avoid inconsistency the Agreement is copyrighted and may only be modified in the following manner. The Agreement Steward reserves the right to publish new versions (including revisions) of this Agreement from time to time. No one other than the Agreement Steward has the right to modify this Agreement. IBM is the initial Agreement Steward. IBM may assign the responsibility to serve as the Agreement Steward to a suitable separate entity. Each new version of the Agreement will be given a distinguishing version number. The Program (including Contributions) may always be distributed subject to the version of the Agreement under which it was received. In addition, after a new version of the Agreement is published, Contributor may elect to distribute the Program (including its Contributions) under the new version. Except as expressly stated in Sections 2(a) and 2(b) above, Recipient receives no rights or licenses to the intellectual property of any Contributor under this Agreement, whether expressly, by implication, estoppel or otherwise. All rights in the Program not expressly granted under this Agreement are reserved.
This Agreement is governed by the laws of the State of New York and the intellectual property laws of the United States of America. No party to this Agreement will bring a legal action under this Agreement more than one year after the cause of action arose. Each party waives its rights to a jury trial in any resulting litigation.