Note

This documentation is for a prior release of Kinetica. For the latest documentation, click here.

Alter Data Source

Alters the properties of an existing data source

Input Parameter Description

NameTypeDescription
namestringName of the data source to be altered. Must be an existing data source.
datasource_updates_mapmap of string to strings

Map containing the properties of the data source to be updated. Error if empty.

Supported Parameters (keys)Parameter Description
locationLocation of the remote storage in 'storage_provider_type://[storage_path[:storage_port]]' format. Supported storage provider types are 'azure','gcs','hdfs','kafka' and 's3'.
user_nameName of the remote system user; may be an empty string
passwordPassword for the remote system user; may be an empty string
skip_validation

Bypass validation of connection to remote source. The default value is false. The supported values are:

  • true
  • false
connection_timeoutTimeout in seconds for connecting to this storage provider
wait_timeoutTimeout in seconds for reading from this storage provider
credentialName of the credential object to be used in data source
s3_bucket_nameName of the Amazon S3 bucket to use as the data source
s3_regionName of the Amazon S3 region where the given bucket is located
s3_aws_role_arnAmazon IAM Role ARN which has required S3 permissions that can be assumed for the given S3 IAM user
s3_encryption_customer_algorithmCustomer encryption algorithm used encrypting data
s3_encryption_customer_keyCustomer encryption key to encrypt or decrypt data
hdfs_kerberos_keytabKerberos keytab file location for the given HDFS user. This may be a KIFS file.
hdfs_delegation_tokenDelegation token for the given HDFS user
hdfs_use_kerberos

Use kerberos authentication for the given HDFS cluster The default value is false. The supported values are:

  • true
  • false
azure_storage_account_nameName of the Azure storage account to use as the data source, this is valid only if tenant_id is specified
azure_container_nameName of the Azure storage container to use as the data source
azure_tenant_idActive Directory tenant ID (or directory ID)
azure_sas_tokenShared access signature token for Azure storage account to use as the data source
azure_oauth_tokenOAuth token to access given storage container
gcs_bucket_nameName of the Google Cloud Storage bucket to use as the data source
gcs_project_idName of the Google Cloud project to use as the data source
gcs_service_account_keysGoogle Cloud service account keys to use for authenticating the data source
kafka_urlThe publicly-accessible full path URL to the Kafka broker, e.g., 'http://172.123.45.67:9300'.
kafka_topic_nameName of the Kafka topic to use as the data source
jdbc_driver_jar_pathJDBC driver jar file location. This may be a KIFS file.
jdbc_driver_class_nameName of the JDBC driver class
anonymous

Create an anonymous connection to the storage provider--DEPRECATED: this is now the default. Specify use_managed_credentials for non-anonymous connection The default value is true. The supported values are:

  • true
  • false
use_managed_credentials

When no credentials are supplied, we use anonymous access by default. If this is set, we will use cloud provider user settings. The default value is false. The supported values are:

  • true
  • false
use_https

Use https to connect to datasource if true, otherwise use http The default value is true. The supported values are:

  • true
  • false
schema_nameUpdates the schema name. If schema_name doesn't exist, an error will be thrown. If schema_name is empty, then the user's default schema will be used.
optionsmap of string to stringsOptional parameters.

Output Parameter Description

NameTypeDescription
updated_properties_mapmap of string to stringsMap of values updated
infomap of string to stringsAdditional information.