7 using System.Collections.Generic;
17 public class AlterDatasinkRequest : KineticaData
23 public struct DatasinkUpdatesMap
51 public const string S3_REGION =
"s3_region";
74 public const string TRUE =
"true";
75 public const string FALSE =
"false";
182 public const string KAFKA_URL =
"kafka_url";
203 public const string ANONYMOUS =
"anonymous";
237 public const string USE_HTTPS =
"use_https";
271 public const string FLAT =
"flat";
274 public const string NESTED =
"nested";
305 public string name {
get;
set; }
689 public IDictionary<string, string>
datasink_updates_map {
get;
set; } =
new Dictionary<string, string>();
692 public IDictionary<string, string>
options {
get;
set; } =
new Dictionary<string, string>();
1088 IDictionary<string, string>
options)
1090 this.name =
name ??
"";
1092 this.options =
options ??
new Dictionary<string, string>();
1099 public class AlterDatasinkResponse : KineticaData
1102 public IDictionary<string, string>
updated_properties_map {
get;
set; } =
new Dictionary<string, string>();
1105 public IDictionary<string, string>
info {
get;
set; } =
new Dictionary<string, string>();
const string JDBC_DRIVER_JAR_PATH
JDBC driver jar file location.
const string AZURE_SAS_TOKEN
Shared access signature token for Azure storage account to use as the data sink
IDictionary< string, string > updated_properties_map
Map of values updated
const string S3_USE_VIRTUAL_ADDRESSING
Whether to use virtual addressing when referencing the Amazon S3 sink.
const string USE_MANAGED_CREDENTIALS
When no credentials are supplied, we use anonymous access by default.
const string AZURE_OAUTH_TOKEN
Oauth token to access given storage container
const string HDFS_USE_KERBEROS
Use kerberos authentication for the given HDFS cluster.
const string HDFS_KERBEROS_KEYTAB
Kerberos keytab file location for the given HDFS user.
const string DESTINATION
Destination for the output data in format 'destination_type://path[:port]'.
const string WAIT_TIMEOUT
Timeout in seconds for waiting for a response from this sink
AlterDatasinkRequest()
Constructs an AlterDatasinkRequest object with default parameters.
const string S3_REGION
Name of the Amazon S3 region where the given bucket is located
const string S3_ENCRYPTION_TYPE
Server side encryption type
const string KAFKA_TOPIC_NAME
Name of the Kafka topic to use for this data sink, if it references a Kafka broker
const string S3_AWS_ROLE_ARN
Amazon IAM Role ARN which has required S3 permissions that can be assumed for the given S3 IAM user
IDictionary< string, string > info
Additional information.
const string JDBC_DRIVER_CLASS_NAME
Name of the JDBC driver class
string name
Name of the data sink to be altered.
const string SCHEMA_NAME
Updates the schema name.
const string NESTED
Records are returned as an array per message
const string S3_ENCRYPTION_CUSTOMER_KEY
Customer encryption key to encrypt or decrypt data
const string USE_HTTPS
Use https to connect to datasink if true, otherwise use http.
const string JSON_FORMAT
The desired format of JSON encoded notifications message.
const string FLAT
A single record is returned per message
IDictionary< string, string > options
Optional parameters.
const string MAX_MESSAGE_SIZE
Maximum size in bytes of each notification message.
const string AZURE_TENANT_ID
Active Directory tenant ID (or directory ID)
const string GCS_BUCKET_NAME
Name of the Google Cloud Storage bucket to use as the data sink
const string CONNECTION_TIMEOUT
Timeout in seconds for connecting to this sink
const string MAX_BATCH_SIZE
Maximum number of records per notification message.
const string S3_VERIFY_SSL
Whether to verify SSL connections.
const string AZURE_STORAGE_ACCOUNT_NAME
Name of the Azure storage account to use as the data sink, this is valid only if tenant_id is specifi...
const string AZURE_CONTAINER_NAME
Name of the Azure storage container to use as the data sink
const string S3_KMS_KEY_ID
KMS key
const string GCS_SERVICE_ACCOUNT_KEYS
Google Cloud service account keys to use for authenticating the data sink
const string S3_BUCKET_NAME
Name of the Amazon S3 bucket to use as the data sink
const string S3_ENCRYPTION_CUSTOMER_ALGORITHM
Customer encryption algorithm used encrypting data
IDictionary< string, string > datasink_updates_map
Map containing the properties of the data sink to be updated.
const string GCS_PROJECT_ID
Name of the Google Cloud project to use as the data sink
AlterDatasinkRequest(string name, IDictionary< string, string > datasink_updates_map, IDictionary< string, string > options)
Constructs an AlterDatasinkRequest object with the specified parameters.
const string SKIP_VALIDATION
Bypass validation of connection to this data sink.
const string KAFKA_URL
The publicly-accessible full path URL to the kafka broker, e.g., 'http://172.123.45....
const string HDFS_DELEGATION_TOKEN
Delegation token for the given HDFS user
const string CREDENTIAL
Name of the credential object to be used in this data sink
const string ANONYMOUS
Create an anonymous connection to the storage provider–DEPRECATED: this is now the default.