A set of parameters for Kinetica.insertRecordsFromQuery(string,string,IDictionary{string, IDictionary{string, string}},IDictionary{string, string},IDictionary{string, string}).
More...
|
| InsertRecordsFromQueryRequest () |
| Constructs an InsertRecordsFromQueryRequest object with default parameters. More...
|
|
| InsertRecordsFromQueryRequest (string table_name, string remote_query, IDictionary< string, IDictionary< string, string >> modify_columns=null, IDictionary< string, string > create_table_options=null, IDictionary< string, string > options=null) |
| Constructs an InsertRecordsFromQueryRequest object with the specified parameters. More...
|
|
| KineticaData (KineticaType type) |
| Constructor from Kinetica Type More...
|
|
| KineticaData (System.Type type=null) |
| Default constructor, with optional System.Type More...
|
|
object | Get (int fieldPos) |
| Retrieve a specific property from this object More...
|
|
void | Put (int fieldPos, object fieldValue) |
| Write a specific property to this object More...
|
|
|
string | table_name [get, set] |
| Name of the table into which the data will be inserted, in [schema_name. More...
|
|
string | remote_query [get, set] |
| Query for which result data needs to be imported More...
|
|
IDictionary< string,
IDictionary< string, string > > | modify_columns [get, set] |
| Not implemented yet. More...
|
|
IDictionary< string, string > | create_table_options = new Dictionary<string, IDictionary<string, string>>() [get, set] |
| Options used when creating the target table. More...
|
|
IDictionary< string, string > | options = new Dictionary<string, string>() [get, set] |
| Optional parameters. More...
|
|
Schema | Schema [get] |
| Avro Schema for this class More...
|
|
A set of parameters for Kinetica.insertRecordsFromQuery(string,string,IDictionary{string, IDictionary{string, string}},IDictionary{string, string},IDictionary{string, string}).
Computes remote query result and inserts the result data into a new or existing table
Definition at line 20 of file InsertRecordsFromQuery.cs.
kinetica.InsertRecordsFromQueryRequest.InsertRecordsFromQueryRequest |
( |
| ) |
|
|
inline |
kinetica.InsertRecordsFromQueryRequest.InsertRecordsFromQueryRequest |
( |
string |
table_name, |
|
|
string |
remote_query, |
|
|
IDictionary< string, IDictionary< string, string >> |
modify_columns = null , |
|
|
IDictionary< string, string > |
create_table_options = null , |
|
|
IDictionary< string, string > |
options = null |
|
) |
| |
|
inline |
Constructs an InsertRecordsFromQueryRequest object with the specified parameters.
- Parameters
-
table_name | Name of the table into which the data will be inserted, in [schema_name.]table_name format, using standard name resolution rules. If the table does not exist, the table will be created using either an existing type_id or the type inferred from the remote query, and the new table name will have to meet standard table naming criteria. |
remote_query | Query for which result data needs to be imported |
modify_columns | Not implemented yet. The default value is an empty Dictionary. |
create_table_options | Options used when creating the target table.
-
TYPE_ID: ID of a currently registered type. The default value is ''.
-
NO_ERROR_IF_EXISTS: If true, prevents an error from occurring if the table already exists and is of the given type. If a table with the same ID but a different type exists, it is still an error. Supported values:
The default value is FALSE.
-
IS_REPLICATED: Affects the distribution scheme for the table's data. If true and the given type has no explicit shard key defined, the table will be replicated. If false, the table will be sharded according to the shard key specified in the given type_id, or randomly sharded, if no shard key is specified. Note that a type containing a shard key cannot be used to create a replicated table. Supported values:
The default value is FALSE.
-
FOREIGN_KEYS: Semicolon-separated list of foreign keys, of the format '(source_column_name [, ...]) references target_table_name(primary_key_column_name [, ...]) [as foreign_key_name]'.
-
FOREIGN_SHARD_KEY: Foreign shard key of the format 'source_column references shard_by_column from target_table(primary_key_column)'.
-
PARTITION_TYPE: Partitioning scheme to use. Supported values:
-
PARTITION_KEYS: Comma-separated list of partition keys, which are the columns or column expressions by which records will be assigned to partitions defined by partition_definitions.
-
PARTITION_DEFINITIONS: Comma-separated list of partition definitions, whose format depends on the choice of partition_type. See range partitioning, interval partitioning, list partitioning, hash partitioning, or series partitioning for example formats.
-
IS_AUTOMATIC_PARTITION: If true, a new partition will be created for values which don't fall into an existing partition. Currently only supported for list partitions. Supported values:
The default value is FALSE.
-
TTL: Sets the TTL of the table specified in .
-
CHUNK_SIZE: Indicates the number of records per chunk to be used for this table.
-
IS_RESULT_TABLE: Indicates whether the table is a memory-only table. A result table cannot contain columns with store_only or text_search data-handling or that are non-charN strings, and it will not be retained if the server is restarted. Supported values:
The default value is FALSE.
-
STRATEGY_DEFINITION: The tier strategy for the table and its columns.
The default value is an empty Dictionary. |
options | Optional parameters.
-
BAD_RECORD_TABLE_NAME: Optional name of a table to which records that were rejected are written. The bad-record-table has the following columns: line_number (long), line_rejected (string), error_message (string). When error handling is Abort, bad records table is not populated.
-
BAD_RECORD_TABLE_LIMIT: A positive integer indicating the maximum number of records that can be written to the bad-record-table. Default value is 10000
-
BATCH_SIZE: Number of records per batch when inserting data.
-
DATASOURCE_NAME: Name of an existing external data source from which table will be loaded
-
ERROR_HANDLING: Specifies how errors should be handled upon insertion. Supported values:
-
PERMISSIVE: Records with missing columns are populated with nulls if possible; otherwise, the malformed records are skipped.
-
IGNORE_BAD_RECORDS: Malformed records are skipped.
-
ABORT: Stops current insertion and aborts entire operation when an error is encountered. Primary key collisions are considered abortable errors in this mode.
The default value is ABORT.
-
IGNORE_EXISTING_PK: Specifies the record collision error-suppression policy for inserting into a table with a primary key, only used when not in upsert mode (upsert mode is disabled when update_on_existing_pk is false). If set to true, any record being inserted that is rejected for having primary key values that match those of an existing table record will be ignored with no error generated. If false, the rejection of any record for having primary key values matching an existing record will result in an error being reported, as determined by error_handling. If the specified table does not have a primary key or if upsert mode is in effect (update_on_existing_pk is true), then this option has no effect. Supported values:
-
TRUE: Ignore new records whose primary key values collide with those of existing records
-
FALSE: Treat as errors any new records whose primary key values collide with those of existing records
The default value is FALSE.
-
INGESTION_MODE: Whether to do a full load, dry run, or perform a type inference on the source data. Supported values:
-
FULL: Run a type inference on the source data (if needed) and ingest
-
DRY_RUN: Does not load data, but walks through the source data and determines the number of valid records, taking into account the current mode of error_handling.
-
TYPE_INFERENCE_ONLY: Infer the type of the source data and return, without ingesting any data. The inferred type is returned in the response.
The default value is FULL.
-
JDBC_FETCH_SIZE: The JDBC fetch size, which determines how many rows to fetch per round trip.
-
JDBC_SESSION_INIT_STATEMENT: Executes the statement per each jdbc session before doing actual load. The default value is ''.
-
NUM_SPLITS_PER_RANK: Optional: number of splits for reading data per rank. Default will be external_file_reader_num_tasks. The default value is ''.
-
NUM_TASKS_PER_RANK: Optional: number of tasks for reading data per rank. Default will be external_file_reader_num_tasks
-
PRIMARY_KEYS: Optional: comma separated list of column names, to set as primary keys, when not specified in the type. The default value is ''.
-
SHARD_KEYS: Optional: comma separated list of column names, to set as primary keys, when not specified in the type. The default value is ''.
-
SUBSCRIBE: Continuously poll the data source to check for new data and load it into the table. Supported values:
The default value is FALSE.
-
TRUNCATE_TABLE: If set to true, truncates the table specified by prior to loading the data. Supported values:
The default value is FALSE.
-
REMOTE_QUERY: Remote SQL query from which data will be sourced
-
REMOTE_QUERY_ORDER_BY: Name of column to be used for splitting the query into multiple sub-queries using ordering of given column. The default value is ''.
-
REMOTE_QUERY_FILTER_COLUMN: Name of column to be used for splitting the query into multiple sub-queries using the data distribution of given column. The default value is ''.
-
REMOTE_QUERY_INCREASING_COLUMN: Column on subscribed remote query result that will increase for new records (e.g., TIMESTAMP). The default value is ''.
-
REMOTE_QUERY_PARTITION_COLUMN: Alias name for remote_query_filter_column. The default value is ''.
-
TRUNCATE_STRINGS: If set to true, truncate string values that are longer than the column's type size. Supported values:
The default value is FALSE.
-
UPDATE_ON_EXISTING_PK: Specifies the record collision policy for inserting into a table with a primary key. If set to true, any existing table record with primary key values that match those of a record being inserted will be replaced by that new record (the new data will be "upserted"). If set to false, any existing table record with primary key values that match those of a record being inserted will remain unchanged, while the new record will be rejected and the error handled as determined by ignore_existing_pk & error_handling. If the specified table does not have a primary key, then this option has no effect. Supported values:
-
TRUE: Upsert new records when primary keys match existing records
-
FALSE: Reject new records when primary keys match existing records
The default value is FALSE.
The default value is an empty Dictionary. |
Definition at line 2142 of file InsertRecordsFromQuery.cs.
IDictionary<string, string> kinetica.InsertRecordsFromQueryRequest.create_table_options = new Dictionary<string, IDictionary<string, string>>() |
|
getset |
Options used when creating the target table.
-
TYPE_ID: ID of a currently registered type. The default value is ''.
-
NO_ERROR_IF_EXISTS: If true, prevents an error from occurring if the table already exists and is of the given type. If a table with the same ID but a different type exists, it is still an error. Supported values:
The default value is FALSE.
-
IS_REPLICATED: Affects the distribution scheme for the table's data. If true and the given type has no explicit shard key defined, the table will be replicated. If false, the table will be sharded according to the shard key specified in the given type_id, or randomly sharded, if no shard key is specified. Note that a type containing a shard key cannot be used to create a replicated table. Supported values:
The default value is FALSE.
-
FOREIGN_KEYS: Semicolon-separated list of foreign keys, of the format '(source_column_name [, ...]) references target_table_name(primary_key_column_name [, ...]) [as foreign_key_name]'.
-
FOREIGN_SHARD_KEY: Foreign shard key of the format 'source_column references shard_by_column from target_table(primary_key_column)'.
-
PARTITION_TYPE: Partitioning scheme to use. Supported values:
-
PARTITION_KEYS: Comma-separated list of partition keys, which are the columns or column expressions by which records will be assigned to partitions defined by partition_definitions.
-
PARTITION_DEFINITIONS: Comma-separated list of partition definitions, whose format depends on the choice of partition_type. See range partitioning, interval partitioning, list partitioning, hash partitioning, or series partitioning for example formats.
-
IS_AUTOMATIC_PARTITION: If true, a new partition will be created for values which don't fall into an existing partition. Currently only supported for list partitions. Supported values:
The default value is FALSE.
-
TTL: Sets the TTL of the table specified in .
-
CHUNK_SIZE: Indicates the number of records per chunk to be used for this table.
-
IS_RESULT_TABLE: Indicates whether the table is a memory-only table. A result table cannot contain columns with store_only or text_search data-handling or that are non-charN strings, and it will not be retained if the server is restarted. Supported values:
The default value is FALSE.
-
STRATEGY_DEFINITION: The tier strategy for the table and its columns.
The default value is an empty Dictionary.
Definition at line 1295 of file InsertRecordsFromQuery.cs.
IDictionary<string, IDictionary<string, string> > kinetica.InsertRecordsFromQueryRequest.modify_columns |
|
getset |
IDictionary<string, string> kinetica.InsertRecordsFromQueryRequest.options = new Dictionary<string, string>() |
|
getset |
Optional parameters.
-
BAD_RECORD_TABLE_NAME: Optional name of a table to which records that were rejected are written. The bad-record-table has the following columns: line_number (long), line_rejected (string), error_message (string). When error handling is Abort, bad records table is not populated.
-
BAD_RECORD_TABLE_LIMIT: A positive integer indicating the maximum number of records that can be written to the bad-record-table. Default value is 10000
-
BATCH_SIZE: Number of records per batch when inserting data.
-
DATASOURCE_NAME: Name of an existing external data source from which table will be loaded
-
ERROR_HANDLING: Specifies how errors should be handled upon insertion. Supported values:
-
PERMISSIVE: Records with missing columns are populated with nulls if possible; otherwise, the malformed records are skipped.
-
IGNORE_BAD_RECORDS: Malformed records are skipped.
-
ABORT: Stops current insertion and aborts entire operation when an error is encountered. Primary key collisions are considered abortable errors in this mode.
The default value is ABORT.
-
IGNORE_EXISTING_PK: Specifies the record collision error-suppression policy for inserting into a table with a primary key, only used when not in upsert mode (upsert mode is disabled when update_on_existing_pk is false). If set to true, any record being inserted that is rejected for having primary key values that match those of an existing table record will be ignored with no error generated. If false, the rejection of any record for having primary key values matching an existing record will result in an error being reported, as determined by error_handling. If the specified table does not have a primary key or if upsert mode is in effect (update_on_existing_pk is true), then this option has no effect. Supported values:
-
TRUE: Ignore new records whose primary key values collide with those of existing records
-
FALSE: Treat as errors any new records whose primary key values collide with those of existing records
The default value is FALSE.
-
INGESTION_MODE: Whether to do a full load, dry run, or perform a type inference on the source data. Supported values:
-
FULL: Run a type inference on the source data (if needed) and ingest
-
DRY_RUN: Does not load data, but walks through the source data and determines the number of valid records, taking into account the current mode of error_handling.
-
TYPE_INFERENCE_ONLY: Infer the type of the source data and return, without ingesting any data. The inferred type is returned in the response.
The default value is FULL.
-
JDBC_FETCH_SIZE: The JDBC fetch size, which determines how many rows to fetch per round trip.
-
JDBC_SESSION_INIT_STATEMENT: Executes the statement per each jdbc session before doing actual load. The default value is ''.
-
NUM_SPLITS_PER_RANK: Optional: number of splits for reading data per rank. Default will be external_file_reader_num_tasks. The default value is ''.
-
NUM_TASKS_PER_RANK: Optional: number of tasks for reading data per rank. Default will be external_file_reader_num_tasks
-
PRIMARY_KEYS: Optional: comma separated list of column names, to set as primary keys, when not specified in the type. The default value is ''.
-
SHARD_KEYS: Optional: comma separated list of column names, to set as primary keys, when not specified in the type. The default value is ''.
-
SUBSCRIBE: Continuously poll the data source to check for new data and load it into the table. Supported values:
The default value is FALSE.
-
TRUNCATE_TABLE: If set to true, truncates the table specified by prior to loading the data. Supported values:
The default value is FALSE.
-
REMOTE_QUERY: Remote SQL query from which data will be sourced
-
REMOTE_QUERY_ORDER_BY: Name of column to be used for splitting the query into multiple sub-queries using ordering of given column. The default value is ''.
-
REMOTE_QUERY_FILTER_COLUMN: Name of column to be used for splitting the query into multiple sub-queries using the data distribution of given column. The default value is ''.
-
REMOTE_QUERY_INCREASING_COLUMN: Column on subscribed remote query result that will increase for new records (e.g., TIMESTAMP). The default value is ''.
-
REMOTE_QUERY_PARTITION_COLUMN: Alias name for remote_query_filter_column. The default value is ''.
-
TRUNCATE_STRINGS: If set to true, truncate string values that are longer than the column's type size. Supported values:
The default value is FALSE.
-
UPDATE_ON_EXISTING_PK: Specifies the record collision policy for inserting into a table with a primary key. If set to true, any existing table record with primary key values that match those of a record being inserted will be replaced by that new record (the new data will be "upserted"). If set to false, any existing table record with primary key values that match those of a record being inserted will remain unchanged, while the new record will be rejected and the error handled as determined by ignore_existing_pk & error_handling. If the specified table does not have a primary key, then this option has no effect. Supported values:
-
TRUE: Upsert new records when primary keys match existing records
-
FALSE: Reject new records when primary keys match existing records
The default value is FALSE.
The default value is an empty Dictionary.
Definition at line 1597 of file InsertRecordsFromQuery.cs.
string kinetica.InsertRecordsFromQueryRequest.remote_query |
|
getset |
string kinetica.InsertRecordsFromQueryRequest.table_name |
|
getset |
Name of the table into which the data will be inserted, in [schema_name.
]table_name format, using standard name resolution rules. If the table does not exist, the table will be created using either an existing type_id or the type inferred from the remote query, and the new table name will have to meet standard table naming criteria.
Definition at line 1067 of file InsertRecordsFromQuery.cs.
The documentation for this class was generated from the following file: