Kafka Stream Import

Using the Kafka Stream Import wizard, data from a Kafka topic can be streamed into Kinetica.

To stream data from Kafka into Kinetica, click on the Kafka Stream panel on the Import landing page in Workbench.

The Kafka Stream Import wizard appears in the right-hand pane.


Kafka Stream Import Wizard

The wizard has two tabs:

  • Import - contains the 5-step process for streaming data from Kafka
  • History - lists all of your previous Kafka stream attempts

Import Overview

When setting up the stream, the following five steps will need to be completed:

At any step, the following actions are available at the bottom of the screen:

  • Next - proceed to the next step
  • Back - return to the previous step
  • Cancel - exit the stream configuration process
  • Import - once enough information has been specified to establish a stream, the Import button will become active


In this step, the data source used to connect to Kafka is selected.

  • Create New Data Source - click to create a new data source that connects to Kafka, then enter its configuration and click Create to create the data source and proceed to the Source selection page:

    • Name - enter a unique name for the data source

    • Host:Port - enter the host & port of the Kafka service to use

    • Topic Name - enter the name of the Kafka topic from which data will be streamed

    • Auth Type - the wizard supports only SASL SSL connections to Kafka services when creating new data sources; enter the Username and Password for the connection


      To authenticate to Kafka services via other authentication schemes, create a Kafka Credential via SQL (in a Workbook, for example) and then create a Kafka Data Source that uses the credential

  • Select Existing Data Source - click to select an existing data source that connects to Kafka and then click Next to proceed to the Source selection page:

    • Data Source - click to open a drop-down of available data sources that connect to Kafka and select one


In this step, the source data options are configured, if necessary.

  • Source - pre-selected as the name of the data source chosen in the previous step
  • Format - pre-selected as JSON
  • Store Points As X/Y - check the box to store GeoJSON data as separate X, Y, & Z columns, instead of a WKT-based format

Once the source configuration is complete, click Next to proceed to the Destination selection page.


In this step, the target table to import into is selected.

  • Schema - name of the schema containing the target table; if blank, the user's default schema will be used
  • Table - name of the target table, which must meet table naming criteria
  • Abort on Error - check, to have the import stop at the first record import failure; any records imported by this point will remain in the target table

Once the destination has been specified, click Next to proceed to the Configure page.


In this step, the target table's structure can be specified, if the table does not exist. Not specifying any structure will cause the import process to infer the table's structure from the source data.

To specify a table structure, click + Add Column once for each field in the source data, then enter the specification for each column, including:

  • Name - name of the column, which must meet the standard naming criteria
  • Type - type of the column, and sub-type, if applicable
  • Nullable - check the box if the column should allow null values
  • Properties - check any properties that should apply to this column:
    • Primary Key - make this column the primary key or part of a composite primary key
    • Shard Key - make this column the shard key or part of a composite shard key
    • Dict. Encoded - apply dictionary encoding to the column's values, reducing storage used for columns with more often repeated data
    • Text Search - make this column full-text searchable, using FILTER_BY_STRING in search mode
    • Init. with Now - replace empty or invalid values inserted into this column with the current date/time
    • Init. with UUID - replace empty values inserted into this column with a universally unique identifier (UUID)

To remove a column from the proposed target table, click the trash can icon at the far right of the column's definition.

Once the table configuration has been established, click Next to proceed to the Summary page.


In this step, the data stream configuration will be displayed.

All Source & Destination configuration will be displayed in their respective sections.

The Generated SQL section will contain the SQL LOAD INTO command corresponding to the stream operation that will take place. The copy-to-clipboard icon can be used to copy the SQL statement for subsequent use, to re-establish the data streaming into the same table.

Once the data stream configuration has been confirmed, click Import to begin streaming data.