Note

This documentation is for a prior release of Kinetica. For the latest documentation, click here.

Load Data from HDFS

Copy and paste tutorial for loading data from HDFS

Loading data from HDFS can be done in three easy steps:

  1. Create a credential - holds HDFS account authentication information
  2. Create a data source - holds HDFS connection information; uses the credential for authentication
  3. Ingest the data - Load data from HDFS; uses the data source to identify the HDFS source

Create Credential

To store the access details for your data source, first create a credential with the CREATE CREDENTIAL command, and reference it in your CREATE DATA SOURCE statement. This will allow you to store the authentication details of your connection separately.

Password
1
2
3
4
CREATE CREDENTIAL hdfs_cred
TYPE = 'hdfs',
IDENTITY = 'jdoe',
SECRET = 'foobaz123'
Kerberos Keytab
1
2
3
4
5
6
7
CREATE CREDENTIAL hdfs_cred
TYPE = 'hdfs',
IDENTITY = 'jdoe'
WITH OPTIONS
(
    KERBEROS KEYTAB = 'kifs://<keytab file path>'
)
Kerberos Token
1
2
3
4
5
6
7
CREATE CREDENTIAL hdfs_cred
TYPE = 'hdfs',
IDENTITY = 'jdoe'
WITH OPTIONS
(
    USE KERBEROS = 'true'
)

Create Data Source

Next, create a data source using the CREATE DATA SOURCE command in Kinetica. The data source defines how Kinetica connects to HDFS.

Credential
1
2
3
4
5
6
CREATE DATA SOURCE hdfs_ds
LOCATION = 'HDFS://example.com:8020'
WITH OPTIONS
(
    CREDENTIAL = 'hdfs_cred'
)

Ingest Data

To initiate the ingest of data into a Kinetica table, use the LOAD INTO command. You can initiate a one-time batch load, or a continuous load using change-data-capture by using the option SUBSCRIBE = TRUE.

Batch Load
1
2
3
4
LOAD DATA INTO example.orders
FROM FILE PATHS 'example_orders.csv'
FORMAT TEXT
WITH OPTIONS (DATA SOURCE = 'hdfs_ds')
Streaming Load
1
2
3
4
5
6
7
LOAD DATA INTO example.orders
FROM FILE PATHS 'example_orders.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    SUBSCRIBE = TRUE
)

Considerations

The following are some common options used when loading. For the full list of options, see LOAD INTO. For copy/paste examples of many of the options, see Loading Data.


Wildcard Ingest

If you want to ingest multiple files inside of an HDFS directory, a "path prefix" can be specified, which will cause all files whose path begins with the given prefix to be included. When using an HDFS data source, the "path prefix" must be the name of an HDFS directory.

For example, a "path prefix" of /data/ for FROM FILE PATHS would match all of the following:

  • /data/geo_0.csv
  • /data/geo_1.csv
  • /data/geo_2.csv

Error Handling

Kinetica has two different error handling modes for dealing with erroneous data. To halt ingestion after a bad record is found, use the ABORT mode. To skip erroneous records and continue the ingest, use the SKIP mode.

To inspect erroneous records, you may use the BAD RECORD TABLE NAME option. All bad records encountered will be stored there for review. The bad records table is limited to 10,000 records by default and may be overridden using the BAD RECORD TABLE LIMIT option.

Abort
1
2
3
4
5
6
7
LOAD INTO ki_home.error_example
FROM FILE PATHS 'error_example.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    ON ERROR = ABORT
)
Log Bad Records
1
2
3
4
5
6
7
8
LOAD INTO ki_home.error_example
FROM FILE PATHS 'error_example.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    BAD RECORD TABLE NAME = 'error_example_invalid',
    ON ERROR = SKIP
)

Load Specific Columns

In some cases, you may only want to store certain columns from your source data. Use the FIELDS MAPPED BY NAME(...) option, which allows you to specify the desired fields to store in Kinetica.

Load Specific Columns Example
1
2
3
4
5
6
7
LOAD DATA INTO example.orders
FROM FILE PATHS 'example_orders.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    FIELDS MAPPED BY NAME(ID, Name, Product_ID, Quantity)
)

DateTime Formatting

Use the COLUMN FORMATS option to format date and time fields into Kinetica Date, Time, and DateTime columns. Time formats are specified using a JSON formatted string. Non-placeholder characters must be wrapped in quotes, which must also be escaped (e.g. '{"dt": {"date": "\"(\"YYYY\")\" Mon, DD"}}'). Alternatively, you can use the ? character as a wildcard character. Note that Kinetica does not handle or store timezones and they will be discarded. See the full list of supported date and time format codes.

ISO 8601 Timestamps
 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
-- Example Data:
-- dt
-- 2022-01-19T15:50:42Z+05:00

LOAD INTO ki_home.timestamp_example
FROM FILE PATHS 'timestamp_example.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    COLUMN FORMATS = '
    {
        "dt": {"datetime": "YYYY-MM-DD?HH:MI:SS"}
    }'
)
Custom Date
 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
-- Example Data:
-- d
-- (2022) Feb, 22

LOAD INTO ki_home.date_example
FROM FILE PATHS 'date_example.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    COLUMN FORMATS = '{"d": {"date": "\"(\"YYYY\")\" Mon, DD"}}'
)
Custom Time
 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
-- Example Data:
-- t
-- 18-27-59.5536

LOAD INTO ki_home.time_example
FROM FILE PATHS 'time_example.csv'
FORMAT TEXT
WITH OPTIONS (
    DATA SOURCE = 'hdfs_ds',
    COLUMN FORMATS = '{"t": {"time": "HH-MI-SS.MS"}}'
)