Multi-Head Ingest

Multi-Head Ingest (Distributed Ingest) is a mechanism that allows sharded data to be ingested directly into cluster nodes, bypassing the overhead of pushing the data through the head node to the cluster nodes.

Operationally, the ingest mechanism calculates the target shard of each record to insert, and sends batches of would-be co-located records to their respective target nodes.

To fully utilize multi-head ingest, the ingestion process should use a multi-node parallel processing framework, such as MapReduce, Storm, or Spark. As data is divided up and flows through the processing nodes, a Kinetica bulk ingest object can be instantiated in each node to push data to the proper Kinetica cluster node. This greatly increases the speed of ingestion by both parallelizing the ingestion process and spreading the network traffic across multiple nodes.

API Support

Language Multi-head Ingest Mechanism
C++ GPUdbIngestor
C# KineticaIngestor
Java BulkInserter
Javascript X
Node.js X
Python GPUdbIngestor
REST X
SQL <automatic>

Configuration

In order for the cluster nodes to receive data directly from an ingest client, the configuration on each node needs to be updated to allow the incoming HTTP requests that carry the data. The /opt/gpudb/core/etc/gpudb.conf file needs to have the following property set for multi-head ingest to work properly:

# Enable worker HTTP servers...
enable_worker_http_servers = true

Both the HA and HTTPD configurations will be taken into account, as well as any public URLs (which override all other settings) defined in /opt/gpudb/core/etc/gpudb.conf when returning multi-head URLs for the worker nodes. With the exception of the Python background multi-head process, the multi-head ingest object requires a list of worker nodes to use to distribute the insert requests, with one entry in the list for each node/process. This list can be auto-populated simply by using a GPUdb connection object, which can retrieve the list of available cluster nodes from the database itself. Below is a code snippet showing an automatically populated worker list:

Java
1
2
3
4
5
import com.gpudb.GPUdb;
import com.gpudb.WorkerList;

GPUdb gpudb = new GPUdb("http://localhost:9191");
WorkerList workers = new WorkerList(gpudb);
Python
1
2
3
4
5
6
from gpudb import GPUdb
from gpudb_multihead_io import GPUdbWorkerList

gpudb = GPUdb("http://localhost:9191");
workers = GPUdbWorkerList(gpudb);
urls = workers.get_worker_urls()

Note that in some cases, workers may be configured to use more than one IP address, not all of which may be accessible to the client; the worker list constructor uses the first IP returned by the server for each worker. In cases where workers may use more than one IP address and public URLs are not configured, a regular expression Pattern or prefix String can be used to match the correct worker IP addresses:

Java
1
2
3
4
//Match 172.X.Y.Z addresses
WorkerList(gpudb, Pattern.compile("172\\..*"));
//or
WorkerList(gpudb, "172.");
Python
1
GPUdbWorkerList(gpudb, "172.");

Considerations

There are several factors to consider when using multi-head ingest:

  • Replicated tables are not supported
  • There is a small performance penalty for calculating the shard key of each record to be inserted into a sharded table.
  • There is an additional per-record performance penalty for primary key collision checks for any target table that has a primary key that is not the same as its shard key (the shard key columns are a proper subset of the primary key columns)
  • Randomly-sharded tables benefit more from multi-head ingest than sharded tables do, as there is no target shard key to calculate for each record and, therefore, no associated performance penalty for that calculation.
  • The batch size used to configure the bulk ingest object determines the record threshold for each of the insert queues targeting the worker nodes, not the record threshold for the bulk ingest object itself. Thus, ingestion into tables with non-uniform record distribution may require periodic flushes of the bulk ingest queues to ensure timely insertion of records in queues that infrequently reach their threshold.

Example

All the functionality for multi-head ingestion is encapsulated in the bulk ingest object. See API Support for chart listing the API-specific object to use for bulk ingestion.

The following is a Java API code block that demonstrates the use of the BulkInserter for ingesting data.

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
// Create a bulk inserter for batch data ingestion, using a
//   try-with-resources to invoke auto closing of the bulk inserter object.
try (
    BulkInserter<MyType> bulkInserter =
        new BulkInserter<MyType>(gpudb, tableName, type, batchSize, null, new WorkerList(gpudb))
)
{
    // Generate data to be inserted into the table, automatically inserting
    // records whenever the batchSize limit is reached
    for (int i = 0; i < numRecords; i++)
    {
        MyType record = new MyType();
        record.put( 0, (i + 0.1) ); // col1
        record.put( 1, ("string " + String.valueOf( i ) ) ); // col2
        record.put( 2, "Group 1" );  // group_id
        bulkInserter.insert( record );
    }

    // To ensure all records are inserted, flush the bulk inserter object.
    bulkInserter.flush();
}

Note

If the BulkInserter is not declared as the resource in a try-with-resources block, the close() method needs to be called explicitly after use to release its resources.

However, for a one-time use, as in this example, the flush() call can be removed, as the close() method will automatically call flush().

Python Background Multihead

In the Python API, the GPUdbTable object has a built-in use_multihead_io parameter, which allows the GPUdbTable to handle all Ingestor interactions with the associated table in the background:

1
ingest_table = gpudb.GPUdbTable(_type=ingest_columns, name="ingest_table", db=h_db, use_multihead_io=True)

Cautions

If using the Java API and MapReduce, there is a conflict between the version of Avro used by Kinetica and the one used by MapReduce. This conflict can be overcome by using the Maven shade plug-in with the relocation tag:

1
2
3
4
<relocation>
  <pattern>org.apache.avro</pattern>
  <shadedPattern>org.shaded.apache.avro</shadedPattern>
</relocation>