The Kinetica File System (KiFS) is a file system interface that's packaged with Kinetica. It provides a repository for users to store and make use of files within the database.
KiFS can be leveraged by several Kinetica features:
- UDFs (in SQL)
- External tables (in SQL)
- /insert/records/fromfiles API calls (SQL LOAD INTO)
KiFS files can be referenced with the following URI:
kifs://<kifs directory><kifs file>
For example, the following URI can be broken down into three components:
kifs://data/geospatial/flights.csv
Component | Value |
---|---|
Scheme | kifs:// |
Directory | data |
File | /geospatial/flights.csv |
The unique KiFS file name, when referenced in the API, is the composite of the directory and file:
data/geospatial/flights.csv
Configuration
In the default configuration, KiFS is enabled and makes use of the standard Kinetica persistence scheme, distributing files among the cluster nodes.
KiFS can also be configured to use any of the following for file storage:
- Local shared storage, mounted and accessible to every node in the Kinetica cluster
- Azure (Microsoft blob storage)
- HDFS (Apache Hadoop Distributed File System)
- S3 (Amazon S3 Bucket)
To configure KiFS to use one of these other storage types, update the
/opt/gpudb/core/etc/gpudb.conf
configuration file in the
KiFS section with one of the following setups,
and then restart the database.
Note
Remote storage configuration parameters mirror those used for defining cold storage tiers. See Cold Storage Tier in the Configuration Reference for the full set of parameters.
|
|
|
|
|
|
|
|
SQL Interface
KiFS is able to be managed via SQL. See Files & Directories (KiFS) for details.
Files can be uploaded to and downloaded from KiFS using Kinetica SQL (KiSQL) or any JDBC client by using the Kinetica JDBC Driver.
Native API Support
The Kinetica Java API provides a streamlined interface for managing files & directories in KiFS. See KiFS API Support for details.