Hadoop Distributed File System
Hadoop Distributed File System (HDFS) is a file system that manages large data sets that can run on commodity hardware.
If you select Hadoop Distributed File System (HDFS) from the Storage section, configure the following details:
Field | Description |
---|---|
Display name | Enter the name to be displayed. |
Thrift URI | Enter the Thrift URI. |
Thrift Port | Enter the Thrift port. |
Kerberos authentication | Use the toggle switch to enable or disable Kerberos authentication. If enabled, enter the following information: a. Enter the following information: i. HDFS principal ii. Hive client principal iii. Hive server principal b. Upload the following files: i. Core site file (.xml) ii. HDFS site (.xml) iii. Kerberos config file (.config) iiii. HDFS keytab file (.keytab) iiiii. Hive keytab file (.keytab) |
Upload core site file (.xml) | Upload core site file (.xml) |
Upload HDFS site file (.xml) | Upload HDFS site file (.xml) |
Associate catalog | Add a catalog for your storage. This catalog is associated with your storage and serves as your query interface with the data stored within. |
Catalog type | The supported catalog is Apache Hive. |
Catalog name | Enter the name of your catalog. |
Associate | Click Associate to create the storage. |
Related API
For information on related API, see