HDFS directories

Feedback


You can use csv data and index files stored in HDFS for distributed analysis.

Register an HDFS directory

Steps:

  1. On the iServer service management page (http://supermapiserver:8090/iserver/manager), click on "Cluster">"Data Registration" to enter data registration page;
  2. Click "Register data storage" to register a data store;
  3. Input "Storage ID", select "Big data file share" in "Data storage type" dropdown arrow, and choose "HDFS directory" in "Shared data type" dropdown arrow;
  4. When configuring the "HDFS directory", you can do the following operations:
  5. If the HDFS directory you're registering enables Kerberos authentication, you need to check "HDFS cluster has Kerberos authentication turned on", and configure the following items:
      1. principle name: username@domain. For example: iServer@SUPERMAP1.COM
        • username: Must be a system user of the HDFS cluster mater node.
        • domain: The domain set by the Kerberos service. (Must be the same as the domain where the HDFS cluster locates)
      2. principle keytab path: The file location on the machine where iServer service locates to which the keytab file is to be copied, which is generated when building the cluster.
      3. config file path: The Kerberos client config file location on the machine where iServer service locates.
  1. Click "Register data Storage" button to complete the registration.

        Now you can start to perform distributed analysis services.

        Note:

1. If you are registering csv files, the data needs to be verified before it can be used for distributed analysis service. For details, see: csv data verification.

2. If you are registering an HDFS directory with Kerberos authentication enabled, it'll be available for distributed analysis only when the distributed analysis service uses the Hadoop Yarn cluster with Kerberos authentication enabled.