Please ensure Javascript is enabled for purposes of website accessibility
Powered by Zoomin Software. For more details please contactZoomin

PI Integrator for Business Analytics

Configure the Apache Hive target

  • Last UpdatedMay 21, 2024
  • 2 minute read

Complete the procedure below to configure the Apache Hive target and test your connection to the Hive database.

Create the Apache Hive target before you configure it. For more information, see Add a publish target.

Tip: Enter the information in the fields in order. After the required information has been specified in the fields, the buttons are enabled. Click the button before continuing to the next field.

  1. On the Administration page, click the Targets tab. Then select the target in the Publish Targets list.

    Note: If you have already selected your target, skip to the next step.

  2. Refer to the following table and enter the required information. Click the buttons as they become enabled.

    Parameter

    Description

    HDFS URL

    URL to access HDFS through WebHDFS directly, or through the Apache Knox gateway or HttpFS.

    Examples:

    HDFS – http://host:50070/webhdfs/v1

    HttpFS – http://host:14000/webhdfs/v1

    Apache Knox – https://host:8443/gateway/default/webhdfs/v1

    Verify TLS/SSL Certificate

    Select the check box if you are using an TLS/SSL certificate.

    Username

    Username used to connect to HDFS. The user must have permissions to read and write to HDFS.

    Password

    (Optional) Password to authenticate the user or when connecting to HttpFS. If a password is provided, then Hadoop HTTP authentication is used. If no password is entered, then PI Integrator for Business Analytics uses Hadoop simple authentication.

    Directory

    Directory where the data files are created. Click Browse to navigate to the directory where you want the files located. The format for specifying the directory location is: /rootfolder/folder1/folder2.

    Hive Hostname

    Host name of the machine on which Thrift is running. This name can be an alias, internal address, or IP address. The default is host name in the HDFS URL.

    Hive Port

    (Optional) Port through which to connect to Thrift.

    Hive Username

    (Optional) Username used to connect to HCatalog.

    Hive Table Format

    Format in which to store tables in Apache Hive. The current supported file formats are text (TEXTFILE) and optimized row columnar (ORC).

    Note: ORC requires Apache Hive 0.11 or later.

    Hive Database

    Name of database in Apache Hive in which the tables are created.

  3. Click Connect to Apache Hive.

    PI Integrator for Business Analytics connects to Apache Hive, queries the database, and populates the Hive Table Format and Hive Database fields with the supported table formats and databases.

  4. Click Verify Apache Hive Writer to verify that Apache Hive writer can write to the Hive database.

    Note: If you are unable to successfully write to the Hive database, refer to Tips for setting up Apache Hive.

  5. Click Save Changes.

TitleResults for “How to create a CRG?”Also Available in