Leverage HDFS as Deep Storage for Real-Time Tables

Configure Hadoop Filesystem Settings

Ensure that the core-site.xml and hdfs-site.xml files are correctly configured and accessible. These files define the Hadoop filesystem settings, including the default filesystem URI and any necessary authentication configurations.

Bash
Copy

Update Pinot Controller Configuration

Modify the controller.conf file to include the following properties.

Bash
Copy
Bash
Copy

Update Pinot Server Configuration

Modify the server.conf file to include the following properties.

Bash
Copy

Restart HDFS and Pinot

Create real-time table for KAFKA- Define Pinot Schema.

Create a file /tmp/pinot/schema-stream.json.

Bash
Copy

Define Pinot Table Config

Create a file /tmp/pinot/table-config-stream.json.

Update the following in table-config-stream.json:

  1. Table and Schema Names.
  2. stream.kafka.broker.listbased on your broker list.
Bash
Copy

Create Pinot Schema and Table

Run the following command:

Bash
Copy

Verify the Segments on HDFS Path

The below image shows the segments of the HDFS path.

Type to search, ESC to discard
Type to search, ESC to discard
Type to search, ESC to discard
  Last updated