Connect Integrations

ClickHouse integrates seamlessly with a wide range of data sources and tools for ingestion, processing, and visualization. The following table lists supported integrations along with reference documentation for setup and usage.

For more information about ClickHouse Integrations, see Integrations.

ODP Ingestion

HDFS Integration

ClickHouse supports reading from and writing to HDFS in parallel. However, the following are not supported:

  • ALTER and SELECT ... SAMPLE queries
  • Indexes
  • Zero-copy replication (technically possible, but not recommended)

Example: Create and Use HDFS Engine Table

  1. Create the table
Bash
Copy
  1. Insert data
Bash
Copy
  1. Query the data
Bash
Copy

Example output:

Bash
Copy
  1. Verify file content on HDFS
Bash
Copy

Expected output:

Bash
Copy

HDFS Engine Configuration

The HDFS engine supports extended configuration through the ClickHouse config.xml file. You can use two configuration scopes:

  • Global scope (hdfs): Applies default settings across all users.
  • User-level scope (hdfs_*): Overrides the global settings for individual users.

ClickHouse first applies the global configuration, followed by user-level settings (if defined).

Bash
Copy

HDFS NameNode High Availability (HA) Support

ClickHouse supports HDFS NameNode HA through libhdfs3.

To enable HDFS Namenode HA in ClickHouse:

  1. Copy **`hdfs-site.xml`** from an HDFS node to /etc/clickhouse-server/.
  2. Update **`config.xml`** in ClickHouse:
Bash
Copy
  1. Use the value of the dfs.nameservices tag from hdfs-site.xml as the namenode address in HDFS URIs.

Example:

Replace:

Bash
Copy

With:

Bash
Copy

This configuration enables ClickHouse to interact with an HA-enabled HDFS setup using logical names instead of direct NameNode IPs.

Kafka Integration

  • Publish or subscribe to data flows.
  • Organize fault-tolerant storage.
  • Process streams as they become available.

Create a Kafka Table

Bash
Copy

Create Storage Table

Bash
Copy

Create a Materialized View to write into Kafka ClickHouse table

Bash
Copy

Create Kafka Pull Table

Bash
Copy

Create a Materialized View to pull Read from the Kafka ClickHouse table

Bash
Copy

Insert Records

Bash
Copy

Read from ClickHouse MAT View

Bash
Copy

Check Published Messages on Kafka Console

Bash
Copy

Hive Integration

Hive Engine – Supported Input Formats

The Hive engine in ClickHouse enables SELECT queries on Hive tables stored in HDFS. It supports the following input formats:

  • Text: Supports only simple scalar types (excluding binary).
  • ORC: Supports simple scalar types (excluding char) and complex array types.
  • Parquet: Supports all simple scalar types and array complex types.
Bash
Copy

Reported Ongoing Issue

A null pointer dereference error has been reported when accessing the unsupported hive table function in ClickHouse. Refer to Issue #82982 on GitHub for updates and investigation status.

Spark 3 Integration

Maven dependency and resolvers for ClickHouse.

Bash
Copy

Example Spark Program

Bash
Copy

NiFi Integration

Below is an implementation example.

Prepare a workflow using Execute SQL

Update Execute SQL

Update and Enable DBCP Pool Service

View Record in NiFi

S3 Integration

Below is an implementation example.

Bash
Copy

For more information, see: S3 Table Engine | ClickHouse Docs Also refer to: Separation of Storage and Compute | ClickHouse Docs

JDBC SQL Clients Integration

To connect with DB Visualizer, follow the below steps:

  • Add nexus resolver URL in Tools → Tool Properties.
Bash
Copy
  • Navigate to Tools → Driver Manager. If the ClickHouse template driver is unavailable, update DBVisualizer to the latest version. Right-click ClickHouse and select Create User Driver from Template.
  • In the Driver Artifacts section, click + and enter the required version details. After applying, the tool automatically downloads and scans the driver classes. You can also download and upload the driver JARs manually if preferred.
  • Choose a new DB connection and assign the driver with the DB connection details

For more information, see JDBC Driver Configuration.

ODBC SQL Client Integration

The ODBC driver is currently not packaged due to ongoing evaluations of environmental requirements. However, you can use the official ClickHouse ODBC clients for your operating system. Download the compressed files from the official ClickHouse ODBC repository to access the latest releases (licensed under Apache 2.0).

Note: Future roadmap includes JDBC and ODBC bridge integrations to enable access to external data sources via ClickHouse storage handler services.

Type to search, ESC to discard
Type to search, ESC to discard
Type to search, ESC to discard
  Last updated