Skip to content

Latest commit

 

History

History
40 lines (27 loc) · 1.69 KB

README.md

File metadata and controls

40 lines (27 loc) · 1.69 KB

Paimon Trino

Because Trino's dependency is JDK 11, it is not possible to include the trino connector in Paimon.

Deploy Paimon Trino Connector

Building Paimon Trino Bundled Jar is by running:

  • Trino 388: mvn clean install -DskipTests , (JDK 11 required).
  • Trino 358: mvn clean install -DskipTests -Ptrino-358 , (JDK 11 required).
  • Trino 391: mvn clean install -DskipTests -Ptrino-391 , (JDK 17 required).
  • Trino 391+: You can change target.java.version to 17, and trino.version to trino version, and mvn clean install -DskipTests. Use JDK 17 and add jvm options like trino 391.

NOTE: For JDK 17, when Deploying Trino, should add jvm options: --add-opens=java.base/sun.nio.ch=ALL-UNNAMED --add-opens=java.base/java.nio=ALL-UNNAMED

Then, copy target/paimon-trino-*.jar and flink-shaded-hadoop-2-uber-2.8.3-10.0.jar to plugin/tablestore.

Configure Paimon Catalog

Catalogs are registered by creating a catalog properties file in the etc/catalog directory. For example, create etc/catalog/tablestore.properties with the following contents to mount the tablestore connector as the tablestore catalog:

connector.name=tablestore
warehouse=file:/tmp/warehouse

If you are using HDFS, choose one of the following ways to configure your HDFS:

  • set environment variable HADOOP_HOME.
  • set environment variable HADOOP_CONF_DIR.
  • configure fs.hdfs.hadoopconf in the properties.

Query

SELECT * FROM tablestore.default.MyTable