Skip to content

Dkbei/paimon-trino

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

41 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Paimon Trino

Because Trino's dependency is JDK 11, it is not possible to include the trino connector in Paimon.

Deploy Paimon Trino Connector

Building Paimon Trino Bundled Jar is by running:

  • Trino 388: mvn clean install -DskipTests , (JDK 11 required).
  • Trino 358: mvn clean install -DskipTests -Ptrino-358 , (JDK 11 required).
  • Trino 391: mvn clean install -DskipTests -Ptrino-391 , (JDK 17 required).
  • Trino 391+: You can change target.java.version to 17, and trino.version to trino version, and mvn clean install -DskipTests. Use JDK 17 and add jvm options like trino 391.

NOTE: For JDK 17, when Deploying Trino, should add jvm options: --add-opens=java.base/sun.nio.ch=ALL-UNNAMED --add-opens=java.base/java.nio=ALL-UNNAMED

Then, copy target/paimon-trino-*.jar and flink-shaded-hadoop-2-uber-2.8.3-10.0.jar to plugin/tablestore.

Configure Paimon Catalog

Catalogs are registered by creating a catalog properties file in the etc/catalog directory. For example, create etc/catalog/tablestore.properties with the following contents to mount the tablestore connector as the tablestore catalog:

connector.name=tablestore
warehouse=file:/tmp/warehouse

If you are using HDFS, choose one of the following ways to configure your HDFS:

  • set environment variable HADOOP_HOME.
  • set environment variable HADOOP_CONF_DIR.
  • configure fs.hdfs.hadoopconf in the properties.

Query

SELECT * FROM tablestore.default.MyTable

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Java 100.0%