From 523727f930673e3c838b9fb451bcbb6c8eff7b93 Mon Sep 17 00:00:00 2001 From: Vinish Reddy Date: Thu, 1 Aug 2024 18:07:25 +0530 Subject: [PATCH] Fix README --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index a150af49..54085408 100644 --- a/README.md +++ b/README.md @@ -110,14 +110,14 @@ catalogOptions: # all other options are passed through in a map key1: value1 key2: value2 ``` -5. run with `java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml [--hadoopConfig hdfs-site.xml] [--convertersConfig converters.yaml] [--icebergCatalogConfig catalog.yaml]` +5. run with `java -jar incubator-xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml [--hadoopConfig hdfs-site.xml] [--convertersConfig converters.yaml] [--icebergCatalogConfig catalog.yaml]` The bundled jar includes hadoop dependencies for AWS, Azure, and GCP. Sample hadoop configurations for configuring the converters can be found in the [xtable-hadoop-defaults.xml](https://github.com/apache/incubator-xtable/blob/main/utilities/src/main/resources/xtable-hadoop-defaults.xml) file. The custom hadoop configurations can be passed in with the `--hadoopConfig [custom-hadoop-config-file]` option. The config in custom hadoop config file will override the default hadoop configurations. For an example of a custom hadoop config file, see [hadoop.xml](https://xtable.apache.org/docs/fabric#step-2-translate-source-table-to-delta-lake-format-using-apache-xtable-incubating). -# Contributing +# Contributingg ## Setup For setting up the repo on IntelliJ, open the project and change the java version to Java11 in File->ProjectStructure ![img.png](style/IDE.png)