From 242309f0a848cbf9f5d36762f558dc406d0b56b1 Mon Sep 17 00:00:00 2001 From: albertshau Date: Fri, 2 Nov 2018 16:57:39 -0700 Subject: [PATCH] CDAP-14442 removed parquet, and orc dependencies Removed parquet and orc dependencies, as the formats have been moved to plugins. Did not remove avro since it is used by the BigQuery plugins. Removed exports for avro, parquet, and orc. Avro is now a private dependency, used only within the source and not required by the program. Also fixed a NPE that would happen if the big query table does not exist when hitting the get schema button. --- pom.xml | 49 ++----------------- .../co/cask/gcp/bigquery/BigQuerySource.java | 10 ++-- .../co/cask/gcp/spanner/sink/SpannerSink.java | 2 +- 3 files changed, 10 insertions(+), 51 deletions(-) diff --git a/pom.xml b/pom.xml index 0f4b9f9c50..859fff36f3 100644 --- a/pom.xml +++ b/pom.xml @@ -57,7 +57,7 @@ UTF-8 5.1.0 - 2.1.0 + 2.1.1-SNAPSHOT 20.0 1.7.5 4.12 @@ -67,11 +67,9 @@ 0.53.0-beta 1.6.1 0.54.0-beta - 1.8.1 3.4.0 1.36.0 - 1.1.0 - 1.7.7 + 1.7.7 2.8.11.1 @@ -97,7 +95,7 @@ org.apache.avro avro - 1.7.7 + ${avro.version} co.cask.cdap @@ -265,17 +263,6 @@ google-cloud-speech ${google.cloud.speech.version} - - org.apache.parquet - parquet-avro - ${parquet.avro.version} - - - org.apache.avro - avro - - - com.google.protobuf protobuf-java @@ -286,34 +273,11 @@ google-cloud-pubsub ${google.cloud.pubsub.version} - - org.apache.orc - orc-mapreduce - ${orc.mapreduce.version} - - - org.apache.hadoop - hadoop-hdfs - - - org.apache.hadoop - hadoop-common - - - org.apache.hadoop - hadoop-mapreduce-client-core - - - org.apache.hive - hive-storage-api - - - org.apache.avro avro-mapred hadoop2 - ${avro.mapred.version} + ${avro.version} com.fasterxml.jackson.core @@ -524,12 +488,7 @@ <_exportcontents> co.cask.gcp.*; - co.cask.hydrator.format.*; com.google.cloud.hadoop.*; - org.apache.avro.mapred.*; - org.apache.avro.mapreduce; - org.apache.parquet.avro.*; - org.apache.parquet.hadoop.*; org.apache.spark.streaming.pubsub*; diff --git a/src/main/java/co/cask/gcp/bigquery/BigQuerySource.java b/src/main/java/co/cask/gcp/bigquery/BigQuerySource.java index ba2fd2afe3..3b7cf1d593 100644 --- a/src/main/java/co/cask/gcp/bigquery/BigQuerySource.java +++ b/src/main/java/co/cask/gcp/bigquery/BigQuerySource.java @@ -165,16 +165,16 @@ public void onRunFinish(boolean succeeded, BatchSourceContext context) { @Path("getSchema") public Schema getSchema(BigQuerySourceConfig request) throws Exception { String dataset = request.getDataset(); - String table = request.getTable(); + String tableName = request.getTable(); String project = request.getDatasetProject(); - Table bqTable = BigQueryUtils.getBigQueryTable(request.getServiceAccountFilePath(), project, dataset, table); + Table table = BigQueryUtils.getBigQueryTable(request.getServiceAccountFilePath(), project, dataset, tableName); if (table == null) { // Table does not exist throw new IllegalArgumentException(String.format("BigQuery table '%s:%s.%s' does not exist", - project, dataset, table)); + project, dataset, tableName)); } - com.google.cloud.bigquery.Schema bgSchema = bqTable.getDefinition().getSchema(); + com.google.cloud.bigquery.Schema bgSchema = table.getDefinition().getSchema(); if (bgSchema == null) { throw new IllegalArgumentException(String.format("Cannot read from table '%s:%s.%s' because it has no schema.", project, dataset, table)); @@ -195,7 +195,7 @@ private void validateOutputSchema() throws IOException { if (table == null) { // Table does not exist throw new IllegalArgumentException(String.format("BigQuery table '%s:%s.%s' does not exist.", - project, dataset, table)); + project, dataset, tableName)); } com.google.cloud.bigquery.Schema bgSchema = table.getDefinition().getSchema(); diff --git a/src/main/java/co/cask/gcp/spanner/sink/SpannerSink.java b/src/main/java/co/cask/gcp/spanner/sink/SpannerSink.java index 1787f5a0eb..42b68d149f 100644 --- a/src/main/java/co/cask/gcp/spanner/sink/SpannerSink.java +++ b/src/main/java/co/cask/gcp/spanner/sink/SpannerSink.java @@ -42,11 +42,11 @@ import com.google.cloud.spanner.Spanner; import com.google.cloud.spanner.SpannerException; import com.google.cloud.spanner.Statement; +import com.google.common.base.Strings; import com.google.spanner.admin.database.v1.CreateDatabaseMetadata; import com.google.spanner.admin.database.v1.UpdateDatabaseDdlMetadata; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.io.NullWritable; -import org.apache.parquet.Strings; import org.slf4j.Logger; import org.slf4j.LoggerFactory;