diff --git a/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteFromInitializer.java b/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteFromInitializer.java index 06f18e70..fd3017be 100644 --- a/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteFromInitializer.java +++ b/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteFromInitializer.java @@ -59,6 +59,14 @@ public Set getJars(InitializerContext context, jars.add(ClassUtils.jarForClass("org.kitesdk.data.Datasets")); jars.add(ClassUtils.jarForClass("com.fasterxml.jackson.databind.JsonNode")); jars.add(ClassUtils.jarForClass("com.fasterxml.jackson.core.TreeNode")); + jars.add(ClassUtils.jarForClass("parquet.hadoop.metadata.CompressionCodecName")); + jars.add(ClassUtils.jarForClass("parquet.format.CompressionCodec")); + jars.add(ClassUtils.jarForClass("parquet.avro.AvroParquetWriter")); + jars.add(ClassUtils.jarForClass("parquet.column.ParquetProperties")); + jars.add(ClassUtils.jarForClass("parquet.Version")); + jars.add(ClassUtils.jarForClass("parquet.org.codehaus.jackson.type.TypeReference")); + jars.add(ClassUtils.jarForClass("parquet.bytes.CapacityByteArrayOutputStream")); + jars.add(ClassUtils.jarForClass("parquet.encoding.Generator")); if (fromJobConfig.fromJobConfig.uri.startsWith("dataset:hive")) { // @TODO(Abe): Remove a deps that aren't used? jars.add(ClassUtils.jarForClass("org.apache.hadoop.hive.conf.HiveConf")); @@ -72,6 +80,7 @@ public Set getJars(InitializerContext context, jars.add(ClassUtils.jarForClass("org.datanucleus.query.compiler.JavaQueryCompiler")); jars.add(ClassUtils.jarForClass("org.datanucleus.query.typesafe.TypesafeSubquery")); jars.add(ClassUtils.jarForClass("org.datanucleus.store.rdbms.sql.SQLStatement")); + jars.add(ClassUtils.jarForClass("parquet.hive.serde.ParquetHiveSerDe")); } return jars; } diff --git a/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteToInitializer.java b/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteToInitializer.java index 39a29892..e27c0b89 100644 --- a/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteToInitializer.java +++ b/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteToInitializer.java @@ -65,6 +65,16 @@ public Set getJars(InitializerContext context, if (FileFormat.CSV.equals(toJobConfig.toJobConfig.fileFormat)) { jars.add(ClassUtils.jarForClass("au.com.bytecode.opencsv.CSVWriter")); } + if (FileFormat.PARQUET.equals(toJobConfig.toJobConfig.fileFormat)) { + jars.add(ClassUtils.jarForClass("parquet.hadoop.metadata.CompressionCodecName")); + jars.add(ClassUtils.jarForClass("parquet.format.CompressionCodec")); + jars.add(ClassUtils.jarForClass("parquet.avro.AvroParquetWriter")); + jars.add(ClassUtils.jarForClass("parquet.column.ParquetProperties")); + jars.add(ClassUtils.jarForClass("parquet.Version")); + jars.add(ClassUtils.jarForClass("parquet.org.codehaus.jackson.type.TypeReference")); + jars.add(ClassUtils.jarForClass("parquet.bytes.CapacityByteArrayOutputStream")); + jars.add(ClassUtils.jarForClass("parquet.encoding.Generator")); + } if (toJobConfig.toJobConfig.uri.startsWith("dataset:hive")) { // @TODO(Abe): Remove a deps that aren't used? jars.add(ClassUtils.jarForClass("org.apache.hadoop.hive.conf.HiveConf")); @@ -78,6 +88,7 @@ public Set getJars(InitializerContext context, jars.add(ClassUtils.jarForClass("org.datanucleus.query.compiler.JavaQueryCompiler")); jars.add(ClassUtils.jarForClass("org.datanucleus.query.typesafe.TypesafeSubquery")); jars.add(ClassUtils.jarForClass("org.datanucleus.store.rdbms.sql.SQLStatement")); + jars.add(ClassUtils.jarForClass("parquet.hive.serde.ParquetHiveSerDe")); } return jars; }