mirror of
https://github.com/apache/sqoop.git
synced 2025-05-03 06:21:11 +08:00
added Support for DB2 DECFLOAT data type when importing to HDFS/Hive - SQOOP-3449
This commit is contained in:
parent
cee4ab15c7
commit
b22270b1cd
362
SQOOP-3449.patch
Normal file
362
SQOOP-3449.patch
Normal file
@ -0,0 +1,362 @@
|
||||
diff --git a/src/java/org/apache/sqoop/manager/Db2Manager.java b/src/java/org/apache/sqoop/manager/Db2Manager.java
|
||||
index c78946e6..693837ef 100644
|
||||
--- a/src/java/org/apache/sqoop/manager/Db2Manager.java
|
||||
+++ b/src/java/org/apache/sqoop/manager/Db2Manager.java
|
||||
@@ -53,7 +53,8 @@
|
||||
public static final Log LOG = LogFactory.getLog(
|
||||
Db2Manager.class.getName());
|
||||
|
||||
- private static final String XML_TO_JAVA_DATA_TYPE = "String";
|
||||
+ private static final int XML_TO_JAVA_DATA_TYPE = 2009;
|
||||
+ private static final int DECFLOAT_TO_JAVA_DATA_TYPE = 1111;
|
||||
|
||||
private Map<String, String> columnTypeNames;
|
||||
|
||||
@@ -266,40 +267,17 @@ public static String getUserSchema(Connection conn) {
|
||||
*/
|
||||
@Override
|
||||
public String toHiveType(String tableName, String columnName, int sqlType) {
|
||||
- String hiveType = super.toHiveType(tableName, columnName, sqlType);
|
||||
- if (hiveType == null) {
|
||||
- hiveType = toDbSpecificHiveType(tableName, columnName);
|
||||
+ String hiveType;
|
||||
+ if (sqlType == XML_TO_JAVA_DATA_TYPE) {
|
||||
+ hiveType = "String";
|
||||
}
|
||||
- return hiveType;
|
||||
- }
|
||||
-
|
||||
- /**
|
||||
- * Resolve a database-specific type to the Hive type that should contain it.
|
||||
- *
|
||||
- * @param tableName
|
||||
- * table name
|
||||
- * @param colName
|
||||
- * column name
|
||||
- * @return the name of a Hive type to hold the sql datatype, or null if
|
||||
- * none.
|
||||
- */
|
||||
- private String toDbSpecificHiveType(String tableName, String colName) {
|
||||
- if (columnTypeNames == null) {
|
||||
- columnTypeNames = getColumnTypeNames(tableName, options.getCall(),
|
||||
- options.getSqlQuery());
|
||||
+ else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) {
|
||||
+ hiveType = "String";
|
||||
}
|
||||
- LOG.debug("database-specific Column Types and names returned = ("
|
||||
- + StringUtils.join(columnTypeNames.keySet(), ",") + ")=>("
|
||||
- + StringUtils.join(columnTypeNames.values(), ",") + ")");
|
||||
-
|
||||
- String colTypeName = columnTypeNames.get(colName);
|
||||
-
|
||||
- if (colTypeName != null) {
|
||||
- if (colTypeName.toUpperCase().startsWith("XML")) {
|
||||
- return XML_TO_JAVA_DATA_TYPE;
|
||||
- }
|
||||
+ else {
|
||||
+ hiveType = super.toHiveType(tableName, columnName, sqlType);
|
||||
}
|
||||
- return null;
|
||||
+ return hiveType;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -315,35 +293,17 @@ private String toDbSpecificHiveType(String tableName, String colName) {
|
||||
*/
|
||||
@Override
|
||||
public String toJavaType(String tableName, String columnName, int sqlType) {
|
||||
- String javaType = super.toJavaType(tableName, columnName, sqlType);
|
||||
- if (javaType == null) {
|
||||
- javaType = toDbSpecificJavaType(tableName, columnName);
|
||||
+ String javaType;
|
||||
+ if (sqlType == XML_TO_JAVA_DATA_TYPE) {
|
||||
+ javaType = "String";
|
||||
}
|
||||
- return javaType;
|
||||
- }
|
||||
-
|
||||
- /**
|
||||
- * Resolve a database-specific type to the Java type that should contain it.
|
||||
- *
|
||||
- * @param tableName
|
||||
- * table name
|
||||
- * @param colName
|
||||
- * column name
|
||||
- * @return the name of a Java type to hold the sql datatype, or null if
|
||||
- * none.
|
||||
- */
|
||||
- private String toDbSpecificJavaType(String tableName, String colName) {
|
||||
- if (columnTypeNames == null) {
|
||||
- columnTypeNames = getColumnTypeNames(tableName, options.getCall(),
|
||||
- options.getSqlQuery());
|
||||
+ else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) {
|
||||
+ javaType = "String";
|
||||
}
|
||||
- String colTypeName = columnTypeNames.get(colName);
|
||||
- if (colTypeName != null) {
|
||||
- if (colTypeName.equalsIgnoreCase("XML")) {
|
||||
- return XML_TO_JAVA_DATA_TYPE;
|
||||
- }
|
||||
+ else {
|
||||
+ javaType = super.toJavaType(tableName, columnName, sqlType);
|
||||
}
|
||||
- return null;
|
||||
+ return javaType;
|
||||
}
|
||||
|
||||
/**
|
||||
diff --git a/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java
|
||||
new file mode 100644
|
||||
index e69de29b..18c4af9b 100644
|
||||
--- /dev/null
|
||||
+++ b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java
|
||||
@@ -0,0 +1,248 @@
|
||||
+/**
|
||||
+ * Licensed to the Apache Software Foundation (ASF) under one
|
||||
+ * or more contributor license agreements. See the NOTICE file
|
||||
+ * distributed with this work for additional information
|
||||
+ * regarding copyright ownership. The ASF licenses this file
|
||||
+ * to you under the Apache License, Version 2.0 (the
|
||||
+ * "License"); you may not use this file except in compliance
|
||||
+ * with the License. You may obtain a copy of the License at
|
||||
+ *
|
||||
+ * http://www.apache.org/licenses/LICENSE-2.0
|
||||
+ *
|
||||
+ * Unless required by applicable law or agreed to in writing, software
|
||||
+ * distributed under the License is distributed on an "AS IS" BASIS,
|
||||
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
+ * See the License for the specific language governing permissions and
|
||||
+ * limitations under the License.
|
||||
+ */
|
||||
+package org.apache.sqoop.manager.db2;
|
||||
+
|
||||
+import java.io.BufferedReader;
|
||||
+import java.io.File;
|
||||
+import java.io.FileInputStream;
|
||||
+import java.io.IOException;
|
||||
+import java.io.InputStreamReader;
|
||||
+import java.sql.Connection;
|
||||
+import java.sql.SQLException;
|
||||
+import java.sql.Statement;
|
||||
+import java.util.ArrayList;
|
||||
+
|
||||
+import org.apache.commons.logging.Log;
|
||||
+import org.apache.commons.logging.LogFactory;
|
||||
+import org.apache.hadoop.fs.Path;
|
||||
+import org.apache.hadoop.io.IOUtils;
|
||||
+import org.apache.sqoop.manager.Db2Manager;
|
||||
+import org.apache.sqoop.testcategories.thirdpartytest.Db2Test;
|
||||
+import org.apache.sqoop.testcategories.sqooptest.ManualTest;
|
||||
+import org.junit.After;
|
||||
+import org.junit.Before;
|
||||
+import org.junit.Test;
|
||||
+
|
||||
+import org.apache.sqoop.SqoopOptions;
|
||||
+import org.apache.sqoop.testutil.CommonArgs;
|
||||
+import org.apache.sqoop.testutil.ImportJobTestCase;
|
||||
+import org.apache.sqoop.util.FileListing;
|
||||
+import org.junit.experimental.categories.Category;
|
||||
+
|
||||
+import static org.junit.Assert.assertEquals;
|
||||
+import static org.junit.Assert.assertTrue;
|
||||
+import static org.junit.Assert.fail;
|
||||
+
|
||||
+/**
|
||||
+ * Test the DB2 DECFLOAT data type.
|
||||
+ *
|
||||
+ * This uses JDBC to import data from an DB2 database into HDFS.
|
||||
+ *
|
||||
+ * Since this requires an DB2 Server installation,
|
||||
+ * this class is named in such a way that Sqoop's default QA process does
|
||||
+ * not run it. You need to run this manually with
|
||||
+ * -Dtestcase=DB2DECFLOATTypeImportManualTest
|
||||
+
|
||||
+ * You need to put DB2 JDBC driver library (db2jcc.jar) in a location
|
||||
+ * where Sqoop will be able to access it (since this library cannot be checked
|
||||
+ * into Apache's tree for licensing reasons).
|
||||
+ *
|
||||
+ * To set up your test environment:
|
||||
+ * Install DB2 Express 9.7 C server.
|
||||
+ * Create a database SQOOP
|
||||
+ * Create a login SQOOP with password PASSWORD and grant all
|
||||
+ * access for database SQOOP to user SQOOP.
|
||||
+ */
|
||||
+public class DB2DECFLOATTypeImportManualTest extends ImportJobTestCase {
|
||||
+
|
||||
+ public static final Log LOG = LogFactory.getLog(
|
||||
+ DB2DECFLOATTypeImportManualTest.class.getName());
|
||||
+
|
||||
+ static final String HOST_URL = System.getProperty(
|
||||
+ "sqoop.test.db2.connectstring.host_url",
|
||||
+ "jdbc:db2://db2host:60000");
|
||||
+
|
||||
+ static final String DATABASE_NAME = System.getProperty(
|
||||
+ "sqoop.test.db2.connectstring.database",
|
||||
+ "SQOOP");
|
||||
+ static final String DATABASE_USER = System.getProperty(
|
||||
+ "sqoop.test.db2.connectstring.username",
|
||||
+ "SQOOP");
|
||||
+ static final String DATABASE_PASSWORD = System.getProperty(
|
||||
+ "sqoop.test.db2.connectstring.password",
|
||||
+ "SQOOP");
|
||||
+
|
||||
+ static final String TABLE_NAME = "DECFLOATTEST";
|
||||
+ static final String CONNECT_STRING = HOST_URL
|
||||
+ + "/" + DATABASE_NAME;
|
||||
+ static final String HIVE_TABLE_NAME = "DECFLOATTESTHIVE";
|
||||
+ static String ExpectedResults =
|
||||
+ "10.123";
|
||||
+
|
||||
+
|
||||
+ static {
|
||||
+ LOG.info("Using DB2 CONNECT_STRING HOST_URL is : "+HOST_URL);
|
||||
+ LOG.info("Using DB2 CONNECT_STRING: " + CONNECT_STRING);
|
||||
+ }
|
||||
+
|
||||
+ // instance variables populated during setUp, used during tests
|
||||
+ private Db2Manager manager;
|
||||
+
|
||||
+ protected String getTableName() {
|
||||
+ return TABLE_NAME;
|
||||
+ }
|
||||
+
|
||||
+
|
||||
+ @Before
|
||||
+ public void setUp() {
|
||||
+ super.setUp();
|
||||
+
|
||||
+ SqoopOptions options = new SqoopOptions(CONNECT_STRING, getTableName());
|
||||
+ options.setUsername(DATABASE_USER);
|
||||
+ options.setPassword(DATABASE_PASSWORD);
|
||||
+
|
||||
+ manager = new Db2Manager(options);
|
||||
+
|
||||
+ // Drop the existing table, if there is one.
|
||||
+ Connection conn = null;
|
||||
+ Statement stmt = null;
|
||||
+ try {
|
||||
+ conn = manager.getConnection();
|
||||
+ stmt = conn.createStatement();
|
||||
+ stmt.execute("DROP TABLE " + getTableName());
|
||||
+ } catch (SQLException sqlE) {
|
||||
+ LOG.info("Table was not dropped: " + sqlE.getMessage());
|
||||
+ } finally {
|
||||
+ try {
|
||||
+ if (null != stmt) {
|
||||
+ stmt.close();
|
||||
+ }
|
||||
+ } catch (Exception ex) {
|
||||
+ LOG.warn("Exception while closing stmt", ex);
|
||||
+ }
|
||||
+ }
|
||||
+
|
||||
+ // Create and populate table
|
||||
+ try {
|
||||
+ conn = manager.getConnection();
|
||||
+ conn.setAutoCommit(false);
|
||||
+ stmt = conn.createStatement();
|
||||
+ String decfloatdata ="10.123";
|
||||
+
|
||||
+
|
||||
+ // create the database table and populate it with data.
|
||||
+ stmt.executeUpdate("CREATE TABLE " + getTableName() + " ("
|
||||
+ + "ID int, "
|
||||
+ + "BALANCE DECFLOAT(16))");
|
||||
+
|
||||
+ stmt.executeUpdate("INSERT INTO " + getTableName() + " VALUES("
|
||||
+ + "1, "
|
||||
+ + decfloatdata
|
||||
+ +" )");
|
||||
+ conn.commit();
|
||||
+ } catch (SQLException sqlE) {
|
||||
+ LOG.error("Encountered SQL Exception: ", sqlE);
|
||||
+ fail("SQLException when running test setUp(): " + sqlE);
|
||||
+ } finally {
|
||||
+ try {
|
||||
+ if (null != stmt) {
|
||||
+ stmt.close();
|
||||
+ }
|
||||
+ } catch (Exception ex) {
|
||||
+ LOG.warn("Exception while closing connection/stmt", ex);
|
||||
+ }
|
||||
+ }
|
||||
+ }
|
||||
+
|
||||
+ @After
|
||||
+ public void tearDown() {
|
||||
+ super.tearDown();
|
||||
+ try {
|
||||
+ manager.close();
|
||||
+ } catch (SQLException sqlE) {
|
||||
+ LOG.error("Got SQLException: " + sqlE);
|
||||
+ }
|
||||
+ }
|
||||
+
|
||||
+ @Test
|
||||
+ public void testDb2Import() throws IOException {
|
||||
+
|
||||
+ runDb2Test(ExpectedResults);
|
||||
+
|
||||
+ }
|
||||
+
|
||||
+ private String [] getArgv() {
|
||||
+ ArrayList<String> args = new ArrayList<String>();
|
||||
+
|
||||
+ CommonArgs.addHadoopFlags(args);
|
||||
+ args.add("--connect");
|
||||
+ args.add(CONNECT_STRING);
|
||||
+ args.add("--username");
|
||||
+ args.add(DATABASE_USER);
|
||||
+ args.add("--password");
|
||||
+ args.add(DATABASE_PASSWORD);
|
||||
+
|
||||
+ args.add("--table");
|
||||
+ args.add(TABLE_NAME);
|
||||
+ args.add("--warehouse-dir");
|
||||
+ args.add(getWarehouseDir());
|
||||
+ args.add("--hive-table");
|
||||
+ args.add(HIVE_TABLE_NAME);
|
||||
+ args.add("--num-mappers");
|
||||
+ args.add("1");
|
||||
+
|
||||
+ return args.toArray(new String[0]);
|
||||
+ }
|
||||
+
|
||||
+ private void runDb2Test(String expectedResults) throws IOException {
|
||||
+
|
||||
+ Path warehousePath = new Path(this.getWarehouseDir());
|
||||
+ Path tablePath = new Path(warehousePath, getTableName());
|
||||
+ Path filePath = new Path(tablePath, "part-m-00000");
|
||||
+
|
||||
+ File tableFile = new File(tablePath.toString());
|
||||
+ if (tableFile.exists() && tableFile.isDirectory()) {
|
||||
+ // remove the directory before running the import.
|
||||
+ FileListing.recursiveDeleteDir(tableFile);
|
||||
+ }
|
||||
+
|
||||
+ String [] argv = getArgv();
|
||||
+ try {
|
||||
+ runImport(argv);
|
||||
+ LOG.info("finish runImport with argv is : "+argv);
|
||||
+ } catch (IOException ioe) {
|
||||
+ LOG.error("Got IOException during import: " + ioe);
|
||||
+ fail(ioe.toString());
|
||||
+ }
|
||||
+
|
||||
+ File f = new File(filePath.toString());
|
||||
+ assertTrue("Could not find imported data file", f.exists());
|
||||
+ BufferedReader r = null;
|
||||
+ try {
|
||||
+ // Read through the file and make sure it's all there.
|
||||
+ r = new BufferedReader(new InputStreamReader(new FileInputStream(f)));
|
||||
+ assertEquals(expectedResults, r.readLine());
|
||||
+ } catch (IOException ioe) {
|
||||
+ LOG.error("Got IOException verifying results: " + ioe);
|
||||
+ fail(ioe.toString());
|
||||
+ } finally {
|
||||
+ IOUtils.closeStream(r);
|
||||
+ }
|
||||
+ }
|
||||
+
|
||||
+}
|
||||
\ No newline at end of file
|
@ -53,7 +53,8 @@ public class Db2Manager
|
||||
public static final Log LOG = LogFactory.getLog(
|
||||
Db2Manager.class.getName());
|
||||
|
||||
private static final String XML_TO_JAVA_DATA_TYPE = "String";
|
||||
private static final int XML_TO_JAVA_DATA_TYPE = 2009;
|
||||
private static final int DECFLOAT_TO_JAVA_DATA_TYPE = 1111;
|
||||
|
||||
private Map<String, String> columnTypeNames;
|
||||
|
||||
@ -266,42 +267,19 @@ public String[] listTables() {
|
||||
*/
|
||||
@Override
|
||||
public String toHiveType(String tableName, String columnName, int sqlType) {
|
||||
String hiveType = super.toHiveType(tableName, columnName, sqlType);
|
||||
if (hiveType == null) {
|
||||
hiveType = toDbSpecificHiveType(tableName, columnName);
|
||||
String hiveType;
|
||||
if (sqlType == XML_TO_JAVA_DATA_TYPE) {
|
||||
hiveType = "String";
|
||||
}
|
||||
else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) {
|
||||
hiveType = "String";
|
||||
}
|
||||
else {
|
||||
hiveType = super.toHiveType(tableName, columnName, sqlType);
|
||||
}
|
||||
return hiveType;
|
||||
}
|
||||
|
||||
/**
|
||||
* Resolve a database-specific type to the Hive type that should contain it.
|
||||
*
|
||||
* @param tableName
|
||||
* table name
|
||||
* @param colName
|
||||
* column name
|
||||
* @return the name of a Hive type to hold the sql datatype, or null if
|
||||
* none.
|
||||
*/
|
||||
private String toDbSpecificHiveType(String tableName, String colName) {
|
||||
if (columnTypeNames == null) {
|
||||
columnTypeNames = getColumnTypeNames(tableName, options.getCall(),
|
||||
options.getSqlQuery());
|
||||
}
|
||||
LOG.debug("database-specific Column Types and names returned = ("
|
||||
+ StringUtils.join(columnTypeNames.keySet(), ",") + ")=>("
|
||||
+ StringUtils.join(columnTypeNames.values(), ",") + ")");
|
||||
|
||||
String colTypeName = columnTypeNames.get(colName);
|
||||
|
||||
if (colTypeName != null) {
|
||||
if (colTypeName.toUpperCase().startsWith("XML")) {
|
||||
return XML_TO_JAVA_DATA_TYPE;
|
||||
}
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Return java type for SQL type.
|
||||
*
|
||||
@ -315,37 +293,19 @@ private String toDbSpecificHiveType(String tableName, String colName) {
|
||||
*/
|
||||
@Override
|
||||
public String toJavaType(String tableName, String columnName, int sqlType) {
|
||||
String javaType = super.toJavaType(tableName, columnName, sqlType);
|
||||
if (javaType == null) {
|
||||
javaType = toDbSpecificJavaType(tableName, columnName);
|
||||
String javaType;
|
||||
if (sqlType == XML_TO_JAVA_DATA_TYPE) {
|
||||
javaType = "String";
|
||||
}
|
||||
else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) {
|
||||
javaType = "String";
|
||||
}
|
||||
else {
|
||||
javaType = super.toJavaType(tableName, columnName, sqlType);
|
||||
}
|
||||
return javaType;
|
||||
}
|
||||
|
||||
/**
|
||||
* Resolve a database-specific type to the Java type that should contain it.
|
||||
*
|
||||
* @param tableName
|
||||
* table name
|
||||
* @param colName
|
||||
* column name
|
||||
* @return the name of a Java type to hold the sql datatype, or null if
|
||||
* none.
|
||||
*/
|
||||
private String toDbSpecificJavaType(String tableName, String colName) {
|
||||
if (columnTypeNames == null) {
|
||||
columnTypeNames = getColumnTypeNames(tableName, options.getCall(),
|
||||
options.getSqlQuery());
|
||||
}
|
||||
String colTypeName = columnTypeNames.get(colName);
|
||||
if (colTypeName != null) {
|
||||
if (colTypeName.equalsIgnoreCase("XML")) {
|
||||
return XML_TO_JAVA_DATA_TYPE;
|
||||
}
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Create related options for PostgreSQL extra parameters.
|
||||
* @return
|
||||
|
@ -0,0 +1,248 @@
|
||||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.sqoop.manager.db2;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.File;
|
||||
import java.io.FileInputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStreamReader;
|
||||
import java.sql.Connection;
|
||||
import java.sql.SQLException;
|
||||
import java.sql.Statement;
|
||||
import java.util.ArrayList;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.apache.sqoop.manager.Db2Manager;
|
||||
import org.apache.sqoop.testcategories.thirdpartytest.Db2Test;
|
||||
import org.apache.sqoop.testcategories.sqooptest.ManualTest;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
|
||||
import org.apache.sqoop.SqoopOptions;
|
||||
import org.apache.sqoop.testutil.CommonArgs;
|
||||
import org.apache.sqoop.testutil.ImportJobTestCase;
|
||||
import org.apache.sqoop.util.FileListing;
|
||||
import org.junit.experimental.categories.Category;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
/**
|
||||
* Test the DB2 DECFLOAT data type.
|
||||
*
|
||||
* This uses JDBC to import data from an DB2 database into HDFS.
|
||||
*
|
||||
* Since this requires an DB2 Server installation,
|
||||
* this class is named in such a way that Sqoop's default QA process does
|
||||
* not run it. You need to run this manually with
|
||||
* -Dtestcase=DB2DECFLOATTypeImportManualTest
|
||||
|
||||
* You need to put DB2 JDBC driver library (db2jcc.jar) in a location
|
||||
* where Sqoop will be able to access it (since this library cannot be checked
|
||||
* into Apache's tree for licensing reasons).
|
||||
*
|
||||
* To set up your test environment:
|
||||
* Install DB2 Express 9.7 C server.
|
||||
* Create a database SQOOP
|
||||
* Create a login SQOOP with password PASSWORD and grant all
|
||||
* access for database SQOOP to user SQOOP.
|
||||
*/
|
||||
public class DB2DECFLOATTypeImportManualTest extends ImportJobTestCase {
|
||||
|
||||
public static final Log LOG = LogFactory.getLog(
|
||||
DB2DECFLOATTypeImportManualTest.class.getName());
|
||||
|
||||
static final String HOST_URL = System.getProperty(
|
||||
"sqoop.test.db2.connectstring.host_url",
|
||||
"jdbc:db2://db2host:60000");
|
||||
|
||||
static final String DATABASE_NAME = System.getProperty(
|
||||
"sqoop.test.db2.connectstring.database",
|
||||
"SQOOP");
|
||||
static final String DATABASE_USER = System.getProperty(
|
||||
"sqoop.test.db2.connectstring.username",
|
||||
"SQOOP");
|
||||
static final String DATABASE_PASSWORD = System.getProperty(
|
||||
"sqoop.test.db2.connectstring.password",
|
||||
"SQOOP");
|
||||
|
||||
static final String TABLE_NAME = "DECFLOATTEST";
|
||||
static final String CONNECT_STRING = HOST_URL
|
||||
+ "/" + DATABASE_NAME;
|
||||
static final String HIVE_TABLE_NAME = "DECFLOATTESTHIVE";
|
||||
static String ExpectedResults =
|
||||
"10.123";
|
||||
|
||||
|
||||
static {
|
||||
LOG.info("Using DB2 CONNECT_STRING HOST_URL is : "+HOST_URL);
|
||||
LOG.info("Using DB2 CONNECT_STRING: " + CONNECT_STRING);
|
||||
}
|
||||
|
||||
// instance variables populated during setUp, used during tests
|
||||
private Db2Manager manager;
|
||||
|
||||
protected String getTableName() {
|
||||
return TABLE_NAME;
|
||||
}
|
||||
|
||||
|
||||
@Before
|
||||
public void setUp() {
|
||||
super.setUp();
|
||||
|
||||
SqoopOptions options = new SqoopOptions(CONNECT_STRING, getTableName());
|
||||
options.setUsername(DATABASE_USER);
|
||||
options.setPassword(DATABASE_PASSWORD);
|
||||
|
||||
manager = new Db2Manager(options);
|
||||
|
||||
// Drop the existing table, if there is one.
|
||||
Connection conn = null;
|
||||
Statement stmt = null;
|
||||
try {
|
||||
conn = manager.getConnection();
|
||||
stmt = conn.createStatement();
|
||||
stmt.execute("DROP TABLE " + getTableName());
|
||||
} catch (SQLException sqlE) {
|
||||
LOG.info("Table was not dropped: " + sqlE.getMessage());
|
||||
} finally {
|
||||
try {
|
||||
if (null != stmt) {
|
||||
stmt.close();
|
||||
}
|
||||
} catch (Exception ex) {
|
||||
LOG.warn("Exception while closing stmt", ex);
|
||||
}
|
||||
}
|
||||
|
||||
// Create and populate table
|
||||
try {
|
||||
conn = manager.getConnection();
|
||||
conn.setAutoCommit(false);
|
||||
stmt = conn.createStatement();
|
||||
String decfloatdata ="10.123";
|
||||
|
||||
|
||||
// create the database table and populate it with data.
|
||||
stmt.executeUpdate("CREATE TABLE " + getTableName() + " ("
|
||||
+ "ID int, "
|
||||
+ "BALANCE DECFLOAT(16))");
|
||||
|
||||
stmt.executeUpdate("INSERT INTO " + getTableName() + " VALUES("
|
||||
+ "1, "
|
||||
+ decfloatdata
|
||||
+" )");
|
||||
conn.commit();
|
||||
} catch (SQLException sqlE) {
|
||||
LOG.error("Encountered SQL Exception: ", sqlE);
|
||||
fail("SQLException when running test setUp(): " + sqlE);
|
||||
} finally {
|
||||
try {
|
||||
if (null != stmt) {
|
||||
stmt.close();
|
||||
}
|
||||
} catch (Exception ex) {
|
||||
LOG.warn("Exception while closing connection/stmt", ex);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@After
|
||||
public void tearDown() {
|
||||
super.tearDown();
|
||||
try {
|
||||
manager.close();
|
||||
} catch (SQLException sqlE) {
|
||||
LOG.error("Got SQLException: " + sqlE);
|
||||
}
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testDb2Import() throws IOException {
|
||||
|
||||
runDb2Test(ExpectedResults);
|
||||
|
||||
}
|
||||
|
||||
private String [] getArgv() {
|
||||
ArrayList<String> args = new ArrayList<String>();
|
||||
|
||||
CommonArgs.addHadoopFlags(args);
|
||||
args.add("--connect");
|
||||
args.add(CONNECT_STRING);
|
||||
args.add("--username");
|
||||
args.add(DATABASE_USER);
|
||||
args.add("--password");
|
||||
args.add(DATABASE_PASSWORD);
|
||||
|
||||
args.add("--table");
|
||||
args.add(TABLE_NAME);
|
||||
args.add("--warehouse-dir");
|
||||
args.add(getWarehouseDir());
|
||||
args.add("--hive-table");
|
||||
args.add(HIVE_TABLE_NAME);
|
||||
args.add("--num-mappers");
|
||||
args.add("1");
|
||||
|
||||
return args.toArray(new String[0]);
|
||||
}
|
||||
|
||||
private void runDb2Test(String expectedResults) throws IOException {
|
||||
|
||||
Path warehousePath = new Path(this.getWarehouseDir());
|
||||
Path tablePath = new Path(warehousePath, getTableName());
|
||||
Path filePath = new Path(tablePath, "part-m-00000");
|
||||
|
||||
File tableFile = new File(tablePath.toString());
|
||||
if (tableFile.exists() && tableFile.isDirectory()) {
|
||||
// remove the directory before running the import.
|
||||
FileListing.recursiveDeleteDir(tableFile);
|
||||
}
|
||||
|
||||
String [] argv = getArgv();
|
||||
try {
|
||||
runImport(argv);
|
||||
LOG.info("finish runImport with argv is : "+argv);
|
||||
} catch (IOException ioe) {
|
||||
LOG.error("Got IOException during import: " + ioe);
|
||||
fail(ioe.toString());
|
||||
}
|
||||
|
||||
File f = new File(filePath.toString());
|
||||
assertTrue("Could not find imported data file", f.exists());
|
||||
BufferedReader r = null;
|
||||
try {
|
||||
// Read through the file and make sure it's all there.
|
||||
r = new BufferedReader(new InputStreamReader(new FileInputStream(f)));
|
||||
assertEquals(expectedResults, r.readLine());
|
||||
} catch (IOException ioe) {
|
||||
LOG.error("Got IOException verifying results: " + ioe);
|
||||
fail(ioe.toString());
|
||||
} finally {
|
||||
IOUtils.closeStream(r);
|
||||
}
|
||||
}
|
||||
|
||||
}
|
Loading…
Reference in New Issue
Block a user