diff --git a/src/java/org/apache/sqoop/manager/Db2Manager.java b/src/java/org/apache/sqoop/manager/Db2Manager.java index c78946e6..693837ef 100644 --- a/src/java/org/apache/sqoop/manager/Db2Manager.java +++ b/src/java/org/apache/sqoop/manager/Db2Manager.java @@ -53,7 +53,8 @@ public static final Log LOG = LogFactory.getLog( Db2Manager.class.getName()); - private static final String XML_TO_JAVA_DATA_TYPE = "String"; + private static final int XML_TO_JAVA_DATA_TYPE = 2009; + private static final int DECFLOAT_TO_JAVA_DATA_TYPE = 1111; private Map columnTypeNames; @@ -266,40 +267,17 @@ public static String getUserSchema(Connection conn) { */ @Override public String toHiveType(String tableName, String columnName, int sqlType) { - String hiveType = super.toHiveType(tableName, columnName, sqlType); - if (hiveType == null) { - hiveType = toDbSpecificHiveType(tableName, columnName); + String hiveType; + if (sqlType == XML_TO_JAVA_DATA_TYPE) { + hiveType = "String"; } - return hiveType; - } - - /** - * Resolve a database-specific type to the Hive type that should contain it. - * - * @param tableName - * table name - * @param colName - * column name - * @return the name of a Hive type to hold the sql datatype, or null if - * none. - */ - private String toDbSpecificHiveType(String tableName, String colName) { - if (columnTypeNames == null) { - columnTypeNames = getColumnTypeNames(tableName, options.getCall(), - options.getSqlQuery()); + else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) { + hiveType = "String"; } - LOG.debug("database-specific Column Types and names returned = (" - + StringUtils.join(columnTypeNames.keySet(), ",") + ")=>(" - + StringUtils.join(columnTypeNames.values(), ",") + ")"); - - String colTypeName = columnTypeNames.get(colName); - - if (colTypeName != null) { - if (colTypeName.toUpperCase().startsWith("XML")) { - return XML_TO_JAVA_DATA_TYPE; - } + else { + hiveType = super.toHiveType(tableName, columnName, sqlType); } - return null; + return hiveType; } /** @@ -315,35 +293,17 @@ private String toDbSpecificHiveType(String tableName, String colName) { */ @Override public String toJavaType(String tableName, String columnName, int sqlType) { - String javaType = super.toJavaType(tableName, columnName, sqlType); - if (javaType == null) { - javaType = toDbSpecificJavaType(tableName, columnName); + String javaType; + if (sqlType == XML_TO_JAVA_DATA_TYPE) { + javaType = "String"; } - return javaType; - } - - /** - * Resolve a database-specific type to the Java type that should contain it. - * - * @param tableName - * table name - * @param colName - * column name - * @return the name of a Java type to hold the sql datatype, or null if - * none. - */ - private String toDbSpecificJavaType(String tableName, String colName) { - if (columnTypeNames == null) { - columnTypeNames = getColumnTypeNames(tableName, options.getCall(), - options.getSqlQuery()); + else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) { + javaType = "String"; } - String colTypeName = columnTypeNames.get(colName); - if (colTypeName != null) { - if (colTypeName.equalsIgnoreCase("XML")) { - return XML_TO_JAVA_DATA_TYPE; - } + else { + javaType = super.toJavaType(tableName, columnName, sqlType); } - return null; + return javaType; } /** diff --git a/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java new file mode 100644 index e69de29b..18c4af9b 100644 --- /dev/null +++ b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java @@ -0,0 +1,248 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.sqoop.manager.db2; + +import java.io.BufferedReader; +import java.io.File; +import java.io.FileInputStream; +import java.io.IOException; +import java.io.InputStreamReader; +import java.sql.Connection; +import java.sql.SQLException; +import java.sql.Statement; +import java.util.ArrayList; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.io.IOUtils; +import org.apache.sqoop.manager.Db2Manager; +import org.apache.sqoop.testcategories.thirdpartytest.Db2Test; +import org.apache.sqoop.testcategories.sqooptest.ManualTest; +import org.junit.After; +import org.junit.Before; +import org.junit.Test; + +import org.apache.sqoop.SqoopOptions; +import org.apache.sqoop.testutil.CommonArgs; +import org.apache.sqoop.testutil.ImportJobTestCase; +import org.apache.sqoop.util.FileListing; +import org.junit.experimental.categories.Category; + +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertTrue; +import static org.junit.Assert.fail; + +/** + * Test the DB2 DECFLOAT data type. + * + * This uses JDBC to import data from an DB2 database into HDFS. + * + * Since this requires an DB2 Server installation, + * this class is named in such a way that Sqoop's default QA process does + * not run it. You need to run this manually with + * -Dtestcase=DB2DECFLOATTypeImportManualTest + + * You need to put DB2 JDBC driver library (db2jcc.jar) in a location + * where Sqoop will be able to access it (since this library cannot be checked + * into Apache's tree for licensing reasons). + * + * To set up your test environment: + * Install DB2 Express 9.7 C server. + * Create a database SQOOP + * Create a login SQOOP with password PASSWORD and grant all + * access for database SQOOP to user SQOOP. + */ +public class DB2DECFLOATTypeImportManualTest extends ImportJobTestCase { + + public static final Log LOG = LogFactory.getLog( + DB2DECFLOATTypeImportManualTest.class.getName()); + + static final String HOST_URL = System.getProperty( + "sqoop.test.db2.connectstring.host_url", + "jdbc:db2://db2host:60000"); + + static final String DATABASE_NAME = System.getProperty( + "sqoop.test.db2.connectstring.database", + "SQOOP"); + static final String DATABASE_USER = System.getProperty( + "sqoop.test.db2.connectstring.username", + "SQOOP"); + static final String DATABASE_PASSWORD = System.getProperty( + "sqoop.test.db2.connectstring.password", + "SQOOP"); + + static final String TABLE_NAME = "DECFLOATTEST"; + static final String CONNECT_STRING = HOST_URL + + "/" + DATABASE_NAME; + static final String HIVE_TABLE_NAME = "DECFLOATTESTHIVE"; + static String ExpectedResults = + "10.123"; + + + static { + LOG.info("Using DB2 CONNECT_STRING HOST_URL is : "+HOST_URL); + LOG.info("Using DB2 CONNECT_STRING: " + CONNECT_STRING); + } + + // instance variables populated during setUp, used during tests + private Db2Manager manager; + + protected String getTableName() { + return TABLE_NAME; + } + + + @Before + public void setUp() { + super.setUp(); + + SqoopOptions options = new SqoopOptions(CONNECT_STRING, getTableName()); + options.setUsername(DATABASE_USER); + options.setPassword(DATABASE_PASSWORD); + + manager = new Db2Manager(options); + + // Drop the existing table, if there is one. + Connection conn = null; + Statement stmt = null; + try { + conn = manager.getConnection(); + stmt = conn.createStatement(); + stmt.execute("DROP TABLE " + getTableName()); + } catch (SQLException sqlE) { + LOG.info("Table was not dropped: " + sqlE.getMessage()); + } finally { + try { + if (null != stmt) { + stmt.close(); + } + } catch (Exception ex) { + LOG.warn("Exception while closing stmt", ex); + } + } + + // Create and populate table + try { + conn = manager.getConnection(); + conn.setAutoCommit(false); + stmt = conn.createStatement(); + String decfloatdata ="10.123"; + + + // create the database table and populate it with data. + stmt.executeUpdate("CREATE TABLE " + getTableName() + " (" + + "ID int, " + + "BALANCE DECFLOAT(16))"); + + stmt.executeUpdate("INSERT INTO " + getTableName() + " VALUES(" + + "1, " + + decfloatdata + +" )"); + conn.commit(); + } catch (SQLException sqlE) { + LOG.error("Encountered SQL Exception: ", sqlE); + fail("SQLException when running test setUp(): " + sqlE); + } finally { + try { + if (null != stmt) { + stmt.close(); + } + } catch (Exception ex) { + LOG.warn("Exception while closing connection/stmt", ex); + } + } + } + + @After + public void tearDown() { + super.tearDown(); + try { + manager.close(); + } catch (SQLException sqlE) { + LOG.error("Got SQLException: " + sqlE); + } + } + + @Test + public void testDb2Import() throws IOException { + + runDb2Test(ExpectedResults); + + } + + private String [] getArgv() { + ArrayList args = new ArrayList(); + + CommonArgs.addHadoopFlags(args); + args.add("--connect"); + args.add(CONNECT_STRING); + args.add("--username"); + args.add(DATABASE_USER); + args.add("--password"); + args.add(DATABASE_PASSWORD); + + args.add("--table"); + args.add(TABLE_NAME); + args.add("--warehouse-dir"); + args.add(getWarehouseDir()); + args.add("--hive-table"); + args.add(HIVE_TABLE_NAME); + args.add("--num-mappers"); + args.add("1"); + + return args.toArray(new String[0]); + } + + private void runDb2Test(String expectedResults) throws IOException { + + Path warehousePath = new Path(this.getWarehouseDir()); + Path tablePath = new Path(warehousePath, getTableName()); + Path filePath = new Path(tablePath, "part-m-00000"); + + File tableFile = new File(tablePath.toString()); + if (tableFile.exists() && tableFile.isDirectory()) { + // remove the directory before running the import. + FileListing.recursiveDeleteDir(tableFile); + } + + String [] argv = getArgv(); + try { + runImport(argv); + LOG.info("finish runImport with argv is : "+argv); + } catch (IOException ioe) { + LOG.error("Got IOException during import: " + ioe); + fail(ioe.toString()); + } + + File f = new File(filePath.toString()); + assertTrue("Could not find imported data file", f.exists()); + BufferedReader r = null; + try { + // Read through the file and make sure it's all there. + r = new BufferedReader(new InputStreamReader(new FileInputStream(f))); + assertEquals(expectedResults, r.readLine()); + } catch (IOException ioe) { + LOG.error("Got IOException verifying results: " + ioe); + fail(ioe.toString()); + } finally { + IOUtils.closeStream(r); + } + } + +} \ No newline at end of file