From b22270b1cd9e33629cee3657f8d1a9fb823ffefe Mon Sep 17 00:00:00 2001 From: Chiran Ravani Date: Fri, 6 Sep 2019 17:08:27 +0530 Subject: [PATCH] added Support for DB2 DECFLOAT data type when importing to HDFS/Hive - SQOOP-3449 --- SQOOP-3449.patch | 362 ++++++++++++++++++ .../org/apache/sqoop/manager/Db2Manager.java | 80 +--- .../db2/DB2DECFLOATTypeImportManualTest.java | 248 ++++++++++++ 3 files changed, 630 insertions(+), 60 deletions(-) create mode 100644 SQOOP-3449.patch create mode 100644 src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java diff --git a/SQOOP-3449.patch b/SQOOP-3449.patch new file mode 100644 index 00000000..40de2114 --- /dev/null +++ b/SQOOP-3449.patch @@ -0,0 +1,362 @@ +diff --git a/src/java/org/apache/sqoop/manager/Db2Manager.java b/src/java/org/apache/sqoop/manager/Db2Manager.java +index c78946e6..693837ef 100644 +--- a/src/java/org/apache/sqoop/manager/Db2Manager.java ++++ b/src/java/org/apache/sqoop/manager/Db2Manager.java +@@ -53,7 +53,8 @@ + public static final Log LOG = LogFactory.getLog( + Db2Manager.class.getName()); + +- private static final String XML_TO_JAVA_DATA_TYPE = "String"; ++ private static final int XML_TO_JAVA_DATA_TYPE = 2009; ++ private static final int DECFLOAT_TO_JAVA_DATA_TYPE = 1111; + + private Map columnTypeNames; + +@@ -266,40 +267,17 @@ public static String getUserSchema(Connection conn) { + */ + @Override + public String toHiveType(String tableName, String columnName, int sqlType) { +- String hiveType = super.toHiveType(tableName, columnName, sqlType); +- if (hiveType == null) { +- hiveType = toDbSpecificHiveType(tableName, columnName); ++ String hiveType; ++ if (sqlType == XML_TO_JAVA_DATA_TYPE) { ++ hiveType = "String"; + } +- return hiveType; +- } +- +- /** +- * Resolve a database-specific type to the Hive type that should contain it. +- * +- * @param tableName +- * table name +- * @param colName +- * column name +- * @return the name of a Hive type to hold the sql datatype, or null if +- * none. +- */ +- private String toDbSpecificHiveType(String tableName, String colName) { +- if (columnTypeNames == null) { +- columnTypeNames = getColumnTypeNames(tableName, options.getCall(), +- options.getSqlQuery()); ++ else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) { ++ hiveType = "String"; + } +- LOG.debug("database-specific Column Types and names returned = (" +- + StringUtils.join(columnTypeNames.keySet(), ",") + ")=>(" +- + StringUtils.join(columnTypeNames.values(), ",") + ")"); +- +- String colTypeName = columnTypeNames.get(colName); +- +- if (colTypeName != null) { +- if (colTypeName.toUpperCase().startsWith("XML")) { +- return XML_TO_JAVA_DATA_TYPE; +- } ++ else { ++ hiveType = super.toHiveType(tableName, columnName, sqlType); + } +- return null; ++ return hiveType; + } + + /** +@@ -315,35 +293,17 @@ private String toDbSpecificHiveType(String tableName, String colName) { + */ + @Override + public String toJavaType(String tableName, String columnName, int sqlType) { +- String javaType = super.toJavaType(tableName, columnName, sqlType); +- if (javaType == null) { +- javaType = toDbSpecificJavaType(tableName, columnName); ++ String javaType; ++ if (sqlType == XML_TO_JAVA_DATA_TYPE) { ++ javaType = "String"; + } +- return javaType; +- } +- +- /** +- * Resolve a database-specific type to the Java type that should contain it. +- * +- * @param tableName +- * table name +- * @param colName +- * column name +- * @return the name of a Java type to hold the sql datatype, or null if +- * none. +- */ +- private String toDbSpecificJavaType(String tableName, String colName) { +- if (columnTypeNames == null) { +- columnTypeNames = getColumnTypeNames(tableName, options.getCall(), +- options.getSqlQuery()); ++ else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) { ++ javaType = "String"; + } +- String colTypeName = columnTypeNames.get(colName); +- if (colTypeName != null) { +- if (colTypeName.equalsIgnoreCase("XML")) { +- return XML_TO_JAVA_DATA_TYPE; +- } ++ else { ++ javaType = super.toJavaType(tableName, columnName, sqlType); + } +- return null; ++ return javaType; + } + + /** +diff --git a/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java +new file mode 100644 +index e69de29b..18c4af9b 100644 +--- /dev/null ++++ b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java +@@ -0,0 +1,248 @@ ++/** ++ * Licensed to the Apache Software Foundation (ASF) under one ++ * or more contributor license agreements. See the NOTICE file ++ * distributed with this work for additional information ++ * regarding copyright ownership. The ASF licenses this file ++ * to you under the Apache License, Version 2.0 (the ++ * "License"); you may not use this file except in compliance ++ * with the License. You may obtain a copy of the License at ++ * ++ * http://www.apache.org/licenses/LICENSE-2.0 ++ * ++ * Unless required by applicable law or agreed to in writing, software ++ * distributed under the License is distributed on an "AS IS" BASIS, ++ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ++ * See the License for the specific language governing permissions and ++ * limitations under the License. ++ */ ++package org.apache.sqoop.manager.db2; ++ ++import java.io.BufferedReader; ++import java.io.File; ++import java.io.FileInputStream; ++import java.io.IOException; ++import java.io.InputStreamReader; ++import java.sql.Connection; ++import java.sql.SQLException; ++import java.sql.Statement; ++import java.util.ArrayList; ++ ++import org.apache.commons.logging.Log; ++import org.apache.commons.logging.LogFactory; ++import org.apache.hadoop.fs.Path; ++import org.apache.hadoop.io.IOUtils; ++import org.apache.sqoop.manager.Db2Manager; ++import org.apache.sqoop.testcategories.thirdpartytest.Db2Test; ++import org.apache.sqoop.testcategories.sqooptest.ManualTest; ++import org.junit.After; ++import org.junit.Before; ++import org.junit.Test; ++ ++import org.apache.sqoop.SqoopOptions; ++import org.apache.sqoop.testutil.CommonArgs; ++import org.apache.sqoop.testutil.ImportJobTestCase; ++import org.apache.sqoop.util.FileListing; ++import org.junit.experimental.categories.Category; ++ ++import static org.junit.Assert.assertEquals; ++import static org.junit.Assert.assertTrue; ++import static org.junit.Assert.fail; ++ ++/** ++ * Test the DB2 DECFLOAT data type. ++ * ++ * This uses JDBC to import data from an DB2 database into HDFS. ++ * ++ * Since this requires an DB2 Server installation, ++ * this class is named in such a way that Sqoop's default QA process does ++ * not run it. You need to run this manually with ++ * -Dtestcase=DB2DECFLOATTypeImportManualTest ++ ++ * You need to put DB2 JDBC driver library (db2jcc.jar) in a location ++ * where Sqoop will be able to access it (since this library cannot be checked ++ * into Apache's tree for licensing reasons). ++ * ++ * To set up your test environment: ++ * Install DB2 Express 9.7 C server. ++ * Create a database SQOOP ++ * Create a login SQOOP with password PASSWORD and grant all ++ * access for database SQOOP to user SQOOP. ++ */ ++public class DB2DECFLOATTypeImportManualTest extends ImportJobTestCase { ++ ++ public static final Log LOG = LogFactory.getLog( ++ DB2DECFLOATTypeImportManualTest.class.getName()); ++ ++ static final String HOST_URL = System.getProperty( ++ "sqoop.test.db2.connectstring.host_url", ++ "jdbc:db2://db2host:60000"); ++ ++ static final String DATABASE_NAME = System.getProperty( ++ "sqoop.test.db2.connectstring.database", ++ "SQOOP"); ++ static final String DATABASE_USER = System.getProperty( ++ "sqoop.test.db2.connectstring.username", ++ "SQOOP"); ++ static final String DATABASE_PASSWORD = System.getProperty( ++ "sqoop.test.db2.connectstring.password", ++ "SQOOP"); ++ ++ static final String TABLE_NAME = "DECFLOATTEST"; ++ static final String CONNECT_STRING = HOST_URL ++ + "/" + DATABASE_NAME; ++ static final String HIVE_TABLE_NAME = "DECFLOATTESTHIVE"; ++ static String ExpectedResults = ++ "10.123"; ++ ++ ++ static { ++ LOG.info("Using DB2 CONNECT_STRING HOST_URL is : "+HOST_URL); ++ LOG.info("Using DB2 CONNECT_STRING: " + CONNECT_STRING); ++ } ++ ++ // instance variables populated during setUp, used during tests ++ private Db2Manager manager; ++ ++ protected String getTableName() { ++ return TABLE_NAME; ++ } ++ ++ ++ @Before ++ public void setUp() { ++ super.setUp(); ++ ++ SqoopOptions options = new SqoopOptions(CONNECT_STRING, getTableName()); ++ options.setUsername(DATABASE_USER); ++ options.setPassword(DATABASE_PASSWORD); ++ ++ manager = new Db2Manager(options); ++ ++ // Drop the existing table, if there is one. ++ Connection conn = null; ++ Statement stmt = null; ++ try { ++ conn = manager.getConnection(); ++ stmt = conn.createStatement(); ++ stmt.execute("DROP TABLE " + getTableName()); ++ } catch (SQLException sqlE) { ++ LOG.info("Table was not dropped: " + sqlE.getMessage()); ++ } finally { ++ try { ++ if (null != stmt) { ++ stmt.close(); ++ } ++ } catch (Exception ex) { ++ LOG.warn("Exception while closing stmt", ex); ++ } ++ } ++ ++ // Create and populate table ++ try { ++ conn = manager.getConnection(); ++ conn.setAutoCommit(false); ++ stmt = conn.createStatement(); ++ String decfloatdata ="10.123"; ++ ++ ++ // create the database table and populate it with data. ++ stmt.executeUpdate("CREATE TABLE " + getTableName() + " (" ++ + "ID int, " ++ + "BALANCE DECFLOAT(16))"); ++ ++ stmt.executeUpdate("INSERT INTO " + getTableName() + " VALUES(" ++ + "1, " ++ + decfloatdata ++ +" )"); ++ conn.commit(); ++ } catch (SQLException sqlE) { ++ LOG.error("Encountered SQL Exception: ", sqlE); ++ fail("SQLException when running test setUp(): " + sqlE); ++ } finally { ++ try { ++ if (null != stmt) { ++ stmt.close(); ++ } ++ } catch (Exception ex) { ++ LOG.warn("Exception while closing connection/stmt", ex); ++ } ++ } ++ } ++ ++ @After ++ public void tearDown() { ++ super.tearDown(); ++ try { ++ manager.close(); ++ } catch (SQLException sqlE) { ++ LOG.error("Got SQLException: " + sqlE); ++ } ++ } ++ ++ @Test ++ public void testDb2Import() throws IOException { ++ ++ runDb2Test(ExpectedResults); ++ ++ } ++ ++ private String [] getArgv() { ++ ArrayList args = new ArrayList(); ++ ++ CommonArgs.addHadoopFlags(args); ++ args.add("--connect"); ++ args.add(CONNECT_STRING); ++ args.add("--username"); ++ args.add(DATABASE_USER); ++ args.add("--password"); ++ args.add(DATABASE_PASSWORD); ++ ++ args.add("--table"); ++ args.add(TABLE_NAME); ++ args.add("--warehouse-dir"); ++ args.add(getWarehouseDir()); ++ args.add("--hive-table"); ++ args.add(HIVE_TABLE_NAME); ++ args.add("--num-mappers"); ++ args.add("1"); ++ ++ return args.toArray(new String[0]); ++ } ++ ++ private void runDb2Test(String expectedResults) throws IOException { ++ ++ Path warehousePath = new Path(this.getWarehouseDir()); ++ Path tablePath = new Path(warehousePath, getTableName()); ++ Path filePath = new Path(tablePath, "part-m-00000"); ++ ++ File tableFile = new File(tablePath.toString()); ++ if (tableFile.exists() && tableFile.isDirectory()) { ++ // remove the directory before running the import. ++ FileListing.recursiveDeleteDir(tableFile); ++ } ++ ++ String [] argv = getArgv(); ++ try { ++ runImport(argv); ++ LOG.info("finish runImport with argv is : "+argv); ++ } catch (IOException ioe) { ++ LOG.error("Got IOException during import: " + ioe); ++ fail(ioe.toString()); ++ } ++ ++ File f = new File(filePath.toString()); ++ assertTrue("Could not find imported data file", f.exists()); ++ BufferedReader r = null; ++ try { ++ // Read through the file and make sure it's all there. ++ r = new BufferedReader(new InputStreamReader(new FileInputStream(f))); ++ assertEquals(expectedResults, r.readLine()); ++ } catch (IOException ioe) { ++ LOG.error("Got IOException verifying results: " + ioe); ++ fail(ioe.toString()); ++ } finally { ++ IOUtils.closeStream(r); ++ } ++ } ++ ++} +\ No newline at end of file diff --git a/src/java/org/apache/sqoop/manager/Db2Manager.java b/src/java/org/apache/sqoop/manager/Db2Manager.java index c78946e6..693837ef 100644 --- a/src/java/org/apache/sqoop/manager/Db2Manager.java +++ b/src/java/org/apache/sqoop/manager/Db2Manager.java @@ -53,7 +53,8 @@ public class Db2Manager public static final Log LOG = LogFactory.getLog( Db2Manager.class.getName()); - private static final String XML_TO_JAVA_DATA_TYPE = "String"; + private static final int XML_TO_JAVA_DATA_TYPE = 2009; + private static final int DECFLOAT_TO_JAVA_DATA_TYPE = 1111; private Map columnTypeNames; @@ -266,42 +267,19 @@ public String[] listTables() { */ @Override public String toHiveType(String tableName, String columnName, int sqlType) { - String hiveType = super.toHiveType(tableName, columnName, sqlType); - if (hiveType == null) { - hiveType = toDbSpecificHiveType(tableName, columnName); + String hiveType; + if (sqlType == XML_TO_JAVA_DATA_TYPE) { + hiveType = "String"; + } + else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) { + hiveType = "String"; + } + else { + hiveType = super.toHiveType(tableName, columnName, sqlType); } return hiveType; } - /** - * Resolve a database-specific type to the Hive type that should contain it. - * - * @param tableName - * table name - * @param colName - * column name - * @return the name of a Hive type to hold the sql datatype, or null if - * none. - */ - private String toDbSpecificHiveType(String tableName, String colName) { - if (columnTypeNames == null) { - columnTypeNames = getColumnTypeNames(tableName, options.getCall(), - options.getSqlQuery()); - } - LOG.debug("database-specific Column Types and names returned = (" - + StringUtils.join(columnTypeNames.keySet(), ",") + ")=>(" - + StringUtils.join(columnTypeNames.values(), ",") + ")"); - - String colTypeName = columnTypeNames.get(colName); - - if (colTypeName != null) { - if (colTypeName.toUpperCase().startsWith("XML")) { - return XML_TO_JAVA_DATA_TYPE; - } - } - return null; - } - /** * Return java type for SQL type. * @@ -315,37 +293,19 @@ private String toDbSpecificHiveType(String tableName, String colName) { */ @Override public String toJavaType(String tableName, String columnName, int sqlType) { - String javaType = super.toJavaType(tableName, columnName, sqlType); - if (javaType == null) { - javaType = toDbSpecificJavaType(tableName, columnName); + String javaType; + if (sqlType == XML_TO_JAVA_DATA_TYPE) { + javaType = "String"; + } + else if (sqlType == DECFLOAT_TO_JAVA_DATA_TYPE) { + javaType = "String"; + } + else { + javaType = super.toJavaType(tableName, columnName, sqlType); } return javaType; } - /** - * Resolve a database-specific type to the Java type that should contain it. - * - * @param tableName - * table name - * @param colName - * column name - * @return the name of a Java type to hold the sql datatype, or null if - * none. - */ - private String toDbSpecificJavaType(String tableName, String colName) { - if (columnTypeNames == null) { - columnTypeNames = getColumnTypeNames(tableName, options.getCall(), - options.getSqlQuery()); - } - String colTypeName = columnTypeNames.get(colName); - if (colTypeName != null) { - if (colTypeName.equalsIgnoreCase("XML")) { - return XML_TO_JAVA_DATA_TYPE; - } - } - return null; - } - /** * Create related options for PostgreSQL extra parameters. * @return diff --git a/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java new file mode 100644 index 00000000..18c4af9b --- /dev/null +++ b/src/test/org/apache/sqoop/manager/db2/DB2DECFLOATTypeImportManualTest.java @@ -0,0 +1,248 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.sqoop.manager.db2; + +import java.io.BufferedReader; +import java.io.File; +import java.io.FileInputStream; +import java.io.IOException; +import java.io.InputStreamReader; +import java.sql.Connection; +import java.sql.SQLException; +import java.sql.Statement; +import java.util.ArrayList; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.io.IOUtils; +import org.apache.sqoop.manager.Db2Manager; +import org.apache.sqoop.testcategories.thirdpartytest.Db2Test; +import org.apache.sqoop.testcategories.sqooptest.ManualTest; +import org.junit.After; +import org.junit.Before; +import org.junit.Test; + +import org.apache.sqoop.SqoopOptions; +import org.apache.sqoop.testutil.CommonArgs; +import org.apache.sqoop.testutil.ImportJobTestCase; +import org.apache.sqoop.util.FileListing; +import org.junit.experimental.categories.Category; + +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertTrue; +import static org.junit.Assert.fail; + +/** + * Test the DB2 DECFLOAT data type. + * + * This uses JDBC to import data from an DB2 database into HDFS. + * + * Since this requires an DB2 Server installation, + * this class is named in such a way that Sqoop's default QA process does + * not run it. You need to run this manually with + * -Dtestcase=DB2DECFLOATTypeImportManualTest + + * You need to put DB2 JDBC driver library (db2jcc.jar) in a location + * where Sqoop will be able to access it (since this library cannot be checked + * into Apache's tree for licensing reasons). + * + * To set up your test environment: + * Install DB2 Express 9.7 C server. + * Create a database SQOOP + * Create a login SQOOP with password PASSWORD and grant all + * access for database SQOOP to user SQOOP. + */ +public class DB2DECFLOATTypeImportManualTest extends ImportJobTestCase { + + public static final Log LOG = LogFactory.getLog( + DB2DECFLOATTypeImportManualTest.class.getName()); + + static final String HOST_URL = System.getProperty( + "sqoop.test.db2.connectstring.host_url", + "jdbc:db2://db2host:60000"); + + static final String DATABASE_NAME = System.getProperty( + "sqoop.test.db2.connectstring.database", + "SQOOP"); + static final String DATABASE_USER = System.getProperty( + "sqoop.test.db2.connectstring.username", + "SQOOP"); + static final String DATABASE_PASSWORD = System.getProperty( + "sqoop.test.db2.connectstring.password", + "SQOOP"); + + static final String TABLE_NAME = "DECFLOATTEST"; + static final String CONNECT_STRING = HOST_URL + + "/" + DATABASE_NAME; + static final String HIVE_TABLE_NAME = "DECFLOATTESTHIVE"; + static String ExpectedResults = + "10.123"; + + + static { + LOG.info("Using DB2 CONNECT_STRING HOST_URL is : "+HOST_URL); + LOG.info("Using DB2 CONNECT_STRING: " + CONNECT_STRING); + } + + // instance variables populated during setUp, used during tests + private Db2Manager manager; + + protected String getTableName() { + return TABLE_NAME; + } + + + @Before + public void setUp() { + super.setUp(); + + SqoopOptions options = new SqoopOptions(CONNECT_STRING, getTableName()); + options.setUsername(DATABASE_USER); + options.setPassword(DATABASE_PASSWORD); + + manager = new Db2Manager(options); + + // Drop the existing table, if there is one. + Connection conn = null; + Statement stmt = null; + try { + conn = manager.getConnection(); + stmt = conn.createStatement(); + stmt.execute("DROP TABLE " + getTableName()); + } catch (SQLException sqlE) { + LOG.info("Table was not dropped: " + sqlE.getMessage()); + } finally { + try { + if (null != stmt) { + stmt.close(); + } + } catch (Exception ex) { + LOG.warn("Exception while closing stmt", ex); + } + } + + // Create and populate table + try { + conn = manager.getConnection(); + conn.setAutoCommit(false); + stmt = conn.createStatement(); + String decfloatdata ="10.123"; + + + // create the database table and populate it with data. + stmt.executeUpdate("CREATE TABLE " + getTableName() + " (" + + "ID int, " + + "BALANCE DECFLOAT(16))"); + + stmt.executeUpdate("INSERT INTO " + getTableName() + " VALUES(" + + "1, " + + decfloatdata + +" )"); + conn.commit(); + } catch (SQLException sqlE) { + LOG.error("Encountered SQL Exception: ", sqlE); + fail("SQLException when running test setUp(): " + sqlE); + } finally { + try { + if (null != stmt) { + stmt.close(); + } + } catch (Exception ex) { + LOG.warn("Exception while closing connection/stmt", ex); + } + } + } + + @After + public void tearDown() { + super.tearDown(); + try { + manager.close(); + } catch (SQLException sqlE) { + LOG.error("Got SQLException: " + sqlE); + } + } + + @Test + public void testDb2Import() throws IOException { + + runDb2Test(ExpectedResults); + + } + + private String [] getArgv() { + ArrayList args = new ArrayList(); + + CommonArgs.addHadoopFlags(args); + args.add("--connect"); + args.add(CONNECT_STRING); + args.add("--username"); + args.add(DATABASE_USER); + args.add("--password"); + args.add(DATABASE_PASSWORD); + + args.add("--table"); + args.add(TABLE_NAME); + args.add("--warehouse-dir"); + args.add(getWarehouseDir()); + args.add("--hive-table"); + args.add(HIVE_TABLE_NAME); + args.add("--num-mappers"); + args.add("1"); + + return args.toArray(new String[0]); + } + + private void runDb2Test(String expectedResults) throws IOException { + + Path warehousePath = new Path(this.getWarehouseDir()); + Path tablePath = new Path(warehousePath, getTableName()); + Path filePath = new Path(tablePath, "part-m-00000"); + + File tableFile = new File(tablePath.toString()); + if (tableFile.exists() && tableFile.isDirectory()) { + // remove the directory before running the import. + FileListing.recursiveDeleteDir(tableFile); + } + + String [] argv = getArgv(); + try { + runImport(argv); + LOG.info("finish runImport with argv is : "+argv); + } catch (IOException ioe) { + LOG.error("Got IOException during import: " + ioe); + fail(ioe.toString()); + } + + File f = new File(filePath.toString()); + assertTrue("Could not find imported data file", f.exists()); + BufferedReader r = null; + try { + // Read through the file and make sure it's all there. + r = new BufferedReader(new InputStreamReader(new FileInputStream(f))); + assertEquals(expectedResults, r.readLine()); + } catch (IOException ioe) { + LOG.error("Got IOException verifying results: " + ioe); + fail(ioe.toString()); + } finally { + IOUtils.closeStream(r); + } + } + +} \ No newline at end of file