diff --git a/test/pom.xml b/test/pom.xml
index a9502d29..3e11f597 100644
--- a/test/pom.xml
+++ b/test/pom.xml
@@ -201,6 +201,21 @@ limitations under the License.
+
+ new-integration-test
+
+ test
+
+ integration-test
+
+
+ src/test/resources/new-integration-tests-suite.xml
+
+
+ new-integration-tests
+
+
+
upgrade-test
diff --git a/test/src/main/java/org/apache/sqoop/test/infrastructure/SqoopTestCase.java b/test/src/main/java/org/apache/sqoop/test/infrastructure/SqoopTestCase.java
index e60caf7a..5cae43b2 100644
--- a/test/src/main/java/org/apache/sqoop/test/infrastructure/SqoopTestCase.java
+++ b/test/src/main/java/org/apache/sqoop/test/infrastructure/SqoopTestCase.java
@@ -21,9 +21,26 @@
import org.apache.log4j.Logger;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.mapred.JobConf;
+import org.apache.sqoop.client.SqoopClient;
+import org.apache.sqoop.client.SubmissionCallback;
+import org.apache.sqoop.common.test.db.DatabaseProvider;
+import org.apache.sqoop.common.test.db.TableName;
+import org.apache.sqoop.connector.hdfs.configuration.ToFormat;
+import org.apache.sqoop.model.MConfigList;
+import org.apache.sqoop.model.MJob;
+import org.apache.sqoop.model.MLink;
+import org.apache.sqoop.model.MPersistableEntity;
+import org.apache.sqoop.model.MSubmission;
+import org.apache.sqoop.submission.SubmissionStatus;
+import org.apache.sqoop.test.data.Cities;
+import org.apache.sqoop.test.data.ShortStories;
+import org.apache.sqoop.test.data.UbuntuReleases;
+import org.apache.sqoop.test.infrastructure.providers.DatabaseInfrastructureProvider;
import org.apache.sqoop.test.infrastructure.providers.HadoopInfrastructureProvider;
import org.apache.sqoop.test.infrastructure.providers.InfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.SqoopInfrastructureProvider;
import org.apache.sqoop.test.utils.HdfsUtils;
+import org.apache.sqoop.validation.Status;
import org.testng.ITest;
import org.testng.ITestContext;
import org.testng.ITestNGMethod;
@@ -38,6 +55,9 @@
import java.util.Map;
import java.util.Set;
+import static org.testng.Assert.assertEquals;
+import static org.testng.Assert.assertNotSame;
+
/**
* Use Infrastructure annotation to boot up miniclusters.
* Order is built-in to code. Hadoop comes first, then
@@ -51,10 +71,32 @@ public class SqoopTestCase implements ITest {
private static final Map PROVIDERS
= new HashMap();
+ /**
+ * Default submission callbacks that are printing various status about the submission.
+ */
+ protected static SubmissionCallback DEFAULT_SUBMISSION_CALLBACKS = new SubmissionCallback() {
+ @Override
+ public void submitted(MSubmission submission) {
+ LOG.info("Submission submitted: " + submission);
+ }
+
+ @Override
+ public void updated(MSubmission submission) {
+ LOG.info("Submission updated: " + submission);
+ }
+
+ @Override
+ public void finished(MSubmission submission) {
+ LOG.info("Submission finished: " + submission);
+ }
+ };
+
private static String suiteName;
private String methodName;
+ private SqoopClient client;
+
@BeforeSuite
public static void findSuiteName(ITestContext context) {
suiteName = context.getSuite().getName();
@@ -79,6 +121,10 @@ public static void startInfrastructureProviders(ITestContext context) {
// Find infrastructure provider classes to be used.
Set> providers = new HashSet>();
for (ITestNGMethod method : context.getSuite().getAllMethods()) {
+ LOG.debug("Looking up dependencies on method ("
+ + method.getConstructorOrMethod().getDeclaringClass().getCanonicalName()
+ + "#" + method.getConstructorOrMethod().getMethod().getName()
+ + ")");
Infrastructure ann;
// If the method has an infrastructure annotation, process it.
@@ -173,4 +219,241 @@ public static T getInfrastructureProvider(Cla
InfrastructureProvider provider = PROVIDERS.get(providerClass.getCanonicalName());
return ((T) provider);
}
+
+ /**
+ * Get the data directory for tests.
+ * @return
+ */
+ public String getMapreduceDirectory() {
+ return HdfsUtils.joinPathFragments(
+ getInfrastructureProvider(HadoopInfrastructureProvider.class).getInstance().getTestDirectory(),
+ getClass().getName(),
+ getTestName());
+ }
+
+ /**
+ * Fill RDBMS Link Configuration with infrastructure provider info.
+ * @param link
+ */
+ public void fillRdbmsLinkConfig(MLink link) {
+ DatabaseProvider provider = getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance();
+
+ MConfigList configs = link.getConnectorLinkConfig();
+ configs.getStringInput("linkConfig.jdbcDriver").setValue(provider.getJdbcDriver());
+ configs.getStringInput("linkConfig.connectionString").setValue(provider.getConnectionUrl());
+ configs.getStringInput("linkConfig.username").setValue(provider.getConnectionUsername());
+ configs.getStringInput("linkConfig.password").setValue(provider.getConnectionPassword());
+ }
+
+ /**
+ * Fill RDBMS FROM Configuration with infrastructure provider info.
+ * @param job
+ * @param partitionColumn
+ */
+ public void fillRdbmsFromConfig(MJob job, String partitionColumn) {
+ DatabaseProvider provider = getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance();
+
+ MConfigList fromConfig = job.getFromJobConfig();
+ fromConfig.getStringInput("fromJobConfig.tableName").setValue(provider.escapeTableName(getTableName().getTableName()));
+ fromConfig.getStringInput("fromJobConfig.partitionColumn").setValue(provider.escapeColumnName(partitionColumn));
+ }
+
+ /**
+ * Fill RDBMS TO Configuration with infrastructure provider info.
+ * @param job
+ */
+ public void fillRdbmsToConfig(MJob job) {
+ DatabaseProvider provider = getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance();
+
+ MConfigList toConfig = job.getToJobConfig();
+ toConfig.getStringInput("toJobConfig.tableName").setValue(provider.escapeTableName(getTableName().getTableName()));
+ }
+
+ /**
+ * Fill HDFS Link Configuration with infrastructure provider info.
+ * @param link
+ */
+ public void fillHdfsLinkConfig(MLink link) {
+ MConfigList configs = link.getConnectorLinkConfig();
+ configs.getStringInput("linkConfig.confDir").setValue(
+ getInfrastructureProvider(SqoopInfrastructureProvider.class).getInstance()
+ .getConfigurationPath());
+ }
+
+ /**
+ * Fill HDFS FROM Configuration with infrastructure provider info.
+ * @param job
+ */
+ public void fillHdfsFromConfig(MJob job) {
+ MConfigList fromConfig = job.getFromJobConfig();
+ fromConfig.getStringInput("fromJobConfig.inputDirectory").setValue(getMapreduceDirectory());
+ }
+
+ /**
+ * Fill HDFS TO Configuration with infrastructure provider info.
+ * @param job
+ * @param output
+ */
+ public void fillHdfsToConfig(MJob job, ToFormat output) {
+ MConfigList toConfig = job.getToJobConfig();
+ toConfig.getEnumInput("toJobConfig.outputFormat").setValue(output);
+ toConfig.getStringInput("toJobConfig.outputDirectory").setValue(getMapreduceDirectory());
+ }
+
+ public String getSqoopServerUrl() {
+ if (getInfrastructureProvider(SqoopInfrastructureProvider.class) == null) {
+ return null;
+ }
+
+ return getInfrastructureProvider(SqoopInfrastructureProvider.class).getInstance()
+ .getServerUrl();
+ }
+
+ /**
+ * Create a sqoop client
+ * @return SqoopClient
+ */
+ public SqoopClient getClient() {
+ if (client == null) {
+ String serverUrl = getSqoopServerUrl();
+
+ if (serverUrl != null) {
+ client = new SqoopClient(serverUrl);
+ }
+ }
+ return client;
+ }
+
+ /**
+ * Create link with asserts to make sure that it was created correctly.
+ *
+ * @param link
+ */
+ public void saveLink(MLink link) {
+ assertEquals(Status.OK, getClient().saveLink(link));
+ assertNotSame(MPersistableEntity.PERSISTANCE_ID_DEFAULT, link.getPersistenceId());
+ }
+
+ /**
+ * Create job with asserts to make sure that it was created correctly.
+ *
+ * @param job
+ */
+ public void saveJob(MJob job) {
+ assertEquals(Status.OK, getClient().saveJob(job));
+ assertNotSame(MPersistableEntity.PERSISTANCE_ID_DEFAULT, job.getPersistenceId());
+ }
+
+ /**
+ * Run job with given jid.
+ *
+ * @param jid Job id
+ * @throws Exception
+ */
+ public void executeJob(long jid) throws Exception {
+ MSubmission finalSubmission = getClient().startJob(jid, DEFAULT_SUBMISSION_CALLBACKS, 100);
+
+ if(finalSubmission.getStatus().isFailure()) {
+ LOG.error("Submission has failed: " + finalSubmission.getError().getErrorSummary());
+ LOG.error("Corresponding error details: " + finalSubmission.getError().getErrorDetails());
+ }
+ assertEquals(SubmissionStatus.SUCCEEDED, finalSubmission.getStatus(), "Submission finished with error: " + finalSubmission.getError().getErrorSummary());
+ }
+
+ /**
+ * Fetch table name to be used by this test.
+ * @return TableName
+ */
+ public TableName getTableName() {
+ return new TableName(getClass().getSimpleName());
+ }
+
+ /**
+ * Create table with table name for this test.
+ * @param primaryKey
+ * @param columns
+ */
+ public void createTable(String primaryKey, String ...columns) {
+ getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance()
+ .createTable(getTableName(), primaryKey, columns);
+ }
+
+ /**
+ * Drop table for this test.
+ */
+ public void dropTable() {
+ getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance()
+ .dropTable(getTableName());
+ }
+
+ /**
+ * Insert row into table for this test.
+ * @param values
+ */
+ public void insertRow(Object ...values) {
+ getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance()
+ .insertRow(getTableName(), values);
+ }
+
+ /**
+ * Insert row into table for this test.
+ * @param escapeValues
+ * @param values
+ */
+ public void insertRow(Boolean escapeValues, Object ...values) {
+ getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance()
+ .insertRow(getTableName(), escapeValues, values);
+ }
+
+ /**
+ * Fetch row count of table for this test.
+ * @return long count
+ */
+ public long rowCount() {
+ return getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance()
+ .rowCount(getTableName());
+ }
+
+ /**
+ * Dump the table for this test.
+ */
+ public void dumpTable() {
+ getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance()
+ .dumpTable(getTableName());
+ }
+
+ /**
+ * Create and load cities data.
+ */
+ public void createAndLoadTableCities() {
+ new Cities(getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance(), getTableName()).createTables().loadBasicData();
+ }
+
+ /**
+ * Create ubuntu releases table.
+ */
+ public void createTableUbuntuReleases() {
+ new UbuntuReleases(getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance(), getTableName()).createTables();
+ }
+
+ /**
+ * Create and load ubuntu releases data.
+ */
+ public void createAndLoadTableUbuntuReleases() {
+ new UbuntuReleases(getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance(), getTableName()).createTables().loadBasicData();
+ }
+
+ /**
+ * Create short stories table.
+ */
+ public void createTableShortStories() {
+ new ShortStories(getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance(), getTableName()).createTables();
+ }
+
+ /**
+ * Create and load short stories data.
+ */
+ public void createAndLoadTableShortStories() {
+ new ShortStories(getInfrastructureProvider(DatabaseInfrastructureProvider.class).getInstance(), getTableName()).createTables().loadBasicData();
+ }
}
diff --git a/test/src/main/java/org/apache/sqoop/test/minicluster/InProcessSqoopMiniCluster.java b/test/src/main/java/org/apache/sqoop/test/minicluster/InProcessSqoopMiniCluster.java
index e11e82c3..ae1fac18 100644
--- a/test/src/main/java/org/apache/sqoop/test/minicluster/InProcessSqoopMiniCluster.java
+++ b/test/src/main/java/org/apache/sqoop/test/minicluster/InProcessSqoopMiniCluster.java
@@ -52,4 +52,9 @@ public void stop() throws Exception {
SqoopServer.destroy();
}
+ /** {@inheritDoc} */
+ @Override
+ public String getServerUrl() {
+ return "http://localhost:12000/sqoop";
+ }
}
diff --git a/test/src/main/java/org/apache/sqoop/test/minicluster/SqoopMiniCluster.java b/test/src/main/java/org/apache/sqoop/test/minicluster/SqoopMiniCluster.java
index 758eb2fa..74400254 100644
--- a/test/src/main/java/org/apache/sqoop/test/minicluster/SqoopMiniCluster.java
+++ b/test/src/main/java/org/apache/sqoop/test/minicluster/SqoopMiniCluster.java
@@ -100,6 +100,11 @@ public String getLogPath() {
*/
abstract public void stop() throws Exception;
+ /**
+ * @return server URL (e.g. http://localhost:12000/sqoop)
+ */
+ abstract public String getServerUrl();
+
/**
* Prepare temporary directory for starting Sqoop server.
*
diff --git a/test/src/main/java/org/apache/sqoop/test/minicluster/TomcatSqoopMiniCluster.java b/test/src/main/java/org/apache/sqoop/test/minicluster/TomcatSqoopMiniCluster.java
index 18a4b123..83f42b65 100644
--- a/test/src/main/java/org/apache/sqoop/test/minicluster/TomcatSqoopMiniCluster.java
+++ b/test/src/main/java/org/apache/sqoop/test/minicluster/TomcatSqoopMiniCluster.java
@@ -189,6 +189,7 @@ protected Map getLoggerConfiguration() {
/**
* Return server URL.
*/
+ @Override
public String getServerUrl() {
// We're not doing any changes, so return default URL
return "http://localhost:" + port + "/sqoop/";
diff --git a/test/src/main/java/org/apache/sqoop/test/utils/HdfsUtils.java b/test/src/main/java/org/apache/sqoop/test/utils/HdfsUtils.java
index 79700a6b..7f83b2df 100644
--- a/test/src/main/java/org/apache/sqoop/test/utils/HdfsUtils.java
+++ b/test/src/main/java/org/apache/sqoop/test/utils/HdfsUtils.java
@@ -22,6 +22,11 @@
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter;
import org.apache.log4j.Logger;
+import org.apache.sqoop.connector.hdfs.configuration.ToFormat;
+import org.apache.sqoop.model.MConfigList;
+import org.apache.sqoop.model.MJob;
+import org.apache.sqoop.model.MLink;
+import org.apache.sqoop.test.hadoop.HadoopRunner;
import java.io.BufferedWriter;
import java.io.FileNotFoundException;
@@ -67,8 +72,7 @@ public boolean accept(Path path) {
* Create HDFS file with given content.
*
* @param fs filesystem object
- * @param directory Directory where the file should be created
- * @param filename File name
+ * @param path path to file be created
* @param lines Individual lines that should be written into the file
* @throws IOException
*/
diff --git a/test/src/test/java/org/apache/sqoop/integration/server/InvalidRESTCallsTest.java b/test/src/test/java/org/apache/sqoop/integration/server/InvalidRESTCallsTest.java
index f1ef3368..0490cf03 100644
--- a/test/src/test/java/org/apache/sqoop/integration/server/InvalidRESTCallsTest.java
+++ b/test/src/test/java/org/apache/sqoop/integration/server/InvalidRESTCallsTest.java
@@ -21,6 +21,11 @@
import com.google.common.collect.Iterables;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthenticatedURL;
+import org.apache.sqoop.test.infrastructure.Infrastructure;
+import org.apache.sqoop.test.infrastructure.SqoopTestCase;
+import org.apache.sqoop.test.infrastructure.providers.DatabaseInfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.HadoopInfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.SqoopInfrastructureProvider;
import org.apache.sqoop.test.testcases.ConnectorTestCase;
import org.apache.sqoop.test.utils.ParametrizedUtils;
import org.testng.ITest;
@@ -35,7 +40,8 @@
import static org.testng.Assert.assertEquals;
-public class InvalidRESTCallsTest extends ConnectorTestCase implements ITest {
+@Infrastructure(dependencies = {HadoopInfrastructureProvider.class, SqoopInfrastructureProvider.class})
+public class InvalidRESTCallsTest extends SqoopTestCase {
private static final Logger LOG = Logger.getLogger(InvalidRESTCallsTest.class);
@@ -120,11 +126,10 @@ public String getTestName() {
public void test() throws Exception {
LOG.info("Start: " + getTestName());
- URL url = new URL(getServerUrl() + desc.rest);
+ URL url = new URL(getSqoopServerUrl() + desc.rest);
HttpURLConnection connection = new DelegationTokenAuthenticatedURL().openConnection(url, new DelegationTokenAuthenticatedURL.Token());
connection.setRequestMethod(desc.method);
-
if(desc.data != null) {
connection.setDoOutput(true);
diff --git a/test/src/test/java/org/apache/sqoop/integration/server/ShowJobInOrderTest.java b/test/src/test/java/org/apache/sqoop/integration/server/ShowJobInOrderTest.java
index 88cfcb5f..6c0622c6 100644
--- a/test/src/test/java/org/apache/sqoop/integration/server/ShowJobInOrderTest.java
+++ b/test/src/test/java/org/apache/sqoop/integration/server/ShowJobInOrderTest.java
@@ -20,6 +20,10 @@
import org.apache.sqoop.connector.hdfs.configuration.ToFormat;
import org.apache.sqoop.model.MJob;
import org.apache.sqoop.model.MLink;
+import org.apache.sqoop.test.infrastructure.Infrastructure;
+import org.apache.sqoop.test.infrastructure.SqoopTestCase;
+import org.apache.sqoop.test.infrastructure.providers.HadoopInfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.SqoopInfrastructureProvider;
import org.apache.sqoop.test.testcases.ConnectorTestCase;
import org.testng.annotations.Test;
@@ -30,7 +34,8 @@
/**
* Ensure that jobs will be shown in order
*/
-public class ShowJobInOrderTest extends ConnectorTestCase {
+@Infrastructure(dependencies = {HadoopInfrastructureProvider.class, SqoopInfrastructureProvider.class})
+public class ShowJobInOrderTest extends SqoopTestCase {
public ShowJobInOrderTest() {
}
@@ -46,7 +51,7 @@ public void testShowJobInOrder() throws Exception {
// HDFS link
MLink hdfsLink = getClient().createLink("hdfs-connector");
- fillHdfsLink(hdfsLink);
+ fillHdfsLinkConfig(hdfsLink);
saveLink(hdfsLink);
// Job creation
diff --git a/test/src/test/java/org/apache/sqoop/integration/server/SubmissionWithDisabledModelObjectsTest.java b/test/src/test/java/org/apache/sqoop/integration/server/SubmissionWithDisabledModelObjectsTest.java
index 199b77c2..93cc7f6a 100644
--- a/test/src/test/java/org/apache/sqoop/integration/server/SubmissionWithDisabledModelObjectsTest.java
+++ b/test/src/test/java/org/apache/sqoop/integration/server/SubmissionWithDisabledModelObjectsTest.java
@@ -18,14 +18,18 @@
package org.apache.sqoop.integration.server;
import org.apache.sqoop.client.ClientError;
-import org.apache.sqoop.common.Direction;
import org.apache.sqoop.common.SqoopException;
import org.apache.sqoop.connector.hdfs.configuration.ToFormat;
import org.apache.sqoop.error.code.DriverError;
import org.apache.sqoop.model.MLink;
-import org.apache.sqoop.model.MConfigList;
import org.apache.sqoop.model.MJob;
-import org.apache.sqoop.test.testcases.ConnectorTestCase;
+import org.apache.sqoop.test.infrastructure.Infrastructure;
+import org.apache.sqoop.test.infrastructure.SqoopTestCase;
+import org.apache.sqoop.test.infrastructure.providers.DatabaseInfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.HadoopInfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.SqoopInfrastructureProvider;
+import org.testng.annotations.AfterMethod;
+import org.testng.annotations.BeforeMethod;
import org.testng.annotations.DataProvider;
import org.testng.annotations.Factory;
import org.testng.annotations.Test;
@@ -39,7 +43,8 @@
* Ensure that server will reject starting job when either job itself
* or corresponding link is disabled.
*/
-public class SubmissionWithDisabledModelObjectsTest extends ConnectorTestCase {
+@Infrastructure(dependencies = {HadoopInfrastructureProvider.class, SqoopInfrastructureProvider.class, DatabaseInfrastructureProvider.class})
+public class SubmissionWithDisabledModelObjectsTest extends SqoopTestCase {
private boolean enabledLink;
private boolean enabledJob;
@@ -59,10 +64,18 @@ public static Object[][] data() {
};
}
+ @BeforeMethod
+ public void setupRdbmsTable() {
+ createAndLoadTableCities();
+ }
+
+ @AfterMethod
+ public void tearDownRdbmsTable() {
+ dropTable();
+ }
+
@Test
public void testWithDisabledObjects() throws Exception {
- createAndLoadTableCities();
-
// RDBMS link
MLink rdbmsLink = getClient().createLink("generic-jdbc-connector");
fillRdbmsLinkConfig(rdbmsLink);
@@ -70,7 +83,7 @@ public void testWithDisabledObjects() throws Exception {
// HDFS link
MLink hdfsLink = getClient().createLink("hdfs-connector");
- fillHdfsLink(hdfsLink);
+ fillHdfsLinkConfig(hdfsLink);
saveLink(hdfsLink);
// Job creation
@@ -90,7 +103,7 @@ public void testWithDisabledObjects() throws Exception {
// Try to execute the job and verify that the it was not executed
try {
- executeJob(job);
+ executeJob(job.getPersistenceId());
fail("Expected exception as the model classes are disabled.");
} catch(SqoopException ex) {
// Top level exception should be CLIENT_0001
@@ -109,8 +122,6 @@ public void testWithDisabledObjects() throws Exception {
} else {
fail("Unexpected expception retrieved from server " + cause);
}
- } finally {
- dropTable();
}
}
}
diff --git a/test/src/test/java/org/apache/sqoop/integration/server/VersionTest.java b/test/src/test/java/org/apache/sqoop/integration/server/VersionTest.java
index bcaf5f5f..ad5a2db3 100644
--- a/test/src/test/java/org/apache/sqoop/integration/server/VersionTest.java
+++ b/test/src/test/java/org/apache/sqoop/integration/server/VersionTest.java
@@ -19,7 +19,10 @@
import org.apache.sqoop.client.request.VersionResourceRequest;
import org.apache.sqoop.common.VersionInfo;
-import org.apache.sqoop.test.testcases.TomcatTestCase;
+import org.apache.sqoop.test.infrastructure.Infrastructure;
+import org.apache.sqoop.test.infrastructure.SqoopTestCase;
+import org.apache.sqoop.test.infrastructure.providers.HadoopInfrastructureProvider;
+import org.apache.sqoop.test.infrastructure.providers.SqoopInfrastructureProvider;
import org.apache.sqoop.json.VersionBean;
import org.testng.annotations.Test;
@@ -28,12 +31,13 @@
/**
* Basic test to check that server is working and returning correct version info.
*/
-public class VersionTest extends TomcatTestCase {
+@Infrastructure(dependencies = {HadoopInfrastructureProvider.class, SqoopInfrastructureProvider.class})
+public class VersionTest extends SqoopTestCase {
@Test
public void testVersion() {
VersionResourceRequest versionRequest = new VersionResourceRequest();
- VersionBean versionBean = versionRequest.read(getServerUrl());
+ VersionBean versionBean = versionRequest.read(getSqoopServerUrl());
assertEquals(versionBean.getBuildVersion(), VersionInfo.getBuildVersion());
assertEquals(versionBean.getBuildDate(), VersionInfo.getBuildDate());
diff --git a/test/src/test/resources/integration-tests-suite.xml b/test/src/test/resources/integration-tests-suite.xml
index c1053298..1cf32995 100644
--- a/test/src/test/resources/integration-tests-suite.xml
+++ b/test/src/test/resources/integration-tests-suite.xml
@@ -29,11 +29,4 @@ limitations under the License.
-
-
-
-
-
-
-
\ No newline at end of file
diff --git a/test/src/test/resources/new-integration-tests-suite.xml b/test/src/test/resources/new-integration-tests-suite.xml
new file mode 100644
index 00000000..8206e524
--- /dev/null
+++ b/test/src/test/resources/new-integration-tests-suite.xml
@@ -0,0 +1,33 @@
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+