mirror of
https://github.com/apache/sqoop.git
synced 2025-05-11 14:30:59 +08:00
SQOOP-692 Move specific error codes to own error codes class
(Jarek Jarcec Cecho)
This commit is contained in:
parent
fa815557c5
commit
5f95f114f6
@ -51,53 +51,6 @@ public enum CoreError implements ErrorCode {
|
|||||||
/** The configuration system has not been initialized correctly. */
|
/** The configuration system has not been initialized correctly. */
|
||||||
CORE_0007("System not initialized"),
|
CORE_0007("System not initialized"),
|
||||||
|
|
||||||
/** Error occurs during job execution. */
|
|
||||||
CORE_0008("Error occurs during job execution"),
|
|
||||||
|
|
||||||
/** The system was unable to load the specified class. */
|
|
||||||
CORE_0009("Unable to load the specified class"),
|
|
||||||
|
|
||||||
/** The system was unable to instantiate the specified class. */
|
|
||||||
CORE_0010("Unable to instantiate the specified class"),
|
|
||||||
|
|
||||||
/** The parameter already exists in the context */
|
|
||||||
CORE_0011("The parameter already exists in the context"),
|
|
||||||
|
|
||||||
/** The type is not supported */
|
|
||||||
CORE_0012("The type is not supported"),
|
|
||||||
|
|
||||||
/** Cannot write to the data writer */
|
|
||||||
CORE_0013("Cannot write to the data writer"),
|
|
||||||
|
|
||||||
/** Cannot read from the data reader */
|
|
||||||
CORE_0014("Cannot read to the data reader"),
|
|
||||||
|
|
||||||
/** Unable to write data due to interrupt */
|
|
||||||
CORE_0015("Unable to write data due to interrupt"),
|
|
||||||
|
|
||||||
/** Unable to read data due to interrupt */
|
|
||||||
CORE_0016("Unable to read data due to interrupt"),
|
|
||||||
|
|
||||||
/** Error occurs during extractor run */
|
|
||||||
CORE_0017("Error occurs during extractor run"),
|
|
||||||
|
|
||||||
/** Error occurs during loader run */
|
|
||||||
CORE_0018("Error occurs during loader run"),
|
|
||||||
|
|
||||||
CORE_0019("Data have not been completely consumed yet"),
|
|
||||||
|
|
||||||
/** The required option has not been set yet */
|
|
||||||
CORE_0020("The required option has not been set yet"),
|
|
||||||
|
|
||||||
/** Error occurs during partitioner run */
|
|
||||||
CORE_0021("Error occurs during partitioner run"),
|
|
||||||
|
|
||||||
/** Unable to parse because it is not properly delimited */
|
|
||||||
CORE_0022("Unable to parse because it is not properly delimited"),
|
|
||||||
|
|
||||||
/** Unknown job type */
|
|
||||||
CORE_0023("Unknown job type"),
|
|
||||||
|
|
||||||
;
|
;
|
||||||
|
|
||||||
private final String message;
|
private final String message;
|
||||||
|
@ -0,0 +1,91 @@
|
|||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.sqoop.job;
|
||||||
|
|
||||||
|
import org.apache.sqoop.common.ErrorCode;
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
public enum MapreduceExecutionError implements ErrorCode {
|
||||||
|
|
||||||
|
MAPRED_EXEC_0000("Unknown error"),
|
||||||
|
|
||||||
|
/** Error occurs during job execution. */
|
||||||
|
MAPRED_EXEC_0008("Error occurs during job execution"),
|
||||||
|
|
||||||
|
/** The system was unable to load the specified class. */
|
||||||
|
MAPRED_EXEC_0009("Unable to load the specified class"),
|
||||||
|
|
||||||
|
/** The system was unable to instantiate the specified class. */
|
||||||
|
MAPRED_EXEC_0010("Unable to instantiate the specified class"),
|
||||||
|
|
||||||
|
/** The parameter already exists in the context */
|
||||||
|
MAPRED_EXEC_0011("The parameter already exists in the context"),
|
||||||
|
|
||||||
|
/** The type is not supported */
|
||||||
|
MAPRED_EXEC_0012("The type is not supported"),
|
||||||
|
|
||||||
|
/** Cannot write to the data writer */
|
||||||
|
MAPRED_EXEC_0013("Cannot write to the data writer"),
|
||||||
|
|
||||||
|
/** Cannot read from the data reader */
|
||||||
|
MAPRED_EXEC_0014("Cannot read to the data reader"),
|
||||||
|
|
||||||
|
/** Unable to write data due to interrupt */
|
||||||
|
MAPRED_EXEC_0015("Unable to write data due to interrupt"),
|
||||||
|
|
||||||
|
/** Unable to read data due to interrupt */
|
||||||
|
MAPRED_EXEC_0016("Unable to read data due to interrupt"),
|
||||||
|
|
||||||
|
/** Error occurs during extractor run */
|
||||||
|
MAPRED_EXEC_0017("Error occurs during extractor run"),
|
||||||
|
|
||||||
|
/** Error occurs during loader run */
|
||||||
|
MAPRED_EXEC_0018("Error occurs during loader run"),
|
||||||
|
|
||||||
|
MAPRED_EXEC_0019("Data have not been completely consumed yet"),
|
||||||
|
|
||||||
|
/** The required option has not been set yet */
|
||||||
|
MAPRED_EXEC_0020("The required option has not been set yet"),
|
||||||
|
|
||||||
|
/** Error occurs during partitioner run */
|
||||||
|
MAPRED_EXEC_0021("Error occurs during partitioner run"),
|
||||||
|
|
||||||
|
/** Unable to parse because it is not properly delimited */
|
||||||
|
MAPRED_EXEC_0022("Unable to parse because it is not properly delimited"),
|
||||||
|
|
||||||
|
/** Unknown job type */
|
||||||
|
MAPRED_EXEC_0023("Unknown job type"),
|
||||||
|
|
||||||
|
;
|
||||||
|
|
||||||
|
private final String message;
|
||||||
|
|
||||||
|
private MapreduceExecutionError(String message) {
|
||||||
|
this.message = message;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getCode() {
|
||||||
|
return name();
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getMessage() {
|
||||||
|
return message;
|
||||||
|
}
|
||||||
|
}
|
@ -29,8 +29,8 @@
|
|||||||
import org.apache.hadoop.io.compress.CompressionCodec;
|
import org.apache.hadoop.io.compress.CompressionCodec;
|
||||||
import org.apache.sqoop.common.ImmutableContext;
|
import org.apache.sqoop.common.ImmutableContext;
|
||||||
import org.apache.sqoop.common.SqoopException;
|
import org.apache.sqoop.common.SqoopException;
|
||||||
import org.apache.sqoop.core.CoreError;
|
|
||||||
import org.apache.sqoop.job.JobConstants;
|
import org.apache.sqoop.job.JobConstants;
|
||||||
|
import org.apache.sqoop.job.MapreduceExecutionError;
|
||||||
import org.apache.sqoop.job.io.Data;
|
import org.apache.sqoop.job.io.Data;
|
||||||
import org.apache.sqoop.job.io.DataReader;
|
import org.apache.sqoop.job.io.DataReader;
|
||||||
import org.apache.sqoop.utils.ClassUtils;
|
import org.apache.sqoop.utils.ClassUtils;
|
||||||
@ -59,7 +59,7 @@ public void run(ImmutableContext context, DataReader reader) {
|
|||||||
if (codecname != null) {
|
if (codecname != null) {
|
||||||
Class<?> clz = ClassUtils.loadClass(codecname);
|
Class<?> clz = ClassUtils.loadClass(codecname);
|
||||||
if (clz == null) {
|
if (clz == null) {
|
||||||
throw new SqoopException(CoreError.CORE_0009, codecname);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0009, codecname);
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
@ -68,7 +68,7 @@ public void run(ImmutableContext context, DataReader reader) {
|
|||||||
((Configurable) codec).setConf(conf);
|
((Configurable) codec).setConf(conf);
|
||||||
}
|
}
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
throw new SqoopException(CoreError.CORE_0010, codecname, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0010, codecname, e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -100,7 +100,7 @@ public void run(ImmutableContext context, DataReader reader) {
|
|||||||
filewriter.close();
|
filewriter.close();
|
||||||
|
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new SqoopException(CoreError.CORE_0018, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0018, e);
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -29,8 +29,8 @@
|
|||||||
import org.apache.hadoop.io.compress.CompressionCodec;
|
import org.apache.hadoop.io.compress.CompressionCodec;
|
||||||
import org.apache.sqoop.common.ImmutableContext;
|
import org.apache.sqoop.common.ImmutableContext;
|
||||||
import org.apache.sqoop.common.SqoopException;
|
import org.apache.sqoop.common.SqoopException;
|
||||||
import org.apache.sqoop.core.CoreError;
|
|
||||||
import org.apache.sqoop.job.JobConstants;
|
import org.apache.sqoop.job.JobConstants;
|
||||||
|
import org.apache.sqoop.job.MapreduceExecutionError;
|
||||||
import org.apache.sqoop.job.io.Data;
|
import org.apache.sqoop.job.io.Data;
|
||||||
import org.apache.sqoop.job.io.DataReader;
|
import org.apache.sqoop.job.io.DataReader;
|
||||||
import org.apache.sqoop.utils.ClassUtils;
|
import org.apache.sqoop.utils.ClassUtils;
|
||||||
@ -58,7 +58,7 @@ public void run(ImmutableContext context, DataReader reader) {
|
|||||||
if (codecname != null) {
|
if (codecname != null) {
|
||||||
Class<?> clz = ClassUtils.loadClass(codecname);
|
Class<?> clz = ClassUtils.loadClass(codecname);
|
||||||
if (clz == null) {
|
if (clz == null) {
|
||||||
throw new SqoopException(CoreError.CORE_0009, codecname);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0009, codecname);
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
@ -67,7 +67,7 @@ public void run(ImmutableContext context, DataReader reader) {
|
|||||||
((Configurable) codec).setConf(conf);
|
((Configurable) codec).setConf(conf);
|
||||||
}
|
}
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
throw new SqoopException(CoreError.CORE_0010, codecname, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0010, codecname, e);
|
||||||
}
|
}
|
||||||
|
|
||||||
filename += codec.getDefaultExtension();
|
filename += codec.getDefaultExtension();
|
||||||
@ -95,7 +95,7 @@ public void run(ImmutableContext context, DataReader reader) {
|
|||||||
filewriter.close();
|
filewriter.close();
|
||||||
|
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new SqoopException(CoreError.CORE_0018, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0018, e);
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -29,7 +29,7 @@
|
|||||||
import org.apache.hadoop.io.WritableComparator;
|
import org.apache.hadoop.io.WritableComparator;
|
||||||
import org.apache.hadoop.io.WritableUtils;
|
import org.apache.hadoop.io.WritableUtils;
|
||||||
import org.apache.sqoop.common.SqoopException;
|
import org.apache.sqoop.common.SqoopException;
|
||||||
import org.apache.sqoop.core.CoreError;
|
import org.apache.sqoop.job.MapreduceExecutionError;
|
||||||
|
|
||||||
public class Data implements WritableComparable<Data> {
|
public class Data implements WritableComparable<Data> {
|
||||||
|
|
||||||
@ -70,7 +70,7 @@ public void setContent(Object content, int type) {
|
|||||||
this.content = content;
|
this.content = content;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(type));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -81,7 +81,7 @@ public Object getContent(int targetType) {
|
|||||||
case ARRAY_RECORD:
|
case ARRAY_RECORD:
|
||||||
return parse();
|
return parse();
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(targetType));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(targetType));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -135,7 +135,7 @@ public int hashCode() {
|
|||||||
}
|
}
|
||||||
return result;
|
return result;
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(type));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -150,7 +150,7 @@ public void readFields(DataInput in) throws IOException {
|
|||||||
readArray(in);
|
readArray(in);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(type));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -165,7 +165,7 @@ public void write(DataOutput out) throws IOException {
|
|||||||
writeArray(out);
|
writeArray(out);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(type));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -243,7 +243,7 @@ private void readArray(DataInput in) throws IOException {
|
|||||||
|
|
||||||
default:
|
default:
|
||||||
throw new IOException(
|
throw new IOException(
|
||||||
new SqoopException(CoreError.CORE_0012, Integer.toString(type))
|
new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, Integer.toString(type))
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -301,8 +301,8 @@ private void writeArray(DataOutput out) throws IOException {
|
|||||||
|
|
||||||
} else {
|
} else {
|
||||||
throw new IOException(
|
throw new IOException(
|
||||||
new SqoopException(
|
new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012,
|
||||||
CoreError.CORE_0012, array[i].getClass().getName()
|
array[i].getClass().getName()
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
@ -345,7 +345,7 @@ private String format() {
|
|||||||
return sb.toString();
|
return sb.toString();
|
||||||
|
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(type));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -363,7 +363,7 @@ private Object[] parse() {
|
|||||||
return (Object[])content;
|
return (Object[])content;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0012, String.valueOf(type));
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0012, String.valueOf(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -25,8 +25,8 @@
|
|||||||
import org.apache.hadoop.io.NullWritable;
|
import org.apache.hadoop.io.NullWritable;
|
||||||
import org.apache.hadoop.mapreduce.Mapper;
|
import org.apache.hadoop.mapreduce.Mapper;
|
||||||
import org.apache.sqoop.common.SqoopException;
|
import org.apache.sqoop.common.SqoopException;
|
||||||
import org.apache.sqoop.core.CoreError;
|
|
||||||
import org.apache.sqoop.job.JobConstants;
|
import org.apache.sqoop.job.JobConstants;
|
||||||
|
import org.apache.sqoop.job.MapreduceExecutionError;
|
||||||
import org.apache.sqoop.job.PrefixContext;
|
import org.apache.sqoop.job.PrefixContext;
|
||||||
import org.apache.sqoop.job.etl.Extractor;
|
import org.apache.sqoop.job.etl.Extractor;
|
||||||
import org.apache.sqoop.job.io.Data;
|
import org.apache.sqoop.job.io.Data;
|
||||||
@ -67,7 +67,7 @@ public void run(Context context) throws IOException, InterruptedException {
|
|||||||
configJob = ConfigurationUtils.getFrameworkJob(conf);
|
configJob = ConfigurationUtils.getFrameworkJob(conf);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
throw new SqoopException(CoreError.CORE_0023);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0023);
|
||||||
}
|
}
|
||||||
|
|
||||||
SqoopSplit split = context.getCurrentKey();
|
SqoopSplit split = context.getCurrentKey();
|
||||||
@ -77,7 +77,7 @@ public void run(Context context) throws IOException, InterruptedException {
|
|||||||
new MapDataWriter(context));
|
new MapDataWriter(context));
|
||||||
|
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
throw new SqoopException(CoreError.CORE_0017, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0017, e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -118,7 +118,7 @@ public void writeContent(Object content, int type) {
|
|||||||
try {
|
try {
|
||||||
context.write(data, NullWritable.get());
|
context.write(data, NullWritable.get());
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
throw new SqoopException(CoreError.CORE_0013, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0013, e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -26,8 +26,8 @@
|
|||||||
import org.apache.hadoop.mapreduce.RecordWriter;
|
import org.apache.hadoop.mapreduce.RecordWriter;
|
||||||
import org.apache.hadoop.mapreduce.TaskAttemptContext;
|
import org.apache.hadoop.mapreduce.TaskAttemptContext;
|
||||||
import org.apache.sqoop.common.SqoopException;
|
import org.apache.sqoop.common.SqoopException;
|
||||||
import org.apache.sqoop.core.CoreError;
|
|
||||||
import org.apache.sqoop.job.JobConstants;
|
import org.apache.sqoop.job.JobConstants;
|
||||||
|
import org.apache.sqoop.job.MapreduceExecutionError;
|
||||||
import org.apache.sqoop.job.PrefixContext;
|
import org.apache.sqoop.job.PrefixContext;
|
||||||
import org.apache.sqoop.job.etl.Loader;
|
import org.apache.sqoop.job.etl.Loader;
|
||||||
import org.apache.sqoop.job.io.Data;
|
import org.apache.sqoop.job.io.Data;
|
||||||
@ -88,7 +88,7 @@ public void write(Data key, NullWritable value) {
|
|||||||
data.notify();
|
data.notify();
|
||||||
|
|
||||||
// throw exception
|
// throw exception
|
||||||
throw new SqoopException(CoreError.CORE_0015, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0015, e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -119,7 +119,7 @@ public void close(TaskAttemptContext context) {
|
|||||||
data.notify();
|
data.notify();
|
||||||
|
|
||||||
// throw exception
|
// throw exception
|
||||||
throw new SqoopException(CoreError.CORE_0015, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0015, e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -170,7 +170,7 @@ public Object readContent(int type) {
|
|||||||
data.notify();
|
data.notify();
|
||||||
|
|
||||||
// throw exception
|
// throw exception
|
||||||
throw new SqoopException(CoreError.CORE_0016, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0016, e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -201,7 +201,7 @@ public void run() {
|
|||||||
try {
|
try {
|
||||||
loader.run(frameworkContext, reader);
|
loader.run(frameworkContext, reader);
|
||||||
} catch (Throwable t) {
|
} catch (Throwable t) {
|
||||||
throw new SqoopException(CoreError.CORE_0018, t);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0018, t);
|
||||||
}
|
}
|
||||||
|
|
||||||
synchronized (data) {
|
synchronized (data) {
|
||||||
@ -214,7 +214,7 @@ public void run() {
|
|||||||
// if no exception happens yet
|
// if no exception happens yet
|
||||||
if (exception == null && !writerFinished) {
|
if (exception == null && !writerFinished) {
|
||||||
// create exception if data are not all consumed
|
// create exception if data are not all consumed
|
||||||
exception = new SqoopException(CoreError.CORE_0019);
|
exception = new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0019);
|
||||||
}
|
}
|
||||||
|
|
||||||
// throw deferred exception if exist
|
// throw deferred exception if exist
|
||||||
|
@ -24,7 +24,7 @@
|
|||||||
import org.apache.hadoop.io.Writable;
|
import org.apache.hadoop.io.Writable;
|
||||||
import org.apache.hadoop.mapreduce.InputSplit;
|
import org.apache.hadoop.mapreduce.InputSplit;
|
||||||
import org.apache.sqoop.common.SqoopException;
|
import org.apache.sqoop.common.SqoopException;
|
||||||
import org.apache.sqoop.core.CoreError;
|
import org.apache.sqoop.job.MapreduceExecutionError;
|
||||||
import org.apache.sqoop.job.etl.Partition;
|
import org.apache.sqoop.job.etl.Partition;
|
||||||
import org.apache.sqoop.utils.ClassUtils;
|
import org.apache.sqoop.utils.ClassUtils;
|
||||||
|
|
||||||
@ -60,12 +60,12 @@ public void readFields(DataInput in) throws IOException {
|
|||||||
// instantiate Partition object
|
// instantiate Partition object
|
||||||
Class<?> clz = ClassUtils.loadClass(className);
|
Class<?> clz = ClassUtils.loadClass(className);
|
||||||
if (clz == null) {
|
if (clz == null) {
|
||||||
throw new SqoopException(CoreError.CORE_0009, className);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0009, className);
|
||||||
}
|
}
|
||||||
try {
|
try {
|
||||||
partition = (Partition) clz.newInstance();
|
partition = (Partition) clz.newInstance();
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
throw new SqoopException(CoreError.CORE_0010, className, e);
|
throw new SqoopException(MapreduceExecutionError.MAPRED_EXEC_0010, className, e);
|
||||||
}
|
}
|
||||||
// read Partition object content
|
// read Partition object content
|
||||||
partition.readFields(in);
|
partition.readFields(in);
|
||||||
|
Loading…
Reference in New Issue
Block a user