diff --git ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAbs.java ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAbs.java index ee7b821..9aeb9ae 100644 --- ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAbs.java +++ ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAbs.java @@ -28,6 +28,7 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.gen.FuncAbsLongToLong; import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.serde2.io.DoubleWritable; +import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector.Category; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorConverters; @@ -36,6 +37,7 @@ import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector.PrimitiveCategory; import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory; import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils; +import org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.LongWritable; @@ -55,6 +57,7 @@ private final DoubleWritable resultDouble = new DoubleWritable(); private final LongWritable resultLong = new LongWritable(); private final IntWritable resultInt = new IntWritable(); + private final HiveDecimalWritable resultDecimal = new HiveDecimalWritable(); private transient PrimitiveObjectInspector argumentOI; private transient Converter inputConverter; @@ -94,9 +97,10 @@ public ObjectInspector initialize(ObjectInspector[] arguments) throws UDFArgumen outputOI = PrimitiveObjectInspectorFactory.writableDoubleObjectInspector; break; case DECIMAL: + outputOI = PrimitiveObjectInspectorFactory.getPrimitiveWritableObjectInspector( + ((PrimitiveObjectInspector) arguments[0]).getTypeInfo()); inputConverter = ObjectInspectorConverters.getConverter(arguments[0], - PrimitiveObjectInspectorFactory.writableHiveDecimalObjectInspector); - outputOI = PrimitiveObjectInspectorFactory.writableHiveDecimalObjectInspector; + outputOI); break; default: throw new UDFArgumentException( @@ -129,11 +133,13 @@ public Object evaluate(DeferredObject[] arguments) throws HiveException { resultDouble.set(Math.abs(((DoubleWritable) valObject).get())); return resultDouble; case DECIMAL: - return PrimitiveObjectInspectorFactory.writableHiveDecimalObjectInspector.set( - PrimitiveObjectInspectorFactory.writableHiveDecimalObjectInspector - .create(HiveDecimal.ZERO), - PrimitiveObjectInspectorUtils.getHiveDecimal(valObject, - argumentOI).abs()); + WritableHiveDecimalObjectInspector decimalOI = + (WritableHiveDecimalObjectInspector) argumentOI; + HiveDecimalWritable val = decimalOI.getPrimitiveWritableObject(valObject); + if (val != null) { + val = (HiveDecimalWritable) decimalOI.set(resultDecimal, val.getHiveDecimal().abs()); + } + return val; default: throw new UDFArgumentException( "ABS only takes SHORT/BYTE/INT/LONG/DOUBLE/FLOAT/STRING/DECIMAL types, got " + inputType); diff --git ql/src/test/org/apache/hadoop/hive/ql/udf/generic/TestGenericUDFAbs.java ql/src/test/org/apache/hadoop/hive/ql/udf/generic/TestGenericUDFAbs.java index 1fe5361..8c531ea 100644 --- ql/src/test/org/apache/hadoop/hive/ql/udf/generic/TestGenericUDFAbs.java +++ ql/src/test/org/apache/hadoop/hive/ql/udf/generic/TestGenericUDFAbs.java @@ -28,7 +28,9 @@ import org.apache.hadoop.hive.serde2.io.DoubleWritable; import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; +import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory; +import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; import org.apache.hadoop.io.FloatWritable; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.LongWritable; @@ -135,10 +137,17 @@ public void testText() throws HiveException { public void testHiveDecimal() throws HiveException { GenericUDFAbs udf = new GenericUDFAbs(); - ObjectInspector valueOI = PrimitiveObjectInspectorFactory.writableHiveDecimalObjectInspector; + int prec = 12; + int scale = 9; + ObjectInspector valueOI = PrimitiveObjectInspectorFactory.getPrimitiveWritableObjectInspector( + TypeInfoFactory.getDecimalTypeInfo(prec, scale)); ObjectInspector[] arguments = {valueOI}; - udf.initialize(arguments); + PrimitiveObjectInspector outputOI = (PrimitiveObjectInspector) udf.initialize(arguments); + // Make sure result precision/scale matches the input prec/scale + assertEquals("result precision for abs()", prec, outputOI.precision()); + assertEquals("result scale for abs()", scale, outputOI.scale()); + DeferredObject valueObj = new DeferredJavaObject(new HiveDecimalWritable(HiveDecimal.create( "107.123456789"))); DeferredObject[] args = {valueObj}; @@ -153,5 +162,15 @@ public void testHiveDecimal() throws HiveException { assertEquals("abs() test for HiveDecimal failed ", 107.123456789, output.getHiveDecimal() .doubleValue()); + + // null input + args[0] = new DeferredJavaObject(null); + output = (HiveDecimalWritable) udf.evaluate(args); + assertEquals("abs(null)", null, output); + + // if value too large, should also be null + args[0] = new DeferredJavaObject(new HiveDecimalWritable(HiveDecimal.create("-1000.123456"))); + output = (HiveDecimalWritable) udf.evaluate(args); + assertEquals("abs() of too large decimal value", null, output); } }