diff --git accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java index 718a5c5..7730783 100644 --- accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java +++ accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java @@ -134,7 +134,7 @@ public static AccumuloPredicateHandler getInstance() { * * @param udfType * GenericUDF classname to lookup matching CompareOpt - * @return Class extends CompareOpt/> + * @return Class<? extends CompareOpt/> */ public Class extends CompareOp> getCompareOpClass(String udfType) throws NoSuchCompareOpException { @@ -166,7 +166,7 @@ public CompareOp getCompareOp(String udfType, IndexSearchCondition sc) * * @param type * String hive column lookup matching PrimitiveCompare - * @return Class extends >?> + * @return Class<? extends ></?> */ public Class extends PrimitiveComparison> getPrimitiveComparisonClass(String type) throws NoSuchPrimitiveComparisonException { diff --git accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloCompositeRowId.java accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloCompositeRowId.java index f3ebbd1..2367ca2 100644 --- accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloCompositeRowId.java +++ accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloCompositeRowId.java @@ -33,21 +33,21 @@ * AccumuloCompositeKey extension of LazyStruct. All complex composite keys should extend this class * and override the {@link LazyStruct#getField(int)} method where fieldID corresponds to the ID of a * key in the composite key. - *
+ *
* For example, for a composite key "/part1/part2/part3", part1 will have an id
* 0, part2 will have an id 1 and part3 will have an id 2. Custom
* implementations of getField(fieldID) should return the value corresponding to that fieldID. So,
- * for the above example, the value returned for getField(0) should be part1,
+ * for the above example, the value returned for getField(0) should be part1,
* getField(1) should be part2 and getField(2) should be part3.
- *
+ *
* All custom implementation are expected to have a constructor of the form:
*
*
* MyCustomCompositeKey(LazySimpleStructObjectInspector oi, Properties tbl, Configuration conf) *- * + *
+ *
* For example, for a composite key "/part1/part2/part3", part1 will have an id
* 0, part2 will have an id 1 and part3 will have an id 2. Custom
* implementations of getField(fieldID) should return the value corresponding to that fieldID. So,
- * for the above example, the value returned for getField(0) should be part1,
+ * for the above example, the value returned for getField(0) should be part1,
* getField(1) should be part2 and getField(2) should be part3.
- *
+ *
+ *
* All custom implementation are expected to have a constructor of the form:
*
*
* MyCustomCompositeKey(LazySimpleStructObjectInspector oi, Properties tbl, Configuration conf) ** - * * * */ public class HBaseCompositeKey extends LazyStruct { diff --git hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDe.java hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDe.java index c2e7808..de6b3a6 100644 --- hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDe.java +++ hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDe.java @@ -154,7 +154,7 @@ public static ColumnMappings parseColumnsMapping( * @param columnsMappingSpec string hbase.columns.mapping specified when creating table * @param doColumnRegexMatching whether to do a regex matching on the columns or not * @param hideColumnPrefix whether to hide a prefix of column mapping in key name in a map (works only if @doColumnRegexMatching is true) - * @return List
+ *
* For example, for a value structure "/part1/part2/part3", part1 will have an id
* 0, part2 will have an id 1 and part3 will have an id 2. Custom
* implementations of getField(fieldID) should return the value corresponding to that fieldID. So,
- * for the above example, the value returned for getField(0) should be part1,
+ * for the above example, the value returned for getField(0) should be part1,
* getField(1) should be part2 and getField(2) should be part3.
- *
+ *
+ *
* All implementation are expected to have a constructor of the form
*
*
* MyCustomStructObject(LazySimpleStructObjectInspector oi, Properties props, Configuration conf, ColumnMapping colMap) ** - * * */ public class HBaseStructValue extends LazyStruct { diff --git hbase-handler/src/java/org/apache/hadoop/hive/hbase/struct/HBaseValueFactory.java hbase-handler/src/java/org/apache/hadoop/hive/hbase/struct/HBaseValueFactory.java index 3fead1e..e660a2a 100644 --- hbase-handler/src/java/org/apache/hadoop/hive/hbase/struct/HBaseValueFactory.java +++ hbase-handler/src/java/org/apache/hadoop/hive/hbase/struct/HBaseValueFactory.java @@ -38,7 +38,7 @@ /** * Initialize factory with properties * - * @param hbaseParam the {@link HBaseParameters hbase parameters} + * @param hbaseParam the HBaseParameters hbase parameters * @param conf the hadoop {@link Configuration configuration} * @param properties the custom {@link Properties} * @throws SerDeException if there was an issue initializing the factory @@ -67,7 +67,7 @@ void init(HBaseSerDeParameters hbaseParam, Configuration conf, Properties proper * @param object the object to be serialized * @param field the {@link StructField} * @return the serialized value - * @throws {@link IOException} if there was an issue serializing the value + * @throws IOException if there was an issue serializing the value */ byte[] serializeValue(Object object, StructField field) throws IOException; } \ No newline at end of file diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatConstants.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatConstants.java index 3998407..924fd7c 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatConstants.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatConstants.java @@ -186,7 +186,7 @@ private HCatConstants() { // restrict instantiation /** * {@value} (default: {@value #HCAT_DATA_CONVERT_BOOLEAN_TO_INTEGER_DEFAULT}). - * Pig < 0.10.0 does not have boolean support, and scripts written for pre-boolean Pig versions + * Pig < 0.10.0 does not have boolean support, and scripts written for pre-boolean Pig versions * will not expect boolean values when upgrading Pig. For integration the option is offered to * convert boolean fields to integers by setting this Hadoop configuration key. */ diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java index 8b927af..4cb4d39 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/common/HCatUtil.java @@ -289,11 +289,11 @@ public static HCatSchema getPartitionColumns(Table table) throws IOException { * Test if the first FsAction is more permissive than the second. This is * useful in cases where we want to ensure that a file owner has more * permissions than the group they belong to, for eg. More completely(but - * potentially more cryptically) owner-r >= group-r >= world-r : bitwise - * and-masked with 0444 => 444 >= 440 >= 400 >= 000 owner-w >= group-w >= - * world-w : bitwise and-masked with &0222 => 222 >= 220 >= 200 >= 000 - * owner-x >= group-x >= world-x : bitwise and-masked with &0111 => 111 >= - * 110 >= 100 >= 000 + * potentially more cryptically) owner-r >= group-r >= world-r : bitwise + * and-masked with 0444 => 444 >= 440 >= 400 >= 000 owner-w >= group-w >= + * world-w : bitwise and-masked with &0222 => 222 >= 220 >= 200 >= 000 + * owner-x >= group-x >= world-x : bitwise and-masked with &0111 => 111 >= + * 110 >= 100 >= 000 * * @return true if first FsAction is more permissive than the second, false * if not. @@ -563,9 +563,9 @@ public static IMetaStoreClient getHiveMetastoreClient(HiveConf hiveConf) /** * Get or create a hive client depending on whether it exits in cache or not. - * @Deprecated : use {@link #getHiveMetastoreClient(HiveConf)} instead. + * @deprecated : use {@link #getHiveMetastoreClient(HiveConf)} instead. * This was deprecated in Hive 1.2, slated for removal in two versions - * (i.e. 1.2 & 1.3(projected) will have it, but it will be removed after that) + * (i.e. 1.2 & 1.3(projected) will have it, but it will be removed after that) * @param hiveConf The hive configuration * @return the client * @throws MetaException When HiveMetaStoreClient couldn't be created diff --git hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java index da6ad5b..694771d 100644 --- hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java +++ hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java @@ -58,7 +58,7 @@ * * @param recordItr * {@link Iterator} records to be written into HCatalog. - * @throws {@link HCatException} + * @throws HCatException */ public abstract void write(final Iterator
The optional capacity bound constructor argument serves as a * way to prevent excessive expansion. The capacity, if unspecified, * is equal to {@link Integer#MAX_VALUE}. - *
+ *This class and its iterator implement all of the * optional methods of the {@link Collection} and {@link * Iterator} interfaces. - *
+ *This method is equivalent to {@link #addLast}.
+ *
+ * This method is equivalent to {@link #addLast}.
*
* @throws IllegalStateException if the element cannot be added at this
* time due to capacity restrictions
@@ -481,8 +481,8 @@ public boolean offer(E e, long timeout, TimeUnit unit)
* Retrieves and removes the head of the queue represented by this deque.
* This method differs from {@link #poll poll} only in that it throws an
* exception if this deque is empty.
- *
This method is equivalent to {@link #removeFirst() removeFirst}.
+ *
+ * This method is equivalent to {@link #removeFirst() removeFirst}.
*
* @return the head of the queue represented by this deque
* @throws NoSuchElementException if this deque is empty
@@ -508,8 +508,8 @@ public E poll(long timeout, TimeUnit unit) throws InterruptedException {
* Retrieves, but does not remove, the head of the queue represented by
* this deque. This method differs from {@link #peek peek} only in that
* it throws an exception if this deque is empty.
- *
This method is equivalent to {@link #getFirst() getFirst}.
+ *
+ * This method is equivalent to {@link #getFirst() getFirst}.
*
* @return the head of the queue represented by this deque
* @throws NoSuchElementException if this deque is empty
@@ -528,8 +528,8 @@ public E peek() {
* (in the absence of memory or resource constraints) accept without
* blocking. This is always equal to the initial capacity of this deque
* less the current size of this deque.
- *
Note that you cannot always tell if an attempt to insert
+ *
+ * Note that you cannot always tell if an attempt to insert
* an element will succeed by inspecting remainingCapacity
* because it may be the case that another thread is about to
* insert or remove an element.
@@ -622,8 +622,8 @@ public E pop() {
* o.equals(e) (if such an element exists).
* Returns true if this deque contained the specified element
* (or equivalently, if this deque changed as a result of the call).
- *
This method is equivalent to
+ *
+ * This method is equivalent to
* {@link #removeFirstOccurrence(Object) removeFirstOccurrence}.
*
* @param o element to be removed from this deque, if present
@@ -671,12 +671,12 @@ public boolean contains(Object o) {
/**
* Returns an array containing all of the elements in this deque, in
* proper sequence (from first to last element).
- *
The returned array will be "safe" in that no references to it are * maintained by this deque. (In other words, this method must allocate * a new array). The caller is thus free to modify the returned array. - *
- *This method acts as bridge between array-based and collection-based + *
+ *If this deque fits in the specified array with room to spare * (i.e., the array has more elements than this deque), the element in * the array immediately following the end of the deque is set to * null. - *
+ * *Like the {@link #toArray()} method, this method acts as bridge between * array-based and collection-based APIs. Further, this method allows * precise control over the runtime type of the output array, and may, * under certain circumstances, be used to save allocation costs. - *
+ * *Suppose x is a deque known to contain only strings. * The following code can be used to dump the deque into a newly * allocated array of String: - *
+ * *
* String[] y = x.toArray(new String[0]);
- *
+ *
+ *
*