.../apache/hadoop/hive/common/jsonexplain/DagJsonParserUtils.java | 2 +- common/src/java/org/apache/hadoop/hive/common/jsonexplain/Op.java | 6 +++--- .../hive/hcatalog/streaming/mutate/worker/GroupingValidator.java | 4 ++-- .../test/org/apache/hive/hcatalog/streaming/TestStreaming.java | 8 ++++---- hplsql/src/main/java/org/apache/hive/hplsql/Utils.java | 2 +- .../hive/ql/security/DummyHiveMetastoreAuthorizationProvider.java | 2 +- .../src/java/org/apache/hadoop/hive/llap/LlapRowRecordReader.java | 4 ++-- llap-common/src/java/org/apache/hadoop/hive/llap/Schema.java | 2 +- .../apache/hadoop/hive/llap/shufflehandler/ShuffleHandler.java | 2 +- .../java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java | 2 +- .../apache/hadoop/hive/metastore/hbase/HBaseFilterPlanUtil.java | 2 +- .../java/org/apache/hadoop/hive/metastore/hbase/HBaseUtils.java | 8 ++++---- .../ql/exec/vector/mapjoin/fast/VectorMapJoinFastValueStore.java | 2 +- ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java | 2 +- ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java | 2 +- .../ql/exec/vector/expressions/TestVectorStringExpressions.java | 6 +++--- .../serde2/binarysortable/fast/BinarySortableDeserializeRead.java | 2 +- .../hadoop/hive/serde2/lazy/fast/LazySimpleDeserializeRead.java | 2 +- .../hive/serde2/lazybinary/fast/LazyBinaryDeserializeRead.java | 2 +- 19 files changed, 31 insertions(+), 31 deletions(-) diff --git a/common/src/java/org/apache/hadoop/hive/common/jsonexplain/DagJsonParserUtils.java b/common/src/java/org/apache/hadoop/hive/common/jsonexplain/DagJsonParserUtils.java index a518ac1..8ea25d5 100644 --- a/common/src/java/org/apache/hadoop/hive/common/jsonexplain/DagJsonParserUtils.java +++ b/common/src/java/org/apache/hadoop/hive/common/jsonexplain/DagJsonParserUtils.java @@ -38,7 +38,7 @@ public static String renameReduceOutputOperator(String operatorName, Vertex vert } public static String attrsToString(Map attrs) { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); boolean first = true; for (Entry entry : attrs.entrySet()) { if (first) { diff --git a/common/src/java/org/apache/hadoop/hive/common/jsonexplain/Op.java b/common/src/java/org/apache/hadoop/hive/common/jsonexplain/Op.java index 2b853ef..39c44f1 100644 --- a/common/src/java/org/apache/hadoop/hive/common/jsonexplain/Op.java +++ b/common/src/java/org/apache/hadoop/hive/common/jsonexplain/Op.java @@ -170,7 +170,7 @@ else if (parentVertexes.size() == 1) { } } this.attrs.remove("keys:"); - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); JSONArray conditionMap = opObject.getJSONArray("condition map:"); for (int index = 0; index < conditionMap.length(); index++) { JSONObject cond = conditionMap.getJSONObject(index); @@ -250,7 +250,7 @@ else if (parentVertexes.size() == 1) { } // update the attrs this.attrs.remove("keys:"); - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); JSONArray conditionMap = opObject.getJSONArray("condition map:"); for (int index = 0; index < conditionMap.length(); index++) { JSONObject cond = conditionMap.getJSONObject(index); @@ -273,7 +273,7 @@ else if (parentVertexes.size() == 1) { } private String getNameWithOpIdStats() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); sb.append(DagJsonParserUtils.renameReduceOutputOperator(name, vertex)); if (operatorId != null) { sb.append(" [" + operatorId + "]"); diff --git a/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/GroupingValidator.java b/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/GroupingValidator.java index 18754f2..f28b8ff 100644 --- a/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/GroupingValidator.java +++ b/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/GroupingValidator.java @@ -31,14 +31,14 @@ class GroupingValidator { private final Map> visited; - private final StringBuffer partitionKeyBuilder; + private final StringBuilder partitionKeyBuilder; private long groups; private String lastPartitionKey; private int lastBucketId = -1; GroupingValidator() { visited = new HashMap>(); - partitionKeyBuilder = new StringBuffer(64); + partitionKeyBuilder = new StringBuilder(64); } /** diff --git a/hcatalog/streaming/src/test/org/apache/hive/hcatalog/streaming/TestStreaming.java b/hcatalog/streaming/src/test/org/apache/hive/hcatalog/streaming/TestStreaming.java index 097de9b..5e8fe62 100644 --- a/hcatalog/streaming/src/test/org/apache/hive/hcatalog/streaming/TestStreaming.java +++ b/hcatalog/streaming/src/test/org/apache/hive/hcatalog/streaming/TestStreaming.java @@ -1952,7 +1952,7 @@ private static Path getPartitionPath(Driver driver, String tableName, String par } private static String getTableColumnsStr(String[] colNames, String[] colTypes) { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); for (int i=0; i < colNames.length; ++i) { sb.append(colNames[i] + " " + colTypes[i]); if (i partVals) { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); for (int i=0; i < partVals.size(); ++i) { sb.append(partNames[i] + " = '" + partVals.get(i) + "'"); if(i < partVals.size()-1) { @@ -1992,7 +1992,7 @@ private static String getPartsSpec(String[] partNames, List partVals) { private static String join(String[] values, String delimiter) { if(values==null) return null; - StringBuffer strbuf = new StringBuffer(); + StringBuilder strbuf = new StringBuilder(); boolean first = true; diff --git a/hplsql/src/main/java/org/apache/hive/hplsql/Utils.java b/hplsql/src/main/java/org/apache/hive/hplsql/Utils.java index a768b00..d261df1 100644 --- a/hplsql/src/main/java/org/apache/hive/hplsql/Utils.java +++ b/hplsql/src/main/java/org/apache/hive/hplsql/Utils.java @@ -32,7 +32,7 @@ public static String unquoteString(String s) { } int len = s.length(); - StringBuffer s2 = new StringBuffer(len); + StringBuilder s2 = new StringBuilder(len); for (int i = 0; i < len; i++) { char ch = s.charAt(i); diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/ql/security/DummyHiveMetastoreAuthorizationProvider.java b/itests/util/src/main/java/org/apache/hadoop/hive/ql/security/DummyHiveMetastoreAuthorizationProvider.java index 346abd6..93637d4 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/ql/security/DummyHiveMetastoreAuthorizationProvider.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/ql/security/DummyHiveMetastoreAuthorizationProvider.java @@ -187,7 +187,7 @@ private void debugLog(String s) { } private String debugPrivPrint(Privilege[] privileges) { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); sb.append("Privileges{"); if (privileges != null){ for (Privilege p : privileges){ diff --git a/llap-client/src/java/org/apache/hadoop/hive/llap/LlapRowRecordReader.java b/llap-client/src/java/org/apache/hadoop/hive/llap/LlapRowRecordReader.java index e3c0955..d21d8f0 100644 --- a/llap-client/src/java/org/apache/hadoop/hive/llap/LlapRowRecordReader.java +++ b/llap-client/src/java/org/apache/hadoop/hive/llap/LlapRowRecordReader.java @@ -229,8 +229,8 @@ static void setRowFromStruct(Row row, Object structVal, StructObjectInspector so protected AbstractSerDe initSerDe(Configuration conf) throws SerDeException { Properties props = new Properties(); - StringBuffer columnsBuffer = new StringBuffer(); - StringBuffer typesBuffer = new StringBuffer(); + StringBuilder columnsBuffer = new StringBuilder(); + StringBuilder typesBuffer = new StringBuilder(); boolean isFirst = true; for (FieldDesc colDesc : schema.getColumns()) { if (!isFirst) { diff --git a/llap-common/src/java/org/apache/hadoop/hive/llap/Schema.java b/llap-common/src/java/org/apache/hadoop/hive/llap/Schema.java index c1bf4ea..08cc54d 100644 --- a/llap-common/src/java/org/apache/hadoop/hive/llap/Schema.java +++ b/llap-common/src/java/org/apache/hadoop/hive/llap/Schema.java @@ -43,7 +43,7 @@ public Schema() { @Override public String toString() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); boolean first = true; for (FieldDesc colDesc : getColumns()) { if (!first) { diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/shufflehandler/ShuffleHandler.java b/llap-server/src/java/org/apache/hadoop/hive/llap/shufflehandler/ShuffleHandler.java index e90f0df..f34b385 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/shufflehandler/ShuffleHandler.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/shufflehandler/ShuffleHandler.java @@ -813,7 +813,7 @@ public void operationComplete(ChannelFuture future) throws Exception { } private String getErrorMessage(Throwable t) { - StringBuffer sb = new StringBuffer(t.getMessage()); + StringBuilder sb = new StringBuilder(t.getMessage()); while (t.getCause() != null) { sb.append(t.getCause().getMessage()); t = t.getCause(); diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java b/metastore/src/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java index b1c8e39..a08c5bd 100644 --- a/metastore/src/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java +++ b/metastore/src/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java @@ -109,7 +109,7 @@ public Object invoke(final Object proxy, final Method method, final Object[] arg error = false; return result.result; } finally { - StringBuffer additionalInfo = new StringBuffer(); + StringBuilder additionalInfo = new StringBuilder(); additionalInfo.append("threadId=").append(threadId).append(" retryCount=").append(retryCount) .append(" error=").append(error); perfLogger.PerfLogEnd(CLASS_NAME, method.getName(), additionalInfo.toString()); diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseFilterPlanUtil.java b/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseFilterPlanUtil.java index 9762309..3c03846 100644 --- a/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseFilterPlanUtil.java +++ b/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseFilterPlanUtil.java @@ -473,7 +473,7 @@ public FilterPlan or(FilterPlan other) { @Override public String toString() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); sb.append("ScanPlan:\n"); for (Map.Entry entry : markers.entrySet()) { sb.append("key=" + entry.getKey() + "[startMarker=" + entry.getValue().startMarker diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseUtils.java b/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseUtils.java index 7dfa38f..64082e8 100644 --- a/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseUtils.java +++ b/metastore/src/java/org/apache/hadoop/hive/metastore/hbase/HBaseUtils.java @@ -956,9 +956,9 @@ static StorageDescriptorParts deserializePartition(String dbName, String tableNa private static List desierliazeDbNameTableNameFromPartitionKey(byte[] key, Configuration conf) { - StringBuffer names = new StringBuffer(); + StringBuilder names = new StringBuilder(); names.append("dbName,tableName,"); - StringBuffer types = new StringBuffer(); + StringBuilder types = new StringBuilder(); types.append("string,string,"); BinarySortableSerDe serDe = new BinarySortableSerDe(); Properties props = new Properties(); @@ -990,9 +990,9 @@ static StorageDescriptorParts deserializePartition(String dbName, String tableNa // Deserialize a partition key and return _only_ the partition values. private static List deserializePartitionKey(List partitions, byte[] key, Configuration conf) { - StringBuffer names = new StringBuffer(); + StringBuilder names = new StringBuilder(); names.append("dbName,tableName,"); - StringBuffer types = new StringBuffer(); + StringBuilder types = new StringBuilder(); types.append("string,string,"); for (int i=0;i cteAlias) ASTNode astNode = (ASTNode) queue.poll(); if (astNode.getToken().getType() == HiveParser.TOK_TABREF) { int aliasIndex = 0; - StringBuffer additionalTabInfo = new StringBuffer(); + StringBuilder additionalTabInfo = new StringBuilder(); for (int index = 1; index < astNode.getChildCount(); index++) { ASTNode ct = (ASTNode) astNode.getChild(index); if (ct.getToken().getType() == HiveParser.TOK_TABLEBUCKETSAMPLE diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java index 3daa83f..04686f7 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java @@ -148,7 +148,7 @@ public String getAlias() { @Explain(explainLevels = { Level.USER }) public String getTbl() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); sb.append(this.tableMetadata.getCompleteName()); sb.append("," + alias); if (isAcidTable()) { diff --git a/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorStringExpressions.java b/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorStringExpressions.java index d97152f..aa60878 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorStringExpressions.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/expressions/TestVectorStringExpressions.java @@ -4344,14 +4344,14 @@ private String randomizePattern(Random control, String value) { return value.toLowerCase(); } case 8: { - StringBuffer sb = new StringBuffer(8); + StringBuilder sb = new StringBuilder(); for (int i = 0; i < control.nextInt(12); i++) { sb.append((char) ('a' + control.nextInt(26))); } return sb.toString(); } case 9: { - StringBuffer sb = new StringBuffer(8); + StringBuilder sb = new StringBuilder(); for (int i = 0; i < control.nextInt(12); i++) { sb.append((char) ('A' + control.nextInt(26))); } @@ -4361,7 +4361,7 @@ private String randomizePattern(Random control, String value) { } private String generateCandidate(Random control, String pattern) { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(); final StringTokenizer tokens = new StringTokenizer(pattern, "%"); final boolean leftAnchor = pattern.startsWith("%"); final boolean rightAnchor = pattern.endsWith("%"); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/fast/BinarySortableDeserializeRead.java b/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/fast/BinarySortableDeserializeRead.java index d9160d7..b5ac842 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/fast/BinarySortableDeserializeRead.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/fast/BinarySortableDeserializeRead.java @@ -175,7 +175,7 @@ private void clearIndex(Field field) { * Get detailed read position information to help diagnose exceptions. */ public String getDetailedReadPositionString() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(64); sb.append("Reading inputByteBuffer of length "); sb.append(inputByteBuffer.getEnd()); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/LazySimpleDeserializeRead.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/LazySimpleDeserializeRead.java index 64e316b..9b3f3d0 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/LazySimpleDeserializeRead.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazy/fast/LazySimpleDeserializeRead.java @@ -360,7 +360,7 @@ public void set(byte[] bytes, int offset, int length) { * Get detailed read position information to help diagnose exceptions. */ public String getDetailedReadPositionString() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(64); sb.append("Reading byte[] of length "); sb.append(bytes.length); diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/fast/LazyBinaryDeserializeRead.java b/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/fast/LazyBinaryDeserializeRead.java index 8e0a499..c0c4d91 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/fast/LazyBinaryDeserializeRead.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/fast/LazyBinaryDeserializeRead.java @@ -173,7 +173,7 @@ private void clearIndex(Field field) { * Get detailed read position information to help diagnose exceptions. */ public String getDetailedReadPositionString() { - StringBuffer sb = new StringBuffer(); + StringBuilder sb = new StringBuilder(64); sb.append("Reading byte[] of length "); sb.append(bytes.length);