diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomPartitionVertex.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomPartitionVertex.java index 26afe90faa..ef148d9b05 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomPartitionVertex.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomPartitionVertex.java @@ -108,14 +108,15 @@ public int compare(InputSplit inp1, InputSplit inp2) { private final Multimap bucketToTaskMap = HashMultimap. create(); private final Map> inputToGroupedSplitMap = - new HashMap>(); + new HashMap<>(); private int numInputsAffectingRootInputSpecUpdate = 1; private int numInputsSeenSoFar = 0; private final Map emMap = Maps.newHashMap(); private final List finalSplits = Lists.newLinkedList(); private final Map inputNameInputSpecMap = - new HashMap(); + new HashMap<>(); + private Map inputToBucketMap; public CustomPartitionVertex(VertexManagerPluginContext context) { super(context); @@ -137,6 +138,7 @@ public void initialize() { this.mainWorkName = vertexConf.getInputName(); this.vertexType = vertexConf.getVertexType(); this.numInputsAffectingRootInputSpecUpdate = vertexConf.getNumInputs(); + this.inputToBucketMap = vertexConf.getInputToBucketMap(); } @Override @@ -242,7 +244,7 @@ public void onRootVertexInitialized(String inputName, InputDescriptor inputDescr } Multimap bucketToInitialSplitMap = - getBucketSplitMapForPath(pathFileSplitsMap); + getBucketSplitMapForPath(inputName, pathFileSplitsMap); try { int totalResource = context.getTotalAvailableResource().getMemory(); @@ -532,20 +534,47 @@ private FileSplit getFileSplitFromEvent(InputDataInformationEvent event) throws /* * This method generates the map of bucket to file splits. */ - private Multimap getBucketSplitMapForPath( + private Multimap getBucketSplitMapForPath(String inputName, Map> pathFileSplitsMap) { - int bucketNum = 0; Multimap bucketToInitialSplitMap = - ArrayListMultimap. create(); + ArrayListMultimap.create(); + boolean fallback = false; + List bucketIds = new ArrayList<>(); for (Map.Entry> entry : pathFileSplitsMap.entrySet()) { - int bucketId = bucketNum % numBuckets; + // Extract the buckedID from pathFilesMap, this is more accurate method, + // however. it may not work in certain cases where buckets are named + // after files used while loading data. In such case, fallback to old + // potential inaccurate method. + // The accepted file names are such as 000000_0, 000001_0_copy_1. + String bucketIdStr = + Utilities.getBucketFileNameFromPathSubString(entry.getKey()); + int bucketId = Utilities.getBucketIdFromFile(bucketIdStr); + if (bucketId == -1) { + fallback = true; + LOG.info("Fallback to using older sort based logic to assign " + + "buckets to splits."); + bucketIds.clear(); + break; + } + bucketIds.add(bucketId); for (FileSplit fsplit : entry.getValue()) { bucketToInitialSplitMap.put(bucketId, fsplit); } - bucketNum++; + } + + int bucketNum = 0; + if (fallback) { + // This is the old logic which assumes that the filenames are sorted in + // alphanumeric order and mapped to appropriate bucket number. + for (Map.Entry> entry : pathFileSplitsMap.entrySet()) { + for (FileSplit fsplit : entry.getValue()) { + bucketToInitialSplitMap.put(bucketNum, fsplit); + } + bucketNum++; + } } // this is just for SMB join use-case. The numBuckets would be equal to that of the big table @@ -553,16 +582,28 @@ private FileSplit getFileSplitFromEvent(InputDataInformationEvent event) throws // data from the right buckets to the big table side. For e.g. Big table has 8 buckets and small // table has 4 buckets, bucket 0 of small table needs to be sent to bucket 4 of the big table as // well. - if (bucketNum < numBuckets) { - int loopedBucketId = 0; - for (; bucketNum < numBuckets; bucketNum++) { - for (InputSplit fsplit : bucketToInitialSplitMap.get(loopedBucketId)) { - bucketToInitialSplitMap.put(bucketNum, fsplit); + if (numInputsAffectingRootInputSpecUpdate != 1 && + inputName.compareTo(mainWorkName) != 0) { + // small table + int inputNumBuckets = inputToBucketMap.get(inputName); + if (fallback && bucketNum != inputNumBuckets) { + // The fallback mechanism kicked in which only works correctly if + // there exists a file for each bucket, else it may result in wrong + // result. Throw an error + + } + if (inputNumBuckets < numBuckets) { + // Need to send the splits to multiple buckets + for (int i = 1; i < numBuckets/inputNumBuckets; i++) { + int bucketIdBase = i * inputNumBuckets; + for (Integer bucketId : bucketIds) { + for (InputSplit fsplit : bucketToInitialSplitMap.get(bucketId)) { + bucketToInitialSplitMap.put(bucketIdBase + bucketId, fsplit); + } + } } - loopedBucketId++; } } - return bucketToInitialSplitMap; } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomVertexConfiguration.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomVertexConfiguration.java index ef5e7edcd6..4301829517 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomVertexConfiguration.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/CustomVertexConfiguration.java @@ -21,7 +21,10 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import java.util.HashMap; +import java.util.Map; +import com.google.common.base.Preconditions; import org.apache.hadoop.hive.ql.plan.TezWork.VertexType; import org.apache.hadoop.io.Writable; @@ -39,22 +42,24 @@ private VertexType vertexType = VertexType.AUTO_INITIALIZED_EDGES; private int numInputs; private String inputName; + private Map inputToBucketMap; public CustomVertexConfiguration() { } // this is the constructor to use for the Bucket map join case. public CustomVertexConfiguration(int numBuckets, VertexType vertexType) { - this(numBuckets, vertexType, "", 1); + this(numBuckets, vertexType, "", 1, null); } // this is the constructor to use for SMB. public CustomVertexConfiguration(int numBuckets, VertexType vertexType, String inputName, - int numInputs) { + int numInputs, Map inputToBucketMap) { this.numBuckets = numBuckets; this.vertexType = vertexType; this.numInputs = numInputs; this.inputName = inputName; + this.inputToBucketMap = inputToBucketMap; } @Override @@ -63,6 +68,14 @@ public void write(DataOutput out) throws IOException { out.writeInt(this.numBuckets); out.writeInt(numInputs); out.writeUTF(inputName); + int sz = inputToBucketMap != null ? inputToBucketMap.size() : 0; + out.writeInt(sz); + if (sz > 0) { + for (Map.Entry entry : inputToBucketMap.entrySet()) { + out.writeUTF(entry.getKey()); + out.writeInt(entry.getValue()); + } + } } @Override @@ -71,6 +84,16 @@ public void readFields(DataInput in) throws IOException { this.numBuckets = in.readInt(); this.numInputs = in.readInt(); this.inputName = in.readUTF(); + int sz = in.readInt(); + Preconditions.checkState(sz >= 0); + if (sz == 0) { + this.inputToBucketMap = null; + } else { + this.inputToBucketMap = new HashMap<>(); + for (int i = 0; i < sz; i++) { + this.inputToBucketMap.put(in.readUTF(), in.readInt()); + } + } } public int getNumBuckets() { @@ -88,4 +111,8 @@ public String getInputName() { public int getNumInputs() { return numInputs; } + + public Map getInputToBucketMap() { + return inputToBucketMap; + } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java index 9885038588..0e75f6e5e8 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java @@ -21,6 +21,7 @@ import java.util.concurrent.ConcurrentHashMap; import com.google.common.base.Function; +import com.google.common.base.Preconditions; import com.google.common.collect.Iterators; import com.google.common.collect.Lists; import javax.security.auth.login.LoginException; @@ -568,13 +569,26 @@ private Vertex createVertex(JobConf conf, MergeJoinWork mergeJoinWork, FileSyste MultiMRInput.createConfigBuilder(conf, HiveInputFormat.class).build()); } + // To be populated for SMB joins only for all the small tables + Map inputToBucketMap = new HashMap<>(); + if (mergeJoinWork.getMergeJoinOperator().getParentOperators().size() == 1 + && mergeJoinWork.getMergeJoinOperator().getOpTraits() != null) { + // This is an SMB join. + for (BaseWork work : mapWorkList) { + MapWork mw = (MapWork) work; + Map> aliasToWork = mw.getAliasToWork(); + Preconditions.checkState(aliasToWork.size() == 1, + "More than 1 alias in SMB mapwork"); + inputToBucketMap.put(mw.getName(), mw.getWorks().get(0).getOpTraits().getNumBuckets()); + } + } VertexManagerPluginDescriptor desc = VertexManagerPluginDescriptor.create(CustomPartitionVertex.class.getName()); // the +1 to the size is because of the main work. CustomVertexConfiguration vertexConf = new CustomVertexConfiguration(mergeJoinWork.getMergeJoinOperator().getConf() .getNumBuckets(), vertexType, mergeJoinWork.getBigTableAlias(), - mapWorkList.size() + 1); + mapWorkList.size() + 1, inputToBucketMap); DataOutputBuffer dob = new DataOutputBuffer(); vertexConf.write(dob); byte[] userPayload = dob.getData(); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java index 632a21390d..4ee0579b4b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java @@ -42,15 +42,7 @@ import org.apache.hadoop.hive.metastore.HiveMetaStoreUtils; import org.apache.hadoop.hive.metastore.TableType; import org.apache.hadoop.hive.metastore.Warehouse; -import org.apache.hadoop.hive.metastore.api.BasicTxnInfo; -import org.apache.hadoop.hive.metastore.api.CreationMetadata; -import org.apache.hadoop.hive.metastore.api.FieldSchema; -import org.apache.hadoop.hive.metastore.api.MetaException; -import org.apache.hadoop.hive.metastore.api.Order; -import org.apache.hadoop.hive.metastore.api.SerDeInfo; -import org.apache.hadoop.hive.metastore.api.SkewedInfo; -import org.apache.hadoop.hive.metastore.api.StorageDescriptor; -import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; +import org.apache.hadoop.hive.metastore.api.*; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.io.HiveFileFormatUtils; @@ -188,6 +180,8 @@ public void setTTable(org.apache.hadoop.hive.metastore.api.Table tTable) { t.setOwner(SessionState.getUserFromAuthenticator()); // set create time t.setCreateTime((int) (System.currentTimeMillis() / 1000)); + t.setBucketingVersion(BucketingVersion.MURMUR_BUCKETING); + t.setLoadInBucketedTable(false); } return t; } @@ -676,6 +670,10 @@ public int getNumBuckets() { return tTable.getSd().getNumBuckets(); } + public BucketingVersion getBucketingVersion() { + return tTable.getBucketingVersion(); + } + public void setInputFormatClass(String name) throws HiveException { if (name == null) { inputFormatClass = null; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConvertJoinMapJoin.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConvertJoinMapJoin.java index dc698c8de8..f30cf4e086 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConvertJoinMapJoin.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConvertJoinMapJoin.java @@ -26,9 +26,11 @@ import java.util.Set; import java.util.Stack; +import com.google.common.base.Preconditions; import org.apache.hadoop.hive.common.JavaUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; +import org.apache.hadoop.hive.metastore.api.BucketingVersion; import org.apache.hadoop.hive.ql.exec.AppMasterEventOperator; import org.apache.hadoop.hive.ql.exec.CommonJoinOperator; import org.apache.hadoop.hive.ql.exec.CommonMergeJoinOperator; @@ -180,7 +182,8 @@ MapJoinOperator mapJoinOp = convertJoinMapJoin(joinOp, context, mapJoinConversionPos, true); // map join operator by default has no bucket cols and num of reduce sinks // reduced by 1 - mapJoinOp.setOpTraits(new OpTraits(null, -1, null, joinOp.getOpTraits().getNumReduceSinks())); + mapJoinOp.setOpTraits(new OpTraits(null, -1, null, + joinOp.getOpTraits().getNumReduceSinks(), joinOp.getOpTraits().getBucketingVersion())); mapJoinOp.setStatistics(joinOp.getStatistics()); // propagate this change till the next RS for (Operator childOp : mapJoinOp.getChildOperators()) { @@ -378,7 +381,7 @@ private void convertJoinSMBJoin(JoinOperator joinOp, OptimizeTezProcContext cont joinOp.getSchema()); int numReduceSinks = joinOp.getOpTraits().getNumReduceSinks(); OpTraits opTraits = new OpTraits(joinOp.getOpTraits().getBucketColNames(), numBuckets, - joinOp.getOpTraits().getSortCols(), numReduceSinks); + joinOp.getOpTraits().getSortCols(), numReduceSinks, joinOp.getOpTraits().getBucketingVersion()); mergeJoinOp.setOpTraits(opTraits); mergeJoinOp.setStatistics(joinOp.getStatistics()); @@ -445,7 +448,8 @@ private void setAllChildrenTraits(Operator currentOp, Op return; } currentOp.setOpTraits(new OpTraits(opTraits.getBucketColNames(), - opTraits.getNumBuckets(), opTraits.getSortCols(), opTraits.getNumReduceSinks())); + opTraits.getNumBuckets(), opTraits.getSortCols(), opTraits.getNumReduceSinks(), + opTraits.getBucketingVersion())); for (Operator childOp : currentOp.getChildOperators()) { if ((childOp instanceof ReduceSinkOperator) || (childOp instanceof GroupByOperator)) { break; @@ -498,7 +502,8 @@ private boolean convertJoinBucketMapJoin(JoinOperator joinOp, OptimizeTezProcCon // we can set the traits for this join operator opTraits = new OpTraits(joinOp.getOpTraits().getBucketColNames(), - tezBucketJoinProcCtx.getNumBuckets(), null, joinOp.getOpTraits().getNumReduceSinks()); + tezBucketJoinProcCtx.getNumBuckets(), null, + joinOp.getOpTraits().getNumReduceSinks(), joinOp.getOpTraits().getBucketingVersion()); mapJoinOp.setOpTraits(opTraits); mapJoinOp.setStatistics(joinOp.getStatistics()); setNumberOfBucketsOnChildren(mapJoinOp); @@ -576,6 +581,13 @@ private boolean checkConvertJoinSMBJoin(JoinOperator joinOp, OptimizeTezProcCont return false; } ReduceSinkOperator rsOp = (ReduceSinkOperator) parentOp; + // If the chosen big table has less number of buckets than any of the + // small tables, then those buckets will have no mapping to any of the + // big table buckets resulting in wrong results. + if (numBuckets > 0 && numBuckets < rsOp.getOpTraits().getNumBuckets()) { + LOG.info("Small table has more buckets than big table."); + return false; + } if (!checkColEquality(rsOp.getParentOperators().get(0).getOpTraits().getSortCols(), rsOp .getOpTraits().getSortCols(), rsOp.getColumnExprMap(), false)) { LOG.info("We cannot convert to SMB because the sort column names do not match."); @@ -593,6 +605,37 @@ private boolean checkConvertJoinSMBJoin(JoinOperator joinOp, OptimizeTezProcCont numBuckets = bigTableRS.getConf().getNumReducers(); } tezBucketJoinProcCtx.setNumBuckets(numBuckets); + + // With bucketing using two different versions. Version 1 for exiting + // tables and version 2 for new tables. All the inputs to the SMB must be + // from same version. This only applies to tables read directly and not + // intermediate outputs of joins/groupbys + BucketingVersion version = BucketingVersion.INVALID_BUCKETING; + for (Operator parentOp : joinOp.getParentOperators()) { + // Check if the parent is coming from a table scan, if so, what is the version of it. + assert parentOp.getParentOperators() != null && parentOp.getParentOperators().size() == 1; + Operator op = parentOp.getParentOperators().get(0); + while(op != null && !(op instanceof TableScanOperator + || op instanceof ReduceSinkOperator + || op instanceof CommonJoinOperator)) { + // If op has parents it is guaranteed to be 1. + List> parents = op.getParentOperators(); + Preconditions.checkState(parents.size() == 0 || parents.size() == 1); + op = parents.size() == 1 ? parents.get(0) : null; + } + + if (op instanceof TableScanOperator) { + BucketingVersion localVersion = ((TableScanOperator)op).getConf(). + getTableMetadata().getBucketingVersion(); + if (version == BucketingVersion.INVALID_BUCKETING) { + version = localVersion; + } else if (version != localVersion) { + // versions dont match, return false. + LOG.debug("SMB Join can't be performed due to bucketing version mismatch"); + return false; + } + } + } LOG.info("We can convert the join to an SMB join."); return true; } @@ -1168,7 +1211,8 @@ private boolean convertJoinDynamicPartitionedHashJoin(JoinOperator joinOp, Optim joinOp.getOpTraits().getBucketColNames(), numReducers, null, - joinOp.getOpTraits().getNumReduceSinks()); + joinOp.getOpTraits().getNumReduceSinks(), + joinOp.getOpTraits().getBucketingVersion()); mapJoinOp.setOpTraits(opTraits); mapJoinOp.setStatistics(joinOp.getStatistics()); // propagate this change till the next RS diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java index 69d9f3125a..76964028bd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java @@ -22,6 +22,7 @@ import java.util.Map.Entry; import org.apache.hadoop.hive.conf.HiveConf; +import org.apache.hadoop.hive.metastore.api.BucketingVersion; import org.apache.hadoop.hive.metastore.api.Order; import org.apache.hadoop.hive.ql.exec.GroupByOperator; import org.apache.hadoop.hive.ql.exec.JoinOperator; @@ -92,10 +93,12 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, List> listBucketCols = new ArrayList>(); int numBuckets = -1; int numReduceSinks = 1; + BucketingVersion bucketingVersion = BucketingVersion.INVALID_BUCKETING; OpTraits parentOpTraits = rs.getParentOperators().get(0).getOpTraits(); if (parentOpTraits != null) { numBuckets = parentOpTraits.getNumBuckets(); numReduceSinks += parentOpTraits.getNumReduceSinks(); + bucketingVersion = parentOpTraits.getBucketingVersion(); } List bucketCols = new ArrayList<>(); @@ -134,7 +137,8 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, } listBucketCols.add(bucketCols); - OpTraits opTraits = new OpTraits(listBucketCols, numBuckets, listBucketCols, numReduceSinks); + OpTraits opTraits = new OpTraits(listBucketCols, numBuckets, + listBucketCols, numReduceSinks, bucketingVersion); rs.setOpTraits(opTraits); return null; } @@ -213,7 +217,8 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, sortedColsList.add(sortCols); } // num reduce sinks hardcoded to 0 because TS has no parents - OpTraits opTraits = new OpTraits(bucketColsList, numBuckets, sortedColsList, 0); + OpTraits opTraits = new OpTraits(bucketColsList, numBuckets, + sortedColsList, 0, table.getBucketingVersion()); ts.setOpTraits(opTraits); return null; } @@ -239,12 +244,15 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, List> listBucketCols = new ArrayList>(); int numReduceSinks = 0; + BucketingVersion bucketingVersion = BucketingVersion.INVALID_BUCKETING; OpTraits parentOpTraits = gbyOp.getParentOperators().get(0).getOpTraits(); if (parentOpTraits != null) { numReduceSinks = parentOpTraits.getNumReduceSinks(); + bucketingVersion = parentOpTraits.getBucketingVersion(); } listBucketCols.add(gbyKeys); - OpTraits opTraits = new OpTraits(listBucketCols, -1, listBucketCols, numReduceSinks); + OpTraits opTraits = new OpTraits(listBucketCols, -1, listBucketCols, + numReduceSinks, bucketingVersion); gbyOp.setOpTraits(opTraits); return null; } @@ -298,12 +306,15 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, int numBuckets = -1; int numReduceSinks = 0; + BucketingVersion bucketingVersion = BucketingVersion.INVALID_BUCKETING; OpTraits parentOpTraits = selOp.getParentOperators().get(0).getOpTraits(); if (parentOpTraits != null) { numBuckets = parentOpTraits.getNumBuckets(); numReduceSinks = parentOpTraits.getNumReduceSinks(); + bucketingVersion = parentOpTraits.getBucketingVersion(); } - OpTraits opTraits = new OpTraits(listBucketCols, numBuckets, listSortCols, numReduceSinks); + OpTraits opTraits = new OpTraits(listBucketCols, numBuckets, listSortCols, + numReduceSinks, bucketingVersion); selOp.setOpTraits(opTraits); return null; } @@ -319,6 +330,7 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, List> sortColsList = new ArrayList>(); byte pos = 0; int numReduceSinks = 0; // will be set to the larger of the parents + boolean bucketingVersionSeen = false; for (Operator parentOp : joinOp.getParentOperators()) { if (!(parentOp instanceof ReduceSinkOperator)) { // can be mux operator @@ -338,7 +350,7 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, pos++; } - joinOp.setOpTraits(new OpTraits(bucketColsList, -1, bucketColsList, numReduceSinks)); + joinOp.setOpTraits(new OpTraits(bucketColsList, -1, bucketColsList, numReduceSinks, BucketingVersion.INVALID_BUCKETING)); return null; } @@ -392,6 +404,8 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, Operator operator = (Operator) nd; int numReduceSinks = 0; + BucketingVersion bucketingVersion = BucketingVersion.INVALID_BUCKETING; + boolean bucketingVersionSeen = false; for (Operator parentOp : operator.getParentOperators()) { if (parentOp.getOpTraits() == null) { continue; @@ -399,8 +413,17 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, if (parentOp.getOpTraits().getNumReduceSinks() > numReduceSinks) { numReduceSinks = parentOp.getOpTraits().getNumReduceSinks(); } + // If there is mismatch in bucketingVersion, then it should be set to + // -1, that way SMB will be disabled. + if (bucketingVersion == BucketingVersion.INVALID_BUCKETING && !bucketingVersionSeen) { + bucketingVersion = parentOp.getOpTraits().getBucketingVersion(); + bucketingVersionSeen = true; + } else if (bucketingVersion != parentOp.getOpTraits().getBucketingVersion()) { + bucketingVersion = BucketingVersion.INVALID_BUCKETING; + } } - OpTraits opTraits = new OpTraits(null, -1, null, numReduceSinks); + OpTraits opTraits = new OpTraits(null, -1, + null, numReduceSinks, bucketingVersion); operator.setOpTraits(opTraits); return null; } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java index bacc44482a..39d2370435 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java @@ -121,7 +121,8 @@ } // we can set the traits for this join operator - OpTraits opTraits = new OpTraits(bucketColNames, numBuckets, null, joinOp.getOpTraits().getNumReduceSinks()); + OpTraits opTraits = new OpTraits(bucketColNames, numBuckets, null, + joinOp.getOpTraits().getNumReduceSinks(), joinOp.getOpTraits().getBucketingVersion()); mapJoinOp.setOpTraits(opTraits); mapJoinOp.setStatistics(joinOp.getStatistics()); setNumberOfBucketsOnChildren(mapJoinOp); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java index 54f5bab6de..361976311d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java @@ -23,6 +23,7 @@ import java.io.Serializable; import java.net.URI; import java.net.URISyntaxException; +import java.util.Arrays; import java.util.LinkedHashMap; import java.util.List; import java.util.Map; @@ -160,6 +161,48 @@ private URI initializeFromURI(String fromPath, boolean isLocal) throws IOExcepti "source contains directory: " + oneSrc.getPath().toString())); } } + // Do another loop if table is bucketed + List bucketCols = table.getBucketCols(); + if (bucketCols != null && !bucketCols.isEmpty()) { + // Hive assumes that user names the files as per the corresponding + // bucket. For e.g, file names should follow the format 000000_0, 000000_1 etc. + // Here the 1st file will belong to bucket 0 and 2nd to bucket 1 and so on. + boolean[] bucketArray = new boolean[table.getNumBuckets()]; + // initialize the array + Arrays.fill(bucketArray, false); + int numBuckets = table.getNumBuckets(); + + for (FileStatus oneSrc : srcs) { + String bucketName = oneSrc.getPath().getName(); + + //get the bucket id + String bucketIdStr = + Utilities.getBucketFileNameFromPathSubString(bucketName); + int bucketId = Utilities.getBucketIdFromFile(bucketIdStr); + LOG.debug("bucket ID for file " + oneSrc.getPath() + " = " + bucketId + + " for table " + table.getFullyQualifiedName()); + if (bucketId == -1) { + throw new SemanticException(ErrorMsg.INVALID_PATH.getMsg( + "The file name is invalid : " + + oneSrc.getPath().toString() + " for table " + + table.getFullyQualifiedName())); + } + if (bucketId >= numBuckets) { + throw new SemanticException(ErrorMsg.INVALID_PATH.getMsg( + "The file name corresponds to invalid bucketId : " + + oneSrc.getPath().toString()) + + ". Maximum number of buckets can be " + numBuckets + + " for table " + table.getFullyQualifiedName()); + } + if (bucketArray[bucketId]) { + throw new SemanticException(ErrorMsg.INVALID_PATH.getMsg( + "Multiple files for same bucket : " + bucketId + + ". Only 1 file per bucket allowed in single load command. To load multiple files for same bucket, use multiple statements for table " + + table.getFullyQualifiedName())); + } + bucketArray[bucketId] = true; + } + } } catch (IOException e) { // Has to use full name to make sure it does not conflict with // org.apache.commons.lang.StringUtils diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/OpTraits.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/OpTraits.java index 9621c3be53..0dcc229e15 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/OpTraits.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/OpTraits.java @@ -18,21 +18,26 @@ package org.apache.hadoop.hive.ql.plan; +import org.apache.hadoop.hive.metastore.api.BucketingVersion; + import java.util.List; public class OpTraits { - List> bucketColNames; - List> sortColNames; - int numBuckets; - int numReduceSinks; + private List> bucketColNames; + private List> sortColNames; + private int numBuckets; + private int numReduceSinks; + private BucketingVersion bucketingVersion; public OpTraits(List> bucketColNames, int numBuckets, - List> sortColNames, int numReduceSinks) { + List> sortColNames, int numReduceSinks, + BucketingVersion bucketingVersion) { this.bucketColNames = bucketColNames; this.numBuckets = numBuckets; this.sortColNames = sortColNames; this.numReduceSinks = numReduceSinks; + this.bucketingVersion = bucketingVersion; } public List> getBucketColNames() { @@ -68,6 +73,13 @@ public int getNumReduceSinks() { return this.numReduceSinks; } + public void setBucketingVersion(BucketingVersion bucketingVersion) { + this.bucketingVersion = bucketingVersion; + } + + public BucketingVersion getBucketingVersion() { + return bucketingVersion; + } @Override public String toString() { diff --git a/ql/src/test/org/apache/hadoop/hive/ql/metadata/TestHive.java b/ql/src/test/org/apache/hadoop/hive/ql/metadata/TestHive.java index b5b478fbad..5355e064f4 100755 --- a/ql/src/test/org/apache/hadoop/hive/ql/metadata/TestHive.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/metadata/TestHive.java @@ -35,12 +35,7 @@ import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.metastore.PartitionDropOptions; import org.apache.hadoop.hive.metastore.Warehouse; -import org.apache.hadoop.hive.metastore.api.BasicTxnInfo; -import org.apache.hadoop.hive.metastore.api.Database; -import org.apache.hadoop.hive.metastore.api.FieldSchema; -import org.apache.hadoop.hive.metastore.api.Index; -import org.apache.hadoop.hive.metastore.api.MetaException; -import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; +import org.apache.hadoop.hive.metastore.api.*; import org.apache.hadoop.hive.ql.index.HiveIndex; import org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat; import org.apache.hadoop.hive.ql.session.SessionState; @@ -171,6 +166,8 @@ public void testTable() throws Throwable { tbl.setStoredAsSubDirectories(false); tbl.setRewriteEnabled(false); + tbl.getTTable().setBucketingVersion(BucketingVersion.MURMUR_BUCKETING); + tbl.getTTable().setLoadInBucketedTable(false); // create table setNullCreateTableGrants(); diff --git a/ql/src/test/queries/clientpositive/auto_sortmerge_join_2.q b/ql/src/test/queries/clientpositive/auto_sortmerge_join_2.q index e5fdcb57e4..b7bd10eed2 100644 --- a/ql/src/test/queries/clientpositive/auto_sortmerge_join_2.q +++ b/ql/src/test/queries/clientpositive/auto_sortmerge_join_2.q @@ -1,19 +1,21 @@ set hive.strict.checks.bucketing=false; set hive.mapred.mode=nonstrict; --- small 1 part, 4 bucket & big 2 part, 2 bucket -CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; +-- small 1 part, 2 bucket & big 2 part, 4 bucket +CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/small/000000_0' INTO TABLE bucket_small partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/small/000001_0' INTO TABLE bucket_small partition(ds='2008-04-08'); -load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08'); -load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08'); -CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; +CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/big/000001_0' INTO TABLE bucket_big partition(ds='2008-04-08'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09'); load data local inpath '../../data/files/auto_sortmerge_join/big/000001_0' INTO TABLE bucket_big partition(ds='2008-04-09'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09'); set hive.auto.convert.join=true; set hive.auto.convert.sortmerge.join=true; diff --git a/ql/src/test/queries/clientpositive/auto_sortmerge_join_4.q b/ql/src/test/queries/clientpositive/auto_sortmerge_join_4.q index abf09e5534..9f719aebb5 100644 --- a/ql/src/test/queries/clientpositive/auto_sortmerge_join_4.q +++ b/ql/src/test/queries/clientpositive/auto_sortmerge_join_4.q @@ -1,7 +1,7 @@ set hive.strict.checks.bucketing=false; set hive.mapred.mode=nonstrict; --- small 2 part, 4 bucket & big 1 part, 2 bucket +-- small 2 part, 4 bucket & big 1 part, 4 bucket CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/small/000000_0' INTO TABLE bucket_small partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/small/000001_0' INTO TABLE bucket_small partition(ds='2008-04-08'); @@ -13,9 +13,11 @@ load data local inpath '../../data/files/auto_sortmerge_join/small/000001_0' INT load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-09'); load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-09'); -CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; +CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/big/000001_0' INTO TABLE bucket_big partition(ds='2008-04-08'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08'); set hive.auto.convert.join=true; set hive.auto.convert.sortmerge.join=true; diff --git a/ql/src/test/queries/clientpositive/auto_sortmerge_join_5.q b/ql/src/test/queries/clientpositive/auto_sortmerge_join_5.q index b85c4a7aa3..c107501d0a 100644 --- a/ql/src/test/queries/clientpositive/auto_sortmerge_join_5.q +++ b/ql/src/test/queries/clientpositive/auto_sortmerge_join_5.q @@ -1,19 +1,19 @@ set hive.strict.checks.bucketing=false; set hive.mapred.mode=nonstrict; --- small no part, 4 bucket & big no part, 2 bucket +-- small no part, 2 bucket & big no part, 4 bucket -- SORT_QUERY_RESULTS -CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; +CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/small/000000_0' INTO TABLE bucket_small; load data local inpath '../../data/files/auto_sortmerge_join/small/000001_0' INTO TABLE bucket_small; -load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small; -load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small; -CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; +CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big; load data local inpath '../../data/files/auto_sortmerge_join/big/000001_0' INTO TABLE bucket_big; +load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big; +load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big; set hive.auto.convert.sortmerge.join=true; set hive.optimize.bucketmapjoin = true; diff --git a/ql/src/test/queries/clientpositive/auto_sortmerge_join_7.q b/ql/src/test/queries/clientpositive/auto_sortmerge_join_7.q index bd780861e3..a5cc04a97f 100644 --- a/ql/src/test/queries/clientpositive/auto_sortmerge_join_7.q +++ b/ql/src/test/queries/clientpositive/auto_sortmerge_join_7.q @@ -1,7 +1,7 @@ set hive.strict.checks.bucketing=false; set hive.mapred.mode=nonstrict; --- small 2 part, 4 bucket & big 2 part, 2 bucket +-- small 2 part, 4 bucket & big 2 part, 4 bucket CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/small/000000_0' INTO TABLE bucket_small partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/small/000001_0' INTO TABLE bucket_small partition(ds='2008-04-08'); @@ -13,12 +13,16 @@ load data local inpath '../../data/files/auto_sortmerge_join/small/000001_0' INT load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-09'); load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-09'); -CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; +CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/big/000001_0' INTO TABLE bucket_big partition(ds='2008-04-08'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08'); load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09'); load data local inpath '../../data/files/auto_sortmerge_join/big/000001_0' INTO TABLE bucket_big partition(ds='2008-04-09'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09'); +load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09'); set hive.auto.convert.join=true; set hive.auto.convert.sortmerge.join=true; diff --git a/ql/src/test/results/clientnegative/bucket_mapjoin_mismatch1.q.out b/ql/src/test/results/clientnegative/bucket_mapjoin_mismatch1.q.out index b9c2e6f827..37dbbf9ff3 100644 --- a/ql/src/test/results/clientnegative/bucket_mapjoin_mismatch1.q.out +++ b/ql/src/test/results/clientnegative/bucket_mapjoin_mismatch1.q.out @@ -53,174 +53,4 @@ POSTHOOK: query: CREATE TABLE srcbucket_mapjoin_part_2 (key int, value string) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@srcbucket_mapjoin_part_2 -PREHOOK: query: load data local inpath '../../data/files/bmj/000002_0' - INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@srcbucket_mapjoin_part_2 -POSTHOOK: query: load data local inpath '../../data/files/bmj/000002_0' - INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@srcbucket_mapjoin_part_2 -POSTHOOK: Output: default@srcbucket_mapjoin_part_2@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/bmj/000003_0' - INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@srcbucket_mapjoin_part_2@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/bmj/000003_0' - INTO TABLE srcbucket_mapjoin_part_2 partition(ds='2008-04-08') -POSTHOOK: type: LOAD #### A masked pattern was here #### -POSTHOOK: Output: default@srcbucket_mapjoin_part_2@ds=2008-04-08 -PREHOOK: query: explain -select a.key, a.value, b.value -from srcbucket_mapjoin_part a join srcbucket_mapjoin_part_2 b -on a.key=b.key and a.ds="2008-04-08" and b.ds="2008-04-08" -PREHOOK: type: QUERY -POSTHOOK: query: explain -select a.key, a.value, b.value -from srcbucket_mapjoin_part a join srcbucket_mapjoin_part_2 b -on a.key=b.key and a.ds="2008-04-08" and b.ds="2008-04-08" -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 108 Data size: 42000 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 108 Data size: 42000 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 108 Data size: 42000 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 108 Data size: 42000 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - TableScan - alias: b - Statistics: Num rows: 78 Data size: 30620 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 78 Data size: 30620 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 78 Data size: 30620 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 78 Data size: 30620 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col4 - Statistics: Num rows: 118 Data size: 46200 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 118 Data size: 46200 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 118 Data size: 46200 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -select /*+mapjoin(b)*/ a.key, a.value, b.value -from srcbucket_mapjoin_part a join srcbucket_mapjoin_part_2 b -on a.key=b.key and a.ds="2008-04-08" and b.ds="2008-04-08" -PREHOOK: type: QUERY -POSTHOOK: query: explain -select /*+mapjoin(b)*/ a.key, a.value, b.value -from srcbucket_mapjoin_part a join srcbucket_mapjoin_part_2 b -on a.key=b.key and a.ds="2008-04-08" and b.ds="2008-04-08" -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-1 depends on stages: Stage-3 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-3 - Map Reduce Local Work - Alias -> Map Local Tables: - b - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - b - TableScan - alias: b - Statistics: Num rows: 102 Data size: 30620 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 102 Data size: 30620 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 key (type: int) - 1 key (type: int) - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 140 Data size: 42000 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 140 Data size: 42000 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: int) - 1 key (type: int) - outputColumnNames: _col0, _col1, _col7 - Statistics: Num rows: 154 Data size: 46200 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col7 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 154 Data size: 46200 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 154 Data size: 46200 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -FAILED: SemanticException [Error 10136]: Bucketed mapjoin cannot be performed. This can be due to multiple reasons: . Join columns dont match bucketed columns. . Number of buckets are not a multiple of each other. If you really want to perform the operation, either remove the mapjoin hint from your query or set hive.enforce.bucketmapjoin to false. diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out index 5cfc35aa73..dda72115c5 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out @@ -1,8 +1,8 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_small @@ -23,27 +23,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -64,6 +48,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09') PREHOOK: type: LOAD #### A masked pattern was here #### @@ -81,6 +81,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -95,16 +111,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -134,7 +150,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -142,7 +158,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -150,7 +166,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -158,7 +174,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -183,7 +199,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -191,7 +207,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -199,7 +215,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -207,7 +223,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -308,7 +324,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -316,7 +332,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 partition_columns ds partition_columns.types string @@ -324,7 +340,7 @@ STAGE PLANS: serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -332,7 +348,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -353,16 +369,16 @@ STAGE PLANS: $hdt$_1:b TableScan alias: b - Statistics: Num rows: 4 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1140 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 4 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 4 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1140 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -374,16 +390,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -414,7 +430,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -422,7 +438,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -430,7 +446,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -438,7 +454,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -463,7 +479,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -471,7 +487,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -479,7 +495,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -487,7 +503,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -511,7 +527,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -519,7 +535,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 partition_columns ds partition_columns.types string @@ -527,7 +543,7 @@ STAGE PLANS: serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -535,7 +551,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -597,7 +613,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -605,7 +621,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -613,7 +629,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -621,7 +637,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -645,7 +661,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -653,7 +669,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -661,7 +677,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -669,7 +685,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -690,16 +706,16 @@ STAGE PLANS: $hdt$_0:a TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -711,16 +727,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 4 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1140 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 4 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 4 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -751,7 +767,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -759,7 +775,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -767,7 +783,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -775,7 +791,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -800,7 +816,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -808,7 +824,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -816,7 +832,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -824,7 +840,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -848,7 +864,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -856,7 +872,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 partition_columns ds partition_columns.types string @@ -864,7 +880,7 @@ STAGE PLANS: serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -872,7 +888,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -924,16 +940,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -963,7 +979,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -971,7 +987,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -979,7 +995,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -987,7 +1003,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1012,7 +1028,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1020,7 +1036,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1028,7 +1044,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1036,7 +1052,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out index 0d586fd26b..b54c574358 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out @@ -72,11 +72,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -97,6 +97,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -111,16 +127,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -150,7 +166,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -158,7 +174,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -166,7 +182,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -174,7 +190,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -243,7 +259,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -258,16 +274,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -297,7 +313,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -305,7 +321,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -313,7 +329,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -321,7 +337,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -390,7 +406,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -534,16 +550,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -574,7 +590,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -582,7 +598,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -590,7 +606,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -598,7 +614,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -755,7 +771,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -763,7 +779,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -771,7 +787,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -779,7 +795,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -800,16 +816,16 @@ STAGE PLANS: $hdt$_0:a TableScan alias: a - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -861,7 +877,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -869,7 +885,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -877,7 +893,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -885,7 +901,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1034,16 +1050,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -1073,7 +1089,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1081,7 +1097,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1089,7 +1105,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1097,7 +1113,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1166,4 +1182,4 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out index 45704d1253..451c3b3353 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out @@ -1,8 +1,8 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_small @@ -22,27 +22,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -62,6 +46,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -76,16 +76,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -114,7 +114,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -122,13 +122,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -136,7 +136,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -144,13 +144,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -216,16 +216,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -254,7 +254,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -262,13 +262,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -276,7 +276,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -284,13 +284,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -369,16 +369,16 @@ STAGE PLANS: $hdt$_1:b TableScan alias: b - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -390,16 +390,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -429,7 +429,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -437,13 +437,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -451,7 +451,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -459,13 +459,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -475,7 +475,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -492,7 +492,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -500,13 +500,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small @@ -551,16 +551,16 @@ STAGE PLANS: $hdt$_0:a TableScan alias: a - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -572,16 +572,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -611,7 +611,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -619,13 +619,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -633,7 +633,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -641,13 +641,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -657,7 +657,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -674,7 +674,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -682,13 +682,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small @@ -728,16 +728,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -766,7 +766,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -774,13 +774,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -788,7 +788,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -796,13 +796,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out index 1959075912..f335142360 100644 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out @@ -72,11 +72,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -97,6 +97,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09') PREHOOK: type: LOAD #### A masked pattern was here #### @@ -114,6 +130,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -128,16 +160,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -167,7 +199,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -175,7 +207,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -183,7 +215,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -191,7 +223,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -216,7 +248,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -224,7 +256,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -232,7 +264,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -240,7 +272,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -312,7 +344,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -327,16 +359,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -366,7 +398,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -374,7 +406,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -382,7 +414,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -390,7 +422,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -415,7 +447,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -423,7 +455,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -431,7 +463,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -439,7 +471,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -511,7 +543,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -655,16 +687,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -695,7 +727,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -703,7 +735,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -711,7 +743,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -719,7 +751,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -744,7 +776,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -752,7 +784,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -760,7 +792,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -768,7 +800,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -926,7 +958,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -934,7 +966,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -942,7 +974,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -950,7 +982,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -974,7 +1006,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -982,7 +1014,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -990,7 +1022,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -998,7 +1030,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1019,16 +1051,16 @@ STAGE PLANS: $hdt$_0:a TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -1080,7 +1112,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1088,7 +1120,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1096,7 +1128,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1104,7 +1136,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1129,7 +1161,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1137,7 +1169,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1145,7 +1177,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1153,7 +1185,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1302,16 +1334,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -1341,7 +1373,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1349,7 +1381,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1357,7 +1389,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1365,7 +1397,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1390,7 +1422,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1398,7 +1430,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1406,7 +1438,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1414,7 +1446,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1486,4 +1518,4 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out index 054b0d00be..d4472cf2a0 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out @@ -1,8 +1,8 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_small @@ -23,27 +23,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -64,6 +48,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09') PREHOOK: type: LOAD #### A masked pattern was here #### @@ -81,6 +81,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -101,16 +117,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 4 Data size: 2996 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1508 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 4 Data size: 2996 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1508 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 4 Data size: 2996 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1508 Basic stats: COMPLETE Column stats: NONE Path -> Alias: #### A masked pattern was here #### Path -> Partition: @@ -121,7 +137,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -129,7 +145,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 partition_columns ds partition_columns.types string @@ -137,7 +153,7 @@ STAGE PLANS: serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -145,7 +161,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -167,16 +183,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 74872 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 158376 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -184,7 +200,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 117 Data size: 78681 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 165502 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -208,7 +224,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -216,7 +232,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -224,7 +240,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -232,7 +248,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -256,7 +272,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -264,7 +280,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -272,7 +288,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -280,7 +296,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -375,16 +391,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 4 Data size: 2996 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1508 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 4 Data size: 2996 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1508 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 4 Data size: 2996 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 1508 Basic stats: COMPLETE Column stats: NONE Path -> Alias: #### A masked pattern was here #### Path -> Partition: @@ -395,7 +411,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -403,7 +419,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 partition_columns ds partition_columns.types string @@ -411,7 +427,7 @@ STAGE PLANS: serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -419,7 +435,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -441,16 +457,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 74872 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 158376 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -458,7 +474,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 117 Data size: 78681 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 165502 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -482,7 +498,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -490,7 +506,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -498,7 +514,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -506,7 +522,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -530,7 +546,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -538,7 +554,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -546,7 +562,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -554,7 +570,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out index 95d329862c..5cd5d798bc 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out @@ -72,11 +72,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -97,6 +97,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -232,16 +248,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 56 Data size: 37620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 79280 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 54 Data size: 36276 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 114 Data size: 75316 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 54 Data size: 36276 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 114 Data size: 75316 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -249,7 +265,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 1 - Statistics: Num rows: 59 Data size: 39903 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 125 Data size: 82847 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -273,7 +289,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -281,7 +297,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -289,7 +305,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -297,7 +313,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -370,7 +386,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -506,16 +522,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 37620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 79280 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 54 Data size: 36276 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 114 Data size: 75316 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 54 Data size: 36276 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 114 Data size: 75316 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -523,7 +539,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 59 Data size: 39903 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 125 Data size: 82847 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -547,7 +563,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -555,7 +571,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -563,7 +579,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -571,7 +587,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -644,7 +660,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -780,16 +796,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 37620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 79280 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 54 Data size: 36276 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 114 Data size: 75316 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 54 Data size: 36276 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 114 Data size: 75316 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -797,7 +813,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 59 Data size: 39903 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 125 Data size: 82847 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -821,7 +837,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -829,7 +845,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -837,7 +853,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -845,7 +861,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -918,4 +934,4 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out index e711715aa5..a18f4b21fc 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out @@ -1,8 +1,8 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_small @@ -22,27 +22,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -62,6 +46,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -101,7 +101,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -109,13 +109,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -123,7 +123,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -131,13 +131,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small @@ -187,7 +187,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -195,13 +195,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -209,7 +209,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -217,13 +217,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -318,7 +318,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -326,13 +326,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -340,7 +340,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -348,13 +348,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small @@ -404,7 +404,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -412,13 +412,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -426,7 +426,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -434,13 +434,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -562,7 +562,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -570,13 +570,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -584,7 +584,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -592,13 +592,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -639,7 +639,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -647,13 +647,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -661,7 +661,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -669,13 +669,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out index 53c685cb11..fdea211fa4 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out @@ -72,11 +72,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -97,6 +97,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09') PREHOOK: type: LOAD #### A masked pattern was here #### @@ -114,6 +130,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -249,16 +281,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 112 Data size: 74872 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 158376 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -266,7 +298,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 1 - Statistics: Num rows: 117 Data size: 78681 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 165502 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -290,7 +322,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -298,7 +330,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -306,7 +338,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -314,7 +346,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -338,7 +370,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -346,7 +378,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -354,7 +386,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -362,7 +394,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -438,7 +470,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -574,16 +606,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 74872 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 158376 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -591,7 +623,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 117 Data size: 78681 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 165502 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -615,7 +647,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -623,7 +655,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -631,7 +663,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -639,7 +671,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -663,7 +695,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -671,7 +703,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -679,7 +711,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -687,7 +719,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -763,7 +795,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -899,16 +931,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 74872 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 158376 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 107 Data size: 71529 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 228 Data size: 150457 Basic stats: COMPLETE Column stats: NONE Merge Join Operator condition map: Inner Join 0 to 1 @@ -916,7 +948,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 117 Data size: 78681 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 165502 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -940,7 +972,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -948,7 +980,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -956,7 +988,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -964,7 +996,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -988,7 +1020,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -996,7 +1028,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -1004,7 +1036,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1012,7 +1044,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -1088,4 +1120,4 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out index 8cfa113794..117ff4aecc 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_2.q.out @@ -1,8 +1,8 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_small @@ -23,27 +23,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -64,6 +48,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09') PREHOOK: type: LOAD #### A masked pattern was here #### @@ -81,6 +81,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -100,16 +116,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -117,7 +133,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 123 Data size: 60500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 264 Data size: 127864 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -142,7 +158,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -150,7 +166,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -158,7 +174,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -166,7 +182,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -191,7 +207,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -199,7 +215,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -207,7 +223,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -215,7 +231,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -308,16 +324,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -325,7 +341,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 123 Data size: 60500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 264 Data size: 127864 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -350,7 +366,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -358,7 +374,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -366,7 +382,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -374,7 +390,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -399,7 +415,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -407,7 +423,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -415,7 +431,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -423,7 +439,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out index fce5e0cfc4..aff5a0d242 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out @@ -72,11 +72,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -97,6 +97,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -116,16 +132,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -133,7 +149,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 1 - Statistics: Num rows: 61 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 132 Data size: 63932 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -158,7 +174,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -166,7 +182,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -174,7 +190,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -182,7 +198,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -254,7 +270,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -274,16 +290,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -291,7 +307,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 61 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 132 Data size: 63932 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -316,7 +332,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -324,7 +340,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -332,7 +348,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -340,7 +356,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -412,7 +428,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -432,16 +448,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 56 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 120 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -449,7 +465,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 61 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 132 Data size: 63932 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -474,7 +490,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -482,7 +498,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -490,7 +506,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -498,7 +514,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -570,4 +586,4 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -38 +78 diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out index 8250eca099..6255dd2819 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out @@ -1,8 +1,8 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_small @@ -22,27 +22,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000002_0' INTO TABLE bucket_small -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/000003_0' INTO TABLE bucket_small -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@bucket_small -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -62,6 +46,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -81,16 +81,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -98,7 +98,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 1 - Statistics: Num rows: 1 Data size: 2486 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1254 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -122,7 +122,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -130,13 +130,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -144,7 +144,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -152,13 +152,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -232,16 +232,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -249,7 +249,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 1 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 63932 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -273,7 +273,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -281,13 +281,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -295,7 +295,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -303,13 +303,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big @@ -382,16 +382,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 2260 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1140 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 _col0 (type: string) @@ -414,7 +414,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -422,13 +422,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -436,7 +436,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 4 + bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value @@ -444,13 +444,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 + numFiles 2 numRows 0 rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 + totalSize 114 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small @@ -468,16 +468,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 58120 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -487,7 +487,7 @@ STAGE PLANS: input vertices: 1 Map 3 Position of Big Table: 0 - Statistics: Num rows: 1 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 63932 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -518,7 +518,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -526,13 +526,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -540,7 +540,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -548,13 +548,13 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out index eb813c1734..ac5cd47fbb 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out @@ -72,11 +72,11 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/sm POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@bucket_big @@ -97,6 +97,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-08 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-08') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-08 PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000000_0' INTO TABLE bucket_big partition(ds='2008-04-09') PREHOOK: type: LOAD #### A masked pattern was here #### @@ -114,6 +130,22 @@ POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/bi POSTHOOK: type: LOAD #### A masked pattern was here #### POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000002_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 +PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@bucket_big@ds=2008-04-09 +POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/000003_0' INTO TABLE bucket_big partition(ds='2008-04-09') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@bucket_big@ds=2008-04-09 PREHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key @@ -133,16 +165,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -150,7 +182,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 1 - Statistics: Num rows: 123 Data size: 60500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 264 Data size: 127864 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -175,7 +207,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -183,7 +215,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -191,7 +223,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -199,7 +231,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -224,7 +256,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -232,7 +264,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -240,7 +272,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -248,7 +280,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -323,7 +355,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -343,16 +375,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -360,7 +392,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 123 Data size: 60500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 264 Data size: 127864 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -385,7 +417,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -393,7 +425,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -401,7 +433,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -409,7 +441,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -434,7 +466,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -442,7 +474,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -450,7 +482,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -458,7 +490,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -533,7 +565,7 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 PREHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key PREHOOK: type: QUERY POSTHOOK: query: explain extended select count(*) FROM bucket_big a JOIN bucket_small b ON a.key = b.key @@ -553,16 +585,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 112 Data size: 55000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 240 Data size: 116240 Basic stats: COMPLETE Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -570,7 +602,7 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) Position of Big Table: 0 - Statistics: Num rows: 123 Data size: 60500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 264 Data size: 127864 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() @@ -595,7 +627,7 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -603,7 +635,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -611,7 +643,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -619,7 +651,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -644,7 +676,7 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -652,7 +684,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 + numFiles 4 numRows 0 partition_columns ds partition_columns.types string @@ -660,7 +692,7 @@ STAGE PLANS: serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -668,7 +700,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 + bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value @@ -743,4 +775,4 @@ POSTHOOK: Input: default@bucket_small POSTHOOK: Input: default@bucket_small@ds=2008-04-08 POSTHOOK: Input: default@bucket_small@ds=2008-04-09 #### A masked pattern was here #### -76 +156 diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp index af0fd6b0e0..615c024ad0 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp +++ b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp @@ -1240,14 +1240,14 @@ uint32_t ThriftHiveMetastore_get_databases_result::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1049; - ::apache::thrift::protocol::TType _etype1052; - xfer += iprot->readListBegin(_etype1052, _size1049); - this->success.resize(_size1049); - uint32_t _i1053; - for (_i1053 = 0; _i1053 < _size1049; ++_i1053) + uint32_t _size1050; + ::apache::thrift::protocol::TType _etype1053; + xfer += iprot->readListBegin(_etype1053, _size1050); + this->success.resize(_size1050); + uint32_t _i1054; + for (_i1054 = 0; _i1054 < _size1050; ++_i1054) { - xfer += iprot->readString(this->success[_i1053]); + xfer += iprot->readString(this->success[_i1054]); } xfer += iprot->readListEnd(); } @@ -1286,10 +1286,10 @@ uint32_t ThriftHiveMetastore_get_databases_result::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1054; - for (_iter1054 = this->success.begin(); _iter1054 != this->success.end(); ++_iter1054) + std::vector ::const_iterator _iter1055; + for (_iter1055 = this->success.begin(); _iter1055 != this->success.end(); ++_iter1055) { - xfer += oprot->writeString((*_iter1054)); + xfer += oprot->writeString((*_iter1055)); } xfer += oprot->writeListEnd(); } @@ -1334,14 +1334,14 @@ uint32_t ThriftHiveMetastore_get_databases_presult::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1055; - ::apache::thrift::protocol::TType _etype1058; - xfer += iprot->readListBegin(_etype1058, _size1055); - (*(this->success)).resize(_size1055); - uint32_t _i1059; - for (_i1059 = 0; _i1059 < _size1055; ++_i1059) + uint32_t _size1056; + ::apache::thrift::protocol::TType _etype1059; + xfer += iprot->readListBegin(_etype1059, _size1056); + (*(this->success)).resize(_size1056); + uint32_t _i1060; + for (_i1060 = 0; _i1060 < _size1056; ++_i1060) { - xfer += iprot->readString((*(this->success))[_i1059]); + xfer += iprot->readString((*(this->success))[_i1060]); } xfer += iprot->readListEnd(); } @@ -1458,14 +1458,14 @@ uint32_t ThriftHiveMetastore_get_all_databases_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1060; - ::apache::thrift::protocol::TType _etype1063; - xfer += iprot->readListBegin(_etype1063, _size1060); - this->success.resize(_size1060); - uint32_t _i1064; - for (_i1064 = 0; _i1064 < _size1060; ++_i1064) + uint32_t _size1061; + ::apache::thrift::protocol::TType _etype1064; + xfer += iprot->readListBegin(_etype1064, _size1061); + this->success.resize(_size1061); + uint32_t _i1065; + for (_i1065 = 0; _i1065 < _size1061; ++_i1065) { - xfer += iprot->readString(this->success[_i1064]); + xfer += iprot->readString(this->success[_i1065]); } xfer += iprot->readListEnd(); } @@ -1504,10 +1504,10 @@ uint32_t ThriftHiveMetastore_get_all_databases_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1065; - for (_iter1065 = this->success.begin(); _iter1065 != this->success.end(); ++_iter1065) + std::vector ::const_iterator _iter1066; + for (_iter1066 = this->success.begin(); _iter1066 != this->success.end(); ++_iter1066) { - xfer += oprot->writeString((*_iter1065)); + xfer += oprot->writeString((*_iter1066)); } xfer += oprot->writeListEnd(); } @@ -1552,14 +1552,14 @@ uint32_t ThriftHiveMetastore_get_all_databases_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1066; - ::apache::thrift::protocol::TType _etype1069; - xfer += iprot->readListBegin(_etype1069, _size1066); - (*(this->success)).resize(_size1066); - uint32_t _i1070; - for (_i1070 = 0; _i1070 < _size1066; ++_i1070) + uint32_t _size1067; + ::apache::thrift::protocol::TType _etype1070; + xfer += iprot->readListBegin(_etype1070, _size1067); + (*(this->success)).resize(_size1067); + uint32_t _i1071; + for (_i1071 = 0; _i1071 < _size1067; ++_i1071) { - xfer += iprot->readString((*(this->success))[_i1070]); + xfer += iprot->readString((*(this->success))[_i1071]); } xfer += iprot->readListEnd(); } @@ -2621,17 +2621,17 @@ uint32_t ThriftHiveMetastore_get_type_all_result::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size1071; - ::apache::thrift::protocol::TType _ktype1072; - ::apache::thrift::protocol::TType _vtype1073; - xfer += iprot->readMapBegin(_ktype1072, _vtype1073, _size1071); - uint32_t _i1075; - for (_i1075 = 0; _i1075 < _size1071; ++_i1075) + uint32_t _size1072; + ::apache::thrift::protocol::TType _ktype1073; + ::apache::thrift::protocol::TType _vtype1074; + xfer += iprot->readMapBegin(_ktype1073, _vtype1074, _size1072); + uint32_t _i1076; + for (_i1076 = 0; _i1076 < _size1072; ++_i1076) { - std::string _key1076; - xfer += iprot->readString(_key1076); - Type& _val1077 = this->success[_key1076]; - xfer += _val1077.read(iprot); + std::string _key1077; + xfer += iprot->readString(_key1077); + Type& _val1078 = this->success[_key1077]; + xfer += _val1078.read(iprot); } xfer += iprot->readMapEnd(); } @@ -2670,11 +2670,11 @@ uint32_t ThriftHiveMetastore_get_type_all_result::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::map ::const_iterator _iter1078; - for (_iter1078 = this->success.begin(); _iter1078 != this->success.end(); ++_iter1078) + std::map ::const_iterator _iter1079; + for (_iter1079 = this->success.begin(); _iter1079 != this->success.end(); ++_iter1079) { - xfer += oprot->writeString(_iter1078->first); - xfer += _iter1078->second.write(oprot); + xfer += oprot->writeString(_iter1079->first); + xfer += _iter1079->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -2719,17 +2719,17 @@ uint32_t ThriftHiveMetastore_get_type_all_presult::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size1079; - ::apache::thrift::protocol::TType _ktype1080; - ::apache::thrift::protocol::TType _vtype1081; - xfer += iprot->readMapBegin(_ktype1080, _vtype1081, _size1079); - uint32_t _i1083; - for (_i1083 = 0; _i1083 < _size1079; ++_i1083) + uint32_t _size1080; + ::apache::thrift::protocol::TType _ktype1081; + ::apache::thrift::protocol::TType _vtype1082; + xfer += iprot->readMapBegin(_ktype1081, _vtype1082, _size1080); + uint32_t _i1084; + for (_i1084 = 0; _i1084 < _size1080; ++_i1084) { - std::string _key1084; - xfer += iprot->readString(_key1084); - Type& _val1085 = (*(this->success))[_key1084]; - xfer += _val1085.read(iprot); + std::string _key1085; + xfer += iprot->readString(_key1085); + Type& _val1086 = (*(this->success))[_key1085]; + xfer += _val1086.read(iprot); } xfer += iprot->readMapEnd(); } @@ -2883,14 +2883,14 @@ uint32_t ThriftHiveMetastore_get_fields_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1086; - ::apache::thrift::protocol::TType _etype1089; - xfer += iprot->readListBegin(_etype1089, _size1086); - this->success.resize(_size1086); - uint32_t _i1090; - for (_i1090 = 0; _i1090 < _size1086; ++_i1090) + uint32_t _size1087; + ::apache::thrift::protocol::TType _etype1090; + xfer += iprot->readListBegin(_etype1090, _size1087); + this->success.resize(_size1087); + uint32_t _i1091; + for (_i1091 = 0; _i1091 < _size1087; ++_i1091) { - xfer += this->success[_i1090].read(iprot); + xfer += this->success[_i1091].read(iprot); } xfer += iprot->readListEnd(); } @@ -2945,10 +2945,10 @@ uint32_t ThriftHiveMetastore_get_fields_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1091; - for (_iter1091 = this->success.begin(); _iter1091 != this->success.end(); ++_iter1091) + std::vector ::const_iterator _iter1092; + for (_iter1092 = this->success.begin(); _iter1092 != this->success.end(); ++_iter1092) { - xfer += (*_iter1091).write(oprot); + xfer += (*_iter1092).write(oprot); } xfer += oprot->writeListEnd(); } @@ -3001,14 +3001,14 @@ uint32_t ThriftHiveMetastore_get_fields_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1092; - ::apache::thrift::protocol::TType _etype1095; - xfer += iprot->readListBegin(_etype1095, _size1092); - (*(this->success)).resize(_size1092); - uint32_t _i1096; - for (_i1096 = 0; _i1096 < _size1092; ++_i1096) + uint32_t _size1093; + ::apache::thrift::protocol::TType _etype1096; + xfer += iprot->readListBegin(_etype1096, _size1093); + (*(this->success)).resize(_size1093); + uint32_t _i1097; + for (_i1097 = 0; _i1097 < _size1093; ++_i1097) { - xfer += (*(this->success))[_i1096].read(iprot); + xfer += (*(this->success))[_i1097].read(iprot); } xfer += iprot->readListEnd(); } @@ -3194,14 +3194,14 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_result::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1097; - ::apache::thrift::protocol::TType _etype1100; - xfer += iprot->readListBegin(_etype1100, _size1097); - this->success.resize(_size1097); - uint32_t _i1101; - for (_i1101 = 0; _i1101 < _size1097; ++_i1101) + uint32_t _size1098; + ::apache::thrift::protocol::TType _etype1101; + xfer += iprot->readListBegin(_etype1101, _size1098); + this->success.resize(_size1098); + uint32_t _i1102; + for (_i1102 = 0; _i1102 < _size1098; ++_i1102) { - xfer += this->success[_i1101].read(iprot); + xfer += this->success[_i1102].read(iprot); } xfer += iprot->readListEnd(); } @@ -3256,10 +3256,10 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_result::write(: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1102; - for (_iter1102 = this->success.begin(); _iter1102 != this->success.end(); ++_iter1102) + std::vector ::const_iterator _iter1103; + for (_iter1103 = this->success.begin(); _iter1103 != this->success.end(); ++_iter1103) { - xfer += (*_iter1102).write(oprot); + xfer += (*_iter1103).write(oprot); } xfer += oprot->writeListEnd(); } @@ -3312,14 +3312,14 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_presult::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1103; - ::apache::thrift::protocol::TType _etype1106; - xfer += iprot->readListBegin(_etype1106, _size1103); - (*(this->success)).resize(_size1103); - uint32_t _i1107; - for (_i1107 = 0; _i1107 < _size1103; ++_i1107) + uint32_t _size1104; + ::apache::thrift::protocol::TType _etype1107; + xfer += iprot->readListBegin(_etype1107, _size1104); + (*(this->success)).resize(_size1104); + uint32_t _i1108; + for (_i1108 = 0; _i1108 < _size1104; ++_i1108) { - xfer += (*(this->success))[_i1107].read(iprot); + xfer += (*(this->success))[_i1108].read(iprot); } xfer += iprot->readListEnd(); } @@ -3489,14 +3489,14 @@ uint32_t ThriftHiveMetastore_get_schema_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1108; - ::apache::thrift::protocol::TType _etype1111; - xfer += iprot->readListBegin(_etype1111, _size1108); - this->success.resize(_size1108); - uint32_t _i1112; - for (_i1112 = 0; _i1112 < _size1108; ++_i1112) + uint32_t _size1109; + ::apache::thrift::protocol::TType _etype1112; + xfer += iprot->readListBegin(_etype1112, _size1109); + this->success.resize(_size1109); + uint32_t _i1113; + for (_i1113 = 0; _i1113 < _size1109; ++_i1113) { - xfer += this->success[_i1112].read(iprot); + xfer += this->success[_i1113].read(iprot); } xfer += iprot->readListEnd(); } @@ -3551,10 +3551,10 @@ uint32_t ThriftHiveMetastore_get_schema_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1113; - for (_iter1113 = this->success.begin(); _iter1113 != this->success.end(); ++_iter1113) + std::vector ::const_iterator _iter1114; + for (_iter1114 = this->success.begin(); _iter1114 != this->success.end(); ++_iter1114) { - xfer += (*_iter1113).write(oprot); + xfer += (*_iter1114).write(oprot); } xfer += oprot->writeListEnd(); } @@ -3607,14 +3607,14 @@ uint32_t ThriftHiveMetastore_get_schema_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1114; - ::apache::thrift::protocol::TType _etype1117; - xfer += iprot->readListBegin(_etype1117, _size1114); - (*(this->success)).resize(_size1114); - uint32_t _i1118; - for (_i1118 = 0; _i1118 < _size1114; ++_i1118) + uint32_t _size1115; + ::apache::thrift::protocol::TType _etype1118; + xfer += iprot->readListBegin(_etype1118, _size1115); + (*(this->success)).resize(_size1115); + uint32_t _i1119; + for (_i1119 = 0; _i1119 < _size1115; ++_i1119) { - xfer += (*(this->success))[_i1118].read(iprot); + xfer += (*(this->success))[_i1119].read(iprot); } xfer += iprot->readListEnd(); } @@ -3800,14 +3800,14 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_result::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1119; - ::apache::thrift::protocol::TType _etype1122; - xfer += iprot->readListBegin(_etype1122, _size1119); - this->success.resize(_size1119); - uint32_t _i1123; - for (_i1123 = 0; _i1123 < _size1119; ++_i1123) + uint32_t _size1120; + ::apache::thrift::protocol::TType _etype1123; + xfer += iprot->readListBegin(_etype1123, _size1120); + this->success.resize(_size1120); + uint32_t _i1124; + for (_i1124 = 0; _i1124 < _size1120; ++_i1124) { - xfer += this->success[_i1123].read(iprot); + xfer += this->success[_i1124].read(iprot); } xfer += iprot->readListEnd(); } @@ -3862,10 +3862,10 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_result::write(: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1124; - for (_iter1124 = this->success.begin(); _iter1124 != this->success.end(); ++_iter1124) + std::vector ::const_iterator _iter1125; + for (_iter1125 = this->success.begin(); _iter1125 != this->success.end(); ++_iter1125) { - xfer += (*_iter1124).write(oprot); + xfer += (*_iter1125).write(oprot); } xfer += oprot->writeListEnd(); } @@ -3918,14 +3918,14 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_presult::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1125; - ::apache::thrift::protocol::TType _etype1128; - xfer += iprot->readListBegin(_etype1128, _size1125); - (*(this->success)).resize(_size1125); - uint32_t _i1129; - for (_i1129 = 0; _i1129 < _size1125; ++_i1129) + uint32_t _size1126; + ::apache::thrift::protocol::TType _etype1129; + xfer += iprot->readListBegin(_etype1129, _size1126); + (*(this->success)).resize(_size1126); + uint32_t _i1130; + for (_i1130 = 0; _i1130 < _size1126; ++_i1130) { - xfer += (*(this->success))[_i1129].read(iprot); + xfer += (*(this->success))[_i1130].read(iprot); } xfer += iprot->readListEnd(); } @@ -4518,14 +4518,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->primaryKeys.clear(); - uint32_t _size1130; - ::apache::thrift::protocol::TType _etype1133; - xfer += iprot->readListBegin(_etype1133, _size1130); - this->primaryKeys.resize(_size1130); - uint32_t _i1134; - for (_i1134 = 0; _i1134 < _size1130; ++_i1134) + uint32_t _size1131; + ::apache::thrift::protocol::TType _etype1134; + xfer += iprot->readListBegin(_etype1134, _size1131); + this->primaryKeys.resize(_size1131); + uint32_t _i1135; + for (_i1135 = 0; _i1135 < _size1131; ++_i1135) { - xfer += this->primaryKeys[_i1134].read(iprot); + xfer += this->primaryKeys[_i1135].read(iprot); } xfer += iprot->readListEnd(); } @@ -4538,14 +4538,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->foreignKeys.clear(); - uint32_t _size1135; - ::apache::thrift::protocol::TType _etype1138; - xfer += iprot->readListBegin(_etype1138, _size1135); - this->foreignKeys.resize(_size1135); - uint32_t _i1139; - for (_i1139 = 0; _i1139 < _size1135; ++_i1139) + uint32_t _size1136; + ::apache::thrift::protocol::TType _etype1139; + xfer += iprot->readListBegin(_etype1139, _size1136); + this->foreignKeys.resize(_size1136); + uint32_t _i1140; + for (_i1140 = 0; _i1140 < _size1136; ++_i1140) { - xfer += this->foreignKeys[_i1139].read(iprot); + xfer += this->foreignKeys[_i1140].read(iprot); } xfer += iprot->readListEnd(); } @@ -4558,14 +4558,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->uniqueConstraints.clear(); - uint32_t _size1140; - ::apache::thrift::protocol::TType _etype1143; - xfer += iprot->readListBegin(_etype1143, _size1140); - this->uniqueConstraints.resize(_size1140); - uint32_t _i1144; - for (_i1144 = 0; _i1144 < _size1140; ++_i1144) + uint32_t _size1141; + ::apache::thrift::protocol::TType _etype1144; + xfer += iprot->readListBegin(_etype1144, _size1141); + this->uniqueConstraints.resize(_size1141); + uint32_t _i1145; + for (_i1145 = 0; _i1145 < _size1141; ++_i1145) { - xfer += this->uniqueConstraints[_i1144].read(iprot); + xfer += this->uniqueConstraints[_i1145].read(iprot); } xfer += iprot->readListEnd(); } @@ -4578,14 +4578,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->notNullConstraints.clear(); - uint32_t _size1145; - ::apache::thrift::protocol::TType _etype1148; - xfer += iprot->readListBegin(_etype1148, _size1145); - this->notNullConstraints.resize(_size1145); - uint32_t _i1149; - for (_i1149 = 0; _i1149 < _size1145; ++_i1149) + uint32_t _size1146; + ::apache::thrift::protocol::TType _etype1149; + xfer += iprot->readListBegin(_etype1149, _size1146); + this->notNullConstraints.resize(_size1146); + uint32_t _i1150; + for (_i1150 = 0; _i1150 < _size1146; ++_i1150) { - xfer += this->notNullConstraints[_i1149].read(iprot); + xfer += this->notNullConstraints[_i1150].read(iprot); } xfer += iprot->readListEnd(); } @@ -4618,10 +4618,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->primaryKeys.size())); - std::vector ::const_iterator _iter1150; - for (_iter1150 = this->primaryKeys.begin(); _iter1150 != this->primaryKeys.end(); ++_iter1150) + std::vector ::const_iterator _iter1151; + for (_iter1151 = this->primaryKeys.begin(); _iter1151 != this->primaryKeys.end(); ++_iter1151) { - xfer += (*_iter1150).write(oprot); + xfer += (*_iter1151).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4630,10 +4630,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->foreignKeys.size())); - std::vector ::const_iterator _iter1151; - for (_iter1151 = this->foreignKeys.begin(); _iter1151 != this->foreignKeys.end(); ++_iter1151) + std::vector ::const_iterator _iter1152; + for (_iter1152 = this->foreignKeys.begin(); _iter1152 != this->foreignKeys.end(); ++_iter1152) { - xfer += (*_iter1151).write(oprot); + xfer += (*_iter1152).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4642,10 +4642,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->uniqueConstraints.size())); - std::vector ::const_iterator _iter1152; - for (_iter1152 = this->uniqueConstraints.begin(); _iter1152 != this->uniqueConstraints.end(); ++_iter1152) + std::vector ::const_iterator _iter1153; + for (_iter1153 = this->uniqueConstraints.begin(); _iter1153 != this->uniqueConstraints.end(); ++_iter1153) { - xfer += (*_iter1152).write(oprot); + xfer += (*_iter1153).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4654,10 +4654,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->notNullConstraints.size())); - std::vector ::const_iterator _iter1153; - for (_iter1153 = this->notNullConstraints.begin(); _iter1153 != this->notNullConstraints.end(); ++_iter1153) + std::vector ::const_iterator _iter1154; + for (_iter1154 = this->notNullConstraints.begin(); _iter1154 != this->notNullConstraints.end(); ++_iter1154) { - xfer += (*_iter1153).write(oprot); + xfer += (*_iter1154).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4685,10 +4685,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->primaryKeys)).size())); - std::vector ::const_iterator _iter1154; - for (_iter1154 = (*(this->primaryKeys)).begin(); _iter1154 != (*(this->primaryKeys)).end(); ++_iter1154) + std::vector ::const_iterator _iter1155; + for (_iter1155 = (*(this->primaryKeys)).begin(); _iter1155 != (*(this->primaryKeys)).end(); ++_iter1155) { - xfer += (*_iter1154).write(oprot); + xfer += (*_iter1155).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4697,10 +4697,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->foreignKeys)).size())); - std::vector ::const_iterator _iter1155; - for (_iter1155 = (*(this->foreignKeys)).begin(); _iter1155 != (*(this->foreignKeys)).end(); ++_iter1155) + std::vector ::const_iterator _iter1156; + for (_iter1156 = (*(this->foreignKeys)).begin(); _iter1156 != (*(this->foreignKeys)).end(); ++_iter1156) { - xfer += (*_iter1155).write(oprot); + xfer += (*_iter1156).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4709,10 +4709,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->uniqueConstraints)).size())); - std::vector ::const_iterator _iter1156; - for (_iter1156 = (*(this->uniqueConstraints)).begin(); _iter1156 != (*(this->uniqueConstraints)).end(); ++_iter1156) + std::vector ::const_iterator _iter1157; + for (_iter1157 = (*(this->uniqueConstraints)).begin(); _iter1157 != (*(this->uniqueConstraints)).end(); ++_iter1157) { - xfer += (*_iter1156).write(oprot); + xfer += (*_iter1157).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4721,10 +4721,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->notNullConstraints)).size())); - std::vector ::const_iterator _iter1157; - for (_iter1157 = (*(this->notNullConstraints)).begin(); _iter1157 != (*(this->notNullConstraints)).end(); ++_iter1157) + std::vector ::const_iterator _iter1158; + for (_iter1158 = (*(this->notNullConstraints)).begin(); _iter1158 != (*(this->notNullConstraints)).end(); ++_iter1158) { - xfer += (*_iter1157).write(oprot); + xfer += (*_iter1158).write(oprot); } xfer += oprot->writeListEnd(); } @@ -6478,14 +6478,14 @@ uint32_t ThriftHiveMetastore_truncate_table_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partNames.clear(); - uint32_t _size1158; - ::apache::thrift::protocol::TType _etype1161; - xfer += iprot->readListBegin(_etype1161, _size1158); - this->partNames.resize(_size1158); - uint32_t _i1162; - for (_i1162 = 0; _i1162 < _size1158; ++_i1162) + uint32_t _size1159; + ::apache::thrift::protocol::TType _etype1162; + xfer += iprot->readListBegin(_etype1162, _size1159); + this->partNames.resize(_size1159); + uint32_t _i1163; + for (_i1163 = 0; _i1163 < _size1159; ++_i1163) { - xfer += iprot->readString(this->partNames[_i1162]); + xfer += iprot->readString(this->partNames[_i1163]); } xfer += iprot->readListEnd(); } @@ -6522,10 +6522,10 @@ uint32_t ThriftHiveMetastore_truncate_table_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partNames.size())); - std::vector ::const_iterator _iter1163; - for (_iter1163 = this->partNames.begin(); _iter1163 != this->partNames.end(); ++_iter1163) + std::vector ::const_iterator _iter1164; + for (_iter1164 = this->partNames.begin(); _iter1164 != this->partNames.end(); ++_iter1164) { - xfer += oprot->writeString((*_iter1163)); + xfer += oprot->writeString((*_iter1164)); } xfer += oprot->writeListEnd(); } @@ -6557,10 +6557,10 @@ uint32_t ThriftHiveMetastore_truncate_table_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->partNames)).size())); - std::vector ::const_iterator _iter1164; - for (_iter1164 = (*(this->partNames)).begin(); _iter1164 != (*(this->partNames)).end(); ++_iter1164) + std::vector ::const_iterator _iter1165; + for (_iter1165 = (*(this->partNames)).begin(); _iter1165 != (*(this->partNames)).end(); ++_iter1165) { - xfer += oprot->writeString((*_iter1164)); + xfer += oprot->writeString((*_iter1165)); } xfer += oprot->writeListEnd(); } @@ -6804,14 +6804,14 @@ uint32_t ThriftHiveMetastore_get_tables_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1165; - ::apache::thrift::protocol::TType _etype1168; - xfer += iprot->readListBegin(_etype1168, _size1165); - this->success.resize(_size1165); - uint32_t _i1169; - for (_i1169 = 0; _i1169 < _size1165; ++_i1169) + uint32_t _size1166; + ::apache::thrift::protocol::TType _etype1169; + xfer += iprot->readListBegin(_etype1169, _size1166); + this->success.resize(_size1166); + uint32_t _i1170; + for (_i1170 = 0; _i1170 < _size1166; ++_i1170) { - xfer += iprot->readString(this->success[_i1169]); + xfer += iprot->readString(this->success[_i1170]); } xfer += iprot->readListEnd(); } @@ -6850,10 +6850,10 @@ uint32_t ThriftHiveMetastore_get_tables_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1170; - for (_iter1170 = this->success.begin(); _iter1170 != this->success.end(); ++_iter1170) + std::vector ::const_iterator _iter1171; + for (_iter1171 = this->success.begin(); _iter1171 != this->success.end(); ++_iter1171) { - xfer += oprot->writeString((*_iter1170)); + xfer += oprot->writeString((*_iter1171)); } xfer += oprot->writeListEnd(); } @@ -6898,14 +6898,14 @@ uint32_t ThriftHiveMetastore_get_tables_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1171; - ::apache::thrift::protocol::TType _etype1174; - xfer += iprot->readListBegin(_etype1174, _size1171); - (*(this->success)).resize(_size1171); - uint32_t _i1175; - for (_i1175 = 0; _i1175 < _size1171; ++_i1175) + uint32_t _size1172; + ::apache::thrift::protocol::TType _etype1175; + xfer += iprot->readListBegin(_etype1175, _size1172); + (*(this->success)).resize(_size1172); + uint32_t _i1176; + for (_i1176 = 0; _i1176 < _size1172; ++_i1176) { - xfer += iprot->readString((*(this->success))[_i1175]); + xfer += iprot->readString((*(this->success))[_i1176]); } xfer += iprot->readListEnd(); } @@ -7075,14 +7075,14 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_result::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1176; - ::apache::thrift::protocol::TType _etype1179; - xfer += iprot->readListBegin(_etype1179, _size1176); - this->success.resize(_size1176); - uint32_t _i1180; - for (_i1180 = 0; _i1180 < _size1176; ++_i1180) + uint32_t _size1177; + ::apache::thrift::protocol::TType _etype1180; + xfer += iprot->readListBegin(_etype1180, _size1177); + this->success.resize(_size1177); + uint32_t _i1181; + for (_i1181 = 0; _i1181 < _size1177; ++_i1181) { - xfer += iprot->readString(this->success[_i1180]); + xfer += iprot->readString(this->success[_i1181]); } xfer += iprot->readListEnd(); } @@ -7121,10 +7121,10 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_result::write(::apache::thrift:: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1181; - for (_iter1181 = this->success.begin(); _iter1181 != this->success.end(); ++_iter1181) + std::vector ::const_iterator _iter1182; + for (_iter1182 = this->success.begin(); _iter1182 != this->success.end(); ++_iter1182) { - xfer += oprot->writeString((*_iter1181)); + xfer += oprot->writeString((*_iter1182)); } xfer += oprot->writeListEnd(); } @@ -7169,14 +7169,14 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_presult::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1182; - ::apache::thrift::protocol::TType _etype1185; - xfer += iprot->readListBegin(_etype1185, _size1182); - (*(this->success)).resize(_size1182); - uint32_t _i1186; - for (_i1186 = 0; _i1186 < _size1182; ++_i1186) + uint32_t _size1183; + ::apache::thrift::protocol::TType _etype1186; + xfer += iprot->readListBegin(_etype1186, _size1183); + (*(this->success)).resize(_size1183); + uint32_t _i1187; + for (_i1187 = 0; _i1187 < _size1183; ++_i1187) { - xfer += iprot->readString((*(this->success))[_i1186]); + xfer += iprot->readString((*(this->success))[_i1187]); } xfer += iprot->readListEnd(); } @@ -7314,14 +7314,14 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_result::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1187; - ::apache::thrift::protocol::TType _etype1190; - xfer += iprot->readListBegin(_etype1190, _size1187); - this->success.resize(_size1187); - uint32_t _i1191; - for (_i1191 = 0; _i1191 < _size1187; ++_i1191) + uint32_t _size1188; + ::apache::thrift::protocol::TType _etype1191; + xfer += iprot->readListBegin(_etype1191, _size1188); + this->success.resize(_size1188); + uint32_t _i1192; + for (_i1192 = 0; _i1192 < _size1188; ++_i1192) { - xfer += iprot->readString(this->success[_i1191]); + xfer += iprot->readString(this->success[_i1192]); } xfer += iprot->readListEnd(); } @@ -7360,10 +7360,10 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_result::write( xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1192; - for (_iter1192 = this->success.begin(); _iter1192 != this->success.end(); ++_iter1192) + std::vector ::const_iterator _iter1193; + for (_iter1193 = this->success.begin(); _iter1193 != this->success.end(); ++_iter1193) { - xfer += oprot->writeString((*_iter1192)); + xfer += oprot->writeString((*_iter1193)); } xfer += oprot->writeListEnd(); } @@ -7408,14 +7408,14 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_presult::read( if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1193; - ::apache::thrift::protocol::TType _etype1196; - xfer += iprot->readListBegin(_etype1196, _size1193); - (*(this->success)).resize(_size1193); - uint32_t _i1197; - for (_i1197 = 0; _i1197 < _size1193; ++_i1197) + uint32_t _size1194; + ::apache::thrift::protocol::TType _etype1197; + xfer += iprot->readListBegin(_etype1197, _size1194); + (*(this->success)).resize(_size1194); + uint32_t _i1198; + for (_i1198 = 0; _i1198 < _size1194; ++_i1198) { - xfer += iprot->readString((*(this->success))[_i1197]); + xfer += iprot->readString((*(this->success))[_i1198]); } xfer += iprot->readListEnd(); } @@ -7490,14 +7490,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tbl_types.clear(); - uint32_t _size1198; - ::apache::thrift::protocol::TType _etype1201; - xfer += iprot->readListBegin(_etype1201, _size1198); - this->tbl_types.resize(_size1198); - uint32_t _i1202; - for (_i1202 = 0; _i1202 < _size1198; ++_i1202) + uint32_t _size1199; + ::apache::thrift::protocol::TType _etype1202; + xfer += iprot->readListBegin(_etype1202, _size1199); + this->tbl_types.resize(_size1199); + uint32_t _i1203; + for (_i1203 = 0; _i1203 < _size1199; ++_i1203) { - xfer += iprot->readString(this->tbl_types[_i1202]); + xfer += iprot->readString(this->tbl_types[_i1203]); } xfer += iprot->readListEnd(); } @@ -7534,10 +7534,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("tbl_types", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tbl_types.size())); - std::vector ::const_iterator _iter1203; - for (_iter1203 = this->tbl_types.begin(); _iter1203 != this->tbl_types.end(); ++_iter1203) + std::vector ::const_iterator _iter1204; + for (_iter1204 = this->tbl_types.begin(); _iter1204 != this->tbl_types.end(); ++_iter1204) { - xfer += oprot->writeString((*_iter1203)); + xfer += oprot->writeString((*_iter1204)); } xfer += oprot->writeListEnd(); } @@ -7569,10 +7569,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("tbl_types", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->tbl_types)).size())); - std::vector ::const_iterator _iter1204; - for (_iter1204 = (*(this->tbl_types)).begin(); _iter1204 != (*(this->tbl_types)).end(); ++_iter1204) + std::vector ::const_iterator _iter1205; + for (_iter1205 = (*(this->tbl_types)).begin(); _iter1205 != (*(this->tbl_types)).end(); ++_iter1205) { - xfer += oprot->writeString((*_iter1204)); + xfer += oprot->writeString((*_iter1205)); } xfer += oprot->writeListEnd(); } @@ -7613,14 +7613,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1205; - ::apache::thrift::protocol::TType _etype1208; - xfer += iprot->readListBegin(_etype1208, _size1205); - this->success.resize(_size1205); - uint32_t _i1209; - for (_i1209 = 0; _i1209 < _size1205; ++_i1209) + uint32_t _size1206; + ::apache::thrift::protocol::TType _etype1209; + xfer += iprot->readListBegin(_etype1209, _size1206); + this->success.resize(_size1206); + uint32_t _i1210; + for (_i1210 = 0; _i1210 < _size1206; ++_i1210) { - xfer += this->success[_i1209].read(iprot); + xfer += this->success[_i1210].read(iprot); } xfer += iprot->readListEnd(); } @@ -7659,10 +7659,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1210; - for (_iter1210 = this->success.begin(); _iter1210 != this->success.end(); ++_iter1210) + std::vector ::const_iterator _iter1211; + for (_iter1211 = this->success.begin(); _iter1211 != this->success.end(); ++_iter1211) { - xfer += (*_iter1210).write(oprot); + xfer += (*_iter1211).write(oprot); } xfer += oprot->writeListEnd(); } @@ -7707,14 +7707,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1211; - ::apache::thrift::protocol::TType _etype1214; - xfer += iprot->readListBegin(_etype1214, _size1211); - (*(this->success)).resize(_size1211); - uint32_t _i1215; - for (_i1215 = 0; _i1215 < _size1211; ++_i1215) + uint32_t _size1212; + ::apache::thrift::protocol::TType _etype1215; + xfer += iprot->readListBegin(_etype1215, _size1212); + (*(this->success)).resize(_size1212); + uint32_t _i1216; + for (_i1216 = 0; _i1216 < _size1212; ++_i1216) { - xfer += (*(this->success))[_i1215].read(iprot); + xfer += (*(this->success))[_i1216].read(iprot); } xfer += iprot->readListEnd(); } @@ -7852,14 +7852,14 @@ uint32_t ThriftHiveMetastore_get_all_tables_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1216; - ::apache::thrift::protocol::TType _etype1219; - xfer += iprot->readListBegin(_etype1219, _size1216); - this->success.resize(_size1216); - uint32_t _i1220; - for (_i1220 = 0; _i1220 < _size1216; ++_i1220) + uint32_t _size1217; + ::apache::thrift::protocol::TType _etype1220; + xfer += iprot->readListBegin(_etype1220, _size1217); + this->success.resize(_size1217); + uint32_t _i1221; + for (_i1221 = 0; _i1221 < _size1217; ++_i1221) { - xfer += iprot->readString(this->success[_i1220]); + xfer += iprot->readString(this->success[_i1221]); } xfer += iprot->readListEnd(); } @@ -7898,10 +7898,10 @@ uint32_t ThriftHiveMetastore_get_all_tables_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1221; - for (_iter1221 = this->success.begin(); _iter1221 != this->success.end(); ++_iter1221) + std::vector ::const_iterator _iter1222; + for (_iter1222 = this->success.begin(); _iter1222 != this->success.end(); ++_iter1222) { - xfer += oprot->writeString((*_iter1221)); + xfer += oprot->writeString((*_iter1222)); } xfer += oprot->writeListEnd(); } @@ -7946,14 +7946,14 @@ uint32_t ThriftHiveMetastore_get_all_tables_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1222; - ::apache::thrift::protocol::TType _etype1225; - xfer += iprot->readListBegin(_etype1225, _size1222); - (*(this->success)).resize(_size1222); - uint32_t _i1226; - for (_i1226 = 0; _i1226 < _size1222; ++_i1226) + uint32_t _size1223; + ::apache::thrift::protocol::TType _etype1226; + xfer += iprot->readListBegin(_etype1226, _size1223); + (*(this->success)).resize(_size1223); + uint32_t _i1227; + for (_i1227 = 0; _i1227 < _size1223; ++_i1227) { - xfer += iprot->readString((*(this->success))[_i1226]); + xfer += iprot->readString((*(this->success))[_i1227]); } xfer += iprot->readListEnd(); } @@ -8263,14 +8263,14 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_args::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tbl_names.clear(); - uint32_t _size1227; - ::apache::thrift::protocol::TType _etype1230; - xfer += iprot->readListBegin(_etype1230, _size1227); - this->tbl_names.resize(_size1227); - uint32_t _i1231; - for (_i1231 = 0; _i1231 < _size1227; ++_i1231) + uint32_t _size1228; + ::apache::thrift::protocol::TType _etype1231; + xfer += iprot->readListBegin(_etype1231, _size1228); + this->tbl_names.resize(_size1228); + uint32_t _i1232; + for (_i1232 = 0; _i1232 < _size1228; ++_i1232) { - xfer += iprot->readString(this->tbl_names[_i1231]); + xfer += iprot->readString(this->tbl_names[_i1232]); } xfer += iprot->readListEnd(); } @@ -8303,10 +8303,10 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_args::write(::apache::thr xfer += oprot->writeFieldBegin("tbl_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tbl_names.size())); - std::vector ::const_iterator _iter1232; - for (_iter1232 = this->tbl_names.begin(); _iter1232 != this->tbl_names.end(); ++_iter1232) + std::vector ::const_iterator _iter1233; + for (_iter1233 = this->tbl_names.begin(); _iter1233 != this->tbl_names.end(); ++_iter1233) { - xfer += oprot->writeString((*_iter1232)); + xfer += oprot->writeString((*_iter1233)); } xfer += oprot->writeListEnd(); } @@ -8334,10 +8334,10 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_pargs::write(::apache::th xfer += oprot->writeFieldBegin("tbl_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->tbl_names)).size())); - std::vector ::const_iterator _iter1233; - for (_iter1233 = (*(this->tbl_names)).begin(); _iter1233 != (*(this->tbl_names)).end(); ++_iter1233) + std::vector ::const_iterator _iter1234; + for (_iter1234 = (*(this->tbl_names)).begin(); _iter1234 != (*(this->tbl_names)).end(); ++_iter1234) { - xfer += oprot->writeString((*_iter1233)); + xfer += oprot->writeString((*_iter1234)); } xfer += oprot->writeListEnd(); } @@ -8378,14 +8378,14 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1234; - ::apache::thrift::protocol::TType _etype1237; - xfer += iprot->readListBegin(_etype1237, _size1234); - this->success.resize(_size1234); - uint32_t _i1238; - for (_i1238 = 0; _i1238 < _size1234; ++_i1238) + uint32_t _size1235; + ::apache::thrift::protocol::TType _etype1238; + xfer += iprot->readListBegin(_etype1238, _size1235); + this->success.resize(_size1235); + uint32_t _i1239; + for (_i1239 = 0; _i1239 < _size1235; ++_i1239) { - xfer += this->success[_i1238].read(iprot); + xfer += this->success[_i1239].read(iprot); } xfer += iprot->readListEnd(); } @@ -8416,10 +8416,10 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1239; - for (_iter1239 = this->success.begin(); _iter1239 != this->success.end(); ++_iter1239) + std::vector
::const_iterator _iter1240; + for (_iter1240 = this->success.begin(); _iter1240 != this->success.end(); ++_iter1240) { - xfer += (*_iter1239).write(oprot); + xfer += (*_iter1240).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8460,14 +8460,14 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1240; - ::apache::thrift::protocol::TType _etype1243; - xfer += iprot->readListBegin(_etype1243, _size1240); - (*(this->success)).resize(_size1240); - uint32_t _i1244; - for (_i1244 = 0; _i1244 < _size1240; ++_i1244) + uint32_t _size1241; + ::apache::thrift::protocol::TType _etype1244; + xfer += iprot->readListBegin(_etype1244, _size1241); + (*(this->success)).resize(_size1241); + uint32_t _i1245; + for (_i1245 = 0; _i1245 < _size1241; ++_i1245) { - xfer += (*(this->success))[_i1244].read(iprot); + xfer += (*(this->success))[_i1245].read(iprot); } xfer += iprot->readListEnd(); } @@ -9000,14 +9000,14 @@ uint32_t ThriftHiveMetastore_get_materialization_invalidation_info_args::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tbl_names.clear(); - uint32_t _size1245; - ::apache::thrift::protocol::TType _etype1248; - xfer += iprot->readListBegin(_etype1248, _size1245); - this->tbl_names.resize(_size1245); - uint32_t _i1249; - for (_i1249 = 0; _i1249 < _size1245; ++_i1249) + uint32_t _size1246; + ::apache::thrift::protocol::TType _etype1249; + xfer += iprot->readListBegin(_etype1249, _size1246); + this->tbl_names.resize(_size1246); + uint32_t _i1250; + for (_i1250 = 0; _i1250 < _size1246; ++_i1250) { - xfer += iprot->readString(this->tbl_names[_i1249]); + xfer += iprot->readString(this->tbl_names[_i1250]); } xfer += iprot->readListEnd(); } @@ -9040,10 +9040,10 @@ uint32_t ThriftHiveMetastore_get_materialization_invalidation_info_args::write(: xfer += oprot->writeFieldBegin("tbl_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tbl_names.size())); - std::vector ::const_iterator _iter1250; - for (_iter1250 = this->tbl_names.begin(); _iter1250 != this->tbl_names.end(); ++_iter1250) + std::vector ::const_iterator _iter1251; + for (_iter1251 = this->tbl_names.begin(); _iter1251 != this->tbl_names.end(); ++_iter1251) { - xfer += oprot->writeString((*_iter1250)); + xfer += oprot->writeString((*_iter1251)); } xfer += oprot->writeListEnd(); } @@ -9071,10 +9071,10 @@ uint32_t ThriftHiveMetastore_get_materialization_invalidation_info_pargs::write( xfer += oprot->writeFieldBegin("tbl_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->tbl_names)).size())); - std::vector ::const_iterator _iter1251; - for (_iter1251 = (*(this->tbl_names)).begin(); _iter1251 != (*(this->tbl_names)).end(); ++_iter1251) + std::vector ::const_iterator _iter1252; + for (_iter1252 = (*(this->tbl_names)).begin(); _iter1252 != (*(this->tbl_names)).end(); ++_iter1252) { - xfer += oprot->writeString((*_iter1251)); + xfer += oprot->writeString((*_iter1252)); } xfer += oprot->writeListEnd(); } @@ -9115,17 +9115,17 @@ uint32_t ThriftHiveMetastore_get_materialization_invalidation_info_result::read( if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size1252; - ::apache::thrift::protocol::TType _ktype1253; - ::apache::thrift::protocol::TType _vtype1254; - xfer += iprot->readMapBegin(_ktype1253, _vtype1254, _size1252); - uint32_t _i1256; - for (_i1256 = 0; _i1256 < _size1252; ++_i1256) + uint32_t _size1253; + ::apache::thrift::protocol::TType _ktype1254; + ::apache::thrift::protocol::TType _vtype1255; + xfer += iprot->readMapBegin(_ktype1254, _vtype1255, _size1253); + uint32_t _i1257; + for (_i1257 = 0; _i1257 < _size1253; ++_i1257) { - std::string _key1257; - xfer += iprot->readString(_key1257); - Materialization& _val1258 = this->success[_key1257]; - xfer += _val1258.read(iprot); + std::string _key1258; + xfer += iprot->readString(_key1258); + Materialization& _val1259 = this->success[_key1258]; + xfer += _val1259.read(iprot); } xfer += iprot->readMapEnd(); } @@ -9180,11 +9180,11 @@ uint32_t ThriftHiveMetastore_get_materialization_invalidation_info_result::write xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::map ::const_iterator _iter1259; - for (_iter1259 = this->success.begin(); _iter1259 != this->success.end(); ++_iter1259) + std::map ::const_iterator _iter1260; + for (_iter1260 = this->success.begin(); _iter1260 != this->success.end(); ++_iter1260) { - xfer += oprot->writeString(_iter1259->first); - xfer += _iter1259->second.write(oprot); + xfer += oprot->writeString(_iter1260->first); + xfer += _iter1260->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -9237,17 +9237,17 @@ uint32_t ThriftHiveMetastore_get_materialization_invalidation_info_presult::read if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size1260; - ::apache::thrift::protocol::TType _ktype1261; - ::apache::thrift::protocol::TType _vtype1262; - xfer += iprot->readMapBegin(_ktype1261, _vtype1262, _size1260); - uint32_t _i1264; - for (_i1264 = 0; _i1264 < _size1260; ++_i1264) + uint32_t _size1261; + ::apache::thrift::protocol::TType _ktype1262; + ::apache::thrift::protocol::TType _vtype1263; + xfer += iprot->readMapBegin(_ktype1262, _vtype1263, _size1261); + uint32_t _i1265; + for (_i1265 = 0; _i1265 < _size1261; ++_i1265) { - std::string _key1265; - xfer += iprot->readString(_key1265); - Materialization& _val1266 = (*(this->success))[_key1265]; - xfer += _val1266.read(iprot); + std::string _key1266; + xfer += iprot->readString(_key1266); + Materialization& _val1267 = (*(this->success))[_key1266]; + xfer += _val1267.read(iprot); } xfer += iprot->readMapEnd(); } @@ -9433,14 +9433,14 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1267; - ::apache::thrift::protocol::TType _etype1270; - xfer += iprot->readListBegin(_etype1270, _size1267); - this->success.resize(_size1267); - uint32_t _i1271; - for (_i1271 = 0; _i1271 < _size1267; ++_i1271) + uint32_t _size1268; + ::apache::thrift::protocol::TType _etype1271; + xfer += iprot->readListBegin(_etype1271, _size1268); + this->success.resize(_size1268); + uint32_t _i1272; + for (_i1272 = 0; _i1272 < _size1268; ++_i1272) { - xfer += iprot->readString(this->success[_i1271]); + xfer += iprot->readString(this->success[_i1272]); } xfer += iprot->readListEnd(); } @@ -9495,10 +9495,10 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1272; - for (_iter1272 = this->success.begin(); _iter1272 != this->success.end(); ++_iter1272) + std::vector ::const_iterator _iter1273; + for (_iter1273 = this->success.begin(); _iter1273 != this->success.end(); ++_iter1273) { - xfer += oprot->writeString((*_iter1272)); + xfer += oprot->writeString((*_iter1273)); } xfer += oprot->writeListEnd(); } @@ -9551,14 +9551,14 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1273; - ::apache::thrift::protocol::TType _etype1276; - xfer += iprot->readListBegin(_etype1276, _size1273); - (*(this->success)).resize(_size1273); - uint32_t _i1277; - for (_i1277 = 0; _i1277 < _size1273; ++_i1277) + uint32_t _size1274; + ::apache::thrift::protocol::TType _etype1277; + xfer += iprot->readListBegin(_etype1277, _size1274); + (*(this->success)).resize(_size1274); + uint32_t _i1278; + for (_i1278 = 0; _i1278 < _size1274; ++_i1278) { - xfer += iprot->readString((*(this->success))[_i1277]); + xfer += iprot->readString((*(this->success))[_i1278]); } xfer += iprot->readListEnd(); } @@ -10892,14 +10892,14 @@ uint32_t ThriftHiveMetastore_add_partitions_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1278; - ::apache::thrift::protocol::TType _etype1281; - xfer += iprot->readListBegin(_etype1281, _size1278); - this->new_parts.resize(_size1278); - uint32_t _i1282; - for (_i1282 = 0; _i1282 < _size1278; ++_i1282) + uint32_t _size1279; + ::apache::thrift::protocol::TType _etype1282; + xfer += iprot->readListBegin(_etype1282, _size1279); + this->new_parts.resize(_size1279); + uint32_t _i1283; + for (_i1283 = 0; _i1283 < _size1279; ++_i1283) { - xfer += this->new_parts[_i1282].read(iprot); + xfer += this->new_parts[_i1283].read(iprot); } xfer += iprot->readListEnd(); } @@ -10928,10 +10928,10 @@ uint32_t ThriftHiveMetastore_add_partitions_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1283; - for (_iter1283 = this->new_parts.begin(); _iter1283 != this->new_parts.end(); ++_iter1283) + std::vector ::const_iterator _iter1284; + for (_iter1284 = this->new_parts.begin(); _iter1284 != this->new_parts.end(); ++_iter1284) { - xfer += (*_iter1283).write(oprot); + xfer += (*_iter1284).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10955,10 +10955,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1284; - for (_iter1284 = (*(this->new_parts)).begin(); _iter1284 != (*(this->new_parts)).end(); ++_iter1284) + std::vector ::const_iterator _iter1285; + for (_iter1285 = (*(this->new_parts)).begin(); _iter1285 != (*(this->new_parts)).end(); ++_iter1285) { - xfer += (*_iter1284).write(oprot); + xfer += (*_iter1285).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11167,14 +11167,14 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_args::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1285; - ::apache::thrift::protocol::TType _etype1288; - xfer += iprot->readListBegin(_etype1288, _size1285); - this->new_parts.resize(_size1285); - uint32_t _i1289; - for (_i1289 = 0; _i1289 < _size1285; ++_i1289) + uint32_t _size1286; + ::apache::thrift::protocol::TType _etype1289; + xfer += iprot->readListBegin(_etype1289, _size1286); + this->new_parts.resize(_size1286); + uint32_t _i1290; + for (_i1290 = 0; _i1290 < _size1286; ++_i1290) { - xfer += this->new_parts[_i1289].read(iprot); + xfer += this->new_parts[_i1290].read(iprot); } xfer += iprot->readListEnd(); } @@ -11203,10 +11203,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_args::write(::apache::thrift:: xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1290; - for (_iter1290 = this->new_parts.begin(); _iter1290 != this->new_parts.end(); ++_iter1290) + std::vector ::const_iterator _iter1291; + for (_iter1291 = this->new_parts.begin(); _iter1291 != this->new_parts.end(); ++_iter1291) { - xfer += (*_iter1290).write(oprot); + xfer += (*_iter1291).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11230,10 +11230,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_pargs::write(::apache::thrift: xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1291; - for (_iter1291 = (*(this->new_parts)).begin(); _iter1291 != (*(this->new_parts)).end(); ++_iter1291) + std::vector ::const_iterator _iter1292; + for (_iter1292 = (*(this->new_parts)).begin(); _iter1292 != (*(this->new_parts)).end(); ++_iter1292) { - xfer += (*_iter1291).write(oprot); + xfer += (*_iter1292).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11458,14 +11458,14 @@ uint32_t ThriftHiveMetastore_append_partition_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1292; - ::apache::thrift::protocol::TType _etype1295; - xfer += iprot->readListBegin(_etype1295, _size1292); - this->part_vals.resize(_size1292); - uint32_t _i1296; - for (_i1296 = 0; _i1296 < _size1292; ++_i1296) + uint32_t _size1293; + ::apache::thrift::protocol::TType _etype1296; + xfer += iprot->readListBegin(_etype1296, _size1293); + this->part_vals.resize(_size1293); + uint32_t _i1297; + for (_i1297 = 0; _i1297 < _size1293; ++_i1297) { - xfer += iprot->readString(this->part_vals[_i1296]); + xfer += iprot->readString(this->part_vals[_i1297]); } xfer += iprot->readListEnd(); } @@ -11502,10 +11502,10 @@ uint32_t ThriftHiveMetastore_append_partition_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1297; - for (_iter1297 = this->part_vals.begin(); _iter1297 != this->part_vals.end(); ++_iter1297) + std::vector ::const_iterator _iter1298; + for (_iter1298 = this->part_vals.begin(); _iter1298 != this->part_vals.end(); ++_iter1298) { - xfer += oprot->writeString((*_iter1297)); + xfer += oprot->writeString((*_iter1298)); } xfer += oprot->writeListEnd(); } @@ -11537,10 +11537,10 @@ uint32_t ThriftHiveMetastore_append_partition_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1298; - for (_iter1298 = (*(this->part_vals)).begin(); _iter1298 != (*(this->part_vals)).end(); ++_iter1298) + std::vector ::const_iterator _iter1299; + for (_iter1299 = (*(this->part_vals)).begin(); _iter1299 != (*(this->part_vals)).end(); ++_iter1299) { - xfer += oprot->writeString((*_iter1298)); + xfer += oprot->writeString((*_iter1299)); } xfer += oprot->writeListEnd(); } @@ -12012,14 +12012,14 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_args::rea if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1299; - ::apache::thrift::protocol::TType _etype1302; - xfer += iprot->readListBegin(_etype1302, _size1299); - this->part_vals.resize(_size1299); - uint32_t _i1303; - for (_i1303 = 0; _i1303 < _size1299; ++_i1303) + uint32_t _size1300; + ::apache::thrift::protocol::TType _etype1303; + xfer += iprot->readListBegin(_etype1303, _size1300); + this->part_vals.resize(_size1300); + uint32_t _i1304; + for (_i1304 = 0; _i1304 < _size1300; ++_i1304) { - xfer += iprot->readString(this->part_vals[_i1303]); + xfer += iprot->readString(this->part_vals[_i1304]); } xfer += iprot->readListEnd(); } @@ -12064,10 +12064,10 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_args::wri xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1304; - for (_iter1304 = this->part_vals.begin(); _iter1304 != this->part_vals.end(); ++_iter1304) + std::vector ::const_iterator _iter1305; + for (_iter1305 = this->part_vals.begin(); _iter1305 != this->part_vals.end(); ++_iter1305) { - xfer += oprot->writeString((*_iter1304)); + xfer += oprot->writeString((*_iter1305)); } xfer += oprot->writeListEnd(); } @@ -12103,10 +12103,10 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_pargs::wr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1305; - for (_iter1305 = (*(this->part_vals)).begin(); _iter1305 != (*(this->part_vals)).end(); ++_iter1305) + std::vector ::const_iterator _iter1306; + for (_iter1306 = (*(this->part_vals)).begin(); _iter1306 != (*(this->part_vals)).end(); ++_iter1306) { - xfer += oprot->writeString((*_iter1305)); + xfer += oprot->writeString((*_iter1306)); } xfer += oprot->writeListEnd(); } @@ -12909,14 +12909,14 @@ uint32_t ThriftHiveMetastore_drop_partition_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1306; - ::apache::thrift::protocol::TType _etype1309; - xfer += iprot->readListBegin(_etype1309, _size1306); - this->part_vals.resize(_size1306); - uint32_t _i1310; - for (_i1310 = 0; _i1310 < _size1306; ++_i1310) + uint32_t _size1307; + ::apache::thrift::protocol::TType _etype1310; + xfer += iprot->readListBegin(_etype1310, _size1307); + this->part_vals.resize(_size1307); + uint32_t _i1311; + for (_i1311 = 0; _i1311 < _size1307; ++_i1311) { - xfer += iprot->readString(this->part_vals[_i1310]); + xfer += iprot->readString(this->part_vals[_i1311]); } xfer += iprot->readListEnd(); } @@ -12961,10 +12961,10 @@ uint32_t ThriftHiveMetastore_drop_partition_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1311; - for (_iter1311 = this->part_vals.begin(); _iter1311 != this->part_vals.end(); ++_iter1311) + std::vector ::const_iterator _iter1312; + for (_iter1312 = this->part_vals.begin(); _iter1312 != this->part_vals.end(); ++_iter1312) { - xfer += oprot->writeString((*_iter1311)); + xfer += oprot->writeString((*_iter1312)); } xfer += oprot->writeListEnd(); } @@ -13000,10 +13000,10 @@ uint32_t ThriftHiveMetastore_drop_partition_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1312; - for (_iter1312 = (*(this->part_vals)).begin(); _iter1312 != (*(this->part_vals)).end(); ++_iter1312) + std::vector ::const_iterator _iter1313; + for (_iter1313 = (*(this->part_vals)).begin(); _iter1313 != (*(this->part_vals)).end(); ++_iter1313) { - xfer += oprot->writeString((*_iter1312)); + xfer += oprot->writeString((*_iter1313)); } xfer += oprot->writeListEnd(); } @@ -13212,14 +13212,14 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_args::read( if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1313; - ::apache::thrift::protocol::TType _etype1316; - xfer += iprot->readListBegin(_etype1316, _size1313); - this->part_vals.resize(_size1313); - uint32_t _i1317; - for (_i1317 = 0; _i1317 < _size1313; ++_i1317) + uint32_t _size1314; + ::apache::thrift::protocol::TType _etype1317; + xfer += iprot->readListBegin(_etype1317, _size1314); + this->part_vals.resize(_size1314); + uint32_t _i1318; + for (_i1318 = 0; _i1318 < _size1314; ++_i1318) { - xfer += iprot->readString(this->part_vals[_i1317]); + xfer += iprot->readString(this->part_vals[_i1318]); } xfer += iprot->readListEnd(); } @@ -13272,10 +13272,10 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_args::write xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1318; - for (_iter1318 = this->part_vals.begin(); _iter1318 != this->part_vals.end(); ++_iter1318) + std::vector ::const_iterator _iter1319; + for (_iter1319 = this->part_vals.begin(); _iter1319 != this->part_vals.end(); ++_iter1319) { - xfer += oprot->writeString((*_iter1318)); + xfer += oprot->writeString((*_iter1319)); } xfer += oprot->writeListEnd(); } @@ -13315,10 +13315,10 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_pargs::writ xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1319; - for (_iter1319 = (*(this->part_vals)).begin(); _iter1319 != (*(this->part_vals)).end(); ++_iter1319) + std::vector ::const_iterator _iter1320; + for (_iter1320 = (*(this->part_vals)).begin(); _iter1320 != (*(this->part_vals)).end(); ++_iter1320) { - xfer += oprot->writeString((*_iter1319)); + xfer += oprot->writeString((*_iter1320)); } xfer += oprot->writeListEnd(); } @@ -14324,14 +14324,14 @@ uint32_t ThriftHiveMetastore_get_partition_args::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1320; - ::apache::thrift::protocol::TType _etype1323; - xfer += iprot->readListBegin(_etype1323, _size1320); - this->part_vals.resize(_size1320); - uint32_t _i1324; - for (_i1324 = 0; _i1324 < _size1320; ++_i1324) + uint32_t _size1321; + ::apache::thrift::protocol::TType _etype1324; + xfer += iprot->readListBegin(_etype1324, _size1321); + this->part_vals.resize(_size1321); + uint32_t _i1325; + for (_i1325 = 0; _i1325 < _size1321; ++_i1325) { - xfer += iprot->readString(this->part_vals[_i1324]); + xfer += iprot->readString(this->part_vals[_i1325]); } xfer += iprot->readListEnd(); } @@ -14368,10 +14368,10 @@ uint32_t ThriftHiveMetastore_get_partition_args::write(::apache::thrift::protoco xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1325; - for (_iter1325 = this->part_vals.begin(); _iter1325 != this->part_vals.end(); ++_iter1325) + std::vector ::const_iterator _iter1326; + for (_iter1326 = this->part_vals.begin(); _iter1326 != this->part_vals.end(); ++_iter1326) { - xfer += oprot->writeString((*_iter1325)); + xfer += oprot->writeString((*_iter1326)); } xfer += oprot->writeListEnd(); } @@ -14403,10 +14403,10 @@ uint32_t ThriftHiveMetastore_get_partition_pargs::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1326; - for (_iter1326 = (*(this->part_vals)).begin(); _iter1326 != (*(this->part_vals)).end(); ++_iter1326) + std::vector ::const_iterator _iter1327; + for (_iter1327 = (*(this->part_vals)).begin(); _iter1327 != (*(this->part_vals)).end(); ++_iter1327) { - xfer += oprot->writeString((*_iter1326)); + xfer += oprot->writeString((*_iter1327)); } xfer += oprot->writeListEnd(); } @@ -14595,17 +14595,17 @@ uint32_t ThriftHiveMetastore_exchange_partition_args::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partitionSpecs.clear(); - uint32_t _size1327; - ::apache::thrift::protocol::TType _ktype1328; - ::apache::thrift::protocol::TType _vtype1329; - xfer += iprot->readMapBegin(_ktype1328, _vtype1329, _size1327); - uint32_t _i1331; - for (_i1331 = 0; _i1331 < _size1327; ++_i1331) + uint32_t _size1328; + ::apache::thrift::protocol::TType _ktype1329; + ::apache::thrift::protocol::TType _vtype1330; + xfer += iprot->readMapBegin(_ktype1329, _vtype1330, _size1328); + uint32_t _i1332; + for (_i1332 = 0; _i1332 < _size1328; ++_i1332) { - std::string _key1332; - xfer += iprot->readString(_key1332); - std::string& _val1333 = this->partitionSpecs[_key1332]; - xfer += iprot->readString(_val1333); + std::string _key1333; + xfer += iprot->readString(_key1333); + std::string& _val1334 = this->partitionSpecs[_key1333]; + xfer += iprot->readString(_val1334); } xfer += iprot->readMapEnd(); } @@ -14666,11 +14666,11 @@ uint32_t ThriftHiveMetastore_exchange_partition_args::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->partitionSpecs.size())); - std::map ::const_iterator _iter1334; - for (_iter1334 = this->partitionSpecs.begin(); _iter1334 != this->partitionSpecs.end(); ++_iter1334) + std::map ::const_iterator _iter1335; + for (_iter1335 = this->partitionSpecs.begin(); _iter1335 != this->partitionSpecs.end(); ++_iter1335) { - xfer += oprot->writeString(_iter1334->first); - xfer += oprot->writeString(_iter1334->second); + xfer += oprot->writeString(_iter1335->first); + xfer += oprot->writeString(_iter1335->second); } xfer += oprot->writeMapEnd(); } @@ -14710,11 +14710,11 @@ uint32_t ThriftHiveMetastore_exchange_partition_pargs::write(::apache::thrift::p xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->partitionSpecs)).size())); - std::map ::const_iterator _iter1335; - for (_iter1335 = (*(this->partitionSpecs)).begin(); _iter1335 != (*(this->partitionSpecs)).end(); ++_iter1335) + std::map ::const_iterator _iter1336; + for (_iter1336 = (*(this->partitionSpecs)).begin(); _iter1336 != (*(this->partitionSpecs)).end(); ++_iter1336) { - xfer += oprot->writeString(_iter1335->first); - xfer += oprot->writeString(_iter1335->second); + xfer += oprot->writeString(_iter1336->first); + xfer += oprot->writeString(_iter1336->second); } xfer += oprot->writeMapEnd(); } @@ -14959,17 +14959,17 @@ uint32_t ThriftHiveMetastore_exchange_partitions_args::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partitionSpecs.clear(); - uint32_t _size1336; - ::apache::thrift::protocol::TType _ktype1337; - ::apache::thrift::protocol::TType _vtype1338; - xfer += iprot->readMapBegin(_ktype1337, _vtype1338, _size1336); - uint32_t _i1340; - for (_i1340 = 0; _i1340 < _size1336; ++_i1340) + uint32_t _size1337; + ::apache::thrift::protocol::TType _ktype1338; + ::apache::thrift::protocol::TType _vtype1339; + xfer += iprot->readMapBegin(_ktype1338, _vtype1339, _size1337); + uint32_t _i1341; + for (_i1341 = 0; _i1341 < _size1337; ++_i1341) { - std::string _key1341; - xfer += iprot->readString(_key1341); - std::string& _val1342 = this->partitionSpecs[_key1341]; - xfer += iprot->readString(_val1342); + std::string _key1342; + xfer += iprot->readString(_key1342); + std::string& _val1343 = this->partitionSpecs[_key1342]; + xfer += iprot->readString(_val1343); } xfer += iprot->readMapEnd(); } @@ -15030,11 +15030,11 @@ uint32_t ThriftHiveMetastore_exchange_partitions_args::write(::apache::thrift::p xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->partitionSpecs.size())); - std::map ::const_iterator _iter1343; - for (_iter1343 = this->partitionSpecs.begin(); _iter1343 != this->partitionSpecs.end(); ++_iter1343) + std::map ::const_iterator _iter1344; + for (_iter1344 = this->partitionSpecs.begin(); _iter1344 != this->partitionSpecs.end(); ++_iter1344) { - xfer += oprot->writeString(_iter1343->first); - xfer += oprot->writeString(_iter1343->second); + xfer += oprot->writeString(_iter1344->first); + xfer += oprot->writeString(_iter1344->second); } xfer += oprot->writeMapEnd(); } @@ -15074,11 +15074,11 @@ uint32_t ThriftHiveMetastore_exchange_partitions_pargs::write(::apache::thrift:: xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->partitionSpecs)).size())); - std::map ::const_iterator _iter1344; - for (_iter1344 = (*(this->partitionSpecs)).begin(); _iter1344 != (*(this->partitionSpecs)).end(); ++_iter1344) + std::map ::const_iterator _iter1345; + for (_iter1345 = (*(this->partitionSpecs)).begin(); _iter1345 != (*(this->partitionSpecs)).end(); ++_iter1345) { - xfer += oprot->writeString(_iter1344->first); - xfer += oprot->writeString(_iter1344->second); + xfer += oprot->writeString(_iter1345->first); + xfer += oprot->writeString(_iter1345->second); } xfer += oprot->writeMapEnd(); } @@ -15135,14 +15135,14 @@ uint32_t ThriftHiveMetastore_exchange_partitions_result::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1345; - ::apache::thrift::protocol::TType _etype1348; - xfer += iprot->readListBegin(_etype1348, _size1345); - this->success.resize(_size1345); - uint32_t _i1349; - for (_i1349 = 0; _i1349 < _size1345; ++_i1349) + uint32_t _size1346; + ::apache::thrift::protocol::TType _etype1349; + xfer += iprot->readListBegin(_etype1349, _size1346); + this->success.resize(_size1346); + uint32_t _i1350; + for (_i1350 = 0; _i1350 < _size1346; ++_i1350) { - xfer += this->success[_i1349].read(iprot); + xfer += this->success[_i1350].read(iprot); } xfer += iprot->readListEnd(); } @@ -15205,10 +15205,10 @@ uint32_t ThriftHiveMetastore_exchange_partitions_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1350; - for (_iter1350 = this->success.begin(); _iter1350 != this->success.end(); ++_iter1350) + std::vector ::const_iterator _iter1351; + for (_iter1351 = this->success.begin(); _iter1351 != this->success.end(); ++_iter1351) { - xfer += (*_iter1350).write(oprot); + xfer += (*_iter1351).write(oprot); } xfer += oprot->writeListEnd(); } @@ -15265,14 +15265,14 @@ uint32_t ThriftHiveMetastore_exchange_partitions_presult::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1351; - ::apache::thrift::protocol::TType _etype1354; - xfer += iprot->readListBegin(_etype1354, _size1351); - (*(this->success)).resize(_size1351); - uint32_t _i1355; - for (_i1355 = 0; _i1355 < _size1351; ++_i1355) + uint32_t _size1352; + ::apache::thrift::protocol::TType _etype1355; + xfer += iprot->readListBegin(_etype1355, _size1352); + (*(this->success)).resize(_size1352); + uint32_t _i1356; + for (_i1356 = 0; _i1356 < _size1352; ++_i1356) { - xfer += (*(this->success))[_i1355].read(iprot); + xfer += (*(this->success))[_i1356].read(iprot); } xfer += iprot->readListEnd(); } @@ -15371,14 +15371,14 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1356; - ::apache::thrift::protocol::TType _etype1359; - xfer += iprot->readListBegin(_etype1359, _size1356); - this->part_vals.resize(_size1356); - uint32_t _i1360; - for (_i1360 = 0; _i1360 < _size1356; ++_i1360) + uint32_t _size1357; + ::apache::thrift::protocol::TType _etype1360; + xfer += iprot->readListBegin(_etype1360, _size1357); + this->part_vals.resize(_size1357); + uint32_t _i1361; + for (_i1361 = 0; _i1361 < _size1357; ++_i1361) { - xfer += iprot->readString(this->part_vals[_i1360]); + xfer += iprot->readString(this->part_vals[_i1361]); } xfer += iprot->readListEnd(); } @@ -15399,14 +15399,14 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1361; - ::apache::thrift::protocol::TType _etype1364; - xfer += iprot->readListBegin(_etype1364, _size1361); - this->group_names.resize(_size1361); - uint32_t _i1365; - for (_i1365 = 0; _i1365 < _size1361; ++_i1365) + uint32_t _size1362; + ::apache::thrift::protocol::TType _etype1365; + xfer += iprot->readListBegin(_etype1365, _size1362); + this->group_names.resize(_size1362); + uint32_t _i1366; + for (_i1366 = 0; _i1366 < _size1362; ++_i1366) { - xfer += iprot->readString(this->group_names[_i1365]); + xfer += iprot->readString(this->group_names[_i1366]); } xfer += iprot->readListEnd(); } @@ -15443,10 +15443,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::write(::apache::thrif xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1366; - for (_iter1366 = this->part_vals.begin(); _iter1366 != this->part_vals.end(); ++_iter1366) + std::vector ::const_iterator _iter1367; + for (_iter1367 = this->part_vals.begin(); _iter1367 != this->part_vals.end(); ++_iter1367) { - xfer += oprot->writeString((*_iter1366)); + xfer += oprot->writeString((*_iter1367)); } xfer += oprot->writeListEnd(); } @@ -15459,10 +15459,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::write(::apache::thrif xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1367; - for (_iter1367 = this->group_names.begin(); _iter1367 != this->group_names.end(); ++_iter1367) + std::vector ::const_iterator _iter1368; + for (_iter1368 = this->group_names.begin(); _iter1368 != this->group_names.end(); ++_iter1368) { - xfer += oprot->writeString((*_iter1367)); + xfer += oprot->writeString((*_iter1368)); } xfer += oprot->writeListEnd(); } @@ -15494,10 +15494,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1368; - for (_iter1368 = (*(this->part_vals)).begin(); _iter1368 != (*(this->part_vals)).end(); ++_iter1368) + std::vector ::const_iterator _iter1369; + for (_iter1369 = (*(this->part_vals)).begin(); _iter1369 != (*(this->part_vals)).end(); ++_iter1369) { - xfer += oprot->writeString((*_iter1368)); + xfer += oprot->writeString((*_iter1369)); } xfer += oprot->writeListEnd(); } @@ -15510,10 +15510,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1369; - for (_iter1369 = (*(this->group_names)).begin(); _iter1369 != (*(this->group_names)).end(); ++_iter1369) + std::vector ::const_iterator _iter1370; + for (_iter1370 = (*(this->group_names)).begin(); _iter1370 != (*(this->group_names)).end(); ++_iter1370) { - xfer += oprot->writeString((*_iter1369)); + xfer += oprot->writeString((*_iter1370)); } xfer += oprot->writeListEnd(); } @@ -16072,14 +16072,14 @@ uint32_t ThriftHiveMetastore_get_partitions_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1370; - ::apache::thrift::protocol::TType _etype1373; - xfer += iprot->readListBegin(_etype1373, _size1370); - this->success.resize(_size1370); - uint32_t _i1374; - for (_i1374 = 0; _i1374 < _size1370; ++_i1374) + uint32_t _size1371; + ::apache::thrift::protocol::TType _etype1374; + xfer += iprot->readListBegin(_etype1374, _size1371); + this->success.resize(_size1371); + uint32_t _i1375; + for (_i1375 = 0; _i1375 < _size1371; ++_i1375) { - xfer += this->success[_i1374].read(iprot); + xfer += this->success[_i1375].read(iprot); } xfer += iprot->readListEnd(); } @@ -16126,10 +16126,10 @@ uint32_t ThriftHiveMetastore_get_partitions_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1375; - for (_iter1375 = this->success.begin(); _iter1375 != this->success.end(); ++_iter1375) + std::vector ::const_iterator _iter1376; + for (_iter1376 = this->success.begin(); _iter1376 != this->success.end(); ++_iter1376) { - xfer += (*_iter1375).write(oprot); + xfer += (*_iter1376).write(oprot); } xfer += oprot->writeListEnd(); } @@ -16178,14 +16178,14 @@ uint32_t ThriftHiveMetastore_get_partitions_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1376; - ::apache::thrift::protocol::TType _etype1379; - xfer += iprot->readListBegin(_etype1379, _size1376); - (*(this->success)).resize(_size1376); - uint32_t _i1380; - for (_i1380 = 0; _i1380 < _size1376; ++_i1380) + uint32_t _size1377; + ::apache::thrift::protocol::TType _etype1380; + xfer += iprot->readListBegin(_etype1380, _size1377); + (*(this->success)).resize(_size1377); + uint32_t _i1381; + for (_i1381 = 0; _i1381 < _size1377; ++_i1381) { - xfer += (*(this->success))[_i1380].read(iprot); + xfer += (*(this->success))[_i1381].read(iprot); } xfer += iprot->readListEnd(); } @@ -16284,14 +16284,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_args::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1381; - ::apache::thrift::protocol::TType _etype1384; - xfer += iprot->readListBegin(_etype1384, _size1381); - this->group_names.resize(_size1381); - uint32_t _i1385; - for (_i1385 = 0; _i1385 < _size1381; ++_i1385) + uint32_t _size1382; + ::apache::thrift::protocol::TType _etype1385; + xfer += iprot->readListBegin(_etype1385, _size1382); + this->group_names.resize(_size1382); + uint32_t _i1386; + for (_i1386 = 0; _i1386 < _size1382; ++_i1386) { - xfer += iprot->readString(this->group_names[_i1385]); + xfer += iprot->readString(this->group_names[_i1386]); } xfer += iprot->readListEnd(); } @@ -16336,10 +16336,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_args::write(::apache::thri xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1386; - for (_iter1386 = this->group_names.begin(); _iter1386 != this->group_names.end(); ++_iter1386) + std::vector ::const_iterator _iter1387; + for (_iter1387 = this->group_names.begin(); _iter1387 != this->group_names.end(); ++_iter1387) { - xfer += oprot->writeString((*_iter1386)); + xfer += oprot->writeString((*_iter1387)); } xfer += oprot->writeListEnd(); } @@ -16379,10 +16379,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_pargs::write(::apache::thr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1387; - for (_iter1387 = (*(this->group_names)).begin(); _iter1387 != (*(this->group_names)).end(); ++_iter1387) + std::vector ::const_iterator _iter1388; + for (_iter1388 = (*(this->group_names)).begin(); _iter1388 != (*(this->group_names)).end(); ++_iter1388) { - xfer += oprot->writeString((*_iter1387)); + xfer += oprot->writeString((*_iter1388)); } xfer += oprot->writeListEnd(); } @@ -16423,14 +16423,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1388; - ::apache::thrift::protocol::TType _etype1391; - xfer += iprot->readListBegin(_etype1391, _size1388); - this->success.resize(_size1388); - uint32_t _i1392; - for (_i1392 = 0; _i1392 < _size1388; ++_i1392) + uint32_t _size1389; + ::apache::thrift::protocol::TType _etype1392; + xfer += iprot->readListBegin(_etype1392, _size1389); + this->success.resize(_size1389); + uint32_t _i1393; + for (_i1393 = 0; _i1393 < _size1389; ++_i1393) { - xfer += this->success[_i1392].read(iprot); + xfer += this->success[_i1393].read(iprot); } xfer += iprot->readListEnd(); } @@ -16477,10 +16477,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1393; - for (_iter1393 = this->success.begin(); _iter1393 != this->success.end(); ++_iter1393) + std::vector ::const_iterator _iter1394; + for (_iter1394 = this->success.begin(); _iter1394 != this->success.end(); ++_iter1394) { - xfer += (*_iter1393).write(oprot); + xfer += (*_iter1394).write(oprot); } xfer += oprot->writeListEnd(); } @@ -16529,14 +16529,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1394; - ::apache::thrift::protocol::TType _etype1397; - xfer += iprot->readListBegin(_etype1397, _size1394); - (*(this->success)).resize(_size1394); - uint32_t _i1398; - for (_i1398 = 0; _i1398 < _size1394; ++_i1398) + uint32_t _size1395; + ::apache::thrift::protocol::TType _etype1398; + xfer += iprot->readListBegin(_etype1398, _size1395); + (*(this->success)).resize(_size1395); + uint32_t _i1399; + for (_i1399 = 0; _i1399 < _size1395; ++_i1399) { - xfer += (*(this->success))[_i1398].read(iprot); + xfer += (*(this->success))[_i1399].read(iprot); } xfer += iprot->readListEnd(); } @@ -16714,14 +16714,14 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_result::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1399; - ::apache::thrift::protocol::TType _etype1402; - xfer += iprot->readListBegin(_etype1402, _size1399); - this->success.resize(_size1399); - uint32_t _i1403; - for (_i1403 = 0; _i1403 < _size1399; ++_i1403) + uint32_t _size1400; + ::apache::thrift::protocol::TType _etype1403; + xfer += iprot->readListBegin(_etype1403, _size1400); + this->success.resize(_size1400); + uint32_t _i1404; + for (_i1404 = 0; _i1404 < _size1400; ++_i1404) { - xfer += this->success[_i1403].read(iprot); + xfer += this->success[_i1404].read(iprot); } xfer += iprot->readListEnd(); } @@ -16768,10 +16768,10 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_result::write(::apache::thrift xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1404; - for (_iter1404 = this->success.begin(); _iter1404 != this->success.end(); ++_iter1404) + std::vector ::const_iterator _iter1405; + for (_iter1405 = this->success.begin(); _iter1405 != this->success.end(); ++_iter1405) { - xfer += (*_iter1404).write(oprot); + xfer += (*_iter1405).write(oprot); } xfer += oprot->writeListEnd(); } @@ -16820,14 +16820,14 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_presult::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1405; - ::apache::thrift::protocol::TType _etype1408; - xfer += iprot->readListBegin(_etype1408, _size1405); - (*(this->success)).resize(_size1405); - uint32_t _i1409; - for (_i1409 = 0; _i1409 < _size1405; ++_i1409) + uint32_t _size1406; + ::apache::thrift::protocol::TType _etype1409; + xfer += iprot->readListBegin(_etype1409, _size1406); + (*(this->success)).resize(_size1406); + uint32_t _i1410; + for (_i1410 = 0; _i1410 < _size1406; ++_i1410) { - xfer += (*(this->success))[_i1409].read(iprot); + xfer += (*(this->success))[_i1410].read(iprot); } xfer += iprot->readListEnd(); } @@ -17005,14 +17005,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_result::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1410; - ::apache::thrift::protocol::TType _etype1413; - xfer += iprot->readListBegin(_etype1413, _size1410); - this->success.resize(_size1410); - uint32_t _i1414; - for (_i1414 = 0; _i1414 < _size1410; ++_i1414) + uint32_t _size1411; + ::apache::thrift::protocol::TType _etype1414; + xfer += iprot->readListBegin(_etype1414, _size1411); + this->success.resize(_size1411); + uint32_t _i1415; + for (_i1415 = 0; _i1415 < _size1411; ++_i1415) { - xfer += iprot->readString(this->success[_i1414]); + xfer += iprot->readString(this->success[_i1415]); } xfer += iprot->readListEnd(); } @@ -17059,10 +17059,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1415; - for (_iter1415 = this->success.begin(); _iter1415 != this->success.end(); ++_iter1415) + std::vector ::const_iterator _iter1416; + for (_iter1416 = this->success.begin(); _iter1416 != this->success.end(); ++_iter1416) { - xfer += oprot->writeString((*_iter1415)); + xfer += oprot->writeString((*_iter1416)); } xfer += oprot->writeListEnd(); } @@ -17111,14 +17111,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_presult::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1416; - ::apache::thrift::protocol::TType _etype1419; - xfer += iprot->readListBegin(_etype1419, _size1416); - (*(this->success)).resize(_size1416); - uint32_t _i1420; - for (_i1420 = 0; _i1420 < _size1416; ++_i1420) + uint32_t _size1417; + ::apache::thrift::protocol::TType _etype1420; + xfer += iprot->readListBegin(_etype1420, _size1417); + (*(this->success)).resize(_size1417); + uint32_t _i1421; + for (_i1421 = 0; _i1421 < _size1417; ++_i1421) { - xfer += iprot->readString((*(this->success))[_i1420]); + xfer += iprot->readString((*(this->success))[_i1421]); } xfer += iprot->readListEnd(); } @@ -17428,14 +17428,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_args::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1421; - ::apache::thrift::protocol::TType _etype1424; - xfer += iprot->readListBegin(_etype1424, _size1421); - this->part_vals.resize(_size1421); - uint32_t _i1425; - for (_i1425 = 0; _i1425 < _size1421; ++_i1425) + uint32_t _size1422; + ::apache::thrift::protocol::TType _etype1425; + xfer += iprot->readListBegin(_etype1425, _size1422); + this->part_vals.resize(_size1422); + uint32_t _i1426; + for (_i1426 = 0; _i1426 < _size1422; ++_i1426) { - xfer += iprot->readString(this->part_vals[_i1425]); + xfer += iprot->readString(this->part_vals[_i1426]); } xfer += iprot->readListEnd(); } @@ -17480,10 +17480,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_args::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1426; - for (_iter1426 = this->part_vals.begin(); _iter1426 != this->part_vals.end(); ++_iter1426) + std::vector ::const_iterator _iter1427; + for (_iter1427 = this->part_vals.begin(); _iter1427 != this->part_vals.end(); ++_iter1427) { - xfer += oprot->writeString((*_iter1426)); + xfer += oprot->writeString((*_iter1427)); } xfer += oprot->writeListEnd(); } @@ -17519,10 +17519,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_pargs::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1427; - for (_iter1427 = (*(this->part_vals)).begin(); _iter1427 != (*(this->part_vals)).end(); ++_iter1427) + std::vector ::const_iterator _iter1428; + for (_iter1428 = (*(this->part_vals)).begin(); _iter1428 != (*(this->part_vals)).end(); ++_iter1428) { - xfer += oprot->writeString((*_iter1427)); + xfer += oprot->writeString((*_iter1428)); } xfer += oprot->writeListEnd(); } @@ -17567,14 +17567,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1428; - ::apache::thrift::protocol::TType _etype1431; - xfer += iprot->readListBegin(_etype1431, _size1428); - this->success.resize(_size1428); - uint32_t _i1432; - for (_i1432 = 0; _i1432 < _size1428; ++_i1432) + uint32_t _size1429; + ::apache::thrift::protocol::TType _etype1432; + xfer += iprot->readListBegin(_etype1432, _size1429); + this->success.resize(_size1429); + uint32_t _i1433; + for (_i1433 = 0; _i1433 < _size1429; ++_i1433) { - xfer += this->success[_i1432].read(iprot); + xfer += this->success[_i1433].read(iprot); } xfer += iprot->readListEnd(); } @@ -17621,10 +17621,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1433; - for (_iter1433 = this->success.begin(); _iter1433 != this->success.end(); ++_iter1433) + std::vector ::const_iterator _iter1434; + for (_iter1434 = this->success.begin(); _iter1434 != this->success.end(); ++_iter1434) { - xfer += (*_iter1433).write(oprot); + xfer += (*_iter1434).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17673,14 +17673,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1434; - ::apache::thrift::protocol::TType _etype1437; - xfer += iprot->readListBegin(_etype1437, _size1434); - (*(this->success)).resize(_size1434); - uint32_t _i1438; - for (_i1438 = 0; _i1438 < _size1434; ++_i1438) + uint32_t _size1435; + ::apache::thrift::protocol::TType _etype1438; + xfer += iprot->readListBegin(_etype1438, _size1435); + (*(this->success)).resize(_size1435); + uint32_t _i1439; + for (_i1439 = 0; _i1439 < _size1435; ++_i1439) { - xfer += (*(this->success))[_i1438].read(iprot); + xfer += (*(this->success))[_i1439].read(iprot); } xfer += iprot->readListEnd(); } @@ -17763,14 +17763,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1439; - ::apache::thrift::protocol::TType _etype1442; - xfer += iprot->readListBegin(_etype1442, _size1439); - this->part_vals.resize(_size1439); - uint32_t _i1443; - for (_i1443 = 0; _i1443 < _size1439; ++_i1443) + uint32_t _size1440; + ::apache::thrift::protocol::TType _etype1443; + xfer += iprot->readListBegin(_etype1443, _size1440); + this->part_vals.resize(_size1440); + uint32_t _i1444; + for (_i1444 = 0; _i1444 < _size1440; ++_i1444) { - xfer += iprot->readString(this->part_vals[_i1443]); + xfer += iprot->readString(this->part_vals[_i1444]); } xfer += iprot->readListEnd(); } @@ -17799,14 +17799,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1444; - ::apache::thrift::protocol::TType _etype1447; - xfer += iprot->readListBegin(_etype1447, _size1444); - this->group_names.resize(_size1444); - uint32_t _i1448; - for (_i1448 = 0; _i1448 < _size1444; ++_i1448) + uint32_t _size1445; + ::apache::thrift::protocol::TType _etype1448; + xfer += iprot->readListBegin(_etype1448, _size1445); + this->group_names.resize(_size1445); + uint32_t _i1449; + for (_i1449 = 0; _i1449 < _size1445; ++_i1449) { - xfer += iprot->readString(this->group_names[_i1448]); + xfer += iprot->readString(this->group_names[_i1449]); } xfer += iprot->readListEnd(); } @@ -17843,10 +17843,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::write(::apache::t xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1449; - for (_iter1449 = this->part_vals.begin(); _iter1449 != this->part_vals.end(); ++_iter1449) + std::vector ::const_iterator _iter1450; + for (_iter1450 = this->part_vals.begin(); _iter1450 != this->part_vals.end(); ++_iter1450) { - xfer += oprot->writeString((*_iter1449)); + xfer += oprot->writeString((*_iter1450)); } xfer += oprot->writeListEnd(); } @@ -17863,10 +17863,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::write(::apache::t xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1450; - for (_iter1450 = this->group_names.begin(); _iter1450 != this->group_names.end(); ++_iter1450) + std::vector ::const_iterator _iter1451; + for (_iter1451 = this->group_names.begin(); _iter1451 != this->group_names.end(); ++_iter1451) { - xfer += oprot->writeString((*_iter1450)); + xfer += oprot->writeString((*_iter1451)); } xfer += oprot->writeListEnd(); } @@ -17898,10 +17898,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_pargs::write(::apache:: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1451; - for (_iter1451 = (*(this->part_vals)).begin(); _iter1451 != (*(this->part_vals)).end(); ++_iter1451) + std::vector ::const_iterator _iter1452; + for (_iter1452 = (*(this->part_vals)).begin(); _iter1452 != (*(this->part_vals)).end(); ++_iter1452) { - xfer += oprot->writeString((*_iter1451)); + xfer += oprot->writeString((*_iter1452)); } xfer += oprot->writeListEnd(); } @@ -17918,10 +17918,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_pargs::write(::apache:: xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1452; - for (_iter1452 = (*(this->group_names)).begin(); _iter1452 != (*(this->group_names)).end(); ++_iter1452) + std::vector ::const_iterator _iter1453; + for (_iter1453 = (*(this->group_names)).begin(); _iter1453 != (*(this->group_names)).end(); ++_iter1453) { - xfer += oprot->writeString((*_iter1452)); + xfer += oprot->writeString((*_iter1453)); } xfer += oprot->writeListEnd(); } @@ -17962,14 +17962,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_result::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1453; - ::apache::thrift::protocol::TType _etype1456; - xfer += iprot->readListBegin(_etype1456, _size1453); - this->success.resize(_size1453); - uint32_t _i1457; - for (_i1457 = 0; _i1457 < _size1453; ++_i1457) + uint32_t _size1454; + ::apache::thrift::protocol::TType _etype1457; + xfer += iprot->readListBegin(_etype1457, _size1454); + this->success.resize(_size1454); + uint32_t _i1458; + for (_i1458 = 0; _i1458 < _size1454; ++_i1458) { - xfer += this->success[_i1457].read(iprot); + xfer += this->success[_i1458].read(iprot); } xfer += iprot->readListEnd(); } @@ -18016,10 +18016,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_result::write(::apache: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1458; - for (_iter1458 = this->success.begin(); _iter1458 != this->success.end(); ++_iter1458) + std::vector ::const_iterator _iter1459; + for (_iter1459 = this->success.begin(); _iter1459 != this->success.end(); ++_iter1459) { - xfer += (*_iter1458).write(oprot); + xfer += (*_iter1459).write(oprot); } xfer += oprot->writeListEnd(); } @@ -18068,14 +18068,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_presult::read(::apache: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1459; - ::apache::thrift::protocol::TType _etype1462; - xfer += iprot->readListBegin(_etype1462, _size1459); - (*(this->success)).resize(_size1459); - uint32_t _i1463; - for (_i1463 = 0; _i1463 < _size1459; ++_i1463) + uint32_t _size1460; + ::apache::thrift::protocol::TType _etype1463; + xfer += iprot->readListBegin(_etype1463, _size1460); + (*(this->success)).resize(_size1460); + uint32_t _i1464; + for (_i1464 = 0; _i1464 < _size1460; ++_i1464) { - xfer += (*(this->success))[_i1463].read(iprot); + xfer += (*(this->success))[_i1464].read(iprot); } xfer += iprot->readListEnd(); } @@ -18158,14 +18158,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_args::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1464; - ::apache::thrift::protocol::TType _etype1467; - xfer += iprot->readListBegin(_etype1467, _size1464); - this->part_vals.resize(_size1464); - uint32_t _i1468; - for (_i1468 = 0; _i1468 < _size1464; ++_i1468) + uint32_t _size1465; + ::apache::thrift::protocol::TType _etype1468; + xfer += iprot->readListBegin(_etype1468, _size1465); + this->part_vals.resize(_size1465); + uint32_t _i1469; + for (_i1469 = 0; _i1469 < _size1465; ++_i1469) { - xfer += iprot->readString(this->part_vals[_i1468]); + xfer += iprot->readString(this->part_vals[_i1469]); } xfer += iprot->readListEnd(); } @@ -18210,10 +18210,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_args::write(::apache::thrift xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1469; - for (_iter1469 = this->part_vals.begin(); _iter1469 != this->part_vals.end(); ++_iter1469) + std::vector ::const_iterator _iter1470; + for (_iter1470 = this->part_vals.begin(); _iter1470 != this->part_vals.end(); ++_iter1470) { - xfer += oprot->writeString((*_iter1469)); + xfer += oprot->writeString((*_iter1470)); } xfer += oprot->writeListEnd(); } @@ -18249,10 +18249,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_pargs::write(::apache::thrif xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1470; - for (_iter1470 = (*(this->part_vals)).begin(); _iter1470 != (*(this->part_vals)).end(); ++_iter1470) + std::vector ::const_iterator _iter1471; + for (_iter1471 = (*(this->part_vals)).begin(); _iter1471 != (*(this->part_vals)).end(); ++_iter1471) { - xfer += oprot->writeString((*_iter1470)); + xfer += oprot->writeString((*_iter1471)); } xfer += oprot->writeListEnd(); } @@ -18297,14 +18297,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1471; - ::apache::thrift::protocol::TType _etype1474; - xfer += iprot->readListBegin(_etype1474, _size1471); - this->success.resize(_size1471); - uint32_t _i1475; - for (_i1475 = 0; _i1475 < _size1471; ++_i1475) + uint32_t _size1472; + ::apache::thrift::protocol::TType _etype1475; + xfer += iprot->readListBegin(_etype1475, _size1472); + this->success.resize(_size1472); + uint32_t _i1476; + for (_i1476 = 0; _i1476 < _size1472; ++_i1476) { - xfer += iprot->readString(this->success[_i1475]); + xfer += iprot->readString(this->success[_i1476]); } xfer += iprot->readListEnd(); } @@ -18351,10 +18351,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1476; - for (_iter1476 = this->success.begin(); _iter1476 != this->success.end(); ++_iter1476) + std::vector ::const_iterator _iter1477; + for (_iter1477 = this->success.begin(); _iter1477 != this->success.end(); ++_iter1477) { - xfer += oprot->writeString((*_iter1476)); + xfer += oprot->writeString((*_iter1477)); } xfer += oprot->writeListEnd(); } @@ -18403,14 +18403,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1477; - ::apache::thrift::protocol::TType _etype1480; - xfer += iprot->readListBegin(_etype1480, _size1477); - (*(this->success)).resize(_size1477); - uint32_t _i1481; - for (_i1481 = 0; _i1481 < _size1477; ++_i1481) + uint32_t _size1478; + ::apache::thrift::protocol::TType _etype1481; + xfer += iprot->readListBegin(_etype1481, _size1478); + (*(this->success)).resize(_size1478); + uint32_t _i1482; + for (_i1482 = 0; _i1482 < _size1478; ++_i1482) { - xfer += iprot->readString((*(this->success))[_i1481]); + xfer += iprot->readString((*(this->success))[_i1482]); } xfer += iprot->readListEnd(); } @@ -18604,14 +18604,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1482; - ::apache::thrift::protocol::TType _etype1485; - xfer += iprot->readListBegin(_etype1485, _size1482); - this->success.resize(_size1482); - uint32_t _i1486; - for (_i1486 = 0; _i1486 < _size1482; ++_i1486) + uint32_t _size1483; + ::apache::thrift::protocol::TType _etype1486; + xfer += iprot->readListBegin(_etype1486, _size1483); + this->success.resize(_size1483); + uint32_t _i1487; + for (_i1487 = 0; _i1487 < _size1483; ++_i1487) { - xfer += this->success[_i1486].read(iprot); + xfer += this->success[_i1487].read(iprot); } xfer += iprot->readListEnd(); } @@ -18658,10 +18658,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1487; - for (_iter1487 = this->success.begin(); _iter1487 != this->success.end(); ++_iter1487) + std::vector ::const_iterator _iter1488; + for (_iter1488 = this->success.begin(); _iter1488 != this->success.end(); ++_iter1488) { - xfer += (*_iter1487).write(oprot); + xfer += (*_iter1488).write(oprot); } xfer += oprot->writeListEnd(); } @@ -18710,14 +18710,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1488; - ::apache::thrift::protocol::TType _etype1491; - xfer += iprot->readListBegin(_etype1491, _size1488); - (*(this->success)).resize(_size1488); - uint32_t _i1492; - for (_i1492 = 0; _i1492 < _size1488; ++_i1492) + uint32_t _size1489; + ::apache::thrift::protocol::TType _etype1492; + xfer += iprot->readListBegin(_etype1492, _size1489); + (*(this->success)).resize(_size1489); + uint32_t _i1493; + for (_i1493 = 0; _i1493 < _size1489; ++_i1493) { - xfer += (*(this->success))[_i1492].read(iprot); + xfer += (*(this->success))[_i1493].read(iprot); } xfer += iprot->readListEnd(); } @@ -18911,14 +18911,14 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1493; - ::apache::thrift::protocol::TType _etype1496; - xfer += iprot->readListBegin(_etype1496, _size1493); - this->success.resize(_size1493); - uint32_t _i1497; - for (_i1497 = 0; _i1497 < _size1493; ++_i1497) + uint32_t _size1494; + ::apache::thrift::protocol::TType _etype1497; + xfer += iprot->readListBegin(_etype1497, _size1494); + this->success.resize(_size1494); + uint32_t _i1498; + for (_i1498 = 0; _i1498 < _size1494; ++_i1498) { - xfer += this->success[_i1497].read(iprot); + xfer += this->success[_i1498].read(iprot); } xfer += iprot->readListEnd(); } @@ -18965,10 +18965,10 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1498; - for (_iter1498 = this->success.begin(); _iter1498 != this->success.end(); ++_iter1498) + std::vector ::const_iterator _iter1499; + for (_iter1499 = this->success.begin(); _iter1499 != this->success.end(); ++_iter1499) { - xfer += (*_iter1498).write(oprot); + xfer += (*_iter1499).write(oprot); } xfer += oprot->writeListEnd(); } @@ -19017,14 +19017,14 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1499; - ::apache::thrift::protocol::TType _etype1502; - xfer += iprot->readListBegin(_etype1502, _size1499); - (*(this->success)).resize(_size1499); - uint32_t _i1503; - for (_i1503 = 0; _i1503 < _size1499; ++_i1503) + uint32_t _size1500; + ::apache::thrift::protocol::TType _etype1503; + xfer += iprot->readListBegin(_etype1503, _size1500); + (*(this->success)).resize(_size1500); + uint32_t _i1504; + for (_i1504 = 0; _i1504 < _size1500; ++_i1504) { - xfer += (*(this->success))[_i1503].read(iprot); + xfer += (*(this->success))[_i1504].read(iprot); } xfer += iprot->readListEnd(); } @@ -19593,14 +19593,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->names.clear(); - uint32_t _size1504; - ::apache::thrift::protocol::TType _etype1507; - xfer += iprot->readListBegin(_etype1507, _size1504); - this->names.resize(_size1504); - uint32_t _i1508; - for (_i1508 = 0; _i1508 < _size1504; ++_i1508) + uint32_t _size1505; + ::apache::thrift::protocol::TType _etype1508; + xfer += iprot->readListBegin(_etype1508, _size1505); + this->names.resize(_size1505); + uint32_t _i1509; + for (_i1509 = 0; _i1509 < _size1505; ++_i1509) { - xfer += iprot->readString(this->names[_i1508]); + xfer += iprot->readString(this->names[_i1509]); } xfer += iprot->readListEnd(); } @@ -19637,10 +19637,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_args::write(::apache::thrif xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->names.size())); - std::vector ::const_iterator _iter1509; - for (_iter1509 = this->names.begin(); _iter1509 != this->names.end(); ++_iter1509) + std::vector ::const_iterator _iter1510; + for (_iter1510 = this->names.begin(); _iter1510 != this->names.end(); ++_iter1510) { - xfer += oprot->writeString((*_iter1509)); + xfer += oprot->writeString((*_iter1510)); } xfer += oprot->writeListEnd(); } @@ -19672,10 +19672,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->names)).size())); - std::vector ::const_iterator _iter1510; - for (_iter1510 = (*(this->names)).begin(); _iter1510 != (*(this->names)).end(); ++_iter1510) + std::vector ::const_iterator _iter1511; + for (_iter1511 = (*(this->names)).begin(); _iter1511 != (*(this->names)).end(); ++_iter1511) { - xfer += oprot->writeString((*_iter1510)); + xfer += oprot->writeString((*_iter1511)); } xfer += oprot->writeListEnd(); } @@ -19716,14 +19716,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1511; - ::apache::thrift::protocol::TType _etype1514; - xfer += iprot->readListBegin(_etype1514, _size1511); - this->success.resize(_size1511); - uint32_t _i1515; - for (_i1515 = 0; _i1515 < _size1511; ++_i1515) + uint32_t _size1512; + ::apache::thrift::protocol::TType _etype1515; + xfer += iprot->readListBegin(_etype1515, _size1512); + this->success.resize(_size1512); + uint32_t _i1516; + for (_i1516 = 0; _i1516 < _size1512; ++_i1516) { - xfer += this->success[_i1515].read(iprot); + xfer += this->success[_i1516].read(iprot); } xfer += iprot->readListEnd(); } @@ -19770,10 +19770,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1516; - for (_iter1516 = this->success.begin(); _iter1516 != this->success.end(); ++_iter1516) + std::vector ::const_iterator _iter1517; + for (_iter1517 = this->success.begin(); _iter1517 != this->success.end(); ++_iter1517) { - xfer += (*_iter1516).write(oprot); + xfer += (*_iter1517).write(oprot); } xfer += oprot->writeListEnd(); } @@ -19822,14 +19822,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1517; - ::apache::thrift::protocol::TType _etype1520; - xfer += iprot->readListBegin(_etype1520, _size1517); - (*(this->success)).resize(_size1517); - uint32_t _i1521; - for (_i1521 = 0; _i1521 < _size1517; ++_i1521) + uint32_t _size1518; + ::apache::thrift::protocol::TType _etype1521; + xfer += iprot->readListBegin(_etype1521, _size1518); + (*(this->success)).resize(_size1518); + uint32_t _i1522; + for (_i1522 = 0; _i1522 < _size1518; ++_i1522) { - xfer += (*(this->success))[_i1521].read(iprot); + xfer += (*(this->success))[_i1522].read(iprot); } xfer += iprot->readListEnd(); } @@ -20151,14 +20151,14 @@ uint32_t ThriftHiveMetastore_alter_partitions_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1522; - ::apache::thrift::protocol::TType _etype1525; - xfer += iprot->readListBegin(_etype1525, _size1522); - this->new_parts.resize(_size1522); - uint32_t _i1526; - for (_i1526 = 0; _i1526 < _size1522; ++_i1526) + uint32_t _size1523; + ::apache::thrift::protocol::TType _etype1526; + xfer += iprot->readListBegin(_etype1526, _size1523); + this->new_parts.resize(_size1523); + uint32_t _i1527; + for (_i1527 = 0; _i1527 < _size1523; ++_i1527) { - xfer += this->new_parts[_i1526].read(iprot); + xfer += this->new_parts[_i1527].read(iprot); } xfer += iprot->readListEnd(); } @@ -20195,10 +20195,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1527; - for (_iter1527 = this->new_parts.begin(); _iter1527 != this->new_parts.end(); ++_iter1527) + std::vector ::const_iterator _iter1528; + for (_iter1528 = this->new_parts.begin(); _iter1528 != this->new_parts.end(); ++_iter1528) { - xfer += (*_iter1527).write(oprot); + xfer += (*_iter1528).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20230,10 +20230,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1528; - for (_iter1528 = (*(this->new_parts)).begin(); _iter1528 != (*(this->new_parts)).end(); ++_iter1528) + std::vector ::const_iterator _iter1529; + for (_iter1529 = (*(this->new_parts)).begin(); _iter1529 != (*(this->new_parts)).end(); ++_iter1529) { - xfer += (*_iter1528).write(oprot); + xfer += (*_iter1529).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20418,14 +20418,14 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_args::rea if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1529; - ::apache::thrift::protocol::TType _etype1532; - xfer += iprot->readListBegin(_etype1532, _size1529); - this->new_parts.resize(_size1529); - uint32_t _i1533; - for (_i1533 = 0; _i1533 < _size1529; ++_i1533) + uint32_t _size1530; + ::apache::thrift::protocol::TType _etype1533; + xfer += iprot->readListBegin(_etype1533, _size1530); + this->new_parts.resize(_size1530); + uint32_t _i1534; + for (_i1534 = 0; _i1534 < _size1530; ++_i1534) { - xfer += this->new_parts[_i1533].read(iprot); + xfer += this->new_parts[_i1534].read(iprot); } xfer += iprot->readListEnd(); } @@ -20470,10 +20470,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_args::wri xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1534; - for (_iter1534 = this->new_parts.begin(); _iter1534 != this->new_parts.end(); ++_iter1534) + std::vector ::const_iterator _iter1535; + for (_iter1535 = this->new_parts.begin(); _iter1535 != this->new_parts.end(); ++_iter1535) { - xfer += (*_iter1534).write(oprot); + xfer += (*_iter1535).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20509,10 +20509,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_pargs::wr xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1535; - for (_iter1535 = (*(this->new_parts)).begin(); _iter1535 != (*(this->new_parts)).end(); ++_iter1535) + std::vector ::const_iterator _iter1536; + for (_iter1536 = (*(this->new_parts)).begin(); _iter1536 != (*(this->new_parts)).end(); ++_iter1536) { - xfer += (*_iter1535).write(oprot); + xfer += (*_iter1536).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20956,14 +20956,14 @@ uint32_t ThriftHiveMetastore_rename_partition_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1536; - ::apache::thrift::protocol::TType _etype1539; - xfer += iprot->readListBegin(_etype1539, _size1536); - this->part_vals.resize(_size1536); - uint32_t _i1540; - for (_i1540 = 0; _i1540 < _size1536; ++_i1540) + uint32_t _size1537; + ::apache::thrift::protocol::TType _etype1540; + xfer += iprot->readListBegin(_etype1540, _size1537); + this->part_vals.resize(_size1537); + uint32_t _i1541; + for (_i1541 = 0; _i1541 < _size1537; ++_i1541) { - xfer += iprot->readString(this->part_vals[_i1540]); + xfer += iprot->readString(this->part_vals[_i1541]); } xfer += iprot->readListEnd(); } @@ -21008,10 +21008,10 @@ uint32_t ThriftHiveMetastore_rename_partition_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1541; - for (_iter1541 = this->part_vals.begin(); _iter1541 != this->part_vals.end(); ++_iter1541) + std::vector ::const_iterator _iter1542; + for (_iter1542 = this->part_vals.begin(); _iter1542 != this->part_vals.end(); ++_iter1542) { - xfer += oprot->writeString((*_iter1541)); + xfer += oprot->writeString((*_iter1542)); } xfer += oprot->writeListEnd(); } @@ -21047,10 +21047,10 @@ uint32_t ThriftHiveMetastore_rename_partition_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1542; - for (_iter1542 = (*(this->part_vals)).begin(); _iter1542 != (*(this->part_vals)).end(); ++_iter1542) + std::vector ::const_iterator _iter1543; + for (_iter1543 = (*(this->part_vals)).begin(); _iter1543 != (*(this->part_vals)).end(); ++_iter1543) { - xfer += oprot->writeString((*_iter1542)); + xfer += oprot->writeString((*_iter1543)); } xfer += oprot->writeListEnd(); } @@ -21223,14 +21223,14 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_args::read(::ap if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1543; - ::apache::thrift::protocol::TType _etype1546; - xfer += iprot->readListBegin(_etype1546, _size1543); - this->part_vals.resize(_size1543); - uint32_t _i1547; - for (_i1547 = 0; _i1547 < _size1543; ++_i1547) + uint32_t _size1544; + ::apache::thrift::protocol::TType _etype1547; + xfer += iprot->readListBegin(_etype1547, _size1544); + this->part_vals.resize(_size1544); + uint32_t _i1548; + for (_i1548 = 0; _i1548 < _size1544; ++_i1548) { - xfer += iprot->readString(this->part_vals[_i1547]); + xfer += iprot->readString(this->part_vals[_i1548]); } xfer += iprot->readListEnd(); } @@ -21267,10 +21267,10 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_args::write(::a xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1548; - for (_iter1548 = this->part_vals.begin(); _iter1548 != this->part_vals.end(); ++_iter1548) + std::vector ::const_iterator _iter1549; + for (_iter1549 = this->part_vals.begin(); _iter1549 != this->part_vals.end(); ++_iter1549) { - xfer += oprot->writeString((*_iter1548)); + xfer += oprot->writeString((*_iter1549)); } xfer += oprot->writeListEnd(); } @@ -21298,10 +21298,10 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_pargs::write(:: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1549; - for (_iter1549 = (*(this->part_vals)).begin(); _iter1549 != (*(this->part_vals)).end(); ++_iter1549) + std::vector ::const_iterator _iter1550; + for (_iter1550 = (*(this->part_vals)).begin(); _iter1550 != (*(this->part_vals)).end(); ++_iter1550) { - xfer += oprot->writeString((*_iter1549)); + xfer += oprot->writeString((*_iter1550)); } xfer += oprot->writeListEnd(); } @@ -21776,14 +21776,14 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1550; - ::apache::thrift::protocol::TType _etype1553; - xfer += iprot->readListBegin(_etype1553, _size1550); - this->success.resize(_size1550); - uint32_t _i1554; - for (_i1554 = 0; _i1554 < _size1550; ++_i1554) + uint32_t _size1551; + ::apache::thrift::protocol::TType _etype1554; + xfer += iprot->readListBegin(_etype1554, _size1551); + this->success.resize(_size1551); + uint32_t _i1555; + for (_i1555 = 0; _i1555 < _size1551; ++_i1555) { - xfer += iprot->readString(this->success[_i1554]); + xfer += iprot->readString(this->success[_i1555]); } xfer += iprot->readListEnd(); } @@ -21822,10 +21822,10 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1555; - for (_iter1555 = this->success.begin(); _iter1555 != this->success.end(); ++_iter1555) + std::vector ::const_iterator _iter1556; + for (_iter1556 = this->success.begin(); _iter1556 != this->success.end(); ++_iter1556) { - xfer += oprot->writeString((*_iter1555)); + xfer += oprot->writeString((*_iter1556)); } xfer += oprot->writeListEnd(); } @@ -21870,14 +21870,14 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1556; - ::apache::thrift::protocol::TType _etype1559; - xfer += iprot->readListBegin(_etype1559, _size1556); - (*(this->success)).resize(_size1556); - uint32_t _i1560; - for (_i1560 = 0; _i1560 < _size1556; ++_i1560) + uint32_t _size1557; + ::apache::thrift::protocol::TType _etype1560; + xfer += iprot->readListBegin(_etype1560, _size1557); + (*(this->success)).resize(_size1557); + uint32_t _i1561; + for (_i1561 = 0; _i1561 < _size1557; ++_i1561) { - xfer += iprot->readString((*(this->success))[_i1560]); + xfer += iprot->readString((*(this->success))[_i1561]); } xfer += iprot->readListEnd(); } @@ -22015,17 +22015,17 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size1561; - ::apache::thrift::protocol::TType _ktype1562; - ::apache::thrift::protocol::TType _vtype1563; - xfer += iprot->readMapBegin(_ktype1562, _vtype1563, _size1561); - uint32_t _i1565; - for (_i1565 = 0; _i1565 < _size1561; ++_i1565) + uint32_t _size1562; + ::apache::thrift::protocol::TType _ktype1563; + ::apache::thrift::protocol::TType _vtype1564; + xfer += iprot->readMapBegin(_ktype1563, _vtype1564, _size1562); + uint32_t _i1566; + for (_i1566 = 0; _i1566 < _size1562; ++_i1566) { - std::string _key1566; - xfer += iprot->readString(_key1566); - std::string& _val1567 = this->success[_key1566]; - xfer += iprot->readString(_val1567); + std::string _key1567; + xfer += iprot->readString(_key1567); + std::string& _val1568 = this->success[_key1567]; + xfer += iprot->readString(_val1568); } xfer += iprot->readMapEnd(); } @@ -22064,11 +22064,11 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::map ::const_iterator _iter1568; - for (_iter1568 = this->success.begin(); _iter1568 != this->success.end(); ++_iter1568) + std::map ::const_iterator _iter1569; + for (_iter1569 = this->success.begin(); _iter1569 != this->success.end(); ++_iter1569) { - xfer += oprot->writeString(_iter1568->first); - xfer += oprot->writeString(_iter1568->second); + xfer += oprot->writeString(_iter1569->first); + xfer += oprot->writeString(_iter1569->second); } xfer += oprot->writeMapEnd(); } @@ -22113,17 +22113,17 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size1569; - ::apache::thrift::protocol::TType _ktype1570; - ::apache::thrift::protocol::TType _vtype1571; - xfer += iprot->readMapBegin(_ktype1570, _vtype1571, _size1569); - uint32_t _i1573; - for (_i1573 = 0; _i1573 < _size1569; ++_i1573) + uint32_t _size1570; + ::apache::thrift::protocol::TType _ktype1571; + ::apache::thrift::protocol::TType _vtype1572; + xfer += iprot->readMapBegin(_ktype1571, _vtype1572, _size1570); + uint32_t _i1574; + for (_i1574 = 0; _i1574 < _size1570; ++_i1574) { - std::string _key1574; - xfer += iprot->readString(_key1574); - std::string& _val1575 = (*(this->success))[_key1574]; - xfer += iprot->readString(_val1575); + std::string _key1575; + xfer += iprot->readString(_key1575); + std::string& _val1576 = (*(this->success))[_key1575]; + xfer += iprot->readString(_val1576); } xfer += iprot->readMapEnd(); } @@ -22198,17 +22198,17 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_MAP) { { this->part_vals.clear(); - uint32_t _size1576; - ::apache::thrift::protocol::TType _ktype1577; - ::apache::thrift::protocol::TType _vtype1578; - xfer += iprot->readMapBegin(_ktype1577, _vtype1578, _size1576); - uint32_t _i1580; - for (_i1580 = 0; _i1580 < _size1576; ++_i1580) + uint32_t _size1577; + ::apache::thrift::protocol::TType _ktype1578; + ::apache::thrift::protocol::TType _vtype1579; + xfer += iprot->readMapBegin(_ktype1578, _vtype1579, _size1577); + uint32_t _i1581; + for (_i1581 = 0; _i1581 < _size1577; ++_i1581) { - std::string _key1581; - xfer += iprot->readString(_key1581); - std::string& _val1582 = this->part_vals[_key1581]; - xfer += iprot->readString(_val1582); + std::string _key1582; + xfer += iprot->readString(_key1582); + std::string& _val1583 = this->part_vals[_key1582]; + xfer += iprot->readString(_val1583); } xfer += iprot->readMapEnd(); } @@ -22219,9 +22219,9 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::read(::apache::thrift:: break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1583; - xfer += iprot->readI32(ecast1583); - this->eventType = (PartitionEventType::type)ecast1583; + int32_t ecast1584; + xfer += iprot->readI32(ecast1584); + this->eventType = (PartitionEventType::type)ecast1584; this->__isset.eventType = true; } else { xfer += iprot->skip(ftype); @@ -22255,11 +22255,11 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::write(::apache::thrift: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::map ::const_iterator _iter1584; - for (_iter1584 = this->part_vals.begin(); _iter1584 != this->part_vals.end(); ++_iter1584) + std::map ::const_iterator _iter1585; + for (_iter1585 = this->part_vals.begin(); _iter1585 != this->part_vals.end(); ++_iter1585) { - xfer += oprot->writeString(_iter1584->first); - xfer += oprot->writeString(_iter1584->second); + xfer += oprot->writeString(_iter1585->first); + xfer += oprot->writeString(_iter1585->second); } xfer += oprot->writeMapEnd(); } @@ -22295,11 +22295,11 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_pargs::write(::apache::thrift xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::map ::const_iterator _iter1585; - for (_iter1585 = (*(this->part_vals)).begin(); _iter1585 != (*(this->part_vals)).end(); ++_iter1585) + std::map ::const_iterator _iter1586; + for (_iter1586 = (*(this->part_vals)).begin(); _iter1586 != (*(this->part_vals)).end(); ++_iter1586) { - xfer += oprot->writeString(_iter1585->first); - xfer += oprot->writeString(_iter1585->second); + xfer += oprot->writeString(_iter1586->first); + xfer += oprot->writeString(_iter1586->second); } xfer += oprot->writeMapEnd(); } @@ -22568,17 +22568,17 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_MAP) { { this->part_vals.clear(); - uint32_t _size1586; - ::apache::thrift::protocol::TType _ktype1587; - ::apache::thrift::protocol::TType _vtype1588; - xfer += iprot->readMapBegin(_ktype1587, _vtype1588, _size1586); - uint32_t _i1590; - for (_i1590 = 0; _i1590 < _size1586; ++_i1590) + uint32_t _size1587; + ::apache::thrift::protocol::TType _ktype1588; + ::apache::thrift::protocol::TType _vtype1589; + xfer += iprot->readMapBegin(_ktype1588, _vtype1589, _size1587); + uint32_t _i1591; + for (_i1591 = 0; _i1591 < _size1587; ++_i1591) { - std::string _key1591; - xfer += iprot->readString(_key1591); - std::string& _val1592 = this->part_vals[_key1591]; - xfer += iprot->readString(_val1592); + std::string _key1592; + xfer += iprot->readString(_key1592); + std::string& _val1593 = this->part_vals[_key1592]; + xfer += iprot->readString(_val1593); } xfer += iprot->readMapEnd(); } @@ -22589,9 +22589,9 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::read(::apache::thri break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1593; - xfer += iprot->readI32(ecast1593); - this->eventType = (PartitionEventType::type)ecast1593; + int32_t ecast1594; + xfer += iprot->readI32(ecast1594); + this->eventType = (PartitionEventType::type)ecast1594; this->__isset.eventType = true; } else { xfer += iprot->skip(ftype); @@ -22625,11 +22625,11 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::write(::apache::thr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::map ::const_iterator _iter1594; - for (_iter1594 = this->part_vals.begin(); _iter1594 != this->part_vals.end(); ++_iter1594) + std::map ::const_iterator _iter1595; + for (_iter1595 = this->part_vals.begin(); _iter1595 != this->part_vals.end(); ++_iter1595) { - xfer += oprot->writeString(_iter1594->first); - xfer += oprot->writeString(_iter1594->second); + xfer += oprot->writeString(_iter1595->first); + xfer += oprot->writeString(_iter1595->second); } xfer += oprot->writeMapEnd(); } @@ -22665,11 +22665,11 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_pargs::write(::apache::th xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::map ::const_iterator _iter1595; - for (_iter1595 = (*(this->part_vals)).begin(); _iter1595 != (*(this->part_vals)).end(); ++_iter1595) + std::map ::const_iterator _iter1596; + for (_iter1596 = (*(this->part_vals)).begin(); _iter1596 != (*(this->part_vals)).end(); ++_iter1596) { - xfer += oprot->writeString(_iter1595->first); - xfer += oprot->writeString(_iter1595->second); + xfer += oprot->writeString(_iter1596->first); + xfer += oprot->writeString(_iter1596->second); } xfer += oprot->writeMapEnd(); } @@ -24105,14 +24105,14 @@ uint32_t ThriftHiveMetastore_get_indexes_result::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1596; - ::apache::thrift::protocol::TType _etype1599; - xfer += iprot->readListBegin(_etype1599, _size1596); - this->success.resize(_size1596); - uint32_t _i1600; - for (_i1600 = 0; _i1600 < _size1596; ++_i1600) + uint32_t _size1597; + ::apache::thrift::protocol::TType _etype1600; + xfer += iprot->readListBegin(_etype1600, _size1597); + this->success.resize(_size1597); + uint32_t _i1601; + for (_i1601 = 0; _i1601 < _size1597; ++_i1601) { - xfer += this->success[_i1600].read(iprot); + xfer += this->success[_i1601].read(iprot); } xfer += iprot->readListEnd(); } @@ -24159,10 +24159,10 @@ uint32_t ThriftHiveMetastore_get_indexes_result::write(::apache::thrift::protoco xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1601; - for (_iter1601 = this->success.begin(); _iter1601 != this->success.end(); ++_iter1601) + std::vector ::const_iterator _iter1602; + for (_iter1602 = this->success.begin(); _iter1602 != this->success.end(); ++_iter1602) { - xfer += (*_iter1601).write(oprot); + xfer += (*_iter1602).write(oprot); } xfer += oprot->writeListEnd(); } @@ -24211,14 +24211,14 @@ uint32_t ThriftHiveMetastore_get_indexes_presult::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1602; - ::apache::thrift::protocol::TType _etype1605; - xfer += iprot->readListBegin(_etype1605, _size1602); - (*(this->success)).resize(_size1602); - uint32_t _i1606; - for (_i1606 = 0; _i1606 < _size1602; ++_i1606) + uint32_t _size1603; + ::apache::thrift::protocol::TType _etype1606; + xfer += iprot->readListBegin(_etype1606, _size1603); + (*(this->success)).resize(_size1603); + uint32_t _i1607; + for (_i1607 = 0; _i1607 < _size1603; ++_i1607) { - xfer += (*(this->success))[_i1606].read(iprot); + xfer += (*(this->success))[_i1607].read(iprot); } xfer += iprot->readListEnd(); } @@ -24396,14 +24396,14 @@ uint32_t ThriftHiveMetastore_get_index_names_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1607; - ::apache::thrift::protocol::TType _etype1610; - xfer += iprot->readListBegin(_etype1610, _size1607); - this->success.resize(_size1607); - uint32_t _i1611; - for (_i1611 = 0; _i1611 < _size1607; ++_i1611) + uint32_t _size1608; + ::apache::thrift::protocol::TType _etype1611; + xfer += iprot->readListBegin(_etype1611, _size1608); + this->success.resize(_size1608); + uint32_t _i1612; + for (_i1612 = 0; _i1612 < _size1608; ++_i1612) { - xfer += iprot->readString(this->success[_i1611]); + xfer += iprot->readString(this->success[_i1612]); } xfer += iprot->readListEnd(); } @@ -24442,10 +24442,10 @@ uint32_t ThriftHiveMetastore_get_index_names_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1612; - for (_iter1612 = this->success.begin(); _iter1612 != this->success.end(); ++_iter1612) + std::vector ::const_iterator _iter1613; + for (_iter1613 = this->success.begin(); _iter1613 != this->success.end(); ++_iter1613) { - xfer += oprot->writeString((*_iter1612)); + xfer += oprot->writeString((*_iter1613)); } xfer += oprot->writeListEnd(); } @@ -24490,14 +24490,14 @@ uint32_t ThriftHiveMetastore_get_index_names_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1613; - ::apache::thrift::protocol::TType _etype1616; - xfer += iprot->readListBegin(_etype1616, _size1613); - (*(this->success)).resize(_size1613); - uint32_t _i1617; - for (_i1617 = 0; _i1617 < _size1613; ++_i1617) + uint32_t _size1614; + ::apache::thrift::protocol::TType _etype1617; + xfer += iprot->readListBegin(_etype1617, _size1614); + (*(this->success)).resize(_size1614); + uint32_t _i1618; + for (_i1618 = 0; _i1618 < _size1614; ++_i1618) { - xfer += iprot->readString((*(this->success))[_i1617]); + xfer += iprot->readString((*(this->success))[_i1618]); } xfer += iprot->readListEnd(); } @@ -28978,14 +28978,14 @@ uint32_t ThriftHiveMetastore_get_functions_result::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1618; - ::apache::thrift::protocol::TType _etype1621; - xfer += iprot->readListBegin(_etype1621, _size1618); - this->success.resize(_size1618); - uint32_t _i1622; - for (_i1622 = 0; _i1622 < _size1618; ++_i1622) + uint32_t _size1619; + ::apache::thrift::protocol::TType _etype1622; + xfer += iprot->readListBegin(_etype1622, _size1619); + this->success.resize(_size1619); + uint32_t _i1623; + for (_i1623 = 0; _i1623 < _size1619; ++_i1623) { - xfer += iprot->readString(this->success[_i1622]); + xfer += iprot->readString(this->success[_i1623]); } xfer += iprot->readListEnd(); } @@ -29024,10 +29024,10 @@ uint32_t ThriftHiveMetastore_get_functions_result::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1623; - for (_iter1623 = this->success.begin(); _iter1623 != this->success.end(); ++_iter1623) + std::vector ::const_iterator _iter1624; + for (_iter1624 = this->success.begin(); _iter1624 != this->success.end(); ++_iter1624) { - xfer += oprot->writeString((*_iter1623)); + xfer += oprot->writeString((*_iter1624)); } xfer += oprot->writeListEnd(); } @@ -29072,14 +29072,14 @@ uint32_t ThriftHiveMetastore_get_functions_presult::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1624; - ::apache::thrift::protocol::TType _etype1627; - xfer += iprot->readListBegin(_etype1627, _size1624); - (*(this->success)).resize(_size1624); - uint32_t _i1628; - for (_i1628 = 0; _i1628 < _size1624; ++_i1628) + uint32_t _size1625; + ::apache::thrift::protocol::TType _etype1628; + xfer += iprot->readListBegin(_etype1628, _size1625); + (*(this->success)).resize(_size1625); + uint32_t _i1629; + for (_i1629 = 0; _i1629 < _size1625; ++_i1629) { - xfer += iprot->readString((*(this->success))[_i1628]); + xfer += iprot->readString((*(this->success))[_i1629]); } xfer += iprot->readListEnd(); } @@ -30039,14 +30039,14 @@ uint32_t ThriftHiveMetastore_get_role_names_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1629; - ::apache::thrift::protocol::TType _etype1632; - xfer += iprot->readListBegin(_etype1632, _size1629); - this->success.resize(_size1629); - uint32_t _i1633; - for (_i1633 = 0; _i1633 < _size1629; ++_i1633) + uint32_t _size1630; + ::apache::thrift::protocol::TType _etype1633; + xfer += iprot->readListBegin(_etype1633, _size1630); + this->success.resize(_size1630); + uint32_t _i1634; + for (_i1634 = 0; _i1634 < _size1630; ++_i1634) { - xfer += iprot->readString(this->success[_i1633]); + xfer += iprot->readString(this->success[_i1634]); } xfer += iprot->readListEnd(); } @@ -30085,10 +30085,10 @@ uint32_t ThriftHiveMetastore_get_role_names_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1634; - for (_iter1634 = this->success.begin(); _iter1634 != this->success.end(); ++_iter1634) + std::vector ::const_iterator _iter1635; + for (_iter1635 = this->success.begin(); _iter1635 != this->success.end(); ++_iter1635) { - xfer += oprot->writeString((*_iter1634)); + xfer += oprot->writeString((*_iter1635)); } xfer += oprot->writeListEnd(); } @@ -30133,14 +30133,14 @@ uint32_t ThriftHiveMetastore_get_role_names_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1635; - ::apache::thrift::protocol::TType _etype1638; - xfer += iprot->readListBegin(_etype1638, _size1635); - (*(this->success)).resize(_size1635); - uint32_t _i1639; - for (_i1639 = 0; _i1639 < _size1635; ++_i1639) + uint32_t _size1636; + ::apache::thrift::protocol::TType _etype1639; + xfer += iprot->readListBegin(_etype1639, _size1636); + (*(this->success)).resize(_size1636); + uint32_t _i1640; + for (_i1640 = 0; _i1640 < _size1636; ++_i1640) { - xfer += iprot->readString((*(this->success))[_i1639]); + xfer += iprot->readString((*(this->success))[_i1640]); } xfer += iprot->readListEnd(); } @@ -30213,9 +30213,9 @@ uint32_t ThriftHiveMetastore_grant_role_args::read(::apache::thrift::protocol::T break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1640; - xfer += iprot->readI32(ecast1640); - this->principal_type = (PrincipalType::type)ecast1640; + int32_t ecast1641; + xfer += iprot->readI32(ecast1641); + this->principal_type = (PrincipalType::type)ecast1641; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -30231,9 +30231,9 @@ uint32_t ThriftHiveMetastore_grant_role_args::read(::apache::thrift::protocol::T break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1641; - xfer += iprot->readI32(ecast1641); - this->grantorType = (PrincipalType::type)ecast1641; + int32_t ecast1642; + xfer += iprot->readI32(ecast1642); + this->grantorType = (PrincipalType::type)ecast1642; this->__isset.grantorType = true; } else { xfer += iprot->skip(ftype); @@ -30504,9 +30504,9 @@ uint32_t ThriftHiveMetastore_revoke_role_args::read(::apache::thrift::protocol:: break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1642; - xfer += iprot->readI32(ecast1642); - this->principal_type = (PrincipalType::type)ecast1642; + int32_t ecast1643; + xfer += iprot->readI32(ecast1643); + this->principal_type = (PrincipalType::type)ecast1643; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -30737,9 +30737,9 @@ uint32_t ThriftHiveMetastore_list_roles_args::read(::apache::thrift::protocol::T break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1643; - xfer += iprot->readI32(ecast1643); - this->principal_type = (PrincipalType::type)ecast1643; + int32_t ecast1644; + xfer += iprot->readI32(ecast1644); + this->principal_type = (PrincipalType::type)ecast1644; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -30828,14 +30828,14 @@ uint32_t ThriftHiveMetastore_list_roles_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1644; - ::apache::thrift::protocol::TType _etype1647; - xfer += iprot->readListBegin(_etype1647, _size1644); - this->success.resize(_size1644); - uint32_t _i1648; - for (_i1648 = 0; _i1648 < _size1644; ++_i1648) + uint32_t _size1645; + ::apache::thrift::protocol::TType _etype1648; + xfer += iprot->readListBegin(_etype1648, _size1645); + this->success.resize(_size1645); + uint32_t _i1649; + for (_i1649 = 0; _i1649 < _size1645; ++_i1649) { - xfer += this->success[_i1648].read(iprot); + xfer += this->success[_i1649].read(iprot); } xfer += iprot->readListEnd(); } @@ -30874,10 +30874,10 @@ uint32_t ThriftHiveMetastore_list_roles_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1649; - for (_iter1649 = this->success.begin(); _iter1649 != this->success.end(); ++_iter1649) + std::vector ::const_iterator _iter1650; + for (_iter1650 = this->success.begin(); _iter1650 != this->success.end(); ++_iter1650) { - xfer += (*_iter1649).write(oprot); + xfer += (*_iter1650).write(oprot); } xfer += oprot->writeListEnd(); } @@ -30922,14 +30922,14 @@ uint32_t ThriftHiveMetastore_list_roles_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1650; - ::apache::thrift::protocol::TType _etype1653; - xfer += iprot->readListBegin(_etype1653, _size1650); - (*(this->success)).resize(_size1650); - uint32_t _i1654; - for (_i1654 = 0; _i1654 < _size1650; ++_i1654) + uint32_t _size1651; + ::apache::thrift::protocol::TType _etype1654; + xfer += iprot->readListBegin(_etype1654, _size1651); + (*(this->success)).resize(_size1651); + uint32_t _i1655; + for (_i1655 = 0; _i1655 < _size1651; ++_i1655) { - xfer += (*(this->success))[_i1654].read(iprot); + xfer += (*(this->success))[_i1655].read(iprot); } xfer += iprot->readListEnd(); } @@ -31625,14 +31625,14 @@ uint32_t ThriftHiveMetastore_get_privilege_set_args::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1655; - ::apache::thrift::protocol::TType _etype1658; - xfer += iprot->readListBegin(_etype1658, _size1655); - this->group_names.resize(_size1655); - uint32_t _i1659; - for (_i1659 = 0; _i1659 < _size1655; ++_i1659) + uint32_t _size1656; + ::apache::thrift::protocol::TType _etype1659; + xfer += iprot->readListBegin(_etype1659, _size1656); + this->group_names.resize(_size1656); + uint32_t _i1660; + for (_i1660 = 0; _i1660 < _size1656; ++_i1660) { - xfer += iprot->readString(this->group_names[_i1659]); + xfer += iprot->readString(this->group_names[_i1660]); } xfer += iprot->readListEnd(); } @@ -31669,10 +31669,10 @@ uint32_t ThriftHiveMetastore_get_privilege_set_args::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1660; - for (_iter1660 = this->group_names.begin(); _iter1660 != this->group_names.end(); ++_iter1660) + std::vector ::const_iterator _iter1661; + for (_iter1661 = this->group_names.begin(); _iter1661 != this->group_names.end(); ++_iter1661) { - xfer += oprot->writeString((*_iter1660)); + xfer += oprot->writeString((*_iter1661)); } xfer += oprot->writeListEnd(); } @@ -31704,10 +31704,10 @@ uint32_t ThriftHiveMetastore_get_privilege_set_pargs::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1661; - for (_iter1661 = (*(this->group_names)).begin(); _iter1661 != (*(this->group_names)).end(); ++_iter1661) + std::vector ::const_iterator _iter1662; + for (_iter1662 = (*(this->group_names)).begin(); _iter1662 != (*(this->group_names)).end(); ++_iter1662) { - xfer += oprot->writeString((*_iter1661)); + xfer += oprot->writeString((*_iter1662)); } xfer += oprot->writeListEnd(); } @@ -31882,9 +31882,9 @@ uint32_t ThriftHiveMetastore_list_privileges_args::read(::apache::thrift::protoc break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1662; - xfer += iprot->readI32(ecast1662); - this->principal_type = (PrincipalType::type)ecast1662; + int32_t ecast1663; + xfer += iprot->readI32(ecast1663); + this->principal_type = (PrincipalType::type)ecast1663; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -31989,14 +31989,14 @@ uint32_t ThriftHiveMetastore_list_privileges_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1663; - ::apache::thrift::protocol::TType _etype1666; - xfer += iprot->readListBegin(_etype1666, _size1663); - this->success.resize(_size1663); - uint32_t _i1667; - for (_i1667 = 0; _i1667 < _size1663; ++_i1667) + uint32_t _size1664; + ::apache::thrift::protocol::TType _etype1667; + xfer += iprot->readListBegin(_etype1667, _size1664); + this->success.resize(_size1664); + uint32_t _i1668; + for (_i1668 = 0; _i1668 < _size1664; ++_i1668) { - xfer += this->success[_i1667].read(iprot); + xfer += this->success[_i1668].read(iprot); } xfer += iprot->readListEnd(); } @@ -32035,10 +32035,10 @@ uint32_t ThriftHiveMetastore_list_privileges_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1668; - for (_iter1668 = this->success.begin(); _iter1668 != this->success.end(); ++_iter1668) + std::vector ::const_iterator _iter1669; + for (_iter1669 = this->success.begin(); _iter1669 != this->success.end(); ++_iter1669) { - xfer += (*_iter1668).write(oprot); + xfer += (*_iter1669).write(oprot); } xfer += oprot->writeListEnd(); } @@ -32083,14 +32083,14 @@ uint32_t ThriftHiveMetastore_list_privileges_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1669; - ::apache::thrift::protocol::TType _etype1672; - xfer += iprot->readListBegin(_etype1672, _size1669); - (*(this->success)).resize(_size1669); - uint32_t _i1673; - for (_i1673 = 0; _i1673 < _size1669; ++_i1673) + uint32_t _size1670; + ::apache::thrift::protocol::TType _etype1673; + xfer += iprot->readListBegin(_etype1673, _size1670); + (*(this->success)).resize(_size1670); + uint32_t _i1674; + for (_i1674 = 0; _i1674 < _size1670; ++_i1674) { - xfer += (*(this->success))[_i1673].read(iprot); + xfer += (*(this->success))[_i1674].read(iprot); } xfer += iprot->readListEnd(); } @@ -32778,14 +32778,14 @@ uint32_t ThriftHiveMetastore_set_ugi_args::read(::apache::thrift::protocol::TPro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1674; - ::apache::thrift::protocol::TType _etype1677; - xfer += iprot->readListBegin(_etype1677, _size1674); - this->group_names.resize(_size1674); - uint32_t _i1678; - for (_i1678 = 0; _i1678 < _size1674; ++_i1678) + uint32_t _size1675; + ::apache::thrift::protocol::TType _etype1678; + xfer += iprot->readListBegin(_etype1678, _size1675); + this->group_names.resize(_size1675); + uint32_t _i1679; + for (_i1679 = 0; _i1679 < _size1675; ++_i1679) { - xfer += iprot->readString(this->group_names[_i1678]); + xfer += iprot->readString(this->group_names[_i1679]); } xfer += iprot->readListEnd(); } @@ -32818,10 +32818,10 @@ uint32_t ThriftHiveMetastore_set_ugi_args::write(::apache::thrift::protocol::TPr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1679; - for (_iter1679 = this->group_names.begin(); _iter1679 != this->group_names.end(); ++_iter1679) + std::vector ::const_iterator _iter1680; + for (_iter1680 = this->group_names.begin(); _iter1680 != this->group_names.end(); ++_iter1680) { - xfer += oprot->writeString((*_iter1679)); + xfer += oprot->writeString((*_iter1680)); } xfer += oprot->writeListEnd(); } @@ -32849,10 +32849,10 @@ uint32_t ThriftHiveMetastore_set_ugi_pargs::write(::apache::thrift::protocol::TP xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1680; - for (_iter1680 = (*(this->group_names)).begin(); _iter1680 != (*(this->group_names)).end(); ++_iter1680) + std::vector ::const_iterator _iter1681; + for (_iter1681 = (*(this->group_names)).begin(); _iter1681 != (*(this->group_names)).end(); ++_iter1681) { - xfer += oprot->writeString((*_iter1680)); + xfer += oprot->writeString((*_iter1681)); } xfer += oprot->writeListEnd(); } @@ -32893,14 +32893,14 @@ uint32_t ThriftHiveMetastore_set_ugi_result::read(::apache::thrift::protocol::TP if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1681; - ::apache::thrift::protocol::TType _etype1684; - xfer += iprot->readListBegin(_etype1684, _size1681); - this->success.resize(_size1681); - uint32_t _i1685; - for (_i1685 = 0; _i1685 < _size1681; ++_i1685) + uint32_t _size1682; + ::apache::thrift::protocol::TType _etype1685; + xfer += iprot->readListBegin(_etype1685, _size1682); + this->success.resize(_size1682); + uint32_t _i1686; + for (_i1686 = 0; _i1686 < _size1682; ++_i1686) { - xfer += iprot->readString(this->success[_i1685]); + xfer += iprot->readString(this->success[_i1686]); } xfer += iprot->readListEnd(); } @@ -32939,10 +32939,10 @@ uint32_t ThriftHiveMetastore_set_ugi_result::write(::apache::thrift::protocol::T xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1686; - for (_iter1686 = this->success.begin(); _iter1686 != this->success.end(); ++_iter1686) + std::vector ::const_iterator _iter1687; + for (_iter1687 = this->success.begin(); _iter1687 != this->success.end(); ++_iter1687) { - xfer += oprot->writeString((*_iter1686)); + xfer += oprot->writeString((*_iter1687)); } xfer += oprot->writeListEnd(); } @@ -32987,14 +32987,14 @@ uint32_t ThriftHiveMetastore_set_ugi_presult::read(::apache::thrift::protocol::T if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1687; - ::apache::thrift::protocol::TType _etype1690; - xfer += iprot->readListBegin(_etype1690, _size1687); - (*(this->success)).resize(_size1687); - uint32_t _i1691; - for (_i1691 = 0; _i1691 < _size1687; ++_i1691) + uint32_t _size1688; + ::apache::thrift::protocol::TType _etype1691; + xfer += iprot->readListBegin(_etype1691, _size1688); + (*(this->success)).resize(_size1688); + uint32_t _i1692; + for (_i1692 = 0; _i1692 < _size1688; ++_i1692) { - xfer += iprot->readString((*(this->success))[_i1691]); + xfer += iprot->readString((*(this->success))[_i1692]); } xfer += iprot->readListEnd(); } @@ -34305,14 +34305,14 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1692; - ::apache::thrift::protocol::TType _etype1695; - xfer += iprot->readListBegin(_etype1695, _size1692); - this->success.resize(_size1692); - uint32_t _i1696; - for (_i1696 = 0; _i1696 < _size1692; ++_i1696) + uint32_t _size1693; + ::apache::thrift::protocol::TType _etype1696; + xfer += iprot->readListBegin(_etype1696, _size1693); + this->success.resize(_size1693); + uint32_t _i1697; + for (_i1697 = 0; _i1697 < _size1693; ++_i1697) { - xfer += iprot->readString(this->success[_i1696]); + xfer += iprot->readString(this->success[_i1697]); } xfer += iprot->readListEnd(); } @@ -34343,10 +34343,10 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1697; - for (_iter1697 = this->success.begin(); _iter1697 != this->success.end(); ++_iter1697) + std::vector ::const_iterator _iter1698; + for (_iter1698 = this->success.begin(); _iter1698 != this->success.end(); ++_iter1698) { - xfer += oprot->writeString((*_iter1697)); + xfer += oprot->writeString((*_iter1698)); } xfer += oprot->writeListEnd(); } @@ -34387,14 +34387,14 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1698; - ::apache::thrift::protocol::TType _etype1701; - xfer += iprot->readListBegin(_etype1701, _size1698); - (*(this->success)).resize(_size1698); - uint32_t _i1702; - for (_i1702 = 0; _i1702 < _size1698; ++_i1702) + uint32_t _size1699; + ::apache::thrift::protocol::TType _etype1702; + xfer += iprot->readListBegin(_etype1702, _size1699); + (*(this->success)).resize(_size1699); + uint32_t _i1703; + for (_i1703 = 0; _i1703 < _size1699; ++_i1703) { - xfer += iprot->readString((*(this->success))[_i1702]); + xfer += iprot->readString((*(this->success))[_i1703]); } xfer += iprot->readListEnd(); } @@ -35120,14 +35120,14 @@ uint32_t ThriftHiveMetastore_get_master_keys_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1703; - ::apache::thrift::protocol::TType _etype1706; - xfer += iprot->readListBegin(_etype1706, _size1703); - this->success.resize(_size1703); - uint32_t _i1707; - for (_i1707 = 0; _i1707 < _size1703; ++_i1707) + uint32_t _size1704; + ::apache::thrift::protocol::TType _etype1707; + xfer += iprot->readListBegin(_etype1707, _size1704); + this->success.resize(_size1704); + uint32_t _i1708; + for (_i1708 = 0; _i1708 < _size1704; ++_i1708) { - xfer += iprot->readString(this->success[_i1707]); + xfer += iprot->readString(this->success[_i1708]); } xfer += iprot->readListEnd(); } @@ -35158,10 +35158,10 @@ uint32_t ThriftHiveMetastore_get_master_keys_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1708; - for (_iter1708 = this->success.begin(); _iter1708 != this->success.end(); ++_iter1708) + std::vector ::const_iterator _iter1709; + for (_iter1709 = this->success.begin(); _iter1709 != this->success.end(); ++_iter1709) { - xfer += oprot->writeString((*_iter1708)); + xfer += oprot->writeString((*_iter1709)); } xfer += oprot->writeListEnd(); } @@ -35202,14 +35202,14 @@ uint32_t ThriftHiveMetastore_get_master_keys_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1709; - ::apache::thrift::protocol::TType _etype1712; - xfer += iprot->readListBegin(_etype1712, _size1709); - (*(this->success)).resize(_size1709); - uint32_t _i1713; - for (_i1713 = 0; _i1713 < _size1709; ++_i1713) + uint32_t _size1710; + ::apache::thrift::protocol::TType _etype1713; + xfer += iprot->readListBegin(_etype1713, _size1710); + (*(this->success)).resize(_size1710); + uint32_t _i1714; + for (_i1714 = 0; _i1714 < _size1710; ++_i1714) { - xfer += iprot->readString((*(this->success))[_i1713]); + xfer += iprot->readString((*(this->success))[_i1714]); } xfer += iprot->readListEnd(); } diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp index aadf8f17c4..216926523c 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp +++ b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp @@ -149,6 +149,18 @@ const char* _kEventRequestTypeNames[] = { }; const std::map _EventRequestType_VALUES_TO_NAMES(::apache::thrift::TEnumIterator(3, _kEventRequestTypeValues, _kEventRequestTypeNames), ::apache::thrift::TEnumIterator(-1, NULL, NULL)); +int _kBucketingVersionValues[] = { + BucketingVersion::INVALID_BUCKETING, + BucketingVersion::JAVA_BUCKETING, + BucketingVersion::MURMUR_BUCKETING +}; +const char* _kBucketingVersionNames[] = { + "INVALID_BUCKETING", + "JAVA_BUCKETING", + "MURMUR_BUCKETING" +}; +const std::map _BucketingVersion_VALUES_TO_NAMES(::apache::thrift::TEnumIterator(3, _kBucketingVersionValues, _kBucketingVersionNames), ::apache::thrift::TEnumIterator(-1, NULL, NULL)); + int _kFunctionTypeValues[] = { FunctionType::JAVA }; @@ -4945,6 +4957,16 @@ void Table::__set_creationMetadata(const CreationMetadata& val) { __isset.creationMetadata = true; } +void Table::__set_bucketingVersion(const BucketingVersion::type val) { + this->bucketingVersion = val; +__isset.bucketingVersion = true; +} + +void Table::__set_loadInBucketedTable(const bool val) { + this->loadInBucketedTable = val; +__isset.loadInBucketedTable = true; +} + uint32_t Table::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); @@ -5121,6 +5143,24 @@ uint32_t Table::read(::apache::thrift::protocol::TProtocol* iprot) { xfer += iprot->skip(ftype); } break; + case 17: + if (ftype == ::apache::thrift::protocol::T_I32) { + int32_t ecast223; + xfer += iprot->readI32(ecast223); + this->bucketingVersion = (BucketingVersion::type)ecast223; + this->__isset.bucketingVersion = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 18: + if (ftype == ::apache::thrift::protocol::T_BOOL) { + xfer += iprot->readBool(this->loadInBucketedTable); + this->__isset.loadInBucketedTable = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -5169,10 +5209,10 @@ uint32_t Table::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("partitionKeys", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitionKeys.size())); - std::vector ::const_iterator _iter223; - for (_iter223 = this->partitionKeys.begin(); _iter223 != this->partitionKeys.end(); ++_iter223) + std::vector ::const_iterator _iter224; + for (_iter224 = this->partitionKeys.begin(); _iter224 != this->partitionKeys.end(); ++_iter224) { - xfer += (*_iter223).write(oprot); + xfer += (*_iter224).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5181,11 +5221,11 @@ uint32_t Table::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("parameters", ::apache::thrift::protocol::T_MAP, 9); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->parameters.size())); - std::map ::const_iterator _iter224; - for (_iter224 = this->parameters.begin(); _iter224 != this->parameters.end(); ++_iter224) + std::map ::const_iterator _iter225; + for (_iter225 = this->parameters.begin(); _iter225 != this->parameters.end(); ++_iter225) { - xfer += oprot->writeString(_iter224->first); - xfer += oprot->writeString(_iter224->second); + xfer += oprot->writeString(_iter225->first); + xfer += oprot->writeString(_iter225->second); } xfer += oprot->writeMapEnd(); } @@ -5223,6 +5263,16 @@ uint32_t Table::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += this->creationMetadata.write(oprot); xfer += oprot->writeFieldEnd(); } + if (this->__isset.bucketingVersion) { + xfer += oprot->writeFieldBegin("bucketingVersion", ::apache::thrift::protocol::T_I32, 17); + xfer += oprot->writeI32((int32_t)this->bucketingVersion); + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.loadInBucketedTable) { + xfer += oprot->writeFieldBegin("loadInBucketedTable", ::apache::thrift::protocol::T_BOOL, 18); + xfer += oprot->writeBool(this->loadInBucketedTable); + xfer += oprot->writeFieldEnd(); + } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; @@ -5246,29 +5296,12 @@ void swap(Table &a, Table &b) { swap(a.temporary, b.temporary); swap(a.rewriteEnabled, b.rewriteEnabled); swap(a.creationMetadata, b.creationMetadata); + swap(a.bucketingVersion, b.bucketingVersion); + swap(a.loadInBucketedTable, b.loadInBucketedTable); swap(a.__isset, b.__isset); } -Table::Table(const Table& other225) { - tableName = other225.tableName; - dbName = other225.dbName; - owner = other225.owner; - createTime = other225.createTime; - lastAccessTime = other225.lastAccessTime; - retention = other225.retention; - sd = other225.sd; - partitionKeys = other225.partitionKeys; - parameters = other225.parameters; - viewOriginalText = other225.viewOriginalText; - viewExpandedText = other225.viewExpandedText; - tableType = other225.tableType; - privileges = other225.privileges; - temporary = other225.temporary; - rewriteEnabled = other225.rewriteEnabled; - creationMetadata = other225.creationMetadata; - __isset = other225.__isset; -} -Table& Table::operator=(const Table& other226) { +Table::Table(const Table& other226) { tableName = other226.tableName; dbName = other226.dbName; owner = other226.owner; @@ -5285,7 +5318,30 @@ Table& Table::operator=(const Table& other226) { temporary = other226.temporary; rewriteEnabled = other226.rewriteEnabled; creationMetadata = other226.creationMetadata; + bucketingVersion = other226.bucketingVersion; + loadInBucketedTable = other226.loadInBucketedTable; __isset = other226.__isset; +} +Table& Table::operator=(const Table& other227) { + tableName = other227.tableName; + dbName = other227.dbName; + owner = other227.owner; + createTime = other227.createTime; + lastAccessTime = other227.lastAccessTime; + retention = other227.retention; + sd = other227.sd; + partitionKeys = other227.partitionKeys; + parameters = other227.parameters; + viewOriginalText = other227.viewOriginalText; + viewExpandedText = other227.viewExpandedText; + tableType = other227.tableType; + privileges = other227.privileges; + temporary = other227.temporary; + rewriteEnabled = other227.rewriteEnabled; + creationMetadata = other227.creationMetadata; + bucketingVersion = other227.bucketingVersion; + loadInBucketedTable = other227.loadInBucketedTable; + __isset = other227.__isset; return *this; } void Table::printTo(std::ostream& out) const { @@ -5307,6 +5363,8 @@ void Table::printTo(std::ostream& out) const { out << ", " << "temporary="; (__isset.temporary ? (out << to_string(temporary)) : (out << "")); out << ", " << "rewriteEnabled="; (__isset.rewriteEnabled ? (out << to_string(rewriteEnabled)) : (out << "")); out << ", " << "creationMetadata="; (__isset.creationMetadata ? (out << to_string(creationMetadata)) : (out << "")); + out << ", " << "bucketingVersion="; (__isset.bucketingVersion ? (out << to_string(bucketingVersion)) : (out << "")); + out << ", " << "loadInBucketedTable="; (__isset.loadInBucketedTable ? (out << to_string(loadInBucketedTable)) : (out << "")); out << ")"; } @@ -5373,14 +5431,14 @@ uint32_t Partition::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->values.clear(); - uint32_t _size227; - ::apache::thrift::protocol::TType _etype230; - xfer += iprot->readListBegin(_etype230, _size227); - this->values.resize(_size227); - uint32_t _i231; - for (_i231 = 0; _i231 < _size227; ++_i231) + uint32_t _size228; + ::apache::thrift::protocol::TType _etype231; + xfer += iprot->readListBegin(_etype231, _size228); + this->values.resize(_size228); + uint32_t _i232; + for (_i232 = 0; _i232 < _size228; ++_i232) { - xfer += iprot->readString(this->values[_i231]); + xfer += iprot->readString(this->values[_i232]); } xfer += iprot->readListEnd(); } @@ -5433,17 +5491,17 @@ uint32_t Partition::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->parameters.clear(); - uint32_t _size232; - ::apache::thrift::protocol::TType _ktype233; - ::apache::thrift::protocol::TType _vtype234; - xfer += iprot->readMapBegin(_ktype233, _vtype234, _size232); - uint32_t _i236; - for (_i236 = 0; _i236 < _size232; ++_i236) + uint32_t _size233; + ::apache::thrift::protocol::TType _ktype234; + ::apache::thrift::protocol::TType _vtype235; + xfer += iprot->readMapBegin(_ktype234, _vtype235, _size233); + uint32_t _i237; + for (_i237 = 0; _i237 < _size233; ++_i237) { - std::string _key237; - xfer += iprot->readString(_key237); - std::string& _val238 = this->parameters[_key237]; - xfer += iprot->readString(_val238); + std::string _key238; + xfer += iprot->readString(_key238); + std::string& _val239 = this->parameters[_key238]; + xfer += iprot->readString(_val239); } xfer += iprot->readMapEnd(); } @@ -5480,10 +5538,10 @@ uint32_t Partition::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("values", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->values.size())); - std::vector ::const_iterator _iter239; - for (_iter239 = this->values.begin(); _iter239 != this->values.end(); ++_iter239) + std::vector ::const_iterator _iter240; + for (_iter240 = this->values.begin(); _iter240 != this->values.end(); ++_iter240) { - xfer += oprot->writeString((*_iter239)); + xfer += oprot->writeString((*_iter240)); } xfer += oprot->writeListEnd(); } @@ -5512,11 +5570,11 @@ uint32_t Partition::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("parameters", ::apache::thrift::protocol::T_MAP, 7); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->parameters.size())); - std::map ::const_iterator _iter240; - for (_iter240 = this->parameters.begin(); _iter240 != this->parameters.end(); ++_iter240) + std::map ::const_iterator _iter241; + for (_iter241 = this->parameters.begin(); _iter241 != this->parameters.end(); ++_iter241) { - xfer += oprot->writeString(_iter240->first); - xfer += oprot->writeString(_iter240->second); + xfer += oprot->writeString(_iter241->first); + xfer += oprot->writeString(_iter241->second); } xfer += oprot->writeMapEnd(); } @@ -5545,18 +5603,7 @@ void swap(Partition &a, Partition &b) { swap(a.__isset, b.__isset); } -Partition::Partition(const Partition& other241) { - values = other241.values; - dbName = other241.dbName; - tableName = other241.tableName; - createTime = other241.createTime; - lastAccessTime = other241.lastAccessTime; - sd = other241.sd; - parameters = other241.parameters; - privileges = other241.privileges; - __isset = other241.__isset; -} -Partition& Partition::operator=(const Partition& other242) { +Partition::Partition(const Partition& other242) { values = other242.values; dbName = other242.dbName; tableName = other242.tableName; @@ -5566,6 +5613,17 @@ Partition& Partition::operator=(const Partition& other242) { parameters = other242.parameters; privileges = other242.privileges; __isset = other242.__isset; +} +Partition& Partition::operator=(const Partition& other243) { + values = other243.values; + dbName = other243.dbName; + tableName = other243.tableName; + createTime = other243.createTime; + lastAccessTime = other243.lastAccessTime; + sd = other243.sd; + parameters = other243.parameters; + privileges = other243.privileges; + __isset = other243.__isset; return *this; } void Partition::printTo(std::ostream& out) const { @@ -5637,14 +5695,14 @@ uint32_t PartitionWithoutSD::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->values.clear(); - uint32_t _size243; - ::apache::thrift::protocol::TType _etype246; - xfer += iprot->readListBegin(_etype246, _size243); - this->values.resize(_size243); - uint32_t _i247; - for (_i247 = 0; _i247 < _size243; ++_i247) + uint32_t _size244; + ::apache::thrift::protocol::TType _etype247; + xfer += iprot->readListBegin(_etype247, _size244); + this->values.resize(_size244); + uint32_t _i248; + for (_i248 = 0; _i248 < _size244; ++_i248) { - xfer += iprot->readString(this->values[_i247]); + xfer += iprot->readString(this->values[_i248]); } xfer += iprot->readListEnd(); } @@ -5681,17 +5739,17 @@ uint32_t PartitionWithoutSD::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_MAP) { { this->parameters.clear(); - uint32_t _size248; - ::apache::thrift::protocol::TType _ktype249; - ::apache::thrift::protocol::TType _vtype250; - xfer += iprot->readMapBegin(_ktype249, _vtype250, _size248); - uint32_t _i252; - for (_i252 = 0; _i252 < _size248; ++_i252) + uint32_t _size249; + ::apache::thrift::protocol::TType _ktype250; + ::apache::thrift::protocol::TType _vtype251; + xfer += iprot->readMapBegin(_ktype250, _vtype251, _size249); + uint32_t _i253; + for (_i253 = 0; _i253 < _size249; ++_i253) { - std::string _key253; - xfer += iprot->readString(_key253); - std::string& _val254 = this->parameters[_key253]; - xfer += iprot->readString(_val254); + std::string _key254; + xfer += iprot->readString(_key254); + std::string& _val255 = this->parameters[_key254]; + xfer += iprot->readString(_val255); } xfer += iprot->readMapEnd(); } @@ -5728,10 +5786,10 @@ uint32_t PartitionWithoutSD::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("values", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->values.size())); - std::vector ::const_iterator _iter255; - for (_iter255 = this->values.begin(); _iter255 != this->values.end(); ++_iter255) + std::vector ::const_iterator _iter256; + for (_iter256 = this->values.begin(); _iter256 != this->values.end(); ++_iter256) { - xfer += oprot->writeString((*_iter255)); + xfer += oprot->writeString((*_iter256)); } xfer += oprot->writeListEnd(); } @@ -5752,11 +5810,11 @@ uint32_t PartitionWithoutSD::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("parameters", ::apache::thrift::protocol::T_MAP, 5); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->parameters.size())); - std::map ::const_iterator _iter256; - for (_iter256 = this->parameters.begin(); _iter256 != this->parameters.end(); ++_iter256) + std::map ::const_iterator _iter257; + for (_iter257 = this->parameters.begin(); _iter257 != this->parameters.end(); ++_iter257) { - xfer += oprot->writeString(_iter256->first); - xfer += oprot->writeString(_iter256->second); + xfer += oprot->writeString(_iter257->first); + xfer += oprot->writeString(_iter257->second); } xfer += oprot->writeMapEnd(); } @@ -5783,16 +5841,7 @@ void swap(PartitionWithoutSD &a, PartitionWithoutSD &b) { swap(a.__isset, b.__isset); } -PartitionWithoutSD::PartitionWithoutSD(const PartitionWithoutSD& other257) { - values = other257.values; - createTime = other257.createTime; - lastAccessTime = other257.lastAccessTime; - relativePath = other257.relativePath; - parameters = other257.parameters; - privileges = other257.privileges; - __isset = other257.__isset; -} -PartitionWithoutSD& PartitionWithoutSD::operator=(const PartitionWithoutSD& other258) { +PartitionWithoutSD::PartitionWithoutSD(const PartitionWithoutSD& other258) { values = other258.values; createTime = other258.createTime; lastAccessTime = other258.lastAccessTime; @@ -5800,6 +5849,15 @@ PartitionWithoutSD& PartitionWithoutSD::operator=(const PartitionWithoutSD& othe parameters = other258.parameters; privileges = other258.privileges; __isset = other258.__isset; +} +PartitionWithoutSD& PartitionWithoutSD::operator=(const PartitionWithoutSD& other259) { + values = other259.values; + createTime = other259.createTime; + lastAccessTime = other259.lastAccessTime; + relativePath = other259.relativePath; + parameters = other259.parameters; + privileges = other259.privileges; + __isset = other259.__isset; return *this; } void PartitionWithoutSD::printTo(std::ostream& out) const { @@ -5852,14 +5910,14 @@ uint32_t PartitionSpecWithSharedSD::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size259; - ::apache::thrift::protocol::TType _etype262; - xfer += iprot->readListBegin(_etype262, _size259); - this->partitions.resize(_size259); - uint32_t _i263; - for (_i263 = 0; _i263 < _size259; ++_i263) + uint32_t _size260; + ::apache::thrift::protocol::TType _etype263; + xfer += iprot->readListBegin(_etype263, _size260); + this->partitions.resize(_size260); + uint32_t _i264; + for (_i264 = 0; _i264 < _size260; ++_i264) { - xfer += this->partitions[_i263].read(iprot); + xfer += this->partitions[_i264].read(iprot); } xfer += iprot->readListEnd(); } @@ -5896,10 +5954,10 @@ uint32_t PartitionSpecWithSharedSD::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter264; - for (_iter264 = this->partitions.begin(); _iter264 != this->partitions.end(); ++_iter264) + std::vector ::const_iterator _iter265; + for (_iter265 = this->partitions.begin(); _iter265 != this->partitions.end(); ++_iter265) { - xfer += (*_iter264).write(oprot); + xfer += (*_iter265).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5921,15 +5979,15 @@ void swap(PartitionSpecWithSharedSD &a, PartitionSpecWithSharedSD &b) { swap(a.__isset, b.__isset); } -PartitionSpecWithSharedSD::PartitionSpecWithSharedSD(const PartitionSpecWithSharedSD& other265) { - partitions = other265.partitions; - sd = other265.sd; - __isset = other265.__isset; -} -PartitionSpecWithSharedSD& PartitionSpecWithSharedSD::operator=(const PartitionSpecWithSharedSD& other266) { +PartitionSpecWithSharedSD::PartitionSpecWithSharedSD(const PartitionSpecWithSharedSD& other266) { partitions = other266.partitions; sd = other266.sd; __isset = other266.__isset; +} +PartitionSpecWithSharedSD& PartitionSpecWithSharedSD::operator=(const PartitionSpecWithSharedSD& other267) { + partitions = other267.partitions; + sd = other267.sd; + __isset = other267.__isset; return *this; } void PartitionSpecWithSharedSD::printTo(std::ostream& out) const { @@ -5974,14 +6032,14 @@ uint32_t PartitionListComposingSpec::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size267; - ::apache::thrift::protocol::TType _etype270; - xfer += iprot->readListBegin(_etype270, _size267); - this->partitions.resize(_size267); - uint32_t _i271; - for (_i271 = 0; _i271 < _size267; ++_i271) + uint32_t _size268; + ::apache::thrift::protocol::TType _etype271; + xfer += iprot->readListBegin(_etype271, _size268); + this->partitions.resize(_size268); + uint32_t _i272; + for (_i272 = 0; _i272 < _size268; ++_i272) { - xfer += this->partitions[_i271].read(iprot); + xfer += this->partitions[_i272].read(iprot); } xfer += iprot->readListEnd(); } @@ -6010,10 +6068,10 @@ uint32_t PartitionListComposingSpec::write(::apache::thrift::protocol::TProtocol xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter272; - for (_iter272 = this->partitions.begin(); _iter272 != this->partitions.end(); ++_iter272) + std::vector ::const_iterator _iter273; + for (_iter273 = this->partitions.begin(); _iter273 != this->partitions.end(); ++_iter273) { - xfer += (*_iter272).write(oprot); + xfer += (*_iter273).write(oprot); } xfer += oprot->writeListEnd(); } @@ -6030,13 +6088,13 @@ void swap(PartitionListComposingSpec &a, PartitionListComposingSpec &b) { swap(a.__isset, b.__isset); } -PartitionListComposingSpec::PartitionListComposingSpec(const PartitionListComposingSpec& other273) { - partitions = other273.partitions; - __isset = other273.__isset; -} -PartitionListComposingSpec& PartitionListComposingSpec::operator=(const PartitionListComposingSpec& other274) { +PartitionListComposingSpec::PartitionListComposingSpec(const PartitionListComposingSpec& other274) { partitions = other274.partitions; __isset = other274.__isset; +} +PartitionListComposingSpec& PartitionListComposingSpec::operator=(const PartitionListComposingSpec& other275) { + partitions = other275.partitions; + __isset = other275.__isset; return *this; } void PartitionListComposingSpec::printTo(std::ostream& out) const { @@ -6188,21 +6246,21 @@ void swap(PartitionSpec &a, PartitionSpec &b) { swap(a.__isset, b.__isset); } -PartitionSpec::PartitionSpec(const PartitionSpec& other275) { - dbName = other275.dbName; - tableName = other275.tableName; - rootPath = other275.rootPath; - sharedSDPartitionSpec = other275.sharedSDPartitionSpec; - partitionList = other275.partitionList; - __isset = other275.__isset; -} -PartitionSpec& PartitionSpec::operator=(const PartitionSpec& other276) { +PartitionSpec::PartitionSpec(const PartitionSpec& other276) { dbName = other276.dbName; tableName = other276.tableName; rootPath = other276.rootPath; sharedSDPartitionSpec = other276.sharedSDPartitionSpec; partitionList = other276.partitionList; __isset = other276.__isset; +} +PartitionSpec& PartitionSpec::operator=(const PartitionSpec& other277) { + dbName = other277.dbName; + tableName = other277.tableName; + rootPath = other277.rootPath; + sharedSDPartitionSpec = other277.sharedSDPartitionSpec; + partitionList = other277.partitionList; + __isset = other277.__isset; return *this; } void PartitionSpec::printTo(std::ostream& out) const { @@ -6350,17 +6408,17 @@ uint32_t Index::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->parameters.clear(); - uint32_t _size277; - ::apache::thrift::protocol::TType _ktype278; - ::apache::thrift::protocol::TType _vtype279; - xfer += iprot->readMapBegin(_ktype278, _vtype279, _size277); - uint32_t _i281; - for (_i281 = 0; _i281 < _size277; ++_i281) + uint32_t _size278; + ::apache::thrift::protocol::TType _ktype279; + ::apache::thrift::protocol::TType _vtype280; + xfer += iprot->readMapBegin(_ktype279, _vtype280, _size278); + uint32_t _i282; + for (_i282 = 0; _i282 < _size278; ++_i282) { - std::string _key282; - xfer += iprot->readString(_key282); - std::string& _val283 = this->parameters[_key282]; - xfer += iprot->readString(_val283); + std::string _key283; + xfer += iprot->readString(_key283); + std::string& _val284 = this->parameters[_key283]; + xfer += iprot->readString(_val284); } xfer += iprot->readMapEnd(); } @@ -6429,11 +6487,11 @@ uint32_t Index::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("parameters", ::apache::thrift::protocol::T_MAP, 9); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->parameters.size())); - std::map ::const_iterator _iter284; - for (_iter284 = this->parameters.begin(); _iter284 != this->parameters.end(); ++_iter284) + std::map ::const_iterator _iter285; + for (_iter285 = this->parameters.begin(); _iter285 != this->parameters.end(); ++_iter285) { - xfer += oprot->writeString(_iter284->first); - xfer += oprot->writeString(_iter284->second); + xfer += oprot->writeString(_iter285->first); + xfer += oprot->writeString(_iter285->second); } xfer += oprot->writeMapEnd(); } @@ -6463,20 +6521,7 @@ void swap(Index &a, Index &b) { swap(a.__isset, b.__isset); } -Index::Index(const Index& other285) { - indexName = other285.indexName; - indexHandlerClass = other285.indexHandlerClass; - dbName = other285.dbName; - origTableName = other285.origTableName; - createTime = other285.createTime; - lastAccessTime = other285.lastAccessTime; - indexTableName = other285.indexTableName; - sd = other285.sd; - parameters = other285.parameters; - deferredRebuild = other285.deferredRebuild; - __isset = other285.__isset; -} -Index& Index::operator=(const Index& other286) { +Index::Index(const Index& other286) { indexName = other286.indexName; indexHandlerClass = other286.indexHandlerClass; dbName = other286.dbName; @@ -6488,6 +6533,19 @@ Index& Index::operator=(const Index& other286) { parameters = other286.parameters; deferredRebuild = other286.deferredRebuild; __isset = other286.__isset; +} +Index& Index::operator=(const Index& other287) { + indexName = other287.indexName; + indexHandlerClass = other287.indexHandlerClass; + dbName = other287.dbName; + origTableName = other287.origTableName; + createTime = other287.createTime; + lastAccessTime = other287.lastAccessTime; + indexTableName = other287.indexTableName; + sd = other287.sd; + parameters = other287.parameters; + deferredRebuild = other287.deferredRebuild; + __isset = other287.__isset; return *this; } void Index::printTo(std::ostream& out) const { @@ -6638,19 +6696,19 @@ void swap(BooleanColumnStatsData &a, BooleanColumnStatsData &b) { swap(a.__isset, b.__isset); } -BooleanColumnStatsData::BooleanColumnStatsData(const BooleanColumnStatsData& other287) { - numTrues = other287.numTrues; - numFalses = other287.numFalses; - numNulls = other287.numNulls; - bitVectors = other287.bitVectors; - __isset = other287.__isset; -} -BooleanColumnStatsData& BooleanColumnStatsData::operator=(const BooleanColumnStatsData& other288) { +BooleanColumnStatsData::BooleanColumnStatsData(const BooleanColumnStatsData& other288) { numTrues = other288.numTrues; numFalses = other288.numFalses; numNulls = other288.numNulls; bitVectors = other288.bitVectors; __isset = other288.__isset; +} +BooleanColumnStatsData& BooleanColumnStatsData::operator=(const BooleanColumnStatsData& other289) { + numTrues = other289.numTrues; + numFalses = other289.numFalses; + numNulls = other289.numNulls; + bitVectors = other289.bitVectors; + __isset = other289.__isset; return *this; } void BooleanColumnStatsData::printTo(std::ostream& out) const { @@ -6813,21 +6871,21 @@ void swap(DoubleColumnStatsData &a, DoubleColumnStatsData &b) { swap(a.__isset, b.__isset); } -DoubleColumnStatsData::DoubleColumnStatsData(const DoubleColumnStatsData& other289) { - lowValue = other289.lowValue; - highValue = other289.highValue; - numNulls = other289.numNulls; - numDVs = other289.numDVs; - bitVectors = other289.bitVectors; - __isset = other289.__isset; -} -DoubleColumnStatsData& DoubleColumnStatsData::operator=(const DoubleColumnStatsData& other290) { +DoubleColumnStatsData::DoubleColumnStatsData(const DoubleColumnStatsData& other290) { lowValue = other290.lowValue; highValue = other290.highValue; numNulls = other290.numNulls; numDVs = other290.numDVs; bitVectors = other290.bitVectors; __isset = other290.__isset; +} +DoubleColumnStatsData& DoubleColumnStatsData::operator=(const DoubleColumnStatsData& other291) { + lowValue = other291.lowValue; + highValue = other291.highValue; + numNulls = other291.numNulls; + numDVs = other291.numDVs; + bitVectors = other291.bitVectors; + __isset = other291.__isset; return *this; } void DoubleColumnStatsData::printTo(std::ostream& out) const { @@ -6991,21 +7049,21 @@ void swap(LongColumnStatsData &a, LongColumnStatsData &b) { swap(a.__isset, b.__isset); } -LongColumnStatsData::LongColumnStatsData(const LongColumnStatsData& other291) { - lowValue = other291.lowValue; - highValue = other291.highValue; - numNulls = other291.numNulls; - numDVs = other291.numDVs; - bitVectors = other291.bitVectors; - __isset = other291.__isset; -} -LongColumnStatsData& LongColumnStatsData::operator=(const LongColumnStatsData& other292) { +LongColumnStatsData::LongColumnStatsData(const LongColumnStatsData& other292) { lowValue = other292.lowValue; highValue = other292.highValue; numNulls = other292.numNulls; numDVs = other292.numDVs; bitVectors = other292.bitVectors; __isset = other292.__isset; +} +LongColumnStatsData& LongColumnStatsData::operator=(const LongColumnStatsData& other293) { + lowValue = other293.lowValue; + highValue = other293.highValue; + numNulls = other293.numNulls; + numDVs = other293.numDVs; + bitVectors = other293.bitVectors; + __isset = other293.__isset; return *this; } void LongColumnStatsData::printTo(std::ostream& out) const { @@ -7171,21 +7229,21 @@ void swap(StringColumnStatsData &a, StringColumnStatsData &b) { swap(a.__isset, b.__isset); } -StringColumnStatsData::StringColumnStatsData(const StringColumnStatsData& other293) { - maxColLen = other293.maxColLen; - avgColLen = other293.avgColLen; - numNulls = other293.numNulls; - numDVs = other293.numDVs; - bitVectors = other293.bitVectors; - __isset = other293.__isset; -} -StringColumnStatsData& StringColumnStatsData::operator=(const StringColumnStatsData& other294) { +StringColumnStatsData::StringColumnStatsData(const StringColumnStatsData& other294) { maxColLen = other294.maxColLen; avgColLen = other294.avgColLen; numNulls = other294.numNulls; numDVs = other294.numDVs; bitVectors = other294.bitVectors; __isset = other294.__isset; +} +StringColumnStatsData& StringColumnStatsData::operator=(const StringColumnStatsData& other295) { + maxColLen = other295.maxColLen; + avgColLen = other295.avgColLen; + numNulls = other295.numNulls; + numDVs = other295.numDVs; + bitVectors = other295.bitVectors; + __isset = other295.__isset; return *this; } void StringColumnStatsData::printTo(std::ostream& out) const { @@ -7331,19 +7389,19 @@ void swap(BinaryColumnStatsData &a, BinaryColumnStatsData &b) { swap(a.__isset, b.__isset); } -BinaryColumnStatsData::BinaryColumnStatsData(const BinaryColumnStatsData& other295) { - maxColLen = other295.maxColLen; - avgColLen = other295.avgColLen; - numNulls = other295.numNulls; - bitVectors = other295.bitVectors; - __isset = other295.__isset; -} -BinaryColumnStatsData& BinaryColumnStatsData::operator=(const BinaryColumnStatsData& other296) { +BinaryColumnStatsData::BinaryColumnStatsData(const BinaryColumnStatsData& other296) { maxColLen = other296.maxColLen; avgColLen = other296.avgColLen; numNulls = other296.numNulls; bitVectors = other296.bitVectors; __isset = other296.__isset; +} +BinaryColumnStatsData& BinaryColumnStatsData::operator=(const BinaryColumnStatsData& other297) { + maxColLen = other297.maxColLen; + avgColLen = other297.avgColLen; + numNulls = other297.numNulls; + bitVectors = other297.bitVectors; + __isset = other297.__isset; return *this; } void BinaryColumnStatsData::printTo(std::ostream& out) const { @@ -7448,13 +7506,13 @@ void swap(Decimal &a, Decimal &b) { swap(a.scale, b.scale); } -Decimal::Decimal(const Decimal& other297) { - unscaled = other297.unscaled; - scale = other297.scale; -} -Decimal& Decimal::operator=(const Decimal& other298) { +Decimal::Decimal(const Decimal& other298) { unscaled = other298.unscaled; scale = other298.scale; +} +Decimal& Decimal::operator=(const Decimal& other299) { + unscaled = other299.unscaled; + scale = other299.scale; return *this; } void Decimal::printTo(std::ostream& out) const { @@ -7615,21 +7673,21 @@ void swap(DecimalColumnStatsData &a, DecimalColumnStatsData &b) { swap(a.__isset, b.__isset); } -DecimalColumnStatsData::DecimalColumnStatsData(const DecimalColumnStatsData& other299) { - lowValue = other299.lowValue; - highValue = other299.highValue; - numNulls = other299.numNulls; - numDVs = other299.numDVs; - bitVectors = other299.bitVectors; - __isset = other299.__isset; -} -DecimalColumnStatsData& DecimalColumnStatsData::operator=(const DecimalColumnStatsData& other300) { +DecimalColumnStatsData::DecimalColumnStatsData(const DecimalColumnStatsData& other300) { lowValue = other300.lowValue; highValue = other300.highValue; numNulls = other300.numNulls; numDVs = other300.numDVs; bitVectors = other300.bitVectors; __isset = other300.__isset; +} +DecimalColumnStatsData& DecimalColumnStatsData::operator=(const DecimalColumnStatsData& other301) { + lowValue = other301.lowValue; + highValue = other301.highValue; + numNulls = other301.numNulls; + numDVs = other301.numDVs; + bitVectors = other301.bitVectors; + __isset = other301.__isset; return *this; } void DecimalColumnStatsData::printTo(std::ostream& out) const { @@ -7715,11 +7773,11 @@ void swap(Date &a, Date &b) { swap(a.daysSinceEpoch, b.daysSinceEpoch); } -Date::Date(const Date& other301) { - daysSinceEpoch = other301.daysSinceEpoch; -} -Date& Date::operator=(const Date& other302) { +Date::Date(const Date& other302) { daysSinceEpoch = other302.daysSinceEpoch; +} +Date& Date::operator=(const Date& other303) { + daysSinceEpoch = other303.daysSinceEpoch; return *this; } void Date::printTo(std::ostream& out) const { @@ -7879,21 +7937,21 @@ void swap(DateColumnStatsData &a, DateColumnStatsData &b) { swap(a.__isset, b.__isset); } -DateColumnStatsData::DateColumnStatsData(const DateColumnStatsData& other303) { - lowValue = other303.lowValue; - highValue = other303.highValue; - numNulls = other303.numNulls; - numDVs = other303.numDVs; - bitVectors = other303.bitVectors; - __isset = other303.__isset; -} -DateColumnStatsData& DateColumnStatsData::operator=(const DateColumnStatsData& other304) { +DateColumnStatsData::DateColumnStatsData(const DateColumnStatsData& other304) { lowValue = other304.lowValue; highValue = other304.highValue; numNulls = other304.numNulls; numDVs = other304.numDVs; bitVectors = other304.bitVectors; __isset = other304.__isset; +} +DateColumnStatsData& DateColumnStatsData::operator=(const DateColumnStatsData& other305) { + lowValue = other305.lowValue; + highValue = other305.highValue; + numNulls = other305.numNulls; + numDVs = other305.numDVs; + bitVectors = other305.bitVectors; + __isset = other305.__isset; return *this; } void DateColumnStatsData::printTo(std::ostream& out) const { @@ -8079,17 +8137,7 @@ void swap(ColumnStatisticsData &a, ColumnStatisticsData &b) { swap(a.__isset, b.__isset); } -ColumnStatisticsData::ColumnStatisticsData(const ColumnStatisticsData& other305) { - booleanStats = other305.booleanStats; - longStats = other305.longStats; - doubleStats = other305.doubleStats; - stringStats = other305.stringStats; - binaryStats = other305.binaryStats; - decimalStats = other305.decimalStats; - dateStats = other305.dateStats; - __isset = other305.__isset; -} -ColumnStatisticsData& ColumnStatisticsData::operator=(const ColumnStatisticsData& other306) { +ColumnStatisticsData::ColumnStatisticsData(const ColumnStatisticsData& other306) { booleanStats = other306.booleanStats; longStats = other306.longStats; doubleStats = other306.doubleStats; @@ -8098,6 +8146,16 @@ ColumnStatisticsData& ColumnStatisticsData::operator=(const ColumnStatisticsData decimalStats = other306.decimalStats; dateStats = other306.dateStats; __isset = other306.__isset; +} +ColumnStatisticsData& ColumnStatisticsData::operator=(const ColumnStatisticsData& other307) { + booleanStats = other307.booleanStats; + longStats = other307.longStats; + doubleStats = other307.doubleStats; + stringStats = other307.stringStats; + binaryStats = other307.binaryStats; + decimalStats = other307.decimalStats; + dateStats = other307.dateStats; + __isset = other307.__isset; return *this; } void ColumnStatisticsData::printTo(std::ostream& out) const { @@ -8225,15 +8283,15 @@ void swap(ColumnStatisticsObj &a, ColumnStatisticsObj &b) { swap(a.statsData, b.statsData); } -ColumnStatisticsObj::ColumnStatisticsObj(const ColumnStatisticsObj& other307) { - colName = other307.colName; - colType = other307.colType; - statsData = other307.statsData; -} -ColumnStatisticsObj& ColumnStatisticsObj::operator=(const ColumnStatisticsObj& other308) { +ColumnStatisticsObj::ColumnStatisticsObj(const ColumnStatisticsObj& other308) { colName = other308.colName; colType = other308.colType; statsData = other308.statsData; +} +ColumnStatisticsObj& ColumnStatisticsObj::operator=(const ColumnStatisticsObj& other309) { + colName = other309.colName; + colType = other309.colType; + statsData = other309.statsData; return *this; } void ColumnStatisticsObj::printTo(std::ostream& out) const { @@ -8396,21 +8454,21 @@ void swap(ColumnStatisticsDesc &a, ColumnStatisticsDesc &b) { swap(a.__isset, b.__isset); } -ColumnStatisticsDesc::ColumnStatisticsDesc(const ColumnStatisticsDesc& other309) { - isTblLevel = other309.isTblLevel; - dbName = other309.dbName; - tableName = other309.tableName; - partName = other309.partName; - lastAnalyzed = other309.lastAnalyzed; - __isset = other309.__isset; -} -ColumnStatisticsDesc& ColumnStatisticsDesc::operator=(const ColumnStatisticsDesc& other310) { +ColumnStatisticsDesc::ColumnStatisticsDesc(const ColumnStatisticsDesc& other310) { isTblLevel = other310.isTblLevel; dbName = other310.dbName; tableName = other310.tableName; partName = other310.partName; lastAnalyzed = other310.lastAnalyzed; __isset = other310.__isset; +} +ColumnStatisticsDesc& ColumnStatisticsDesc::operator=(const ColumnStatisticsDesc& other311) { + isTblLevel = other311.isTblLevel; + dbName = other311.dbName; + tableName = other311.tableName; + partName = other311.partName; + lastAnalyzed = other311.lastAnalyzed; + __isset = other311.__isset; return *this; } void ColumnStatisticsDesc::printTo(std::ostream& out) const { @@ -8472,14 +8530,14 @@ uint32_t ColumnStatistics::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->statsObj.clear(); - uint32_t _size311; - ::apache::thrift::protocol::TType _etype314; - xfer += iprot->readListBegin(_etype314, _size311); - this->statsObj.resize(_size311); - uint32_t _i315; - for (_i315 = 0; _i315 < _size311; ++_i315) + uint32_t _size312; + ::apache::thrift::protocol::TType _etype315; + xfer += iprot->readListBegin(_etype315, _size312); + this->statsObj.resize(_size312); + uint32_t _i316; + for (_i316 = 0; _i316 < _size312; ++_i316) { - xfer += this->statsObj[_i315].read(iprot); + xfer += this->statsObj[_i316].read(iprot); } xfer += iprot->readListEnd(); } @@ -8516,10 +8574,10 @@ uint32_t ColumnStatistics::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("statsObj", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->statsObj.size())); - std::vector ::const_iterator _iter316; - for (_iter316 = this->statsObj.begin(); _iter316 != this->statsObj.end(); ++_iter316) + std::vector ::const_iterator _iter317; + for (_iter317 = this->statsObj.begin(); _iter317 != this->statsObj.end(); ++_iter317) { - xfer += (*_iter316).write(oprot); + xfer += (*_iter317).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8536,13 +8594,13 @@ void swap(ColumnStatistics &a, ColumnStatistics &b) { swap(a.statsObj, b.statsObj); } -ColumnStatistics::ColumnStatistics(const ColumnStatistics& other317) { - statsDesc = other317.statsDesc; - statsObj = other317.statsObj; -} -ColumnStatistics& ColumnStatistics::operator=(const ColumnStatistics& other318) { +ColumnStatistics::ColumnStatistics(const ColumnStatistics& other318) { statsDesc = other318.statsDesc; statsObj = other318.statsObj; +} +ColumnStatistics& ColumnStatistics::operator=(const ColumnStatistics& other319) { + statsDesc = other319.statsDesc; + statsObj = other319.statsObj; return *this; } void ColumnStatistics::printTo(std::ostream& out) const { @@ -8593,14 +8651,14 @@ uint32_t AggrStats::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->colStats.clear(); - uint32_t _size319; - ::apache::thrift::protocol::TType _etype322; - xfer += iprot->readListBegin(_etype322, _size319); - this->colStats.resize(_size319); - uint32_t _i323; - for (_i323 = 0; _i323 < _size319; ++_i323) + uint32_t _size320; + ::apache::thrift::protocol::TType _etype323; + xfer += iprot->readListBegin(_etype323, _size320); + this->colStats.resize(_size320); + uint32_t _i324; + for (_i324 = 0; _i324 < _size320; ++_i324) { - xfer += this->colStats[_i323].read(iprot); + xfer += this->colStats[_i324].read(iprot); } xfer += iprot->readListEnd(); } @@ -8641,10 +8699,10 @@ uint32_t AggrStats::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("colStats", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->colStats.size())); - std::vector ::const_iterator _iter324; - for (_iter324 = this->colStats.begin(); _iter324 != this->colStats.end(); ++_iter324) + std::vector ::const_iterator _iter325; + for (_iter325 = this->colStats.begin(); _iter325 != this->colStats.end(); ++_iter325) { - xfer += (*_iter324).write(oprot); + xfer += (*_iter325).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8665,13 +8723,13 @@ void swap(AggrStats &a, AggrStats &b) { swap(a.partsFound, b.partsFound); } -AggrStats::AggrStats(const AggrStats& other325) { - colStats = other325.colStats; - partsFound = other325.partsFound; -} -AggrStats& AggrStats::operator=(const AggrStats& other326) { +AggrStats::AggrStats(const AggrStats& other326) { colStats = other326.colStats; partsFound = other326.partsFound; +} +AggrStats& AggrStats::operator=(const AggrStats& other327) { + colStats = other327.colStats; + partsFound = other327.partsFound; return *this; } void AggrStats::printTo(std::ostream& out) const { @@ -8722,14 +8780,14 @@ uint32_t SetPartitionsStatsRequest::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->colStats.clear(); - uint32_t _size327; - ::apache::thrift::protocol::TType _etype330; - xfer += iprot->readListBegin(_etype330, _size327); - this->colStats.resize(_size327); - uint32_t _i331; - for (_i331 = 0; _i331 < _size327; ++_i331) + uint32_t _size328; + ::apache::thrift::protocol::TType _etype331; + xfer += iprot->readListBegin(_etype331, _size328); + this->colStats.resize(_size328); + uint32_t _i332; + for (_i332 = 0; _i332 < _size328; ++_i332) { - xfer += this->colStats[_i331].read(iprot); + xfer += this->colStats[_i332].read(iprot); } xfer += iprot->readListEnd(); } @@ -8768,10 +8826,10 @@ uint32_t SetPartitionsStatsRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("colStats", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->colStats.size())); - std::vector ::const_iterator _iter332; - for (_iter332 = this->colStats.begin(); _iter332 != this->colStats.end(); ++_iter332) + std::vector ::const_iterator _iter333; + for (_iter333 = this->colStats.begin(); _iter333 != this->colStats.end(); ++_iter333) { - xfer += (*_iter332).write(oprot); + xfer += (*_iter333).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8794,15 +8852,15 @@ void swap(SetPartitionsStatsRequest &a, SetPartitionsStatsRequest &b) { swap(a.__isset, b.__isset); } -SetPartitionsStatsRequest::SetPartitionsStatsRequest(const SetPartitionsStatsRequest& other333) { - colStats = other333.colStats; - needMerge = other333.needMerge; - __isset = other333.__isset; -} -SetPartitionsStatsRequest& SetPartitionsStatsRequest::operator=(const SetPartitionsStatsRequest& other334) { +SetPartitionsStatsRequest::SetPartitionsStatsRequest(const SetPartitionsStatsRequest& other334) { colStats = other334.colStats; needMerge = other334.needMerge; __isset = other334.__isset; +} +SetPartitionsStatsRequest& SetPartitionsStatsRequest::operator=(const SetPartitionsStatsRequest& other335) { + colStats = other335.colStats; + needMerge = other335.needMerge; + __isset = other335.__isset; return *this; } void SetPartitionsStatsRequest::printTo(std::ostream& out) const { @@ -8851,14 +8909,14 @@ uint32_t Schema::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fieldSchemas.clear(); - uint32_t _size335; - ::apache::thrift::protocol::TType _etype338; - xfer += iprot->readListBegin(_etype338, _size335); - this->fieldSchemas.resize(_size335); - uint32_t _i339; - for (_i339 = 0; _i339 < _size335; ++_i339) + uint32_t _size336; + ::apache::thrift::protocol::TType _etype339; + xfer += iprot->readListBegin(_etype339, _size336); + this->fieldSchemas.resize(_size336); + uint32_t _i340; + for (_i340 = 0; _i340 < _size336; ++_i340) { - xfer += this->fieldSchemas[_i339].read(iprot); + xfer += this->fieldSchemas[_i340].read(iprot); } xfer += iprot->readListEnd(); } @@ -8871,17 +8929,17 @@ uint32_t Schema::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->properties.clear(); - uint32_t _size340; - ::apache::thrift::protocol::TType _ktype341; - ::apache::thrift::protocol::TType _vtype342; - xfer += iprot->readMapBegin(_ktype341, _vtype342, _size340); - uint32_t _i344; - for (_i344 = 0; _i344 < _size340; ++_i344) + uint32_t _size341; + ::apache::thrift::protocol::TType _ktype342; + ::apache::thrift::protocol::TType _vtype343; + xfer += iprot->readMapBegin(_ktype342, _vtype343, _size341); + uint32_t _i345; + for (_i345 = 0; _i345 < _size341; ++_i345) { - std::string _key345; - xfer += iprot->readString(_key345); - std::string& _val346 = this->properties[_key345]; - xfer += iprot->readString(_val346); + std::string _key346; + xfer += iprot->readString(_key346); + std::string& _val347 = this->properties[_key346]; + xfer += iprot->readString(_val347); } xfer += iprot->readMapEnd(); } @@ -8910,10 +8968,10 @@ uint32_t Schema::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("fieldSchemas", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->fieldSchemas.size())); - std::vector ::const_iterator _iter347; - for (_iter347 = this->fieldSchemas.begin(); _iter347 != this->fieldSchemas.end(); ++_iter347) + std::vector ::const_iterator _iter348; + for (_iter348 = this->fieldSchemas.begin(); _iter348 != this->fieldSchemas.end(); ++_iter348) { - xfer += (*_iter347).write(oprot); + xfer += (*_iter348).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8922,11 +8980,11 @@ uint32_t Schema::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("properties", ::apache::thrift::protocol::T_MAP, 2); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->properties.size())); - std::map ::const_iterator _iter348; - for (_iter348 = this->properties.begin(); _iter348 != this->properties.end(); ++_iter348) + std::map ::const_iterator _iter349; + for (_iter349 = this->properties.begin(); _iter349 != this->properties.end(); ++_iter349) { - xfer += oprot->writeString(_iter348->first); - xfer += oprot->writeString(_iter348->second); + xfer += oprot->writeString(_iter349->first); + xfer += oprot->writeString(_iter349->second); } xfer += oprot->writeMapEnd(); } @@ -8944,15 +9002,15 @@ void swap(Schema &a, Schema &b) { swap(a.__isset, b.__isset); } -Schema::Schema(const Schema& other349) { - fieldSchemas = other349.fieldSchemas; - properties = other349.properties; - __isset = other349.__isset; -} -Schema& Schema::operator=(const Schema& other350) { +Schema::Schema(const Schema& other350) { fieldSchemas = other350.fieldSchemas; properties = other350.properties; __isset = other350.__isset; +} +Schema& Schema::operator=(const Schema& other351) { + fieldSchemas = other351.fieldSchemas; + properties = other351.properties; + __isset = other351.__isset; return *this; } void Schema::printTo(std::ostream& out) const { @@ -8997,17 +9055,17 @@ uint32_t EnvironmentContext::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_MAP) { { this->properties.clear(); - uint32_t _size351; - ::apache::thrift::protocol::TType _ktype352; - ::apache::thrift::protocol::TType _vtype353; - xfer += iprot->readMapBegin(_ktype352, _vtype353, _size351); - uint32_t _i355; - for (_i355 = 0; _i355 < _size351; ++_i355) + uint32_t _size352; + ::apache::thrift::protocol::TType _ktype353; + ::apache::thrift::protocol::TType _vtype354; + xfer += iprot->readMapBegin(_ktype353, _vtype354, _size352); + uint32_t _i356; + for (_i356 = 0; _i356 < _size352; ++_i356) { - std::string _key356; - xfer += iprot->readString(_key356); - std::string& _val357 = this->properties[_key356]; - xfer += iprot->readString(_val357); + std::string _key357; + xfer += iprot->readString(_key357); + std::string& _val358 = this->properties[_key357]; + xfer += iprot->readString(_val358); } xfer += iprot->readMapEnd(); } @@ -9036,11 +9094,11 @@ uint32_t EnvironmentContext::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("properties", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->properties.size())); - std::map ::const_iterator _iter358; - for (_iter358 = this->properties.begin(); _iter358 != this->properties.end(); ++_iter358) + std::map ::const_iterator _iter359; + for (_iter359 = this->properties.begin(); _iter359 != this->properties.end(); ++_iter359) { - xfer += oprot->writeString(_iter358->first); - xfer += oprot->writeString(_iter358->second); + xfer += oprot->writeString(_iter359->first); + xfer += oprot->writeString(_iter359->second); } xfer += oprot->writeMapEnd(); } @@ -9057,13 +9115,13 @@ void swap(EnvironmentContext &a, EnvironmentContext &b) { swap(a.__isset, b.__isset); } -EnvironmentContext::EnvironmentContext(const EnvironmentContext& other359) { - properties = other359.properties; - __isset = other359.__isset; -} -EnvironmentContext& EnvironmentContext::operator=(const EnvironmentContext& other360) { +EnvironmentContext::EnvironmentContext(const EnvironmentContext& other360) { properties = other360.properties; __isset = other360.__isset; +} +EnvironmentContext& EnvironmentContext::operator=(const EnvironmentContext& other361) { + properties = other361.properties; + __isset = other361.__isset; return *this; } void EnvironmentContext::printTo(std::ostream& out) const { @@ -9165,13 +9223,13 @@ void swap(PrimaryKeysRequest &a, PrimaryKeysRequest &b) { swap(a.tbl_name, b.tbl_name); } -PrimaryKeysRequest::PrimaryKeysRequest(const PrimaryKeysRequest& other361) { - db_name = other361.db_name; - tbl_name = other361.tbl_name; -} -PrimaryKeysRequest& PrimaryKeysRequest::operator=(const PrimaryKeysRequest& other362) { +PrimaryKeysRequest::PrimaryKeysRequest(const PrimaryKeysRequest& other362) { db_name = other362.db_name; tbl_name = other362.tbl_name; +} +PrimaryKeysRequest& PrimaryKeysRequest::operator=(const PrimaryKeysRequest& other363) { + db_name = other363.db_name; + tbl_name = other363.tbl_name; return *this; } void PrimaryKeysRequest::printTo(std::ostream& out) const { @@ -9217,14 +9275,14 @@ uint32_t PrimaryKeysResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->primaryKeys.clear(); - uint32_t _size363; - ::apache::thrift::protocol::TType _etype366; - xfer += iprot->readListBegin(_etype366, _size363); - this->primaryKeys.resize(_size363); - uint32_t _i367; - for (_i367 = 0; _i367 < _size363; ++_i367) + uint32_t _size364; + ::apache::thrift::protocol::TType _etype367; + xfer += iprot->readListBegin(_etype367, _size364); + this->primaryKeys.resize(_size364); + uint32_t _i368; + for (_i368 = 0; _i368 < _size364; ++_i368) { - xfer += this->primaryKeys[_i367].read(iprot); + xfer += this->primaryKeys[_i368].read(iprot); } xfer += iprot->readListEnd(); } @@ -9255,10 +9313,10 @@ uint32_t PrimaryKeysResponse::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->primaryKeys.size())); - std::vector ::const_iterator _iter368; - for (_iter368 = this->primaryKeys.begin(); _iter368 != this->primaryKeys.end(); ++_iter368) + std::vector ::const_iterator _iter369; + for (_iter369 = this->primaryKeys.begin(); _iter369 != this->primaryKeys.end(); ++_iter369) { - xfer += (*_iter368).write(oprot); + xfer += (*_iter369).write(oprot); } xfer += oprot->writeListEnd(); } @@ -9274,11 +9332,11 @@ void swap(PrimaryKeysResponse &a, PrimaryKeysResponse &b) { swap(a.primaryKeys, b.primaryKeys); } -PrimaryKeysResponse::PrimaryKeysResponse(const PrimaryKeysResponse& other369) { - primaryKeys = other369.primaryKeys; -} -PrimaryKeysResponse& PrimaryKeysResponse::operator=(const PrimaryKeysResponse& other370) { +PrimaryKeysResponse::PrimaryKeysResponse(const PrimaryKeysResponse& other370) { primaryKeys = other370.primaryKeys; +} +PrimaryKeysResponse& PrimaryKeysResponse::operator=(const PrimaryKeysResponse& other371) { + primaryKeys = other371.primaryKeys; return *this; } void PrimaryKeysResponse::printTo(std::ostream& out) const { @@ -9409,19 +9467,19 @@ void swap(ForeignKeysRequest &a, ForeignKeysRequest &b) { swap(a.__isset, b.__isset); } -ForeignKeysRequest::ForeignKeysRequest(const ForeignKeysRequest& other371) { - parent_db_name = other371.parent_db_name; - parent_tbl_name = other371.parent_tbl_name; - foreign_db_name = other371.foreign_db_name; - foreign_tbl_name = other371.foreign_tbl_name; - __isset = other371.__isset; -} -ForeignKeysRequest& ForeignKeysRequest::operator=(const ForeignKeysRequest& other372) { +ForeignKeysRequest::ForeignKeysRequest(const ForeignKeysRequest& other372) { parent_db_name = other372.parent_db_name; parent_tbl_name = other372.parent_tbl_name; foreign_db_name = other372.foreign_db_name; foreign_tbl_name = other372.foreign_tbl_name; __isset = other372.__isset; +} +ForeignKeysRequest& ForeignKeysRequest::operator=(const ForeignKeysRequest& other373) { + parent_db_name = other373.parent_db_name; + parent_tbl_name = other373.parent_tbl_name; + foreign_db_name = other373.foreign_db_name; + foreign_tbl_name = other373.foreign_tbl_name; + __isset = other373.__isset; return *this; } void ForeignKeysRequest::printTo(std::ostream& out) const { @@ -9469,14 +9527,14 @@ uint32_t ForeignKeysResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->foreignKeys.clear(); - uint32_t _size373; - ::apache::thrift::protocol::TType _etype376; - xfer += iprot->readListBegin(_etype376, _size373); - this->foreignKeys.resize(_size373); - uint32_t _i377; - for (_i377 = 0; _i377 < _size373; ++_i377) + uint32_t _size374; + ::apache::thrift::protocol::TType _etype377; + xfer += iprot->readListBegin(_etype377, _size374); + this->foreignKeys.resize(_size374); + uint32_t _i378; + for (_i378 = 0; _i378 < _size374; ++_i378) { - xfer += this->foreignKeys[_i377].read(iprot); + xfer += this->foreignKeys[_i378].read(iprot); } xfer += iprot->readListEnd(); } @@ -9507,10 +9565,10 @@ uint32_t ForeignKeysResponse::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->foreignKeys.size())); - std::vector ::const_iterator _iter378; - for (_iter378 = this->foreignKeys.begin(); _iter378 != this->foreignKeys.end(); ++_iter378) + std::vector ::const_iterator _iter379; + for (_iter379 = this->foreignKeys.begin(); _iter379 != this->foreignKeys.end(); ++_iter379) { - xfer += (*_iter378).write(oprot); + xfer += (*_iter379).write(oprot); } xfer += oprot->writeListEnd(); } @@ -9526,11 +9584,11 @@ void swap(ForeignKeysResponse &a, ForeignKeysResponse &b) { swap(a.foreignKeys, b.foreignKeys); } -ForeignKeysResponse::ForeignKeysResponse(const ForeignKeysResponse& other379) { - foreignKeys = other379.foreignKeys; -} -ForeignKeysResponse& ForeignKeysResponse::operator=(const ForeignKeysResponse& other380) { +ForeignKeysResponse::ForeignKeysResponse(const ForeignKeysResponse& other380) { foreignKeys = other380.foreignKeys; +} +ForeignKeysResponse& ForeignKeysResponse::operator=(const ForeignKeysResponse& other381) { + foreignKeys = other381.foreignKeys; return *this; } void ForeignKeysResponse::printTo(std::ostream& out) const { @@ -9632,13 +9690,13 @@ void swap(UniqueConstraintsRequest &a, UniqueConstraintsRequest &b) { swap(a.tbl_name, b.tbl_name); } -UniqueConstraintsRequest::UniqueConstraintsRequest(const UniqueConstraintsRequest& other381) { - db_name = other381.db_name; - tbl_name = other381.tbl_name; -} -UniqueConstraintsRequest& UniqueConstraintsRequest::operator=(const UniqueConstraintsRequest& other382) { +UniqueConstraintsRequest::UniqueConstraintsRequest(const UniqueConstraintsRequest& other382) { db_name = other382.db_name; tbl_name = other382.tbl_name; +} +UniqueConstraintsRequest& UniqueConstraintsRequest::operator=(const UniqueConstraintsRequest& other383) { + db_name = other383.db_name; + tbl_name = other383.tbl_name; return *this; } void UniqueConstraintsRequest::printTo(std::ostream& out) const { @@ -9684,14 +9742,14 @@ uint32_t UniqueConstraintsResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->uniqueConstraints.clear(); - uint32_t _size383; - ::apache::thrift::protocol::TType _etype386; - xfer += iprot->readListBegin(_etype386, _size383); - this->uniqueConstraints.resize(_size383); - uint32_t _i387; - for (_i387 = 0; _i387 < _size383; ++_i387) + uint32_t _size384; + ::apache::thrift::protocol::TType _etype387; + xfer += iprot->readListBegin(_etype387, _size384); + this->uniqueConstraints.resize(_size384); + uint32_t _i388; + for (_i388 = 0; _i388 < _size384; ++_i388) { - xfer += this->uniqueConstraints[_i387].read(iprot); + xfer += this->uniqueConstraints[_i388].read(iprot); } xfer += iprot->readListEnd(); } @@ -9722,10 +9780,10 @@ uint32_t UniqueConstraintsResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->uniqueConstraints.size())); - std::vector ::const_iterator _iter388; - for (_iter388 = this->uniqueConstraints.begin(); _iter388 != this->uniqueConstraints.end(); ++_iter388) + std::vector ::const_iterator _iter389; + for (_iter389 = this->uniqueConstraints.begin(); _iter389 != this->uniqueConstraints.end(); ++_iter389) { - xfer += (*_iter388).write(oprot); + xfer += (*_iter389).write(oprot); } xfer += oprot->writeListEnd(); } @@ -9741,11 +9799,11 @@ void swap(UniqueConstraintsResponse &a, UniqueConstraintsResponse &b) { swap(a.uniqueConstraints, b.uniqueConstraints); } -UniqueConstraintsResponse::UniqueConstraintsResponse(const UniqueConstraintsResponse& other389) { - uniqueConstraints = other389.uniqueConstraints; -} -UniqueConstraintsResponse& UniqueConstraintsResponse::operator=(const UniqueConstraintsResponse& other390) { +UniqueConstraintsResponse::UniqueConstraintsResponse(const UniqueConstraintsResponse& other390) { uniqueConstraints = other390.uniqueConstraints; +} +UniqueConstraintsResponse& UniqueConstraintsResponse::operator=(const UniqueConstraintsResponse& other391) { + uniqueConstraints = other391.uniqueConstraints; return *this; } void UniqueConstraintsResponse::printTo(std::ostream& out) const { @@ -9847,13 +9905,13 @@ void swap(NotNullConstraintsRequest &a, NotNullConstraintsRequest &b) { swap(a.tbl_name, b.tbl_name); } -NotNullConstraintsRequest::NotNullConstraintsRequest(const NotNullConstraintsRequest& other391) { - db_name = other391.db_name; - tbl_name = other391.tbl_name; -} -NotNullConstraintsRequest& NotNullConstraintsRequest::operator=(const NotNullConstraintsRequest& other392) { +NotNullConstraintsRequest::NotNullConstraintsRequest(const NotNullConstraintsRequest& other392) { db_name = other392.db_name; tbl_name = other392.tbl_name; +} +NotNullConstraintsRequest& NotNullConstraintsRequest::operator=(const NotNullConstraintsRequest& other393) { + db_name = other393.db_name; + tbl_name = other393.tbl_name; return *this; } void NotNullConstraintsRequest::printTo(std::ostream& out) const { @@ -9899,14 +9957,14 @@ uint32_t NotNullConstraintsResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->notNullConstraints.clear(); - uint32_t _size393; - ::apache::thrift::protocol::TType _etype396; - xfer += iprot->readListBegin(_etype396, _size393); - this->notNullConstraints.resize(_size393); - uint32_t _i397; - for (_i397 = 0; _i397 < _size393; ++_i397) + uint32_t _size394; + ::apache::thrift::protocol::TType _etype397; + xfer += iprot->readListBegin(_etype397, _size394); + this->notNullConstraints.resize(_size394); + uint32_t _i398; + for (_i398 = 0; _i398 < _size394; ++_i398) { - xfer += this->notNullConstraints[_i397].read(iprot); + xfer += this->notNullConstraints[_i398].read(iprot); } xfer += iprot->readListEnd(); } @@ -9937,10 +9995,10 @@ uint32_t NotNullConstraintsResponse::write(::apache::thrift::protocol::TProtocol xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->notNullConstraints.size())); - std::vector ::const_iterator _iter398; - for (_iter398 = this->notNullConstraints.begin(); _iter398 != this->notNullConstraints.end(); ++_iter398) + std::vector ::const_iterator _iter399; + for (_iter399 = this->notNullConstraints.begin(); _iter399 != this->notNullConstraints.end(); ++_iter399) { - xfer += (*_iter398).write(oprot); + xfer += (*_iter399).write(oprot); } xfer += oprot->writeListEnd(); } @@ -9956,11 +10014,11 @@ void swap(NotNullConstraintsResponse &a, NotNullConstraintsResponse &b) { swap(a.notNullConstraints, b.notNullConstraints); } -NotNullConstraintsResponse::NotNullConstraintsResponse(const NotNullConstraintsResponse& other399) { - notNullConstraints = other399.notNullConstraints; -} -NotNullConstraintsResponse& NotNullConstraintsResponse::operator=(const NotNullConstraintsResponse& other400) { +NotNullConstraintsResponse::NotNullConstraintsResponse(const NotNullConstraintsResponse& other400) { notNullConstraints = other400.notNullConstraints; +} +NotNullConstraintsResponse& NotNullConstraintsResponse::operator=(const NotNullConstraintsResponse& other401) { + notNullConstraints = other401.notNullConstraints; return *this; } void NotNullConstraintsResponse::printTo(std::ostream& out) const { @@ -10082,15 +10140,15 @@ void swap(DropConstraintRequest &a, DropConstraintRequest &b) { swap(a.constraintname, b.constraintname); } -DropConstraintRequest::DropConstraintRequest(const DropConstraintRequest& other401) { - dbname = other401.dbname; - tablename = other401.tablename; - constraintname = other401.constraintname; -} -DropConstraintRequest& DropConstraintRequest::operator=(const DropConstraintRequest& other402) { +DropConstraintRequest::DropConstraintRequest(const DropConstraintRequest& other402) { dbname = other402.dbname; tablename = other402.tablename; constraintname = other402.constraintname; +} +DropConstraintRequest& DropConstraintRequest::operator=(const DropConstraintRequest& other403) { + dbname = other403.dbname; + tablename = other403.tablename; + constraintname = other403.constraintname; return *this; } void DropConstraintRequest::printTo(std::ostream& out) const { @@ -10137,14 +10195,14 @@ uint32_t AddPrimaryKeyRequest::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->primaryKeyCols.clear(); - uint32_t _size403; - ::apache::thrift::protocol::TType _etype406; - xfer += iprot->readListBegin(_etype406, _size403); - this->primaryKeyCols.resize(_size403); - uint32_t _i407; - for (_i407 = 0; _i407 < _size403; ++_i407) + uint32_t _size404; + ::apache::thrift::protocol::TType _etype407; + xfer += iprot->readListBegin(_etype407, _size404); + this->primaryKeyCols.resize(_size404); + uint32_t _i408; + for (_i408 = 0; _i408 < _size404; ++_i408) { - xfer += this->primaryKeyCols[_i407].read(iprot); + xfer += this->primaryKeyCols[_i408].read(iprot); } xfer += iprot->readListEnd(); } @@ -10175,10 +10233,10 @@ uint32_t AddPrimaryKeyRequest::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("primaryKeyCols", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->primaryKeyCols.size())); - std::vector ::const_iterator _iter408; - for (_iter408 = this->primaryKeyCols.begin(); _iter408 != this->primaryKeyCols.end(); ++_iter408) + std::vector ::const_iterator _iter409; + for (_iter409 = this->primaryKeyCols.begin(); _iter409 != this->primaryKeyCols.end(); ++_iter409) { - xfer += (*_iter408).write(oprot); + xfer += (*_iter409).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10194,11 +10252,11 @@ void swap(AddPrimaryKeyRequest &a, AddPrimaryKeyRequest &b) { swap(a.primaryKeyCols, b.primaryKeyCols); } -AddPrimaryKeyRequest::AddPrimaryKeyRequest(const AddPrimaryKeyRequest& other409) { - primaryKeyCols = other409.primaryKeyCols; -} -AddPrimaryKeyRequest& AddPrimaryKeyRequest::operator=(const AddPrimaryKeyRequest& other410) { +AddPrimaryKeyRequest::AddPrimaryKeyRequest(const AddPrimaryKeyRequest& other410) { primaryKeyCols = other410.primaryKeyCols; +} +AddPrimaryKeyRequest& AddPrimaryKeyRequest::operator=(const AddPrimaryKeyRequest& other411) { + primaryKeyCols = other411.primaryKeyCols; return *this; } void AddPrimaryKeyRequest::printTo(std::ostream& out) const { @@ -10243,14 +10301,14 @@ uint32_t AddForeignKeyRequest::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->foreignKeyCols.clear(); - uint32_t _size411; - ::apache::thrift::protocol::TType _etype414; - xfer += iprot->readListBegin(_etype414, _size411); - this->foreignKeyCols.resize(_size411); - uint32_t _i415; - for (_i415 = 0; _i415 < _size411; ++_i415) + uint32_t _size412; + ::apache::thrift::protocol::TType _etype415; + xfer += iprot->readListBegin(_etype415, _size412); + this->foreignKeyCols.resize(_size412); + uint32_t _i416; + for (_i416 = 0; _i416 < _size412; ++_i416) { - xfer += this->foreignKeyCols[_i415].read(iprot); + xfer += this->foreignKeyCols[_i416].read(iprot); } xfer += iprot->readListEnd(); } @@ -10281,10 +10339,10 @@ uint32_t AddForeignKeyRequest::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("foreignKeyCols", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->foreignKeyCols.size())); - std::vector ::const_iterator _iter416; - for (_iter416 = this->foreignKeyCols.begin(); _iter416 != this->foreignKeyCols.end(); ++_iter416) + std::vector ::const_iterator _iter417; + for (_iter417 = this->foreignKeyCols.begin(); _iter417 != this->foreignKeyCols.end(); ++_iter417) { - xfer += (*_iter416).write(oprot); + xfer += (*_iter417).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10300,11 +10358,11 @@ void swap(AddForeignKeyRequest &a, AddForeignKeyRequest &b) { swap(a.foreignKeyCols, b.foreignKeyCols); } -AddForeignKeyRequest::AddForeignKeyRequest(const AddForeignKeyRequest& other417) { - foreignKeyCols = other417.foreignKeyCols; -} -AddForeignKeyRequest& AddForeignKeyRequest::operator=(const AddForeignKeyRequest& other418) { +AddForeignKeyRequest::AddForeignKeyRequest(const AddForeignKeyRequest& other418) { foreignKeyCols = other418.foreignKeyCols; +} +AddForeignKeyRequest& AddForeignKeyRequest::operator=(const AddForeignKeyRequest& other419) { + foreignKeyCols = other419.foreignKeyCols; return *this; } void AddForeignKeyRequest::printTo(std::ostream& out) const { @@ -10349,14 +10407,14 @@ uint32_t AddUniqueConstraintRequest::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->uniqueConstraintCols.clear(); - uint32_t _size419; - ::apache::thrift::protocol::TType _etype422; - xfer += iprot->readListBegin(_etype422, _size419); - this->uniqueConstraintCols.resize(_size419); - uint32_t _i423; - for (_i423 = 0; _i423 < _size419; ++_i423) + uint32_t _size420; + ::apache::thrift::protocol::TType _etype423; + xfer += iprot->readListBegin(_etype423, _size420); + this->uniqueConstraintCols.resize(_size420); + uint32_t _i424; + for (_i424 = 0; _i424 < _size420; ++_i424) { - xfer += this->uniqueConstraintCols[_i423].read(iprot); + xfer += this->uniqueConstraintCols[_i424].read(iprot); } xfer += iprot->readListEnd(); } @@ -10387,10 +10445,10 @@ uint32_t AddUniqueConstraintRequest::write(::apache::thrift::protocol::TProtocol xfer += oprot->writeFieldBegin("uniqueConstraintCols", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->uniqueConstraintCols.size())); - std::vector ::const_iterator _iter424; - for (_iter424 = this->uniqueConstraintCols.begin(); _iter424 != this->uniqueConstraintCols.end(); ++_iter424) + std::vector ::const_iterator _iter425; + for (_iter425 = this->uniqueConstraintCols.begin(); _iter425 != this->uniqueConstraintCols.end(); ++_iter425) { - xfer += (*_iter424).write(oprot); + xfer += (*_iter425).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10406,11 +10464,11 @@ void swap(AddUniqueConstraintRequest &a, AddUniqueConstraintRequest &b) { swap(a.uniqueConstraintCols, b.uniqueConstraintCols); } -AddUniqueConstraintRequest::AddUniqueConstraintRequest(const AddUniqueConstraintRequest& other425) { - uniqueConstraintCols = other425.uniqueConstraintCols; -} -AddUniqueConstraintRequest& AddUniqueConstraintRequest::operator=(const AddUniqueConstraintRequest& other426) { +AddUniqueConstraintRequest::AddUniqueConstraintRequest(const AddUniqueConstraintRequest& other426) { uniqueConstraintCols = other426.uniqueConstraintCols; +} +AddUniqueConstraintRequest& AddUniqueConstraintRequest::operator=(const AddUniqueConstraintRequest& other427) { + uniqueConstraintCols = other427.uniqueConstraintCols; return *this; } void AddUniqueConstraintRequest::printTo(std::ostream& out) const { @@ -10455,14 +10513,14 @@ uint32_t AddNotNullConstraintRequest::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->notNullConstraintCols.clear(); - uint32_t _size427; - ::apache::thrift::protocol::TType _etype430; - xfer += iprot->readListBegin(_etype430, _size427); - this->notNullConstraintCols.resize(_size427); - uint32_t _i431; - for (_i431 = 0; _i431 < _size427; ++_i431) + uint32_t _size428; + ::apache::thrift::protocol::TType _etype431; + xfer += iprot->readListBegin(_etype431, _size428); + this->notNullConstraintCols.resize(_size428); + uint32_t _i432; + for (_i432 = 0; _i432 < _size428; ++_i432) { - xfer += this->notNullConstraintCols[_i431].read(iprot); + xfer += this->notNullConstraintCols[_i432].read(iprot); } xfer += iprot->readListEnd(); } @@ -10493,10 +10551,10 @@ uint32_t AddNotNullConstraintRequest::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("notNullConstraintCols", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->notNullConstraintCols.size())); - std::vector ::const_iterator _iter432; - for (_iter432 = this->notNullConstraintCols.begin(); _iter432 != this->notNullConstraintCols.end(); ++_iter432) + std::vector ::const_iterator _iter433; + for (_iter433 = this->notNullConstraintCols.begin(); _iter433 != this->notNullConstraintCols.end(); ++_iter433) { - xfer += (*_iter432).write(oprot); + xfer += (*_iter433).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10512,11 +10570,11 @@ void swap(AddNotNullConstraintRequest &a, AddNotNullConstraintRequest &b) { swap(a.notNullConstraintCols, b.notNullConstraintCols); } -AddNotNullConstraintRequest::AddNotNullConstraintRequest(const AddNotNullConstraintRequest& other433) { - notNullConstraintCols = other433.notNullConstraintCols; -} -AddNotNullConstraintRequest& AddNotNullConstraintRequest::operator=(const AddNotNullConstraintRequest& other434) { +AddNotNullConstraintRequest::AddNotNullConstraintRequest(const AddNotNullConstraintRequest& other434) { notNullConstraintCols = other434.notNullConstraintCols; +} +AddNotNullConstraintRequest& AddNotNullConstraintRequest::operator=(const AddNotNullConstraintRequest& other435) { + notNullConstraintCols = other435.notNullConstraintCols; return *this; } void AddNotNullConstraintRequest::printTo(std::ostream& out) const { @@ -10566,14 +10624,14 @@ uint32_t PartitionsByExprResult::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size435; - ::apache::thrift::protocol::TType _etype438; - xfer += iprot->readListBegin(_etype438, _size435); - this->partitions.resize(_size435); - uint32_t _i439; - for (_i439 = 0; _i439 < _size435; ++_i439) + uint32_t _size436; + ::apache::thrift::protocol::TType _etype439; + xfer += iprot->readListBegin(_etype439, _size436); + this->partitions.resize(_size436); + uint32_t _i440; + for (_i440 = 0; _i440 < _size436; ++_i440) { - xfer += this->partitions[_i439].read(iprot); + xfer += this->partitions[_i440].read(iprot); } xfer += iprot->readListEnd(); } @@ -10614,10 +10672,10 @@ uint32_t PartitionsByExprResult::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter440; - for (_iter440 = this->partitions.begin(); _iter440 != this->partitions.end(); ++_iter440) + std::vector ::const_iterator _iter441; + for (_iter441 = this->partitions.begin(); _iter441 != this->partitions.end(); ++_iter441) { - xfer += (*_iter440).write(oprot); + xfer += (*_iter441).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10638,13 +10696,13 @@ void swap(PartitionsByExprResult &a, PartitionsByExprResult &b) { swap(a.hasUnknownPartitions, b.hasUnknownPartitions); } -PartitionsByExprResult::PartitionsByExprResult(const PartitionsByExprResult& other441) { - partitions = other441.partitions; - hasUnknownPartitions = other441.hasUnknownPartitions; -} -PartitionsByExprResult& PartitionsByExprResult::operator=(const PartitionsByExprResult& other442) { +PartitionsByExprResult::PartitionsByExprResult(const PartitionsByExprResult& other442) { partitions = other442.partitions; hasUnknownPartitions = other442.hasUnknownPartitions; +} +PartitionsByExprResult& PartitionsByExprResult::operator=(const PartitionsByExprResult& other443) { + partitions = other443.partitions; + hasUnknownPartitions = other443.hasUnknownPartitions; return *this; } void PartitionsByExprResult::printTo(std::ostream& out) const { @@ -10806,21 +10864,21 @@ void swap(PartitionsByExprRequest &a, PartitionsByExprRequest &b) { swap(a.__isset, b.__isset); } -PartitionsByExprRequest::PartitionsByExprRequest(const PartitionsByExprRequest& other443) { - dbName = other443.dbName; - tblName = other443.tblName; - expr = other443.expr; - defaultPartitionName = other443.defaultPartitionName; - maxParts = other443.maxParts; - __isset = other443.__isset; -} -PartitionsByExprRequest& PartitionsByExprRequest::operator=(const PartitionsByExprRequest& other444) { +PartitionsByExprRequest::PartitionsByExprRequest(const PartitionsByExprRequest& other444) { dbName = other444.dbName; tblName = other444.tblName; expr = other444.expr; defaultPartitionName = other444.defaultPartitionName; maxParts = other444.maxParts; __isset = other444.__isset; +} +PartitionsByExprRequest& PartitionsByExprRequest::operator=(const PartitionsByExprRequest& other445) { + dbName = other445.dbName; + tblName = other445.tblName; + expr = other445.expr; + defaultPartitionName = other445.defaultPartitionName; + maxParts = other445.maxParts; + __isset = other445.__isset; return *this; } void PartitionsByExprRequest::printTo(std::ostream& out) const { @@ -10869,14 +10927,14 @@ uint32_t TableStatsResult::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tableStats.clear(); - uint32_t _size445; - ::apache::thrift::protocol::TType _etype448; - xfer += iprot->readListBegin(_etype448, _size445); - this->tableStats.resize(_size445); - uint32_t _i449; - for (_i449 = 0; _i449 < _size445; ++_i449) + uint32_t _size446; + ::apache::thrift::protocol::TType _etype449; + xfer += iprot->readListBegin(_etype449, _size446); + this->tableStats.resize(_size446); + uint32_t _i450; + for (_i450 = 0; _i450 < _size446; ++_i450) { - xfer += this->tableStats[_i449].read(iprot); + xfer += this->tableStats[_i450].read(iprot); } xfer += iprot->readListEnd(); } @@ -10907,10 +10965,10 @@ uint32_t TableStatsResult::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tableStats", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->tableStats.size())); - std::vector ::const_iterator _iter450; - for (_iter450 = this->tableStats.begin(); _iter450 != this->tableStats.end(); ++_iter450) + std::vector ::const_iterator _iter451; + for (_iter451 = this->tableStats.begin(); _iter451 != this->tableStats.end(); ++_iter451) { - xfer += (*_iter450).write(oprot); + xfer += (*_iter451).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10926,11 +10984,11 @@ void swap(TableStatsResult &a, TableStatsResult &b) { swap(a.tableStats, b.tableStats); } -TableStatsResult::TableStatsResult(const TableStatsResult& other451) { - tableStats = other451.tableStats; -} -TableStatsResult& TableStatsResult::operator=(const TableStatsResult& other452) { +TableStatsResult::TableStatsResult(const TableStatsResult& other452) { tableStats = other452.tableStats; +} +TableStatsResult& TableStatsResult::operator=(const TableStatsResult& other453) { + tableStats = other453.tableStats; return *this; } void TableStatsResult::printTo(std::ostream& out) const { @@ -10975,26 +11033,26 @@ uint32_t PartitionsStatsResult::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partStats.clear(); - uint32_t _size453; - ::apache::thrift::protocol::TType _ktype454; - ::apache::thrift::protocol::TType _vtype455; - xfer += iprot->readMapBegin(_ktype454, _vtype455, _size453); - uint32_t _i457; - for (_i457 = 0; _i457 < _size453; ++_i457) + uint32_t _size454; + ::apache::thrift::protocol::TType _ktype455; + ::apache::thrift::protocol::TType _vtype456; + xfer += iprot->readMapBegin(_ktype455, _vtype456, _size454); + uint32_t _i458; + for (_i458 = 0; _i458 < _size454; ++_i458) { - std::string _key458; - xfer += iprot->readString(_key458); - std::vector & _val459 = this->partStats[_key458]; + std::string _key459; + xfer += iprot->readString(_key459); + std::vector & _val460 = this->partStats[_key459]; { - _val459.clear(); - uint32_t _size460; - ::apache::thrift::protocol::TType _etype463; - xfer += iprot->readListBegin(_etype463, _size460); - _val459.resize(_size460); - uint32_t _i464; - for (_i464 = 0; _i464 < _size460; ++_i464) + _val460.clear(); + uint32_t _size461; + ::apache::thrift::protocol::TType _etype464; + xfer += iprot->readListBegin(_etype464, _size461); + _val460.resize(_size461); + uint32_t _i465; + for (_i465 = 0; _i465 < _size461; ++_i465) { - xfer += _val459[_i464].read(iprot); + xfer += _val460[_i465].read(iprot); } xfer += iprot->readListEnd(); } @@ -11028,16 +11086,16 @@ uint32_t PartitionsStatsResult::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("partStats", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_LIST, static_cast(this->partStats.size())); - std::map > ::const_iterator _iter465; - for (_iter465 = this->partStats.begin(); _iter465 != this->partStats.end(); ++_iter465) + std::map > ::const_iterator _iter466; + for (_iter466 = this->partStats.begin(); _iter466 != this->partStats.end(); ++_iter466) { - xfer += oprot->writeString(_iter465->first); + xfer += oprot->writeString(_iter466->first); { - xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(_iter465->second.size())); - std::vector ::const_iterator _iter466; - for (_iter466 = _iter465->second.begin(); _iter466 != _iter465->second.end(); ++_iter466) + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(_iter466->second.size())); + std::vector ::const_iterator _iter467; + for (_iter467 = _iter466->second.begin(); _iter467 != _iter466->second.end(); ++_iter467) { - xfer += (*_iter466).write(oprot); + xfer += (*_iter467).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11056,11 +11114,11 @@ void swap(PartitionsStatsResult &a, PartitionsStatsResult &b) { swap(a.partStats, b.partStats); } -PartitionsStatsResult::PartitionsStatsResult(const PartitionsStatsResult& other467) { - partStats = other467.partStats; -} -PartitionsStatsResult& PartitionsStatsResult::operator=(const PartitionsStatsResult& other468) { +PartitionsStatsResult::PartitionsStatsResult(const PartitionsStatsResult& other468) { partStats = other468.partStats; +} +PartitionsStatsResult& PartitionsStatsResult::operator=(const PartitionsStatsResult& other469) { + partStats = other469.partStats; return *this; } void PartitionsStatsResult::printTo(std::ostream& out) const { @@ -11131,14 +11189,14 @@ uint32_t TableStatsRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->colNames.clear(); - uint32_t _size469; - ::apache::thrift::protocol::TType _etype472; - xfer += iprot->readListBegin(_etype472, _size469); - this->colNames.resize(_size469); - uint32_t _i473; - for (_i473 = 0; _i473 < _size469; ++_i473) + uint32_t _size470; + ::apache::thrift::protocol::TType _etype473; + xfer += iprot->readListBegin(_etype473, _size470); + this->colNames.resize(_size470); + uint32_t _i474; + for (_i474 = 0; _i474 < _size470; ++_i474) { - xfer += iprot->readString(this->colNames[_i473]); + xfer += iprot->readString(this->colNames[_i474]); } xfer += iprot->readListEnd(); } @@ -11181,10 +11239,10 @@ uint32_t TableStatsRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("colNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->colNames.size())); - std::vector ::const_iterator _iter474; - for (_iter474 = this->colNames.begin(); _iter474 != this->colNames.end(); ++_iter474) + std::vector ::const_iterator _iter475; + for (_iter475 = this->colNames.begin(); _iter475 != this->colNames.end(); ++_iter475) { - xfer += oprot->writeString((*_iter474)); + xfer += oprot->writeString((*_iter475)); } xfer += oprot->writeListEnd(); } @@ -11202,15 +11260,15 @@ void swap(TableStatsRequest &a, TableStatsRequest &b) { swap(a.colNames, b.colNames); } -TableStatsRequest::TableStatsRequest(const TableStatsRequest& other475) { - dbName = other475.dbName; - tblName = other475.tblName; - colNames = other475.colNames; -} -TableStatsRequest& TableStatsRequest::operator=(const TableStatsRequest& other476) { +TableStatsRequest::TableStatsRequest(const TableStatsRequest& other476) { dbName = other476.dbName; tblName = other476.tblName; colNames = other476.colNames; +} +TableStatsRequest& TableStatsRequest::operator=(const TableStatsRequest& other477) { + dbName = other477.dbName; + tblName = other477.tblName; + colNames = other477.colNames; return *this; } void TableStatsRequest::printTo(std::ostream& out) const { @@ -11288,14 +11346,14 @@ uint32_t PartitionsStatsRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->colNames.clear(); - uint32_t _size477; - ::apache::thrift::protocol::TType _etype480; - xfer += iprot->readListBegin(_etype480, _size477); - this->colNames.resize(_size477); - uint32_t _i481; - for (_i481 = 0; _i481 < _size477; ++_i481) + uint32_t _size478; + ::apache::thrift::protocol::TType _etype481; + xfer += iprot->readListBegin(_etype481, _size478); + this->colNames.resize(_size478); + uint32_t _i482; + for (_i482 = 0; _i482 < _size478; ++_i482) { - xfer += iprot->readString(this->colNames[_i481]); + xfer += iprot->readString(this->colNames[_i482]); } xfer += iprot->readListEnd(); } @@ -11308,14 +11366,14 @@ uint32_t PartitionsStatsRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partNames.clear(); - uint32_t _size482; - ::apache::thrift::protocol::TType _etype485; - xfer += iprot->readListBegin(_etype485, _size482); - this->partNames.resize(_size482); - uint32_t _i486; - for (_i486 = 0; _i486 < _size482; ++_i486) + uint32_t _size483; + ::apache::thrift::protocol::TType _etype486; + xfer += iprot->readListBegin(_etype486, _size483); + this->partNames.resize(_size483); + uint32_t _i487; + for (_i487 = 0; _i487 < _size483; ++_i487) { - xfer += iprot->readString(this->partNames[_i486]); + xfer += iprot->readString(this->partNames[_i487]); } xfer += iprot->readListEnd(); } @@ -11360,10 +11418,10 @@ uint32_t PartitionsStatsRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("colNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->colNames.size())); - std::vector ::const_iterator _iter487; - for (_iter487 = this->colNames.begin(); _iter487 != this->colNames.end(); ++_iter487) + std::vector ::const_iterator _iter488; + for (_iter488 = this->colNames.begin(); _iter488 != this->colNames.end(); ++_iter488) { - xfer += oprot->writeString((*_iter487)); + xfer += oprot->writeString((*_iter488)); } xfer += oprot->writeListEnd(); } @@ -11372,10 +11430,10 @@ uint32_t PartitionsStatsRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partNames.size())); - std::vector ::const_iterator _iter488; - for (_iter488 = this->partNames.begin(); _iter488 != this->partNames.end(); ++_iter488) + std::vector ::const_iterator _iter489; + for (_iter489 = this->partNames.begin(); _iter489 != this->partNames.end(); ++_iter489) { - xfer += oprot->writeString((*_iter488)); + xfer += oprot->writeString((*_iter489)); } xfer += oprot->writeListEnd(); } @@ -11394,17 +11452,17 @@ void swap(PartitionsStatsRequest &a, PartitionsStatsRequest &b) { swap(a.partNames, b.partNames); } -PartitionsStatsRequest::PartitionsStatsRequest(const PartitionsStatsRequest& other489) { - dbName = other489.dbName; - tblName = other489.tblName; - colNames = other489.colNames; - partNames = other489.partNames; -} -PartitionsStatsRequest& PartitionsStatsRequest::operator=(const PartitionsStatsRequest& other490) { +PartitionsStatsRequest::PartitionsStatsRequest(const PartitionsStatsRequest& other490) { dbName = other490.dbName; tblName = other490.tblName; colNames = other490.colNames; partNames = other490.partNames; +} +PartitionsStatsRequest& PartitionsStatsRequest::operator=(const PartitionsStatsRequest& other491) { + dbName = other491.dbName; + tblName = other491.tblName; + colNames = other491.colNames; + partNames = other491.partNames; return *this; } void PartitionsStatsRequest::printTo(std::ostream& out) const { @@ -11452,14 +11510,14 @@ uint32_t AddPartitionsResult::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size491; - ::apache::thrift::protocol::TType _etype494; - xfer += iprot->readListBegin(_etype494, _size491); - this->partitions.resize(_size491); - uint32_t _i495; - for (_i495 = 0; _i495 < _size491; ++_i495) + uint32_t _size492; + ::apache::thrift::protocol::TType _etype495; + xfer += iprot->readListBegin(_etype495, _size492); + this->partitions.resize(_size492); + uint32_t _i496; + for (_i496 = 0; _i496 < _size492; ++_i496) { - xfer += this->partitions[_i495].read(iprot); + xfer += this->partitions[_i496].read(iprot); } xfer += iprot->readListEnd(); } @@ -11489,10 +11547,10 @@ uint32_t AddPartitionsResult::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter496; - for (_iter496 = this->partitions.begin(); _iter496 != this->partitions.end(); ++_iter496) + std::vector ::const_iterator _iter497; + for (_iter497 = this->partitions.begin(); _iter497 != this->partitions.end(); ++_iter497) { - xfer += (*_iter496).write(oprot); + xfer += (*_iter497).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11509,13 +11567,13 @@ void swap(AddPartitionsResult &a, AddPartitionsResult &b) { swap(a.__isset, b.__isset); } -AddPartitionsResult::AddPartitionsResult(const AddPartitionsResult& other497) { - partitions = other497.partitions; - __isset = other497.__isset; -} -AddPartitionsResult& AddPartitionsResult::operator=(const AddPartitionsResult& other498) { +AddPartitionsResult::AddPartitionsResult(const AddPartitionsResult& other498) { partitions = other498.partitions; __isset = other498.__isset; +} +AddPartitionsResult& AddPartitionsResult::operator=(const AddPartitionsResult& other499) { + partitions = other499.partitions; + __isset = other499.__isset; return *this; } void AddPartitionsResult::printTo(std::ostream& out) const { @@ -11596,14 +11654,14 @@ uint32_t AddPartitionsRequest::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->parts.clear(); - uint32_t _size499; - ::apache::thrift::protocol::TType _etype502; - xfer += iprot->readListBegin(_etype502, _size499); - this->parts.resize(_size499); - uint32_t _i503; - for (_i503 = 0; _i503 < _size499; ++_i503) + uint32_t _size500; + ::apache::thrift::protocol::TType _etype503; + xfer += iprot->readListBegin(_etype503, _size500); + this->parts.resize(_size500); + uint32_t _i504; + for (_i504 = 0; _i504 < _size500; ++_i504) { - xfer += this->parts[_i503].read(iprot); + xfer += this->parts[_i504].read(iprot); } xfer += iprot->readListEnd(); } @@ -11664,10 +11722,10 @@ uint32_t AddPartitionsRequest::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->parts.size())); - std::vector ::const_iterator _iter504; - for (_iter504 = this->parts.begin(); _iter504 != this->parts.end(); ++_iter504) + std::vector ::const_iterator _iter505; + for (_iter505 = this->parts.begin(); _iter505 != this->parts.end(); ++_iter505) { - xfer += (*_iter504).write(oprot); + xfer += (*_iter505).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11697,21 +11755,21 @@ void swap(AddPartitionsRequest &a, AddPartitionsRequest &b) { swap(a.__isset, b.__isset); } -AddPartitionsRequest::AddPartitionsRequest(const AddPartitionsRequest& other505) { - dbName = other505.dbName; - tblName = other505.tblName; - parts = other505.parts; - ifNotExists = other505.ifNotExists; - needResult = other505.needResult; - __isset = other505.__isset; -} -AddPartitionsRequest& AddPartitionsRequest::operator=(const AddPartitionsRequest& other506) { +AddPartitionsRequest::AddPartitionsRequest(const AddPartitionsRequest& other506) { dbName = other506.dbName; tblName = other506.tblName; parts = other506.parts; ifNotExists = other506.ifNotExists; needResult = other506.needResult; __isset = other506.__isset; +} +AddPartitionsRequest& AddPartitionsRequest::operator=(const AddPartitionsRequest& other507) { + dbName = other507.dbName; + tblName = other507.tblName; + parts = other507.parts; + ifNotExists = other507.ifNotExists; + needResult = other507.needResult; + __isset = other507.__isset; return *this; } void AddPartitionsRequest::printTo(std::ostream& out) const { @@ -11760,14 +11818,14 @@ uint32_t DropPartitionsResult::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitions.clear(); - uint32_t _size507; - ::apache::thrift::protocol::TType _etype510; - xfer += iprot->readListBegin(_etype510, _size507); - this->partitions.resize(_size507); - uint32_t _i511; - for (_i511 = 0; _i511 < _size507; ++_i511) + uint32_t _size508; + ::apache::thrift::protocol::TType _etype511; + xfer += iprot->readListBegin(_etype511, _size508); + this->partitions.resize(_size508); + uint32_t _i512; + for (_i512 = 0; _i512 < _size508; ++_i512) { - xfer += this->partitions[_i511].read(iprot); + xfer += this->partitions[_i512].read(iprot); } xfer += iprot->readListEnd(); } @@ -11797,10 +11855,10 @@ uint32_t DropPartitionsResult::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("partitions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitions.size())); - std::vector ::const_iterator _iter512; - for (_iter512 = this->partitions.begin(); _iter512 != this->partitions.end(); ++_iter512) + std::vector ::const_iterator _iter513; + for (_iter513 = this->partitions.begin(); _iter513 != this->partitions.end(); ++_iter513) { - xfer += (*_iter512).write(oprot); + xfer += (*_iter513).write(oprot); } xfer += oprot->writeListEnd(); } @@ -11817,13 +11875,13 @@ void swap(DropPartitionsResult &a, DropPartitionsResult &b) { swap(a.__isset, b.__isset); } -DropPartitionsResult::DropPartitionsResult(const DropPartitionsResult& other513) { - partitions = other513.partitions; - __isset = other513.__isset; -} -DropPartitionsResult& DropPartitionsResult::operator=(const DropPartitionsResult& other514) { +DropPartitionsResult::DropPartitionsResult(const DropPartitionsResult& other514) { partitions = other514.partitions; __isset = other514.__isset; +} +DropPartitionsResult& DropPartitionsResult::operator=(const DropPartitionsResult& other515) { + partitions = other515.partitions; + __isset = other515.__isset; return *this; } void DropPartitionsResult::printTo(std::ostream& out) const { @@ -11925,15 +11983,15 @@ void swap(DropPartitionsExpr &a, DropPartitionsExpr &b) { swap(a.__isset, b.__isset); } -DropPartitionsExpr::DropPartitionsExpr(const DropPartitionsExpr& other515) { - expr = other515.expr; - partArchiveLevel = other515.partArchiveLevel; - __isset = other515.__isset; -} -DropPartitionsExpr& DropPartitionsExpr::operator=(const DropPartitionsExpr& other516) { +DropPartitionsExpr::DropPartitionsExpr(const DropPartitionsExpr& other516) { expr = other516.expr; partArchiveLevel = other516.partArchiveLevel; __isset = other516.__isset; +} +DropPartitionsExpr& DropPartitionsExpr::operator=(const DropPartitionsExpr& other517) { + expr = other517.expr; + partArchiveLevel = other517.partArchiveLevel; + __isset = other517.__isset; return *this; } void DropPartitionsExpr::printTo(std::ostream& out) const { @@ -11982,14 +12040,14 @@ uint32_t RequestPartsSpec::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->names.clear(); - uint32_t _size517; - ::apache::thrift::protocol::TType _etype520; - xfer += iprot->readListBegin(_etype520, _size517); - this->names.resize(_size517); - uint32_t _i521; - for (_i521 = 0; _i521 < _size517; ++_i521) + uint32_t _size518; + ::apache::thrift::protocol::TType _etype521; + xfer += iprot->readListBegin(_etype521, _size518); + this->names.resize(_size518); + uint32_t _i522; + for (_i522 = 0; _i522 < _size518; ++_i522) { - xfer += iprot->readString(this->names[_i521]); + xfer += iprot->readString(this->names[_i522]); } xfer += iprot->readListEnd(); } @@ -12002,14 +12060,14 @@ uint32_t RequestPartsSpec::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->exprs.clear(); - uint32_t _size522; - ::apache::thrift::protocol::TType _etype525; - xfer += iprot->readListBegin(_etype525, _size522); - this->exprs.resize(_size522); - uint32_t _i526; - for (_i526 = 0; _i526 < _size522; ++_i526) + uint32_t _size523; + ::apache::thrift::protocol::TType _etype526; + xfer += iprot->readListBegin(_etype526, _size523); + this->exprs.resize(_size523); + uint32_t _i527; + for (_i527 = 0; _i527 < _size523; ++_i527) { - xfer += this->exprs[_i526].read(iprot); + xfer += this->exprs[_i527].read(iprot); } xfer += iprot->readListEnd(); } @@ -12038,10 +12096,10 @@ uint32_t RequestPartsSpec::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->names.size())); - std::vector ::const_iterator _iter527; - for (_iter527 = this->names.begin(); _iter527 != this->names.end(); ++_iter527) + std::vector ::const_iterator _iter528; + for (_iter528 = this->names.begin(); _iter528 != this->names.end(); ++_iter528) { - xfer += oprot->writeString((*_iter527)); + xfer += oprot->writeString((*_iter528)); } xfer += oprot->writeListEnd(); } @@ -12050,10 +12108,10 @@ uint32_t RequestPartsSpec::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("exprs", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->exprs.size())); - std::vector ::const_iterator _iter528; - for (_iter528 = this->exprs.begin(); _iter528 != this->exprs.end(); ++_iter528) + std::vector ::const_iterator _iter529; + for (_iter529 = this->exprs.begin(); _iter529 != this->exprs.end(); ++_iter529) { - xfer += (*_iter528).write(oprot); + xfer += (*_iter529).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12071,15 +12129,15 @@ void swap(RequestPartsSpec &a, RequestPartsSpec &b) { swap(a.__isset, b.__isset); } -RequestPartsSpec::RequestPartsSpec(const RequestPartsSpec& other529) { - names = other529.names; - exprs = other529.exprs; - __isset = other529.__isset; -} -RequestPartsSpec& RequestPartsSpec::operator=(const RequestPartsSpec& other530) { +RequestPartsSpec::RequestPartsSpec(const RequestPartsSpec& other530) { names = other530.names; exprs = other530.exprs; __isset = other530.__isset; +} +RequestPartsSpec& RequestPartsSpec::operator=(const RequestPartsSpec& other531) { + names = other531.names; + exprs = other531.exprs; + __isset = other531.__isset; return *this; } void RequestPartsSpec::printTo(std::ostream& out) const { @@ -12298,18 +12356,7 @@ void swap(DropPartitionsRequest &a, DropPartitionsRequest &b) { swap(a.__isset, b.__isset); } -DropPartitionsRequest::DropPartitionsRequest(const DropPartitionsRequest& other531) { - dbName = other531.dbName; - tblName = other531.tblName; - parts = other531.parts; - deleteData = other531.deleteData; - ifExists = other531.ifExists; - ignoreProtection = other531.ignoreProtection; - environmentContext = other531.environmentContext; - needResult = other531.needResult; - __isset = other531.__isset; -} -DropPartitionsRequest& DropPartitionsRequest::operator=(const DropPartitionsRequest& other532) { +DropPartitionsRequest::DropPartitionsRequest(const DropPartitionsRequest& other532) { dbName = other532.dbName; tblName = other532.tblName; parts = other532.parts; @@ -12319,6 +12366,17 @@ DropPartitionsRequest& DropPartitionsRequest::operator=(const DropPartitionsRequ environmentContext = other532.environmentContext; needResult = other532.needResult; __isset = other532.__isset; +} +DropPartitionsRequest& DropPartitionsRequest::operator=(const DropPartitionsRequest& other533) { + dbName = other533.dbName; + tblName = other533.tblName; + parts = other533.parts; + deleteData = other533.deleteData; + ifExists = other533.ifExists; + ignoreProtection = other533.ignoreProtection; + environmentContext = other533.environmentContext; + needResult = other533.needResult; + __isset = other533.__isset; return *this; } void DropPartitionsRequest::printTo(std::ostream& out) const { @@ -12421,14 +12479,14 @@ uint32_t PartitionValuesRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionKeys.clear(); - uint32_t _size533; - ::apache::thrift::protocol::TType _etype536; - xfer += iprot->readListBegin(_etype536, _size533); - this->partitionKeys.resize(_size533); - uint32_t _i537; - for (_i537 = 0; _i537 < _size533; ++_i537) + uint32_t _size534; + ::apache::thrift::protocol::TType _etype537; + xfer += iprot->readListBegin(_etype537, _size534); + this->partitionKeys.resize(_size534); + uint32_t _i538; + for (_i538 = 0; _i538 < _size534; ++_i538) { - xfer += this->partitionKeys[_i537].read(iprot); + xfer += this->partitionKeys[_i538].read(iprot); } xfer += iprot->readListEnd(); } @@ -12457,14 +12515,14 @@ uint32_t PartitionValuesRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionOrder.clear(); - uint32_t _size538; - ::apache::thrift::protocol::TType _etype541; - xfer += iprot->readListBegin(_etype541, _size538); - this->partitionOrder.resize(_size538); - uint32_t _i542; - for (_i542 = 0; _i542 < _size538; ++_i542) + uint32_t _size539; + ::apache::thrift::protocol::TType _etype542; + xfer += iprot->readListBegin(_etype542, _size539); + this->partitionOrder.resize(_size539); + uint32_t _i543; + for (_i543 = 0; _i543 < _size539; ++_i543) { - xfer += this->partitionOrder[_i542].read(iprot); + xfer += this->partitionOrder[_i543].read(iprot); } xfer += iprot->readListEnd(); } @@ -12523,10 +12581,10 @@ uint32_t PartitionValuesRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partitionKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitionKeys.size())); - std::vector ::const_iterator _iter543; - for (_iter543 = this->partitionKeys.begin(); _iter543 != this->partitionKeys.end(); ++_iter543) + std::vector ::const_iterator _iter544; + for (_iter544 = this->partitionKeys.begin(); _iter544 != this->partitionKeys.end(); ++_iter544) { - xfer += (*_iter543).write(oprot); + xfer += (*_iter544).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12546,10 +12604,10 @@ uint32_t PartitionValuesRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("partitionOrder", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitionOrder.size())); - std::vector ::const_iterator _iter544; - for (_iter544 = this->partitionOrder.begin(); _iter544 != this->partitionOrder.end(); ++_iter544) + std::vector ::const_iterator _iter545; + for (_iter545 = this->partitionOrder.begin(); _iter545 != this->partitionOrder.end(); ++_iter545) { - xfer += (*_iter544).write(oprot); + xfer += (*_iter545).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12583,18 +12641,7 @@ void swap(PartitionValuesRequest &a, PartitionValuesRequest &b) { swap(a.__isset, b.__isset); } -PartitionValuesRequest::PartitionValuesRequest(const PartitionValuesRequest& other545) { - dbName = other545.dbName; - tblName = other545.tblName; - partitionKeys = other545.partitionKeys; - applyDistinct = other545.applyDistinct; - filter = other545.filter; - partitionOrder = other545.partitionOrder; - ascending = other545.ascending; - maxParts = other545.maxParts; - __isset = other545.__isset; -} -PartitionValuesRequest& PartitionValuesRequest::operator=(const PartitionValuesRequest& other546) { +PartitionValuesRequest::PartitionValuesRequest(const PartitionValuesRequest& other546) { dbName = other546.dbName; tblName = other546.tblName; partitionKeys = other546.partitionKeys; @@ -12604,6 +12651,17 @@ PartitionValuesRequest& PartitionValuesRequest::operator=(const PartitionValuesR ascending = other546.ascending; maxParts = other546.maxParts; __isset = other546.__isset; +} +PartitionValuesRequest& PartitionValuesRequest::operator=(const PartitionValuesRequest& other547) { + dbName = other547.dbName; + tblName = other547.tblName; + partitionKeys = other547.partitionKeys; + applyDistinct = other547.applyDistinct; + filter = other547.filter; + partitionOrder = other547.partitionOrder; + ascending = other547.ascending; + maxParts = other547.maxParts; + __isset = other547.__isset; return *this; } void PartitionValuesRequest::printTo(std::ostream& out) const { @@ -12655,14 +12713,14 @@ uint32_t PartitionValuesRow::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->row.clear(); - uint32_t _size547; - ::apache::thrift::protocol::TType _etype550; - xfer += iprot->readListBegin(_etype550, _size547); - this->row.resize(_size547); - uint32_t _i551; - for (_i551 = 0; _i551 < _size547; ++_i551) + uint32_t _size548; + ::apache::thrift::protocol::TType _etype551; + xfer += iprot->readListBegin(_etype551, _size548); + this->row.resize(_size548); + uint32_t _i552; + for (_i552 = 0; _i552 < _size548; ++_i552) { - xfer += iprot->readString(this->row[_i551]); + xfer += iprot->readString(this->row[_i552]); } xfer += iprot->readListEnd(); } @@ -12693,10 +12751,10 @@ uint32_t PartitionValuesRow::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("row", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->row.size())); - std::vector ::const_iterator _iter552; - for (_iter552 = this->row.begin(); _iter552 != this->row.end(); ++_iter552) + std::vector ::const_iterator _iter553; + for (_iter553 = this->row.begin(); _iter553 != this->row.end(); ++_iter553) { - xfer += oprot->writeString((*_iter552)); + xfer += oprot->writeString((*_iter553)); } xfer += oprot->writeListEnd(); } @@ -12712,11 +12770,11 @@ void swap(PartitionValuesRow &a, PartitionValuesRow &b) { swap(a.row, b.row); } -PartitionValuesRow::PartitionValuesRow(const PartitionValuesRow& other553) { - row = other553.row; -} -PartitionValuesRow& PartitionValuesRow::operator=(const PartitionValuesRow& other554) { +PartitionValuesRow::PartitionValuesRow(const PartitionValuesRow& other554) { row = other554.row; +} +PartitionValuesRow& PartitionValuesRow::operator=(const PartitionValuesRow& other555) { + row = other555.row; return *this; } void PartitionValuesRow::printTo(std::ostream& out) const { @@ -12761,14 +12819,14 @@ uint32_t PartitionValuesResponse::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionValues.clear(); - uint32_t _size555; - ::apache::thrift::protocol::TType _etype558; - xfer += iprot->readListBegin(_etype558, _size555); - this->partitionValues.resize(_size555); - uint32_t _i559; - for (_i559 = 0; _i559 < _size555; ++_i559) + uint32_t _size556; + ::apache::thrift::protocol::TType _etype559; + xfer += iprot->readListBegin(_etype559, _size556); + this->partitionValues.resize(_size556); + uint32_t _i560; + for (_i560 = 0; _i560 < _size556; ++_i560) { - xfer += this->partitionValues[_i559].read(iprot); + xfer += this->partitionValues[_i560].read(iprot); } xfer += iprot->readListEnd(); } @@ -12799,10 +12857,10 @@ uint32_t PartitionValuesResponse::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("partitionValues", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->partitionValues.size())); - std::vector ::const_iterator _iter560; - for (_iter560 = this->partitionValues.begin(); _iter560 != this->partitionValues.end(); ++_iter560) + std::vector ::const_iterator _iter561; + for (_iter561 = this->partitionValues.begin(); _iter561 != this->partitionValues.end(); ++_iter561) { - xfer += (*_iter560).write(oprot); + xfer += (*_iter561).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12818,11 +12876,11 @@ void swap(PartitionValuesResponse &a, PartitionValuesResponse &b) { swap(a.partitionValues, b.partitionValues); } -PartitionValuesResponse::PartitionValuesResponse(const PartitionValuesResponse& other561) { - partitionValues = other561.partitionValues; -} -PartitionValuesResponse& PartitionValuesResponse::operator=(const PartitionValuesResponse& other562) { +PartitionValuesResponse::PartitionValuesResponse(const PartitionValuesResponse& other562) { partitionValues = other562.partitionValues; +} +PartitionValuesResponse& PartitionValuesResponse::operator=(const PartitionValuesResponse& other563) { + partitionValues = other563.partitionValues; return *this; } void PartitionValuesResponse::printTo(std::ostream& out) const { @@ -12868,9 +12926,9 @@ uint32_t ResourceUri::read(::apache::thrift::protocol::TProtocol* iprot) { { case 1: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast563; - xfer += iprot->readI32(ecast563); - this->resourceType = (ResourceType::type)ecast563; + int32_t ecast564; + xfer += iprot->readI32(ecast564); + this->resourceType = (ResourceType::type)ecast564; this->__isset.resourceType = true; } else { xfer += iprot->skip(ftype); @@ -12921,15 +12979,15 @@ void swap(ResourceUri &a, ResourceUri &b) { swap(a.__isset, b.__isset); } -ResourceUri::ResourceUri(const ResourceUri& other564) { - resourceType = other564.resourceType; - uri = other564.uri; - __isset = other564.__isset; -} -ResourceUri& ResourceUri::operator=(const ResourceUri& other565) { +ResourceUri::ResourceUri(const ResourceUri& other565) { resourceType = other565.resourceType; uri = other565.uri; __isset = other565.__isset; +} +ResourceUri& ResourceUri::operator=(const ResourceUri& other566) { + resourceType = other566.resourceType; + uri = other566.uri; + __isset = other566.__isset; return *this; } void ResourceUri::printTo(std::ostream& out) const { @@ -13032,9 +13090,9 @@ uint32_t Function::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast566; - xfer += iprot->readI32(ecast566); - this->ownerType = (PrincipalType::type)ecast566; + int32_t ecast567; + xfer += iprot->readI32(ecast567); + this->ownerType = (PrincipalType::type)ecast567; this->__isset.ownerType = true; } else { xfer += iprot->skip(ftype); @@ -13050,9 +13108,9 @@ uint32_t Function::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 7: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast567; - xfer += iprot->readI32(ecast567); - this->functionType = (FunctionType::type)ecast567; + int32_t ecast568; + xfer += iprot->readI32(ecast568); + this->functionType = (FunctionType::type)ecast568; this->__isset.functionType = true; } else { xfer += iprot->skip(ftype); @@ -13062,14 +13120,14 @@ uint32_t Function::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->resourceUris.clear(); - uint32_t _size568; - ::apache::thrift::protocol::TType _etype571; - xfer += iprot->readListBegin(_etype571, _size568); - this->resourceUris.resize(_size568); - uint32_t _i572; - for (_i572 = 0; _i572 < _size568; ++_i572) + uint32_t _size569; + ::apache::thrift::protocol::TType _etype572; + xfer += iprot->readListBegin(_etype572, _size569); + this->resourceUris.resize(_size569); + uint32_t _i573; + for (_i573 = 0; _i573 < _size569; ++_i573) { - xfer += this->resourceUris[_i572].read(iprot); + xfer += this->resourceUris[_i573].read(iprot); } xfer += iprot->readListEnd(); } @@ -13126,10 +13184,10 @@ uint32_t Function::write(::apache::thrift::protocol::TProtocol* oprot) const { xfer += oprot->writeFieldBegin("resourceUris", ::apache::thrift::protocol::T_LIST, 8); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->resourceUris.size())); - std::vector ::const_iterator _iter573; - for (_iter573 = this->resourceUris.begin(); _iter573 != this->resourceUris.end(); ++_iter573) + std::vector ::const_iterator _iter574; + for (_iter574 = this->resourceUris.begin(); _iter574 != this->resourceUris.end(); ++_iter574) { - xfer += (*_iter573).write(oprot); + xfer += (*_iter574).write(oprot); } xfer += oprot->writeListEnd(); } @@ -13153,18 +13211,7 @@ void swap(Function &a, Function &b) { swap(a.__isset, b.__isset); } -Function::Function(const Function& other574) { - functionName = other574.functionName; - dbName = other574.dbName; - className = other574.className; - ownerName = other574.ownerName; - ownerType = other574.ownerType; - createTime = other574.createTime; - functionType = other574.functionType; - resourceUris = other574.resourceUris; - __isset = other574.__isset; -} -Function& Function::operator=(const Function& other575) { +Function::Function(const Function& other575) { functionName = other575.functionName; dbName = other575.dbName; className = other575.className; @@ -13174,6 +13221,17 @@ Function& Function::operator=(const Function& other575) { functionType = other575.functionType; resourceUris = other575.resourceUris; __isset = other575.__isset; +} +Function& Function::operator=(const Function& other576) { + functionName = other576.functionName; + dbName = other576.dbName; + className = other576.className; + ownerName = other576.ownerName; + ownerType = other576.ownerType; + createTime = other576.createTime; + functionType = other576.functionType; + resourceUris = other576.resourceUris; + __isset = other576.__isset; return *this; } void Function::printTo(std::ostream& out) const { @@ -13271,9 +13329,9 @@ uint32_t TxnInfo::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast576; - xfer += iprot->readI32(ecast576); - this->state = (TxnState::type)ecast576; + int32_t ecast577; + xfer += iprot->readI32(ecast577); + this->state = (TxnState::type)ecast577; isset_state = true; } else { xfer += iprot->skip(ftype); @@ -13420,19 +13478,7 @@ void swap(TxnInfo &a, TxnInfo &b) { swap(a.__isset, b.__isset); } -TxnInfo::TxnInfo(const TxnInfo& other577) { - id = other577.id; - state = other577.state; - user = other577.user; - hostname = other577.hostname; - agentInfo = other577.agentInfo; - heartbeatCount = other577.heartbeatCount; - metaInfo = other577.metaInfo; - startedTime = other577.startedTime; - lastHeartbeatTime = other577.lastHeartbeatTime; - __isset = other577.__isset; -} -TxnInfo& TxnInfo::operator=(const TxnInfo& other578) { +TxnInfo::TxnInfo(const TxnInfo& other578) { id = other578.id; state = other578.state; user = other578.user; @@ -13443,6 +13489,18 @@ TxnInfo& TxnInfo::operator=(const TxnInfo& other578) { startedTime = other578.startedTime; lastHeartbeatTime = other578.lastHeartbeatTime; __isset = other578.__isset; +} +TxnInfo& TxnInfo::operator=(const TxnInfo& other579) { + id = other579.id; + state = other579.state; + user = other579.user; + hostname = other579.hostname; + agentInfo = other579.agentInfo; + heartbeatCount = other579.heartbeatCount; + metaInfo = other579.metaInfo; + startedTime = other579.startedTime; + lastHeartbeatTime = other579.lastHeartbeatTime; + __isset = other579.__isset; return *this; } void TxnInfo::printTo(std::ostream& out) const { @@ -13508,14 +13566,14 @@ uint32_t GetOpenTxnsInfoResponse::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->open_txns.clear(); - uint32_t _size579; - ::apache::thrift::protocol::TType _etype582; - xfer += iprot->readListBegin(_etype582, _size579); - this->open_txns.resize(_size579); - uint32_t _i583; - for (_i583 = 0; _i583 < _size579; ++_i583) + uint32_t _size580; + ::apache::thrift::protocol::TType _etype583; + xfer += iprot->readListBegin(_etype583, _size580); + this->open_txns.resize(_size580); + uint32_t _i584; + for (_i584 = 0; _i584 < _size580; ++_i584) { - xfer += this->open_txns[_i583].read(iprot); + xfer += this->open_txns[_i584].read(iprot); } xfer += iprot->readListEnd(); } @@ -13552,10 +13610,10 @@ uint32_t GetOpenTxnsInfoResponse::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("open_txns", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->open_txns.size())); - std::vector ::const_iterator _iter584; - for (_iter584 = this->open_txns.begin(); _iter584 != this->open_txns.end(); ++_iter584) + std::vector ::const_iterator _iter585; + for (_iter585 = this->open_txns.begin(); _iter585 != this->open_txns.end(); ++_iter585) { - xfer += (*_iter584).write(oprot); + xfer += (*_iter585).write(oprot); } xfer += oprot->writeListEnd(); } @@ -13572,13 +13630,13 @@ void swap(GetOpenTxnsInfoResponse &a, GetOpenTxnsInfoResponse &b) { swap(a.open_txns, b.open_txns); } -GetOpenTxnsInfoResponse::GetOpenTxnsInfoResponse(const GetOpenTxnsInfoResponse& other585) { - txn_high_water_mark = other585.txn_high_water_mark; - open_txns = other585.open_txns; -} -GetOpenTxnsInfoResponse& GetOpenTxnsInfoResponse::operator=(const GetOpenTxnsInfoResponse& other586) { +GetOpenTxnsInfoResponse::GetOpenTxnsInfoResponse(const GetOpenTxnsInfoResponse& other586) { txn_high_water_mark = other586.txn_high_water_mark; open_txns = other586.open_txns; +} +GetOpenTxnsInfoResponse& GetOpenTxnsInfoResponse::operator=(const GetOpenTxnsInfoResponse& other587) { + txn_high_water_mark = other587.txn_high_water_mark; + open_txns = other587.open_txns; return *this; } void GetOpenTxnsInfoResponse::printTo(std::ostream& out) const { @@ -13647,14 +13705,14 @@ uint32_t GetOpenTxnsResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->open_txns.clear(); - uint32_t _size587; - ::apache::thrift::protocol::TType _etype590; - xfer += iprot->readListBegin(_etype590, _size587); - this->open_txns.resize(_size587); - uint32_t _i591; - for (_i591 = 0; _i591 < _size587; ++_i591) + uint32_t _size588; + ::apache::thrift::protocol::TType _etype591; + xfer += iprot->readListBegin(_etype591, _size588); + this->open_txns.resize(_size588); + uint32_t _i592; + for (_i592 = 0; _i592 < _size588; ++_i592) { - xfer += iprot->readI64(this->open_txns[_i591]); + xfer += iprot->readI64(this->open_txns[_i592]); } xfer += iprot->readListEnd(); } @@ -13709,10 +13767,10 @@ uint32_t GetOpenTxnsResponse::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("open_txns", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->open_txns.size())); - std::vector ::const_iterator _iter592; - for (_iter592 = this->open_txns.begin(); _iter592 != this->open_txns.end(); ++_iter592) + std::vector ::const_iterator _iter593; + for (_iter593 = this->open_txns.begin(); _iter593 != this->open_txns.end(); ++_iter593) { - xfer += oprot->writeI64((*_iter592)); + xfer += oprot->writeI64((*_iter593)); } xfer += oprot->writeListEnd(); } @@ -13741,19 +13799,19 @@ void swap(GetOpenTxnsResponse &a, GetOpenTxnsResponse &b) { swap(a.__isset, b.__isset); } -GetOpenTxnsResponse::GetOpenTxnsResponse(const GetOpenTxnsResponse& other593) { - txn_high_water_mark = other593.txn_high_water_mark; - open_txns = other593.open_txns; - min_open_txn = other593.min_open_txn; - abortedBits = other593.abortedBits; - __isset = other593.__isset; -} -GetOpenTxnsResponse& GetOpenTxnsResponse::operator=(const GetOpenTxnsResponse& other594) { +GetOpenTxnsResponse::GetOpenTxnsResponse(const GetOpenTxnsResponse& other594) { txn_high_water_mark = other594.txn_high_water_mark; open_txns = other594.open_txns; min_open_txn = other594.min_open_txn; abortedBits = other594.abortedBits; __isset = other594.__isset; +} +GetOpenTxnsResponse& GetOpenTxnsResponse::operator=(const GetOpenTxnsResponse& other595) { + txn_high_water_mark = other595.txn_high_water_mark; + open_txns = other595.open_txns; + min_open_txn = other595.min_open_txn; + abortedBits = other595.abortedBits; + __isset = other595.__isset; return *this; } void GetOpenTxnsResponse::printTo(std::ostream& out) const { @@ -13898,19 +13956,19 @@ void swap(OpenTxnRequest &a, OpenTxnRequest &b) { swap(a.__isset, b.__isset); } -OpenTxnRequest::OpenTxnRequest(const OpenTxnRequest& other595) { - num_txns = other595.num_txns; - user = other595.user; - hostname = other595.hostname; - agentInfo = other595.agentInfo; - __isset = other595.__isset; -} -OpenTxnRequest& OpenTxnRequest::operator=(const OpenTxnRequest& other596) { +OpenTxnRequest::OpenTxnRequest(const OpenTxnRequest& other596) { num_txns = other596.num_txns; user = other596.user; hostname = other596.hostname; agentInfo = other596.agentInfo; __isset = other596.__isset; +} +OpenTxnRequest& OpenTxnRequest::operator=(const OpenTxnRequest& other597) { + num_txns = other597.num_txns; + user = other597.user; + hostname = other597.hostname; + agentInfo = other597.agentInfo; + __isset = other597.__isset; return *this; } void OpenTxnRequest::printTo(std::ostream& out) const { @@ -13958,14 +14016,14 @@ uint32_t OpenTxnsResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->txn_ids.clear(); - uint32_t _size597; - ::apache::thrift::protocol::TType _etype600; - xfer += iprot->readListBegin(_etype600, _size597); - this->txn_ids.resize(_size597); - uint32_t _i601; - for (_i601 = 0; _i601 < _size597; ++_i601) + uint32_t _size598; + ::apache::thrift::protocol::TType _etype601; + xfer += iprot->readListBegin(_etype601, _size598); + this->txn_ids.resize(_size598); + uint32_t _i602; + for (_i602 = 0; _i602 < _size598; ++_i602) { - xfer += iprot->readI64(this->txn_ids[_i601]); + xfer += iprot->readI64(this->txn_ids[_i602]); } xfer += iprot->readListEnd(); } @@ -13996,10 +14054,10 @@ uint32_t OpenTxnsResponse::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("txn_ids", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->txn_ids.size())); - std::vector ::const_iterator _iter602; - for (_iter602 = this->txn_ids.begin(); _iter602 != this->txn_ids.end(); ++_iter602) + std::vector ::const_iterator _iter603; + for (_iter603 = this->txn_ids.begin(); _iter603 != this->txn_ids.end(); ++_iter603) { - xfer += oprot->writeI64((*_iter602)); + xfer += oprot->writeI64((*_iter603)); } xfer += oprot->writeListEnd(); } @@ -14015,11 +14073,11 @@ void swap(OpenTxnsResponse &a, OpenTxnsResponse &b) { swap(a.txn_ids, b.txn_ids); } -OpenTxnsResponse::OpenTxnsResponse(const OpenTxnsResponse& other603) { - txn_ids = other603.txn_ids; -} -OpenTxnsResponse& OpenTxnsResponse::operator=(const OpenTxnsResponse& other604) { +OpenTxnsResponse::OpenTxnsResponse(const OpenTxnsResponse& other604) { txn_ids = other604.txn_ids; +} +OpenTxnsResponse& OpenTxnsResponse::operator=(const OpenTxnsResponse& other605) { + txn_ids = other605.txn_ids; return *this; } void OpenTxnsResponse::printTo(std::ostream& out) const { @@ -14101,11 +14159,11 @@ void swap(AbortTxnRequest &a, AbortTxnRequest &b) { swap(a.txnid, b.txnid); } -AbortTxnRequest::AbortTxnRequest(const AbortTxnRequest& other605) { - txnid = other605.txnid; -} -AbortTxnRequest& AbortTxnRequest::operator=(const AbortTxnRequest& other606) { +AbortTxnRequest::AbortTxnRequest(const AbortTxnRequest& other606) { txnid = other606.txnid; +} +AbortTxnRequest& AbortTxnRequest::operator=(const AbortTxnRequest& other607) { + txnid = other607.txnid; return *this; } void AbortTxnRequest::printTo(std::ostream& out) const { @@ -14150,14 +14208,14 @@ uint32_t AbortTxnsRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->txn_ids.clear(); - uint32_t _size607; - ::apache::thrift::protocol::TType _etype610; - xfer += iprot->readListBegin(_etype610, _size607); - this->txn_ids.resize(_size607); - uint32_t _i611; - for (_i611 = 0; _i611 < _size607; ++_i611) + uint32_t _size608; + ::apache::thrift::protocol::TType _etype611; + xfer += iprot->readListBegin(_etype611, _size608); + this->txn_ids.resize(_size608); + uint32_t _i612; + for (_i612 = 0; _i612 < _size608; ++_i612) { - xfer += iprot->readI64(this->txn_ids[_i611]); + xfer += iprot->readI64(this->txn_ids[_i612]); } xfer += iprot->readListEnd(); } @@ -14188,10 +14246,10 @@ uint32_t AbortTxnsRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("txn_ids", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->txn_ids.size())); - std::vector ::const_iterator _iter612; - for (_iter612 = this->txn_ids.begin(); _iter612 != this->txn_ids.end(); ++_iter612) + std::vector ::const_iterator _iter613; + for (_iter613 = this->txn_ids.begin(); _iter613 != this->txn_ids.end(); ++_iter613) { - xfer += oprot->writeI64((*_iter612)); + xfer += oprot->writeI64((*_iter613)); } xfer += oprot->writeListEnd(); } @@ -14207,11 +14265,11 @@ void swap(AbortTxnsRequest &a, AbortTxnsRequest &b) { swap(a.txn_ids, b.txn_ids); } -AbortTxnsRequest::AbortTxnsRequest(const AbortTxnsRequest& other613) { - txn_ids = other613.txn_ids; -} -AbortTxnsRequest& AbortTxnsRequest::operator=(const AbortTxnsRequest& other614) { +AbortTxnsRequest::AbortTxnsRequest(const AbortTxnsRequest& other614) { txn_ids = other614.txn_ids; +} +AbortTxnsRequest& AbortTxnsRequest::operator=(const AbortTxnsRequest& other615) { + txn_ids = other615.txn_ids; return *this; } void AbortTxnsRequest::printTo(std::ostream& out) const { @@ -14293,11 +14351,11 @@ void swap(CommitTxnRequest &a, CommitTxnRequest &b) { swap(a.txnid, b.txnid); } -CommitTxnRequest::CommitTxnRequest(const CommitTxnRequest& other615) { - txnid = other615.txnid; -} -CommitTxnRequest& CommitTxnRequest::operator=(const CommitTxnRequest& other616) { +CommitTxnRequest::CommitTxnRequest(const CommitTxnRequest& other616) { txnid = other616.txnid; +} +CommitTxnRequest& CommitTxnRequest::operator=(const CommitTxnRequest& other617) { + txnid = other617.txnid; return *this; } void CommitTxnRequest::printTo(std::ostream& out) const { @@ -14375,9 +14433,9 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { { case 1: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast617; - xfer += iprot->readI32(ecast617); - this->type = (LockType::type)ecast617; + int32_t ecast618; + xfer += iprot->readI32(ecast618); + this->type = (LockType::type)ecast618; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -14385,9 +14443,9 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast618; - xfer += iprot->readI32(ecast618); - this->level = (LockLevel::type)ecast618; + int32_t ecast619; + xfer += iprot->readI32(ecast619); + this->level = (LockLevel::type)ecast619; isset_level = true; } else { xfer += iprot->skip(ftype); @@ -14419,9 +14477,9 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast619; - xfer += iprot->readI32(ecast619); - this->operationType = (DataOperationType::type)ecast619; + int32_t ecast620; + xfer += iprot->readI32(ecast620); + this->operationType = (DataOperationType::type)ecast620; this->__isset.operationType = true; } else { xfer += iprot->skip(ftype); @@ -14521,18 +14579,7 @@ void swap(LockComponent &a, LockComponent &b) { swap(a.__isset, b.__isset); } -LockComponent::LockComponent(const LockComponent& other620) { - type = other620.type; - level = other620.level; - dbname = other620.dbname; - tablename = other620.tablename; - partitionname = other620.partitionname; - operationType = other620.operationType; - isAcid = other620.isAcid; - isDynamicPartitionWrite = other620.isDynamicPartitionWrite; - __isset = other620.__isset; -} -LockComponent& LockComponent::operator=(const LockComponent& other621) { +LockComponent::LockComponent(const LockComponent& other621) { type = other621.type; level = other621.level; dbname = other621.dbname; @@ -14542,6 +14589,17 @@ LockComponent& LockComponent::operator=(const LockComponent& other621) { isAcid = other621.isAcid; isDynamicPartitionWrite = other621.isDynamicPartitionWrite; __isset = other621.__isset; +} +LockComponent& LockComponent::operator=(const LockComponent& other622) { + type = other622.type; + level = other622.level; + dbname = other622.dbname; + tablename = other622.tablename; + partitionname = other622.partitionname; + operationType = other622.operationType; + isAcid = other622.isAcid; + isDynamicPartitionWrite = other622.isDynamicPartitionWrite; + __isset = other622.__isset; return *this; } void LockComponent::printTo(std::ostream& out) const { @@ -14613,14 +14671,14 @@ uint32_t LockRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->component.clear(); - uint32_t _size622; - ::apache::thrift::protocol::TType _etype625; - xfer += iprot->readListBegin(_etype625, _size622); - this->component.resize(_size622); - uint32_t _i626; - for (_i626 = 0; _i626 < _size622; ++_i626) + uint32_t _size623; + ::apache::thrift::protocol::TType _etype626; + xfer += iprot->readListBegin(_etype626, _size623); + this->component.resize(_size623); + uint32_t _i627; + for (_i627 = 0; _i627 < _size623; ++_i627) { - xfer += this->component[_i626].read(iprot); + xfer += this->component[_i627].read(iprot); } xfer += iprot->readListEnd(); } @@ -14687,10 +14745,10 @@ uint32_t LockRequest::write(::apache::thrift::protocol::TProtocol* oprot) const xfer += oprot->writeFieldBegin("component", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->component.size())); - std::vector ::const_iterator _iter627; - for (_iter627 = this->component.begin(); _iter627 != this->component.end(); ++_iter627) + std::vector ::const_iterator _iter628; + for (_iter628 = this->component.begin(); _iter628 != this->component.end(); ++_iter628) { - xfer += (*_iter627).write(oprot); + xfer += (*_iter628).write(oprot); } xfer += oprot->writeListEnd(); } @@ -14729,21 +14787,21 @@ void swap(LockRequest &a, LockRequest &b) { swap(a.__isset, b.__isset); } -LockRequest::LockRequest(const LockRequest& other628) { - component = other628.component; - txnid = other628.txnid; - user = other628.user; - hostname = other628.hostname; - agentInfo = other628.agentInfo; - __isset = other628.__isset; -} -LockRequest& LockRequest::operator=(const LockRequest& other629) { +LockRequest::LockRequest(const LockRequest& other629) { component = other629.component; txnid = other629.txnid; user = other629.user; hostname = other629.hostname; agentInfo = other629.agentInfo; __isset = other629.__isset; +} +LockRequest& LockRequest::operator=(const LockRequest& other630) { + component = other630.component; + txnid = other630.txnid; + user = other630.user; + hostname = other630.hostname; + agentInfo = other630.agentInfo; + __isset = other630.__isset; return *this; } void LockRequest::printTo(std::ostream& out) const { @@ -14803,9 +14861,9 @@ uint32_t LockResponse::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast630; - xfer += iprot->readI32(ecast630); - this->state = (LockState::type)ecast630; + int32_t ecast631; + xfer += iprot->readI32(ecast631); + this->state = (LockState::type)ecast631; isset_state = true; } else { xfer += iprot->skip(ftype); @@ -14851,13 +14909,13 @@ void swap(LockResponse &a, LockResponse &b) { swap(a.state, b.state); } -LockResponse::LockResponse(const LockResponse& other631) { - lockid = other631.lockid; - state = other631.state; -} -LockResponse& LockResponse::operator=(const LockResponse& other632) { +LockResponse::LockResponse(const LockResponse& other632) { lockid = other632.lockid; state = other632.state; +} +LockResponse& LockResponse::operator=(const LockResponse& other633) { + lockid = other633.lockid; + state = other633.state; return *this; } void LockResponse::printTo(std::ostream& out) const { @@ -14979,17 +15037,17 @@ void swap(CheckLockRequest &a, CheckLockRequest &b) { swap(a.__isset, b.__isset); } -CheckLockRequest::CheckLockRequest(const CheckLockRequest& other633) { - lockid = other633.lockid; - txnid = other633.txnid; - elapsed_ms = other633.elapsed_ms; - __isset = other633.__isset; -} -CheckLockRequest& CheckLockRequest::operator=(const CheckLockRequest& other634) { +CheckLockRequest::CheckLockRequest(const CheckLockRequest& other634) { lockid = other634.lockid; txnid = other634.txnid; elapsed_ms = other634.elapsed_ms; __isset = other634.__isset; +} +CheckLockRequest& CheckLockRequest::operator=(const CheckLockRequest& other635) { + lockid = other635.lockid; + txnid = other635.txnid; + elapsed_ms = other635.elapsed_ms; + __isset = other635.__isset; return *this; } void CheckLockRequest::printTo(std::ostream& out) const { @@ -15073,11 +15131,11 @@ void swap(UnlockRequest &a, UnlockRequest &b) { swap(a.lockid, b.lockid); } -UnlockRequest::UnlockRequest(const UnlockRequest& other635) { - lockid = other635.lockid; -} -UnlockRequest& UnlockRequest::operator=(const UnlockRequest& other636) { +UnlockRequest::UnlockRequest(const UnlockRequest& other636) { lockid = other636.lockid; +} +UnlockRequest& UnlockRequest::operator=(const UnlockRequest& other637) { + lockid = other637.lockid; return *this; } void UnlockRequest::printTo(std::ostream& out) const { @@ -15216,19 +15274,19 @@ void swap(ShowLocksRequest &a, ShowLocksRequest &b) { swap(a.__isset, b.__isset); } -ShowLocksRequest::ShowLocksRequest(const ShowLocksRequest& other637) { - dbname = other637.dbname; - tablename = other637.tablename; - partname = other637.partname; - isExtended = other637.isExtended; - __isset = other637.__isset; -} -ShowLocksRequest& ShowLocksRequest::operator=(const ShowLocksRequest& other638) { +ShowLocksRequest::ShowLocksRequest(const ShowLocksRequest& other638) { dbname = other638.dbname; tablename = other638.tablename; partname = other638.partname; isExtended = other638.isExtended; __isset = other638.__isset; +} +ShowLocksRequest& ShowLocksRequest::operator=(const ShowLocksRequest& other639) { + dbname = other639.dbname; + tablename = other639.tablename; + partname = other639.partname; + isExtended = other639.isExtended; + __isset = other639.__isset; return *this; } void ShowLocksRequest::printTo(std::ostream& out) const { @@ -15381,9 +15439,9 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast639; - xfer += iprot->readI32(ecast639); - this->state = (LockState::type)ecast639; + int32_t ecast640; + xfer += iprot->readI32(ecast640); + this->state = (LockState::type)ecast640; isset_state = true; } else { xfer += iprot->skip(ftype); @@ -15391,9 +15449,9 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast640; - xfer += iprot->readI32(ecast640); - this->type = (LockType::type)ecast640; + int32_t ecast641; + xfer += iprot->readI32(ecast641); + this->type = (LockType::type)ecast641; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -15609,26 +15667,7 @@ void swap(ShowLocksResponseElement &a, ShowLocksResponseElement &b) { swap(a.__isset, b.__isset); } -ShowLocksResponseElement::ShowLocksResponseElement(const ShowLocksResponseElement& other641) { - lockid = other641.lockid; - dbname = other641.dbname; - tablename = other641.tablename; - partname = other641.partname; - state = other641.state; - type = other641.type; - txnid = other641.txnid; - lastheartbeat = other641.lastheartbeat; - acquiredat = other641.acquiredat; - user = other641.user; - hostname = other641.hostname; - heartbeatCount = other641.heartbeatCount; - agentInfo = other641.agentInfo; - blockedByExtId = other641.blockedByExtId; - blockedByIntId = other641.blockedByIntId; - lockIdInternal = other641.lockIdInternal; - __isset = other641.__isset; -} -ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksResponseElement& other642) { +ShowLocksResponseElement::ShowLocksResponseElement(const ShowLocksResponseElement& other642) { lockid = other642.lockid; dbname = other642.dbname; tablename = other642.tablename; @@ -15646,6 +15685,25 @@ ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksRes blockedByIntId = other642.blockedByIntId; lockIdInternal = other642.lockIdInternal; __isset = other642.__isset; +} +ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksResponseElement& other643) { + lockid = other643.lockid; + dbname = other643.dbname; + tablename = other643.tablename; + partname = other643.partname; + state = other643.state; + type = other643.type; + txnid = other643.txnid; + lastheartbeat = other643.lastheartbeat; + acquiredat = other643.acquiredat; + user = other643.user; + hostname = other643.hostname; + heartbeatCount = other643.heartbeatCount; + agentInfo = other643.agentInfo; + blockedByExtId = other643.blockedByExtId; + blockedByIntId = other643.blockedByIntId; + lockIdInternal = other643.lockIdInternal; + __isset = other643.__isset; return *this; } void ShowLocksResponseElement::printTo(std::ostream& out) const { @@ -15704,14 +15762,14 @@ uint32_t ShowLocksResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->locks.clear(); - uint32_t _size643; - ::apache::thrift::protocol::TType _etype646; - xfer += iprot->readListBegin(_etype646, _size643); - this->locks.resize(_size643); - uint32_t _i647; - for (_i647 = 0; _i647 < _size643; ++_i647) + uint32_t _size644; + ::apache::thrift::protocol::TType _etype647; + xfer += iprot->readListBegin(_etype647, _size644); + this->locks.resize(_size644); + uint32_t _i648; + for (_i648 = 0; _i648 < _size644; ++_i648) { - xfer += this->locks[_i647].read(iprot); + xfer += this->locks[_i648].read(iprot); } xfer += iprot->readListEnd(); } @@ -15740,10 +15798,10 @@ uint32_t ShowLocksResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("locks", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->locks.size())); - std::vector ::const_iterator _iter648; - for (_iter648 = this->locks.begin(); _iter648 != this->locks.end(); ++_iter648) + std::vector ::const_iterator _iter649; + for (_iter649 = this->locks.begin(); _iter649 != this->locks.end(); ++_iter649) { - xfer += (*_iter648).write(oprot); + xfer += (*_iter649).write(oprot); } xfer += oprot->writeListEnd(); } @@ -15760,13 +15818,13 @@ void swap(ShowLocksResponse &a, ShowLocksResponse &b) { swap(a.__isset, b.__isset); } -ShowLocksResponse::ShowLocksResponse(const ShowLocksResponse& other649) { - locks = other649.locks; - __isset = other649.__isset; -} -ShowLocksResponse& ShowLocksResponse::operator=(const ShowLocksResponse& other650) { +ShowLocksResponse::ShowLocksResponse(const ShowLocksResponse& other650) { locks = other650.locks; __isset = other650.__isset; +} +ShowLocksResponse& ShowLocksResponse::operator=(const ShowLocksResponse& other651) { + locks = other651.locks; + __isset = other651.__isset; return *this; } void ShowLocksResponse::printTo(std::ostream& out) const { @@ -15867,15 +15925,15 @@ void swap(HeartbeatRequest &a, HeartbeatRequest &b) { swap(a.__isset, b.__isset); } -HeartbeatRequest::HeartbeatRequest(const HeartbeatRequest& other651) { - lockid = other651.lockid; - txnid = other651.txnid; - __isset = other651.__isset; -} -HeartbeatRequest& HeartbeatRequest::operator=(const HeartbeatRequest& other652) { +HeartbeatRequest::HeartbeatRequest(const HeartbeatRequest& other652) { lockid = other652.lockid; txnid = other652.txnid; __isset = other652.__isset; +} +HeartbeatRequest& HeartbeatRequest::operator=(const HeartbeatRequest& other653) { + lockid = other653.lockid; + txnid = other653.txnid; + __isset = other653.__isset; return *this; } void HeartbeatRequest::printTo(std::ostream& out) const { @@ -15978,13 +16036,13 @@ void swap(HeartbeatTxnRangeRequest &a, HeartbeatTxnRangeRequest &b) { swap(a.max, b.max); } -HeartbeatTxnRangeRequest::HeartbeatTxnRangeRequest(const HeartbeatTxnRangeRequest& other653) { - min = other653.min; - max = other653.max; -} -HeartbeatTxnRangeRequest& HeartbeatTxnRangeRequest::operator=(const HeartbeatTxnRangeRequest& other654) { +HeartbeatTxnRangeRequest::HeartbeatTxnRangeRequest(const HeartbeatTxnRangeRequest& other654) { min = other654.min; max = other654.max; +} +HeartbeatTxnRangeRequest& HeartbeatTxnRangeRequest::operator=(const HeartbeatTxnRangeRequest& other655) { + min = other655.min; + max = other655.max; return *this; } void HeartbeatTxnRangeRequest::printTo(std::ostream& out) const { @@ -16035,15 +16093,15 @@ uint32_t HeartbeatTxnRangeResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_SET) { { this->aborted.clear(); - uint32_t _size655; - ::apache::thrift::protocol::TType _etype658; - xfer += iprot->readSetBegin(_etype658, _size655); - uint32_t _i659; - for (_i659 = 0; _i659 < _size655; ++_i659) + uint32_t _size656; + ::apache::thrift::protocol::TType _etype659; + xfer += iprot->readSetBegin(_etype659, _size656); + uint32_t _i660; + for (_i660 = 0; _i660 < _size656; ++_i660) { - int64_t _elem660; - xfer += iprot->readI64(_elem660); - this->aborted.insert(_elem660); + int64_t _elem661; + xfer += iprot->readI64(_elem661); + this->aborted.insert(_elem661); } xfer += iprot->readSetEnd(); } @@ -16056,15 +16114,15 @@ uint32_t HeartbeatTxnRangeResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_SET) { { this->nosuch.clear(); - uint32_t _size661; - ::apache::thrift::protocol::TType _etype664; - xfer += iprot->readSetBegin(_etype664, _size661); - uint32_t _i665; - for (_i665 = 0; _i665 < _size661; ++_i665) + uint32_t _size662; + ::apache::thrift::protocol::TType _etype665; + xfer += iprot->readSetBegin(_etype665, _size662); + uint32_t _i666; + for (_i666 = 0; _i666 < _size662; ++_i666) { - int64_t _elem666; - xfer += iprot->readI64(_elem666); - this->nosuch.insert(_elem666); + int64_t _elem667; + xfer += iprot->readI64(_elem667); + this->nosuch.insert(_elem667); } xfer += iprot->readSetEnd(); } @@ -16097,10 +16155,10 @@ uint32_t HeartbeatTxnRangeResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("aborted", ::apache::thrift::protocol::T_SET, 1); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_I64, static_cast(this->aborted.size())); - std::set ::const_iterator _iter667; - for (_iter667 = this->aborted.begin(); _iter667 != this->aborted.end(); ++_iter667) + std::set ::const_iterator _iter668; + for (_iter668 = this->aborted.begin(); _iter668 != this->aborted.end(); ++_iter668) { - xfer += oprot->writeI64((*_iter667)); + xfer += oprot->writeI64((*_iter668)); } xfer += oprot->writeSetEnd(); } @@ -16109,10 +16167,10 @@ uint32_t HeartbeatTxnRangeResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("nosuch", ::apache::thrift::protocol::T_SET, 2); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_I64, static_cast(this->nosuch.size())); - std::set ::const_iterator _iter668; - for (_iter668 = this->nosuch.begin(); _iter668 != this->nosuch.end(); ++_iter668) + std::set ::const_iterator _iter669; + for (_iter669 = this->nosuch.begin(); _iter669 != this->nosuch.end(); ++_iter669) { - xfer += oprot->writeI64((*_iter668)); + xfer += oprot->writeI64((*_iter669)); } xfer += oprot->writeSetEnd(); } @@ -16129,13 +16187,13 @@ void swap(HeartbeatTxnRangeResponse &a, HeartbeatTxnRangeResponse &b) { swap(a.nosuch, b.nosuch); } -HeartbeatTxnRangeResponse::HeartbeatTxnRangeResponse(const HeartbeatTxnRangeResponse& other669) { - aborted = other669.aborted; - nosuch = other669.nosuch; -} -HeartbeatTxnRangeResponse& HeartbeatTxnRangeResponse::operator=(const HeartbeatTxnRangeResponse& other670) { +HeartbeatTxnRangeResponse::HeartbeatTxnRangeResponse(const HeartbeatTxnRangeResponse& other670) { aborted = other670.aborted; nosuch = other670.nosuch; +} +HeartbeatTxnRangeResponse& HeartbeatTxnRangeResponse::operator=(const HeartbeatTxnRangeResponse& other671) { + aborted = other671.aborted; + nosuch = other671.nosuch; return *this; } void HeartbeatTxnRangeResponse::printTo(std::ostream& out) const { @@ -16228,9 +16286,9 @@ uint32_t CompactionRequest::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast671; - xfer += iprot->readI32(ecast671); - this->type = (CompactionType::type)ecast671; + int32_t ecast672; + xfer += iprot->readI32(ecast672); + this->type = (CompactionType::type)ecast672; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -16248,17 +16306,17 @@ uint32_t CompactionRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->properties.clear(); - uint32_t _size672; - ::apache::thrift::protocol::TType _ktype673; - ::apache::thrift::protocol::TType _vtype674; - xfer += iprot->readMapBegin(_ktype673, _vtype674, _size672); - uint32_t _i676; - for (_i676 = 0; _i676 < _size672; ++_i676) + uint32_t _size673; + ::apache::thrift::protocol::TType _ktype674; + ::apache::thrift::protocol::TType _vtype675; + xfer += iprot->readMapBegin(_ktype674, _vtype675, _size673); + uint32_t _i677; + for (_i677 = 0; _i677 < _size673; ++_i677) { - std::string _key677; - xfer += iprot->readString(_key677); - std::string& _val678 = this->properties[_key677]; - xfer += iprot->readString(_val678); + std::string _key678; + xfer += iprot->readString(_key678); + std::string& _val679 = this->properties[_key678]; + xfer += iprot->readString(_val679); } xfer += iprot->readMapEnd(); } @@ -16316,11 +16374,11 @@ uint32_t CompactionRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("properties", ::apache::thrift::protocol::T_MAP, 6); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->properties.size())); - std::map ::const_iterator _iter679; - for (_iter679 = this->properties.begin(); _iter679 != this->properties.end(); ++_iter679) + std::map ::const_iterator _iter680; + for (_iter680 = this->properties.begin(); _iter680 != this->properties.end(); ++_iter680) { - xfer += oprot->writeString(_iter679->first); - xfer += oprot->writeString(_iter679->second); + xfer += oprot->writeString(_iter680->first); + xfer += oprot->writeString(_iter680->second); } xfer += oprot->writeMapEnd(); } @@ -16342,16 +16400,7 @@ void swap(CompactionRequest &a, CompactionRequest &b) { swap(a.__isset, b.__isset); } -CompactionRequest::CompactionRequest(const CompactionRequest& other680) { - dbname = other680.dbname; - tablename = other680.tablename; - partitionname = other680.partitionname; - type = other680.type; - runas = other680.runas; - properties = other680.properties; - __isset = other680.__isset; -} -CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other681) { +CompactionRequest::CompactionRequest(const CompactionRequest& other681) { dbname = other681.dbname; tablename = other681.tablename; partitionname = other681.partitionname; @@ -16359,6 +16408,15 @@ CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other68 runas = other681.runas; properties = other681.properties; __isset = other681.__isset; +} +CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other682) { + dbname = other682.dbname; + tablename = other682.tablename; + partitionname = other682.partitionname; + type = other682.type; + runas = other682.runas; + properties = other682.properties; + __isset = other682.__isset; return *this; } void CompactionRequest::printTo(std::ostream& out) const { @@ -16485,15 +16543,15 @@ void swap(CompactionResponse &a, CompactionResponse &b) { swap(a.accepted, b.accepted); } -CompactionResponse::CompactionResponse(const CompactionResponse& other682) { - id = other682.id; - state = other682.state; - accepted = other682.accepted; -} -CompactionResponse& CompactionResponse::operator=(const CompactionResponse& other683) { +CompactionResponse::CompactionResponse(const CompactionResponse& other683) { id = other683.id; state = other683.state; accepted = other683.accepted; +} +CompactionResponse& CompactionResponse::operator=(const CompactionResponse& other684) { + id = other684.id; + state = other684.state; + accepted = other684.accepted; return *this; } void CompactionResponse::printTo(std::ostream& out) const { @@ -16554,11 +16612,11 @@ void swap(ShowCompactRequest &a, ShowCompactRequest &b) { (void) b; } -ShowCompactRequest::ShowCompactRequest(const ShowCompactRequest& other684) { - (void) other684; -} -ShowCompactRequest& ShowCompactRequest::operator=(const ShowCompactRequest& other685) { +ShowCompactRequest::ShowCompactRequest(const ShowCompactRequest& other685) { (void) other685; +} +ShowCompactRequest& ShowCompactRequest::operator=(const ShowCompactRequest& other686) { + (void) other686; return *this; } void ShowCompactRequest::printTo(std::ostream& out) const { @@ -16684,9 +16742,9 @@ uint32_t ShowCompactResponseElement::read(::apache::thrift::protocol::TProtocol* break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast686; - xfer += iprot->readI32(ecast686); - this->type = (CompactionType::type)ecast686; + int32_t ecast687; + xfer += iprot->readI32(ecast687); + this->type = (CompactionType::type)ecast687; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -16873,23 +16931,7 @@ void swap(ShowCompactResponseElement &a, ShowCompactResponseElement &b) { swap(a.__isset, b.__isset); } -ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponseElement& other687) { - dbname = other687.dbname; - tablename = other687.tablename; - partitionname = other687.partitionname; - type = other687.type; - state = other687.state; - workerid = other687.workerid; - start = other687.start; - runAs = other687.runAs; - hightestTxnId = other687.hightestTxnId; - metaInfo = other687.metaInfo; - endTime = other687.endTime; - hadoopJobId = other687.hadoopJobId; - id = other687.id; - __isset = other687.__isset; -} -ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other688) { +ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponseElement& other688) { dbname = other688.dbname; tablename = other688.tablename; partitionname = other688.partitionname; @@ -16904,6 +16946,22 @@ ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowComp hadoopJobId = other688.hadoopJobId; id = other688.id; __isset = other688.__isset; +} +ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other689) { + dbname = other689.dbname; + tablename = other689.tablename; + partitionname = other689.partitionname; + type = other689.type; + state = other689.state; + workerid = other689.workerid; + start = other689.start; + runAs = other689.runAs; + hightestTxnId = other689.hightestTxnId; + metaInfo = other689.metaInfo; + endTime = other689.endTime; + hadoopJobId = other689.hadoopJobId; + id = other689.id; + __isset = other689.__isset; return *this; } void ShowCompactResponseElement::printTo(std::ostream& out) const { @@ -16960,14 +17018,14 @@ uint32_t ShowCompactResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->compacts.clear(); - uint32_t _size689; - ::apache::thrift::protocol::TType _etype692; - xfer += iprot->readListBegin(_etype692, _size689); - this->compacts.resize(_size689); - uint32_t _i693; - for (_i693 = 0; _i693 < _size689; ++_i693) + uint32_t _size690; + ::apache::thrift::protocol::TType _etype693; + xfer += iprot->readListBegin(_etype693, _size690); + this->compacts.resize(_size690); + uint32_t _i694; + for (_i694 = 0; _i694 < _size690; ++_i694) { - xfer += this->compacts[_i693].read(iprot); + xfer += this->compacts[_i694].read(iprot); } xfer += iprot->readListEnd(); } @@ -16998,10 +17056,10 @@ uint32_t ShowCompactResponse::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("compacts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->compacts.size())); - std::vector ::const_iterator _iter694; - for (_iter694 = this->compacts.begin(); _iter694 != this->compacts.end(); ++_iter694) + std::vector ::const_iterator _iter695; + for (_iter695 = this->compacts.begin(); _iter695 != this->compacts.end(); ++_iter695) { - xfer += (*_iter694).write(oprot); + xfer += (*_iter695).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17017,11 +17075,11 @@ void swap(ShowCompactResponse &a, ShowCompactResponse &b) { swap(a.compacts, b.compacts); } -ShowCompactResponse::ShowCompactResponse(const ShowCompactResponse& other695) { - compacts = other695.compacts; -} -ShowCompactResponse& ShowCompactResponse::operator=(const ShowCompactResponse& other696) { +ShowCompactResponse::ShowCompactResponse(const ShowCompactResponse& other696) { compacts = other696.compacts; +} +ShowCompactResponse& ShowCompactResponse::operator=(const ShowCompactResponse& other697) { + compacts = other697.compacts; return *this; } void ShowCompactResponse::printTo(std::ostream& out) const { @@ -17110,14 +17168,14 @@ uint32_t AddDynamicPartitions::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionnames.clear(); - uint32_t _size697; - ::apache::thrift::protocol::TType _etype700; - xfer += iprot->readListBegin(_etype700, _size697); - this->partitionnames.resize(_size697); - uint32_t _i701; - for (_i701 = 0; _i701 < _size697; ++_i701) + uint32_t _size698; + ::apache::thrift::protocol::TType _etype701; + xfer += iprot->readListBegin(_etype701, _size698); + this->partitionnames.resize(_size698); + uint32_t _i702; + for (_i702 = 0; _i702 < _size698; ++_i702) { - xfer += iprot->readString(this->partitionnames[_i701]); + xfer += iprot->readString(this->partitionnames[_i702]); } xfer += iprot->readListEnd(); } @@ -17128,9 +17186,9 @@ uint32_t AddDynamicPartitions::read(::apache::thrift::protocol::TProtocol* iprot break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast702; - xfer += iprot->readI32(ecast702); - this->operationType = (DataOperationType::type)ecast702; + int32_t ecast703; + xfer += iprot->readI32(ecast703); + this->operationType = (DataOperationType::type)ecast703; this->__isset.operationType = true; } else { xfer += iprot->skip(ftype); @@ -17176,10 +17234,10 @@ uint32_t AddDynamicPartitions::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("partitionnames", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionnames.size())); - std::vector ::const_iterator _iter703; - for (_iter703 = this->partitionnames.begin(); _iter703 != this->partitionnames.end(); ++_iter703) + std::vector ::const_iterator _iter704; + for (_iter704 = this->partitionnames.begin(); _iter704 != this->partitionnames.end(); ++_iter704) { - xfer += oprot->writeString((*_iter703)); + xfer += oprot->writeString((*_iter704)); } xfer += oprot->writeListEnd(); } @@ -17205,26 +17263,26 @@ void swap(AddDynamicPartitions &a, AddDynamicPartitions &b) { swap(a.__isset, b.__isset); } -AddDynamicPartitions::AddDynamicPartitions(const AddDynamicPartitions& other704) { - txnid = other704.txnid; - dbname = other704.dbname; - tablename = other704.tablename; - partitionnames = other704.partitionnames; - operationType = other704.operationType; - __isset = other704.__isset; -} -AddDynamicPartitions& AddDynamicPartitions::operator=(const AddDynamicPartitions& other705) { +AddDynamicPartitions::AddDynamicPartitions(const AddDynamicPartitions& other705) { txnid = other705.txnid; dbname = other705.dbname; tablename = other705.tablename; partitionnames = other705.partitionnames; operationType = other705.operationType; __isset = other705.__isset; - return *this; } -void AddDynamicPartitions::printTo(std::ostream& out) const { - using ::apache::thrift::to_string; - out << "AddDynamicPartitions("; +AddDynamicPartitions& AddDynamicPartitions::operator=(const AddDynamicPartitions& other706) { + txnid = other706.txnid; + dbname = other706.dbname; + tablename = other706.tablename; + partitionnames = other706.partitionnames; + operationType = other706.operationType; + __isset = other706.__isset; + return *this; +} +void AddDynamicPartitions::printTo(std::ostream& out) const { + using ::apache::thrift::to_string; + out << "AddDynamicPartitions("; out << "txnid=" << to_string(txnid); out << ", " << "dbname=" << to_string(dbname); out << ", " << "tablename=" << to_string(tablename); @@ -17401,16 +17459,7 @@ void swap(BasicTxnInfo &a, BasicTxnInfo &b) { swap(a.__isset, b.__isset); } -BasicTxnInfo::BasicTxnInfo(const BasicTxnInfo& other706) { - isnull = other706.isnull; - time = other706.time; - txnid = other706.txnid; - dbname = other706.dbname; - tablename = other706.tablename; - partitionname = other706.partitionname; - __isset = other706.__isset; -} -BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other707) { +BasicTxnInfo::BasicTxnInfo(const BasicTxnInfo& other707) { isnull = other707.isnull; time = other707.time; txnid = other707.txnid; @@ -17418,6 +17467,15 @@ BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other707) { tablename = other707.tablename; partitionname = other707.partitionname; __isset = other707.__isset; +} +BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other708) { + isnull = other708.isnull; + time = other708.time; + txnid = other708.txnid; + dbname = other708.dbname; + tablename = other708.tablename; + partitionname = other708.partitionname; + __isset = other708.__isset; return *this; } void BasicTxnInfo::printTo(std::ostream& out) const { @@ -17498,15 +17556,15 @@ uint32_t CreationMetadata::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_SET) { { this->tablesUsed.clear(); - uint32_t _size708; - ::apache::thrift::protocol::TType _etype711; - xfer += iprot->readSetBegin(_etype711, _size708); - uint32_t _i712; - for (_i712 = 0; _i712 < _size708; ++_i712) + uint32_t _size709; + ::apache::thrift::protocol::TType _etype712; + xfer += iprot->readSetBegin(_etype712, _size709); + uint32_t _i713; + for (_i713 = 0; _i713 < _size709; ++_i713) { - std::string _elem713; - xfer += iprot->readString(_elem713); - this->tablesUsed.insert(_elem713); + std::string _elem714; + xfer += iprot->readString(_elem714); + this->tablesUsed.insert(_elem714); } xfer += iprot->readSetEnd(); } @@ -17557,10 +17615,10 @@ uint32_t CreationMetadata::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tablesUsed", ::apache::thrift::protocol::T_SET, 3); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tablesUsed.size())); - std::set ::const_iterator _iter714; - for (_iter714 = this->tablesUsed.begin(); _iter714 != this->tablesUsed.end(); ++_iter714) + std::set ::const_iterator _iter715; + for (_iter715 = this->tablesUsed.begin(); _iter715 != this->tablesUsed.end(); ++_iter715) { - xfer += oprot->writeString((*_iter714)); + xfer += oprot->writeString((*_iter715)); } xfer += oprot->writeSetEnd(); } @@ -17585,19 +17643,19 @@ void swap(CreationMetadata &a, CreationMetadata &b) { swap(a.__isset, b.__isset); } -CreationMetadata::CreationMetadata(const CreationMetadata& other715) { - dbName = other715.dbName; - tblName = other715.tblName; - tablesUsed = other715.tablesUsed; - validTxnList = other715.validTxnList; - __isset = other715.__isset; -} -CreationMetadata& CreationMetadata::operator=(const CreationMetadata& other716) { +CreationMetadata::CreationMetadata(const CreationMetadata& other716) { dbName = other716.dbName; tblName = other716.tblName; tablesUsed = other716.tablesUsed; validTxnList = other716.validTxnList; __isset = other716.__isset; +} +CreationMetadata& CreationMetadata::operator=(const CreationMetadata& other717) { + dbName = other717.dbName; + tblName = other717.tblName; + tablesUsed = other717.tablesUsed; + validTxnList = other717.validTxnList; + __isset = other717.__isset; return *this; } void CreationMetadata::printTo(std::ostream& out) const { @@ -17702,15 +17760,15 @@ void swap(NotificationEventRequest &a, NotificationEventRequest &b) { swap(a.__isset, b.__isset); } -NotificationEventRequest::NotificationEventRequest(const NotificationEventRequest& other717) { - lastEvent = other717.lastEvent; - maxEvents = other717.maxEvents; - __isset = other717.__isset; -} -NotificationEventRequest& NotificationEventRequest::operator=(const NotificationEventRequest& other718) { +NotificationEventRequest::NotificationEventRequest(const NotificationEventRequest& other718) { lastEvent = other718.lastEvent; maxEvents = other718.maxEvents; __isset = other718.__isset; +} +NotificationEventRequest& NotificationEventRequest::operator=(const NotificationEventRequest& other719) { + lastEvent = other719.lastEvent; + maxEvents = other719.maxEvents; + __isset = other719.__isset; return *this; } void NotificationEventRequest::printTo(std::ostream& out) const { @@ -17911,17 +17969,7 @@ void swap(NotificationEvent &a, NotificationEvent &b) { swap(a.__isset, b.__isset); } -NotificationEvent::NotificationEvent(const NotificationEvent& other719) { - eventId = other719.eventId; - eventTime = other719.eventTime; - eventType = other719.eventType; - dbName = other719.dbName; - tableName = other719.tableName; - message = other719.message; - messageFormat = other719.messageFormat; - __isset = other719.__isset; -} -NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other720) { +NotificationEvent::NotificationEvent(const NotificationEvent& other720) { eventId = other720.eventId; eventTime = other720.eventTime; eventType = other720.eventType; @@ -17930,6 +17978,16 @@ NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other72 message = other720.message; messageFormat = other720.messageFormat; __isset = other720.__isset; +} +NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other721) { + eventId = other721.eventId; + eventTime = other721.eventTime; + eventType = other721.eventType; + dbName = other721.dbName; + tableName = other721.tableName; + message = other721.message; + messageFormat = other721.messageFormat; + __isset = other721.__isset; return *this; } void NotificationEvent::printTo(std::ostream& out) const { @@ -17980,14 +18038,14 @@ uint32_t NotificationEventResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->events.clear(); - uint32_t _size721; - ::apache::thrift::protocol::TType _etype724; - xfer += iprot->readListBegin(_etype724, _size721); - this->events.resize(_size721); - uint32_t _i725; - for (_i725 = 0; _i725 < _size721; ++_i725) + uint32_t _size722; + ::apache::thrift::protocol::TType _etype725; + xfer += iprot->readListBegin(_etype725, _size722); + this->events.resize(_size722); + uint32_t _i726; + for (_i726 = 0; _i726 < _size722; ++_i726) { - xfer += this->events[_i725].read(iprot); + xfer += this->events[_i726].read(iprot); } xfer += iprot->readListEnd(); } @@ -18018,10 +18076,10 @@ uint32_t NotificationEventResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("events", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->events.size())); - std::vector ::const_iterator _iter726; - for (_iter726 = this->events.begin(); _iter726 != this->events.end(); ++_iter726) + std::vector ::const_iterator _iter727; + for (_iter727 = this->events.begin(); _iter727 != this->events.end(); ++_iter727) { - xfer += (*_iter726).write(oprot); + xfer += (*_iter727).write(oprot); } xfer += oprot->writeListEnd(); } @@ -18037,11 +18095,11 @@ void swap(NotificationEventResponse &a, NotificationEventResponse &b) { swap(a.events, b.events); } -NotificationEventResponse::NotificationEventResponse(const NotificationEventResponse& other727) { - events = other727.events; -} -NotificationEventResponse& NotificationEventResponse::operator=(const NotificationEventResponse& other728) { +NotificationEventResponse::NotificationEventResponse(const NotificationEventResponse& other728) { events = other728.events; +} +NotificationEventResponse& NotificationEventResponse::operator=(const NotificationEventResponse& other729) { + events = other729.events; return *this; } void NotificationEventResponse::printTo(std::ostream& out) const { @@ -18123,11 +18181,11 @@ void swap(CurrentNotificationEventId &a, CurrentNotificationEventId &b) { swap(a.eventId, b.eventId); } -CurrentNotificationEventId::CurrentNotificationEventId(const CurrentNotificationEventId& other729) { - eventId = other729.eventId; -} -CurrentNotificationEventId& CurrentNotificationEventId::operator=(const CurrentNotificationEventId& other730) { +CurrentNotificationEventId::CurrentNotificationEventId(const CurrentNotificationEventId& other730) { eventId = other730.eventId; +} +CurrentNotificationEventId& CurrentNotificationEventId::operator=(const CurrentNotificationEventId& other731) { + eventId = other731.eventId; return *this; } void CurrentNotificationEventId::printTo(std::ostream& out) const { @@ -18229,13 +18287,13 @@ void swap(NotificationEventsCountRequest &a, NotificationEventsCountRequest &b) swap(a.dbName, b.dbName); } -NotificationEventsCountRequest::NotificationEventsCountRequest(const NotificationEventsCountRequest& other731) { - fromEventId = other731.fromEventId; - dbName = other731.dbName; -} -NotificationEventsCountRequest& NotificationEventsCountRequest::operator=(const NotificationEventsCountRequest& other732) { +NotificationEventsCountRequest::NotificationEventsCountRequest(const NotificationEventsCountRequest& other732) { fromEventId = other732.fromEventId; dbName = other732.dbName; +} +NotificationEventsCountRequest& NotificationEventsCountRequest::operator=(const NotificationEventsCountRequest& other733) { + fromEventId = other733.fromEventId; + dbName = other733.dbName; return *this; } void NotificationEventsCountRequest::printTo(std::ostream& out) const { @@ -18318,11 +18376,11 @@ void swap(NotificationEventsCountResponse &a, NotificationEventsCountResponse &b swap(a.eventsCount, b.eventsCount); } -NotificationEventsCountResponse::NotificationEventsCountResponse(const NotificationEventsCountResponse& other733) { - eventsCount = other733.eventsCount; -} -NotificationEventsCountResponse& NotificationEventsCountResponse::operator=(const NotificationEventsCountResponse& other734) { +NotificationEventsCountResponse::NotificationEventsCountResponse(const NotificationEventsCountResponse& other734) { eventsCount = other734.eventsCount; +} +NotificationEventsCountResponse& NotificationEventsCountResponse::operator=(const NotificationEventsCountResponse& other735) { + eventsCount = other735.eventsCount; return *this; } void NotificationEventsCountResponse::printTo(std::ostream& out) const { @@ -18385,14 +18443,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filesAdded.clear(); - uint32_t _size735; - ::apache::thrift::protocol::TType _etype738; - xfer += iprot->readListBegin(_etype738, _size735); - this->filesAdded.resize(_size735); - uint32_t _i739; - for (_i739 = 0; _i739 < _size735; ++_i739) + uint32_t _size736; + ::apache::thrift::protocol::TType _etype739; + xfer += iprot->readListBegin(_etype739, _size736); + this->filesAdded.resize(_size736); + uint32_t _i740; + for (_i740 = 0; _i740 < _size736; ++_i740) { - xfer += iprot->readString(this->filesAdded[_i739]); + xfer += iprot->readString(this->filesAdded[_i740]); } xfer += iprot->readListEnd(); } @@ -18405,14 +18463,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filesAddedChecksum.clear(); - uint32_t _size740; - ::apache::thrift::protocol::TType _etype743; - xfer += iprot->readListBegin(_etype743, _size740); - this->filesAddedChecksum.resize(_size740); - uint32_t _i744; - for (_i744 = 0; _i744 < _size740; ++_i744) + uint32_t _size741; + ::apache::thrift::protocol::TType _etype744; + xfer += iprot->readListBegin(_etype744, _size741); + this->filesAddedChecksum.resize(_size741); + uint32_t _i745; + for (_i745 = 0; _i745 < _size741; ++_i745) { - xfer += iprot->readString(this->filesAddedChecksum[_i744]); + xfer += iprot->readString(this->filesAddedChecksum[_i745]); } xfer += iprot->readListEnd(); } @@ -18448,10 +18506,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("filesAdded", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filesAdded.size())); - std::vector ::const_iterator _iter745; - for (_iter745 = this->filesAdded.begin(); _iter745 != this->filesAdded.end(); ++_iter745) + std::vector ::const_iterator _iter746; + for (_iter746 = this->filesAdded.begin(); _iter746 != this->filesAdded.end(); ++_iter746) { - xfer += oprot->writeString((*_iter745)); + xfer += oprot->writeString((*_iter746)); } xfer += oprot->writeListEnd(); } @@ -18461,10 +18519,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("filesAddedChecksum", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filesAddedChecksum.size())); - std::vector ::const_iterator _iter746; - for (_iter746 = this->filesAddedChecksum.begin(); _iter746 != this->filesAddedChecksum.end(); ++_iter746) + std::vector ::const_iterator _iter747; + for (_iter747 = this->filesAddedChecksum.begin(); _iter747 != this->filesAddedChecksum.end(); ++_iter747) { - xfer += oprot->writeString((*_iter746)); + xfer += oprot->writeString((*_iter747)); } xfer += oprot->writeListEnd(); } @@ -18483,17 +18541,17 @@ void swap(InsertEventRequestData &a, InsertEventRequestData &b) { swap(a.__isset, b.__isset); } -InsertEventRequestData::InsertEventRequestData(const InsertEventRequestData& other747) { - replace = other747.replace; - filesAdded = other747.filesAdded; - filesAddedChecksum = other747.filesAddedChecksum; - __isset = other747.__isset; -} -InsertEventRequestData& InsertEventRequestData::operator=(const InsertEventRequestData& other748) { +InsertEventRequestData::InsertEventRequestData(const InsertEventRequestData& other748) { replace = other748.replace; filesAdded = other748.filesAdded; filesAddedChecksum = other748.filesAddedChecksum; __isset = other748.__isset; +} +InsertEventRequestData& InsertEventRequestData::operator=(const InsertEventRequestData& other749) { + replace = other749.replace; + filesAdded = other749.filesAdded; + filesAddedChecksum = other749.filesAddedChecksum; + __isset = other749.__isset; return *this; } void InsertEventRequestData::printTo(std::ostream& out) const { @@ -18575,13 +18633,13 @@ void swap(FireEventRequestData &a, FireEventRequestData &b) { swap(a.__isset, b.__isset); } -FireEventRequestData::FireEventRequestData(const FireEventRequestData& other749) { - insertData = other749.insertData; - __isset = other749.__isset; -} -FireEventRequestData& FireEventRequestData::operator=(const FireEventRequestData& other750) { +FireEventRequestData::FireEventRequestData(const FireEventRequestData& other750) { insertData = other750.insertData; __isset = other750.__isset; +} +FireEventRequestData& FireEventRequestData::operator=(const FireEventRequestData& other751) { + insertData = other751.insertData; + __isset = other751.__isset; return *this; } void FireEventRequestData::printTo(std::ostream& out) const { @@ -18678,14 +18736,14 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionVals.clear(); - uint32_t _size751; - ::apache::thrift::protocol::TType _etype754; - xfer += iprot->readListBegin(_etype754, _size751); - this->partitionVals.resize(_size751); - uint32_t _i755; - for (_i755 = 0; _i755 < _size751; ++_i755) + uint32_t _size752; + ::apache::thrift::protocol::TType _etype755; + xfer += iprot->readListBegin(_etype755, _size752); + this->partitionVals.resize(_size752); + uint32_t _i756; + for (_i756 = 0; _i756 < _size752; ++_i756) { - xfer += iprot->readString(this->partitionVals[_i755]); + xfer += iprot->readString(this->partitionVals[_i756]); } xfer += iprot->readListEnd(); } @@ -18737,10 +18795,10 @@ uint32_t FireEventRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("partitionVals", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionVals.size())); - std::vector ::const_iterator _iter756; - for (_iter756 = this->partitionVals.begin(); _iter756 != this->partitionVals.end(); ++_iter756) + std::vector ::const_iterator _iter757; + for (_iter757 = this->partitionVals.begin(); _iter757 != this->partitionVals.end(); ++_iter757) { - xfer += oprot->writeString((*_iter756)); + xfer += oprot->writeString((*_iter757)); } xfer += oprot->writeListEnd(); } @@ -18761,21 +18819,21 @@ void swap(FireEventRequest &a, FireEventRequest &b) { swap(a.__isset, b.__isset); } -FireEventRequest::FireEventRequest(const FireEventRequest& other757) { - successful = other757.successful; - data = other757.data; - dbName = other757.dbName; - tableName = other757.tableName; - partitionVals = other757.partitionVals; - __isset = other757.__isset; -} -FireEventRequest& FireEventRequest::operator=(const FireEventRequest& other758) { +FireEventRequest::FireEventRequest(const FireEventRequest& other758) { successful = other758.successful; data = other758.data; dbName = other758.dbName; tableName = other758.tableName; partitionVals = other758.partitionVals; __isset = other758.__isset; +} +FireEventRequest& FireEventRequest::operator=(const FireEventRequest& other759) { + successful = other759.successful; + data = other759.data; + dbName = other759.dbName; + tableName = other759.tableName; + partitionVals = other759.partitionVals; + __isset = other759.__isset; return *this; } void FireEventRequest::printTo(std::ostream& out) const { @@ -18838,11 +18896,11 @@ void swap(FireEventResponse &a, FireEventResponse &b) { (void) b; } -FireEventResponse::FireEventResponse(const FireEventResponse& other759) { - (void) other759; -} -FireEventResponse& FireEventResponse::operator=(const FireEventResponse& other760) { +FireEventResponse::FireEventResponse(const FireEventResponse& other760) { (void) other760; +} +FireEventResponse& FireEventResponse::operator=(const FireEventResponse& other761) { + (void) other761; return *this; } void FireEventResponse::printTo(std::ostream& out) const { @@ -18942,15 +19000,15 @@ void swap(MetadataPpdResult &a, MetadataPpdResult &b) { swap(a.__isset, b.__isset); } -MetadataPpdResult::MetadataPpdResult(const MetadataPpdResult& other761) { - metadata = other761.metadata; - includeBitset = other761.includeBitset; - __isset = other761.__isset; -} -MetadataPpdResult& MetadataPpdResult::operator=(const MetadataPpdResult& other762) { +MetadataPpdResult::MetadataPpdResult(const MetadataPpdResult& other762) { metadata = other762.metadata; includeBitset = other762.includeBitset; __isset = other762.__isset; +} +MetadataPpdResult& MetadataPpdResult::operator=(const MetadataPpdResult& other763) { + metadata = other763.metadata; + includeBitset = other763.includeBitset; + __isset = other763.__isset; return *this; } void MetadataPpdResult::printTo(std::ostream& out) const { @@ -19001,17 +19059,17 @@ uint32_t GetFileMetadataByExprResult::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_MAP) { { this->metadata.clear(); - uint32_t _size763; - ::apache::thrift::protocol::TType _ktype764; - ::apache::thrift::protocol::TType _vtype765; - xfer += iprot->readMapBegin(_ktype764, _vtype765, _size763); - uint32_t _i767; - for (_i767 = 0; _i767 < _size763; ++_i767) + uint32_t _size764; + ::apache::thrift::protocol::TType _ktype765; + ::apache::thrift::protocol::TType _vtype766; + xfer += iprot->readMapBegin(_ktype765, _vtype766, _size764); + uint32_t _i768; + for (_i768 = 0; _i768 < _size764; ++_i768) { - int64_t _key768; - xfer += iprot->readI64(_key768); - MetadataPpdResult& _val769 = this->metadata[_key768]; - xfer += _val769.read(iprot); + int64_t _key769; + xfer += iprot->readI64(_key769); + MetadataPpdResult& _val770 = this->metadata[_key769]; + xfer += _val770.read(iprot); } xfer += iprot->readMapEnd(); } @@ -19052,11 +19110,11 @@ uint32_t GetFileMetadataByExprResult::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRUCT, static_cast(this->metadata.size())); - std::map ::const_iterator _iter770; - for (_iter770 = this->metadata.begin(); _iter770 != this->metadata.end(); ++_iter770) + std::map ::const_iterator _iter771; + for (_iter771 = this->metadata.begin(); _iter771 != this->metadata.end(); ++_iter771) { - xfer += oprot->writeI64(_iter770->first); - xfer += _iter770->second.write(oprot); + xfer += oprot->writeI64(_iter771->first); + xfer += _iter771->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -19077,13 +19135,13 @@ void swap(GetFileMetadataByExprResult &a, GetFileMetadataByExprResult &b) { swap(a.isSupported, b.isSupported); } -GetFileMetadataByExprResult::GetFileMetadataByExprResult(const GetFileMetadataByExprResult& other771) { - metadata = other771.metadata; - isSupported = other771.isSupported; -} -GetFileMetadataByExprResult& GetFileMetadataByExprResult::operator=(const GetFileMetadataByExprResult& other772) { +GetFileMetadataByExprResult::GetFileMetadataByExprResult(const GetFileMetadataByExprResult& other772) { metadata = other772.metadata; isSupported = other772.isSupported; +} +GetFileMetadataByExprResult& GetFileMetadataByExprResult::operator=(const GetFileMetadataByExprResult& other773) { + metadata = other773.metadata; + isSupported = other773.isSupported; return *this; } void GetFileMetadataByExprResult::printTo(std::ostream& out) const { @@ -19144,14 +19202,14 @@ uint32_t GetFileMetadataByExprRequest::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size773; - ::apache::thrift::protocol::TType _etype776; - xfer += iprot->readListBegin(_etype776, _size773); - this->fileIds.resize(_size773); - uint32_t _i777; - for (_i777 = 0; _i777 < _size773; ++_i777) + uint32_t _size774; + ::apache::thrift::protocol::TType _etype777; + xfer += iprot->readListBegin(_etype777, _size774); + this->fileIds.resize(_size774); + uint32_t _i778; + for (_i778 = 0; _i778 < _size774; ++_i778) { - xfer += iprot->readI64(this->fileIds[_i777]); + xfer += iprot->readI64(this->fileIds[_i778]); } xfer += iprot->readListEnd(); } @@ -19178,9 +19236,9 @@ uint32_t GetFileMetadataByExprRequest::read(::apache::thrift::protocol::TProtoco break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast778; - xfer += iprot->readI32(ecast778); - this->type = (FileMetadataExprType::type)ecast778; + int32_t ecast779; + xfer += iprot->readI32(ecast779); + this->type = (FileMetadataExprType::type)ecast779; this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -19210,10 +19268,10 @@ uint32_t GetFileMetadataByExprRequest::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter779; - for (_iter779 = this->fileIds.begin(); _iter779 != this->fileIds.end(); ++_iter779) + std::vector ::const_iterator _iter780; + for (_iter780 = this->fileIds.begin(); _iter780 != this->fileIds.end(); ++_iter780) { - xfer += oprot->writeI64((*_iter779)); + xfer += oprot->writeI64((*_iter780)); } xfer += oprot->writeListEnd(); } @@ -19247,19 +19305,19 @@ void swap(GetFileMetadataByExprRequest &a, GetFileMetadataByExprRequest &b) { swap(a.__isset, b.__isset); } -GetFileMetadataByExprRequest::GetFileMetadataByExprRequest(const GetFileMetadataByExprRequest& other780) { - fileIds = other780.fileIds; - expr = other780.expr; - doGetFooters = other780.doGetFooters; - type = other780.type; - __isset = other780.__isset; -} -GetFileMetadataByExprRequest& GetFileMetadataByExprRequest::operator=(const GetFileMetadataByExprRequest& other781) { +GetFileMetadataByExprRequest::GetFileMetadataByExprRequest(const GetFileMetadataByExprRequest& other781) { fileIds = other781.fileIds; expr = other781.expr; doGetFooters = other781.doGetFooters; type = other781.type; __isset = other781.__isset; +} +GetFileMetadataByExprRequest& GetFileMetadataByExprRequest::operator=(const GetFileMetadataByExprRequest& other782) { + fileIds = other782.fileIds; + expr = other782.expr; + doGetFooters = other782.doGetFooters; + type = other782.type; + __isset = other782.__isset; return *this; } void GetFileMetadataByExprRequest::printTo(std::ostream& out) const { @@ -19312,17 +19370,17 @@ uint32_t GetFileMetadataResult::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->metadata.clear(); - uint32_t _size782; - ::apache::thrift::protocol::TType _ktype783; - ::apache::thrift::protocol::TType _vtype784; - xfer += iprot->readMapBegin(_ktype783, _vtype784, _size782); - uint32_t _i786; - for (_i786 = 0; _i786 < _size782; ++_i786) + uint32_t _size783; + ::apache::thrift::protocol::TType _ktype784; + ::apache::thrift::protocol::TType _vtype785; + xfer += iprot->readMapBegin(_ktype784, _vtype785, _size783); + uint32_t _i787; + for (_i787 = 0; _i787 < _size783; ++_i787) { - int64_t _key787; - xfer += iprot->readI64(_key787); - std::string& _val788 = this->metadata[_key787]; - xfer += iprot->readBinary(_val788); + int64_t _key788; + xfer += iprot->readI64(_key788); + std::string& _val789 = this->metadata[_key788]; + xfer += iprot->readBinary(_val789); } xfer += iprot->readMapEnd(); } @@ -19363,11 +19421,11 @@ uint32_t GetFileMetadataResult::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRING, static_cast(this->metadata.size())); - std::map ::const_iterator _iter789; - for (_iter789 = this->metadata.begin(); _iter789 != this->metadata.end(); ++_iter789) + std::map ::const_iterator _iter790; + for (_iter790 = this->metadata.begin(); _iter790 != this->metadata.end(); ++_iter790) { - xfer += oprot->writeI64(_iter789->first); - xfer += oprot->writeBinary(_iter789->second); + xfer += oprot->writeI64(_iter790->first); + xfer += oprot->writeBinary(_iter790->second); } xfer += oprot->writeMapEnd(); } @@ -19388,13 +19446,13 @@ void swap(GetFileMetadataResult &a, GetFileMetadataResult &b) { swap(a.isSupported, b.isSupported); } -GetFileMetadataResult::GetFileMetadataResult(const GetFileMetadataResult& other790) { - metadata = other790.metadata; - isSupported = other790.isSupported; -} -GetFileMetadataResult& GetFileMetadataResult::operator=(const GetFileMetadataResult& other791) { +GetFileMetadataResult::GetFileMetadataResult(const GetFileMetadataResult& other791) { metadata = other791.metadata; isSupported = other791.isSupported; +} +GetFileMetadataResult& GetFileMetadataResult::operator=(const GetFileMetadataResult& other792) { + metadata = other792.metadata; + isSupported = other792.isSupported; return *this; } void GetFileMetadataResult::printTo(std::ostream& out) const { @@ -19440,14 +19498,14 @@ uint32_t GetFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size792; - ::apache::thrift::protocol::TType _etype795; - xfer += iprot->readListBegin(_etype795, _size792); - this->fileIds.resize(_size792); - uint32_t _i796; - for (_i796 = 0; _i796 < _size792; ++_i796) + uint32_t _size793; + ::apache::thrift::protocol::TType _etype796; + xfer += iprot->readListBegin(_etype796, _size793); + this->fileIds.resize(_size793); + uint32_t _i797; + for (_i797 = 0; _i797 < _size793; ++_i797) { - xfer += iprot->readI64(this->fileIds[_i796]); + xfer += iprot->readI64(this->fileIds[_i797]); } xfer += iprot->readListEnd(); } @@ -19478,10 +19536,10 @@ uint32_t GetFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter797; - for (_iter797 = this->fileIds.begin(); _iter797 != this->fileIds.end(); ++_iter797) + std::vector ::const_iterator _iter798; + for (_iter798 = this->fileIds.begin(); _iter798 != this->fileIds.end(); ++_iter798) { - xfer += oprot->writeI64((*_iter797)); + xfer += oprot->writeI64((*_iter798)); } xfer += oprot->writeListEnd(); } @@ -19497,11 +19555,11 @@ void swap(GetFileMetadataRequest &a, GetFileMetadataRequest &b) { swap(a.fileIds, b.fileIds); } -GetFileMetadataRequest::GetFileMetadataRequest(const GetFileMetadataRequest& other798) { - fileIds = other798.fileIds; -} -GetFileMetadataRequest& GetFileMetadataRequest::operator=(const GetFileMetadataRequest& other799) { +GetFileMetadataRequest::GetFileMetadataRequest(const GetFileMetadataRequest& other799) { fileIds = other799.fileIds; +} +GetFileMetadataRequest& GetFileMetadataRequest::operator=(const GetFileMetadataRequest& other800) { + fileIds = other800.fileIds; return *this; } void GetFileMetadataRequest::printTo(std::ostream& out) const { @@ -19560,11 +19618,11 @@ void swap(PutFileMetadataResult &a, PutFileMetadataResult &b) { (void) b; } -PutFileMetadataResult::PutFileMetadataResult(const PutFileMetadataResult& other800) { - (void) other800; -} -PutFileMetadataResult& PutFileMetadataResult::operator=(const PutFileMetadataResult& other801) { +PutFileMetadataResult::PutFileMetadataResult(const PutFileMetadataResult& other801) { (void) other801; +} +PutFileMetadataResult& PutFileMetadataResult::operator=(const PutFileMetadataResult& other802) { + (void) other802; return *this; } void PutFileMetadataResult::printTo(std::ostream& out) const { @@ -19618,14 +19676,14 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size802; - ::apache::thrift::protocol::TType _etype805; - xfer += iprot->readListBegin(_etype805, _size802); - this->fileIds.resize(_size802); - uint32_t _i806; - for (_i806 = 0; _i806 < _size802; ++_i806) + uint32_t _size803; + ::apache::thrift::protocol::TType _etype806; + xfer += iprot->readListBegin(_etype806, _size803); + this->fileIds.resize(_size803); + uint32_t _i807; + for (_i807 = 0; _i807 < _size803; ++_i807) { - xfer += iprot->readI64(this->fileIds[_i806]); + xfer += iprot->readI64(this->fileIds[_i807]); } xfer += iprot->readListEnd(); } @@ -19638,14 +19696,14 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->metadata.clear(); - uint32_t _size807; - ::apache::thrift::protocol::TType _etype810; - xfer += iprot->readListBegin(_etype810, _size807); - this->metadata.resize(_size807); - uint32_t _i811; - for (_i811 = 0; _i811 < _size807; ++_i811) + uint32_t _size808; + ::apache::thrift::protocol::TType _etype811; + xfer += iprot->readListBegin(_etype811, _size808); + this->metadata.resize(_size808); + uint32_t _i812; + for (_i812 = 0; _i812 < _size808; ++_i812) { - xfer += iprot->readBinary(this->metadata[_i811]); + xfer += iprot->readBinary(this->metadata[_i812]); } xfer += iprot->readListEnd(); } @@ -19656,9 +19714,9 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast812; - xfer += iprot->readI32(ecast812); - this->type = (FileMetadataExprType::type)ecast812; + int32_t ecast813; + xfer += iprot->readI32(ecast813); + this->type = (FileMetadataExprType::type)ecast813; this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -19688,10 +19746,10 @@ uint32_t PutFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter813; - for (_iter813 = this->fileIds.begin(); _iter813 != this->fileIds.end(); ++_iter813) + std::vector ::const_iterator _iter814; + for (_iter814 = this->fileIds.begin(); _iter814 != this->fileIds.end(); ++_iter814) { - xfer += oprot->writeI64((*_iter813)); + xfer += oprot->writeI64((*_iter814)); } xfer += oprot->writeListEnd(); } @@ -19700,10 +19758,10 @@ uint32_t PutFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->metadata.size())); - std::vector ::const_iterator _iter814; - for (_iter814 = this->metadata.begin(); _iter814 != this->metadata.end(); ++_iter814) + std::vector ::const_iterator _iter815; + for (_iter815 = this->metadata.begin(); _iter815 != this->metadata.end(); ++_iter815) { - xfer += oprot->writeBinary((*_iter814)); + xfer += oprot->writeBinary((*_iter815)); } xfer += oprot->writeListEnd(); } @@ -19727,17 +19785,17 @@ void swap(PutFileMetadataRequest &a, PutFileMetadataRequest &b) { swap(a.__isset, b.__isset); } -PutFileMetadataRequest::PutFileMetadataRequest(const PutFileMetadataRequest& other815) { - fileIds = other815.fileIds; - metadata = other815.metadata; - type = other815.type; - __isset = other815.__isset; -} -PutFileMetadataRequest& PutFileMetadataRequest::operator=(const PutFileMetadataRequest& other816) { +PutFileMetadataRequest::PutFileMetadataRequest(const PutFileMetadataRequest& other816) { fileIds = other816.fileIds; metadata = other816.metadata; type = other816.type; __isset = other816.__isset; +} +PutFileMetadataRequest& PutFileMetadataRequest::operator=(const PutFileMetadataRequest& other817) { + fileIds = other817.fileIds; + metadata = other817.metadata; + type = other817.type; + __isset = other817.__isset; return *this; } void PutFileMetadataRequest::printTo(std::ostream& out) const { @@ -19798,11 +19856,11 @@ void swap(ClearFileMetadataResult &a, ClearFileMetadataResult &b) { (void) b; } -ClearFileMetadataResult::ClearFileMetadataResult(const ClearFileMetadataResult& other817) { - (void) other817; -} -ClearFileMetadataResult& ClearFileMetadataResult::operator=(const ClearFileMetadataResult& other818) { +ClearFileMetadataResult::ClearFileMetadataResult(const ClearFileMetadataResult& other818) { (void) other818; +} +ClearFileMetadataResult& ClearFileMetadataResult::operator=(const ClearFileMetadataResult& other819) { + (void) other819; return *this; } void ClearFileMetadataResult::printTo(std::ostream& out) const { @@ -19846,14 +19904,14 @@ uint32_t ClearFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size819; - ::apache::thrift::protocol::TType _etype822; - xfer += iprot->readListBegin(_etype822, _size819); - this->fileIds.resize(_size819); - uint32_t _i823; - for (_i823 = 0; _i823 < _size819; ++_i823) + uint32_t _size820; + ::apache::thrift::protocol::TType _etype823; + xfer += iprot->readListBegin(_etype823, _size820); + this->fileIds.resize(_size820); + uint32_t _i824; + for (_i824 = 0; _i824 < _size820; ++_i824) { - xfer += iprot->readI64(this->fileIds[_i823]); + xfer += iprot->readI64(this->fileIds[_i824]); } xfer += iprot->readListEnd(); } @@ -19884,10 +19942,10 @@ uint32_t ClearFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter824; - for (_iter824 = this->fileIds.begin(); _iter824 != this->fileIds.end(); ++_iter824) + std::vector ::const_iterator _iter825; + for (_iter825 = this->fileIds.begin(); _iter825 != this->fileIds.end(); ++_iter825) { - xfer += oprot->writeI64((*_iter824)); + xfer += oprot->writeI64((*_iter825)); } xfer += oprot->writeListEnd(); } @@ -19903,11 +19961,11 @@ void swap(ClearFileMetadataRequest &a, ClearFileMetadataRequest &b) { swap(a.fileIds, b.fileIds); } -ClearFileMetadataRequest::ClearFileMetadataRequest(const ClearFileMetadataRequest& other825) { - fileIds = other825.fileIds; -} -ClearFileMetadataRequest& ClearFileMetadataRequest::operator=(const ClearFileMetadataRequest& other826) { +ClearFileMetadataRequest::ClearFileMetadataRequest(const ClearFileMetadataRequest& other826) { fileIds = other826.fileIds; +} +ClearFileMetadataRequest& ClearFileMetadataRequest::operator=(const ClearFileMetadataRequest& other827) { + fileIds = other827.fileIds; return *this; } void ClearFileMetadataRequest::printTo(std::ostream& out) const { @@ -19989,11 +20047,11 @@ void swap(CacheFileMetadataResult &a, CacheFileMetadataResult &b) { swap(a.isSupported, b.isSupported); } -CacheFileMetadataResult::CacheFileMetadataResult(const CacheFileMetadataResult& other827) { - isSupported = other827.isSupported; -} -CacheFileMetadataResult& CacheFileMetadataResult::operator=(const CacheFileMetadataResult& other828) { +CacheFileMetadataResult::CacheFileMetadataResult(const CacheFileMetadataResult& other828) { isSupported = other828.isSupported; +} +CacheFileMetadataResult& CacheFileMetadataResult::operator=(const CacheFileMetadataResult& other829) { + isSupported = other829.isSupported; return *this; } void CacheFileMetadataResult::printTo(std::ostream& out) const { @@ -20134,19 +20192,19 @@ void swap(CacheFileMetadataRequest &a, CacheFileMetadataRequest &b) { swap(a.__isset, b.__isset); } -CacheFileMetadataRequest::CacheFileMetadataRequest(const CacheFileMetadataRequest& other829) { - dbName = other829.dbName; - tblName = other829.tblName; - partName = other829.partName; - isAllParts = other829.isAllParts; - __isset = other829.__isset; -} -CacheFileMetadataRequest& CacheFileMetadataRequest::operator=(const CacheFileMetadataRequest& other830) { +CacheFileMetadataRequest::CacheFileMetadataRequest(const CacheFileMetadataRequest& other830) { dbName = other830.dbName; tblName = other830.tblName; partName = other830.partName; isAllParts = other830.isAllParts; __isset = other830.__isset; +} +CacheFileMetadataRequest& CacheFileMetadataRequest::operator=(const CacheFileMetadataRequest& other831) { + dbName = other831.dbName; + tblName = other831.tblName; + partName = other831.partName; + isAllParts = other831.isAllParts; + __isset = other831.__isset; return *this; } void CacheFileMetadataRequest::printTo(std::ostream& out) const { @@ -20194,14 +20252,14 @@ uint32_t GetAllFunctionsResponse::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->functions.clear(); - uint32_t _size831; - ::apache::thrift::protocol::TType _etype834; - xfer += iprot->readListBegin(_etype834, _size831); - this->functions.resize(_size831); - uint32_t _i835; - for (_i835 = 0; _i835 < _size831; ++_i835) + uint32_t _size832; + ::apache::thrift::protocol::TType _etype835; + xfer += iprot->readListBegin(_etype835, _size832); + this->functions.resize(_size832); + uint32_t _i836; + for (_i836 = 0; _i836 < _size832; ++_i836) { - xfer += this->functions[_i835].read(iprot); + xfer += this->functions[_i836].read(iprot); } xfer += iprot->readListEnd(); } @@ -20231,10 +20289,10 @@ uint32_t GetAllFunctionsResponse::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("functions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->functions.size())); - std::vector ::const_iterator _iter836; - for (_iter836 = this->functions.begin(); _iter836 != this->functions.end(); ++_iter836) + std::vector ::const_iterator _iter837; + for (_iter837 = this->functions.begin(); _iter837 != this->functions.end(); ++_iter837) { - xfer += (*_iter836).write(oprot); + xfer += (*_iter837).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20251,13 +20309,13 @@ void swap(GetAllFunctionsResponse &a, GetAllFunctionsResponse &b) { swap(a.__isset, b.__isset); } -GetAllFunctionsResponse::GetAllFunctionsResponse(const GetAllFunctionsResponse& other837) { - functions = other837.functions; - __isset = other837.__isset; -} -GetAllFunctionsResponse& GetAllFunctionsResponse::operator=(const GetAllFunctionsResponse& other838) { +GetAllFunctionsResponse::GetAllFunctionsResponse(const GetAllFunctionsResponse& other838) { functions = other838.functions; __isset = other838.__isset; +} +GetAllFunctionsResponse& GetAllFunctionsResponse::operator=(const GetAllFunctionsResponse& other839) { + functions = other839.functions; + __isset = other839.__isset; return *this; } void GetAllFunctionsResponse::printTo(std::ostream& out) const { @@ -20302,16 +20360,16 @@ uint32_t ClientCapabilities::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->values.clear(); - uint32_t _size839; - ::apache::thrift::protocol::TType _etype842; - xfer += iprot->readListBegin(_etype842, _size839); - this->values.resize(_size839); - uint32_t _i843; - for (_i843 = 0; _i843 < _size839; ++_i843) + uint32_t _size840; + ::apache::thrift::protocol::TType _etype843; + xfer += iprot->readListBegin(_etype843, _size840); + this->values.resize(_size840); + uint32_t _i844; + for (_i844 = 0; _i844 < _size840; ++_i844) { - int32_t ecast844; - xfer += iprot->readI32(ecast844); - this->values[_i843] = (ClientCapability::type)ecast844; + int32_t ecast845; + xfer += iprot->readI32(ecast845); + this->values[_i844] = (ClientCapability::type)ecast845; } xfer += iprot->readListEnd(); } @@ -20342,10 +20400,10 @@ uint32_t ClientCapabilities::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("values", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I32, static_cast(this->values.size())); - std::vector ::const_iterator _iter845; - for (_iter845 = this->values.begin(); _iter845 != this->values.end(); ++_iter845) + std::vector ::const_iterator _iter846; + for (_iter846 = this->values.begin(); _iter846 != this->values.end(); ++_iter846) { - xfer += oprot->writeI32((int32_t)(*_iter845)); + xfer += oprot->writeI32((int32_t)(*_iter846)); } xfer += oprot->writeListEnd(); } @@ -20361,11 +20419,11 @@ void swap(ClientCapabilities &a, ClientCapabilities &b) { swap(a.values, b.values); } -ClientCapabilities::ClientCapabilities(const ClientCapabilities& other846) { - values = other846.values; -} -ClientCapabilities& ClientCapabilities::operator=(const ClientCapabilities& other847) { +ClientCapabilities::ClientCapabilities(const ClientCapabilities& other847) { values = other847.values; +} +ClientCapabilities& ClientCapabilities::operator=(const ClientCapabilities& other848) { + values = other848.values; return *this; } void ClientCapabilities::printTo(std::ostream& out) const { @@ -20487,17 +20545,17 @@ void swap(GetTableRequest &a, GetTableRequest &b) { swap(a.__isset, b.__isset); } -GetTableRequest::GetTableRequest(const GetTableRequest& other848) { - dbName = other848.dbName; - tblName = other848.tblName; - capabilities = other848.capabilities; - __isset = other848.__isset; -} -GetTableRequest& GetTableRequest::operator=(const GetTableRequest& other849) { +GetTableRequest::GetTableRequest(const GetTableRequest& other849) { dbName = other849.dbName; tblName = other849.tblName; capabilities = other849.capabilities; __isset = other849.__isset; +} +GetTableRequest& GetTableRequest::operator=(const GetTableRequest& other850) { + dbName = other850.dbName; + tblName = other850.tblName; + capabilities = other850.capabilities; + __isset = other850.__isset; return *this; } void GetTableRequest::printTo(std::ostream& out) const { @@ -20581,11 +20639,11 @@ void swap(GetTableResult &a, GetTableResult &b) { swap(a.table, b.table); } -GetTableResult::GetTableResult(const GetTableResult& other850) { - table = other850.table; -} -GetTableResult& GetTableResult::operator=(const GetTableResult& other851) { +GetTableResult::GetTableResult(const GetTableResult& other851) { table = other851.table; +} +GetTableResult& GetTableResult::operator=(const GetTableResult& other852) { + table = other852.table; return *this; } void GetTableResult::printTo(std::ostream& out) const { @@ -20648,14 +20706,14 @@ uint32_t GetTablesRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tblNames.clear(); - uint32_t _size852; - ::apache::thrift::protocol::TType _etype855; - xfer += iprot->readListBegin(_etype855, _size852); - this->tblNames.resize(_size852); - uint32_t _i856; - for (_i856 = 0; _i856 < _size852; ++_i856) + uint32_t _size853; + ::apache::thrift::protocol::TType _etype856; + xfer += iprot->readListBegin(_etype856, _size853); + this->tblNames.resize(_size853); + uint32_t _i857; + for (_i857 = 0; _i857 < _size853; ++_i857) { - xfer += iprot->readString(this->tblNames[_i856]); + xfer += iprot->readString(this->tblNames[_i857]); } xfer += iprot->readListEnd(); } @@ -20699,10 +20757,10 @@ uint32_t GetTablesRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tblNames", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tblNames.size())); - std::vector ::const_iterator _iter857; - for (_iter857 = this->tblNames.begin(); _iter857 != this->tblNames.end(); ++_iter857) + std::vector ::const_iterator _iter858; + for (_iter858 = this->tblNames.begin(); _iter858 != this->tblNames.end(); ++_iter858) { - xfer += oprot->writeString((*_iter857)); + xfer += oprot->writeString((*_iter858)); } xfer += oprot->writeListEnd(); } @@ -20726,17 +20784,17 @@ void swap(GetTablesRequest &a, GetTablesRequest &b) { swap(a.__isset, b.__isset); } -GetTablesRequest::GetTablesRequest(const GetTablesRequest& other858) { - dbName = other858.dbName; - tblNames = other858.tblNames; - capabilities = other858.capabilities; - __isset = other858.__isset; -} -GetTablesRequest& GetTablesRequest::operator=(const GetTablesRequest& other859) { +GetTablesRequest::GetTablesRequest(const GetTablesRequest& other859) { dbName = other859.dbName; tblNames = other859.tblNames; capabilities = other859.capabilities; __isset = other859.__isset; +} +GetTablesRequest& GetTablesRequest::operator=(const GetTablesRequest& other860) { + dbName = other860.dbName; + tblNames = other860.tblNames; + capabilities = other860.capabilities; + __isset = other860.__isset; return *this; } void GetTablesRequest::printTo(std::ostream& out) const { @@ -20783,14 +20841,14 @@ uint32_t GetTablesResult::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tables.clear(); - uint32_t _size860; - ::apache::thrift::protocol::TType _etype863; - xfer += iprot->readListBegin(_etype863, _size860); - this->tables.resize(_size860); - uint32_t _i864; - for (_i864 = 0; _i864 < _size860; ++_i864) + uint32_t _size861; + ::apache::thrift::protocol::TType _etype864; + xfer += iprot->readListBegin(_etype864, _size861); + this->tables.resize(_size861); + uint32_t _i865; + for (_i865 = 0; _i865 < _size861; ++_i865) { - xfer += this->tables[_i864].read(iprot); + xfer += this->tables[_i865].read(iprot); } xfer += iprot->readListEnd(); } @@ -20821,10 +20879,10 @@ uint32_t GetTablesResult::write(::apache::thrift::protocol::TProtocol* oprot) co xfer += oprot->writeFieldBegin("tables", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->tables.size())); - std::vector
::const_iterator _iter865; - for (_iter865 = this->tables.begin(); _iter865 != this->tables.end(); ++_iter865) + std::vector
::const_iterator _iter866; + for (_iter866 = this->tables.begin(); _iter866 != this->tables.end(); ++_iter866) { - xfer += (*_iter865).write(oprot); + xfer += (*_iter866).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20840,11 +20898,11 @@ void swap(GetTablesResult &a, GetTablesResult &b) { swap(a.tables, b.tables); } -GetTablesResult::GetTablesResult(const GetTablesResult& other866) { - tables = other866.tables; -} -GetTablesResult& GetTablesResult::operator=(const GetTablesResult& other867) { +GetTablesResult::GetTablesResult(const GetTablesResult& other867) { tables = other867.tables; +} +GetTablesResult& GetTablesResult::operator=(const GetTablesResult& other868) { + tables = other868.tables; return *this; } void GetTablesResult::printTo(std::ostream& out) const { @@ -20946,13 +21004,13 @@ void swap(CmRecycleRequest &a, CmRecycleRequest &b) { swap(a.purge, b.purge); } -CmRecycleRequest::CmRecycleRequest(const CmRecycleRequest& other868) { - dataPath = other868.dataPath; - purge = other868.purge; -} -CmRecycleRequest& CmRecycleRequest::operator=(const CmRecycleRequest& other869) { +CmRecycleRequest::CmRecycleRequest(const CmRecycleRequest& other869) { dataPath = other869.dataPath; purge = other869.purge; +} +CmRecycleRequest& CmRecycleRequest::operator=(const CmRecycleRequest& other870) { + dataPath = other870.dataPath; + purge = other870.purge; return *this; } void CmRecycleRequest::printTo(std::ostream& out) const { @@ -21012,11 +21070,11 @@ void swap(CmRecycleResponse &a, CmRecycleResponse &b) { (void) b; } -CmRecycleResponse::CmRecycleResponse(const CmRecycleResponse& other870) { - (void) other870; -} -CmRecycleResponse& CmRecycleResponse::operator=(const CmRecycleResponse& other871) { +CmRecycleResponse::CmRecycleResponse(const CmRecycleResponse& other871) { (void) other871; +} +CmRecycleResponse& CmRecycleResponse::operator=(const CmRecycleResponse& other872) { + (void) other872; return *this; } void CmRecycleResponse::printTo(std::ostream& out) const { @@ -21157,19 +21215,19 @@ void swap(TableMeta &a, TableMeta &b) { swap(a.__isset, b.__isset); } -TableMeta::TableMeta(const TableMeta& other872) { - dbName = other872.dbName; - tableName = other872.tableName; - tableType = other872.tableType; - comments = other872.comments; - __isset = other872.__isset; -} -TableMeta& TableMeta::operator=(const TableMeta& other873) { +TableMeta::TableMeta(const TableMeta& other873) { dbName = other873.dbName; tableName = other873.tableName; tableType = other873.tableType; comments = other873.comments; __isset = other873.__isset; +} +TableMeta& TableMeta::operator=(const TableMeta& other874) { + dbName = other874.dbName; + tableName = other874.tableName; + tableType = other874.tableType; + comments = other874.comments; + __isset = other874.__isset; return *this; } void TableMeta::printTo(std::ostream& out) const { @@ -21235,15 +21293,15 @@ uint32_t Materialization::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_SET) { { this->tablesUsed.clear(); - uint32_t _size874; - ::apache::thrift::protocol::TType _etype877; - xfer += iprot->readSetBegin(_etype877, _size874); - uint32_t _i878; - for (_i878 = 0; _i878 < _size874; ++_i878) + uint32_t _size875; + ::apache::thrift::protocol::TType _etype878; + xfer += iprot->readSetBegin(_etype878, _size875); + uint32_t _i879; + for (_i879 = 0; _i879 < _size875; ++_i879) { - std::string _elem879; - xfer += iprot->readString(_elem879); - this->tablesUsed.insert(_elem879); + std::string _elem880; + xfer += iprot->readString(_elem880); + this->tablesUsed.insert(_elem880); } xfer += iprot->readSetEnd(); } @@ -21290,10 +21348,10 @@ uint32_t Materialization::write(::apache::thrift::protocol::TProtocol* oprot) co xfer += oprot->writeFieldBegin("tablesUsed", ::apache::thrift::protocol::T_SET, 2); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tablesUsed.size())); - std::set ::const_iterator _iter880; - for (_iter880 = this->tablesUsed.begin(); _iter880 != this->tablesUsed.end(); ++_iter880) + std::set ::const_iterator _iter881; + for (_iter881 = this->tablesUsed.begin(); _iter881 != this->tablesUsed.end(); ++_iter881) { - xfer += oprot->writeString((*_iter880)); + xfer += oprot->writeString((*_iter881)); } xfer += oprot->writeSetEnd(); } @@ -21315,15 +21373,15 @@ void swap(Materialization &a, Materialization &b) { swap(a.invalidationTime, b.invalidationTime); } -Materialization::Materialization(const Materialization& other881) { - materializationTable = other881.materializationTable; - tablesUsed = other881.tablesUsed; - invalidationTime = other881.invalidationTime; -} -Materialization& Materialization::operator=(const Materialization& other882) { +Materialization::Materialization(const Materialization& other882) { materializationTable = other882.materializationTable; tablesUsed = other882.tablesUsed; invalidationTime = other882.invalidationTime; +} +Materialization& Materialization::operator=(const Materialization& other883) { + materializationTable = other883.materializationTable; + tablesUsed = other883.tablesUsed; + invalidationTime = other883.invalidationTime; return *this; } void Materialization::printTo(std::ostream& out) const { @@ -21391,9 +21449,9 @@ uint32_t WMResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast883; - xfer += iprot->readI32(ecast883); - this->status = (WMResourcePlanStatus::type)ecast883; + int32_t ecast884; + xfer += iprot->readI32(ecast884); + this->status = (WMResourcePlanStatus::type)ecast884; this->__isset.status = true; } else { xfer += iprot->skip(ftype); @@ -21467,19 +21525,19 @@ void swap(WMResourcePlan &a, WMResourcePlan &b) { swap(a.__isset, b.__isset); } -WMResourcePlan::WMResourcePlan(const WMResourcePlan& other884) { - name = other884.name; - status = other884.status; - queryParallelism = other884.queryParallelism; - defaultPoolPath = other884.defaultPoolPath; - __isset = other884.__isset; -} -WMResourcePlan& WMResourcePlan::operator=(const WMResourcePlan& other885) { +WMResourcePlan::WMResourcePlan(const WMResourcePlan& other885) { name = other885.name; status = other885.status; queryParallelism = other885.queryParallelism; defaultPoolPath = other885.defaultPoolPath; __isset = other885.__isset; +} +WMResourcePlan& WMResourcePlan::operator=(const WMResourcePlan& other886) { + name = other886.name; + status = other886.status; + queryParallelism = other886.queryParallelism; + defaultPoolPath = other886.defaultPoolPath; + __isset = other886.__isset; return *this; } void WMResourcePlan::printTo(std::ostream& out) const { @@ -21558,9 +21616,9 @@ uint32_t WMNullableResourcePlan::read(::apache::thrift::protocol::TProtocol* ipr break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast886; - xfer += iprot->readI32(ecast886); - this->status = (WMResourcePlanStatus::type)ecast886; + int32_t ecast887; + xfer += iprot->readI32(ecast887); + this->status = (WMResourcePlanStatus::type)ecast887; this->__isset.status = true; } else { xfer += iprot->skip(ftype); @@ -21662,16 +21720,7 @@ void swap(WMNullableResourcePlan &a, WMNullableResourcePlan &b) { swap(a.__isset, b.__isset); } -WMNullableResourcePlan::WMNullableResourcePlan(const WMNullableResourcePlan& other887) { - name = other887.name; - status = other887.status; - queryParallelism = other887.queryParallelism; - isSetQueryParallelism = other887.isSetQueryParallelism; - defaultPoolPath = other887.defaultPoolPath; - isSetDefaultPoolPath = other887.isSetDefaultPoolPath; - __isset = other887.__isset; -} -WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResourcePlan& other888) { +WMNullableResourcePlan::WMNullableResourcePlan(const WMNullableResourcePlan& other888) { name = other888.name; status = other888.status; queryParallelism = other888.queryParallelism; @@ -21679,6 +21728,15 @@ WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResour defaultPoolPath = other888.defaultPoolPath; isSetDefaultPoolPath = other888.isSetDefaultPoolPath; __isset = other888.__isset; +} +WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResourcePlan& other889) { + name = other889.name; + status = other889.status; + queryParallelism = other889.queryParallelism; + isSetQueryParallelism = other889.isSetQueryParallelism; + defaultPoolPath = other889.defaultPoolPath; + isSetDefaultPoolPath = other889.isSetDefaultPoolPath; + __isset = other889.__isset; return *this; } void WMNullableResourcePlan::printTo(std::ostream& out) const { @@ -21843,21 +21901,21 @@ void swap(WMPool &a, WMPool &b) { swap(a.__isset, b.__isset); } -WMPool::WMPool(const WMPool& other889) { - resourcePlanName = other889.resourcePlanName; - poolPath = other889.poolPath; - allocFraction = other889.allocFraction; - queryParallelism = other889.queryParallelism; - schedulingPolicy = other889.schedulingPolicy; - __isset = other889.__isset; -} -WMPool& WMPool::operator=(const WMPool& other890) { +WMPool::WMPool(const WMPool& other890) { resourcePlanName = other890.resourcePlanName; poolPath = other890.poolPath; allocFraction = other890.allocFraction; queryParallelism = other890.queryParallelism; schedulingPolicy = other890.schedulingPolicy; __isset = other890.__isset; +} +WMPool& WMPool::operator=(const WMPool& other891) { + resourcePlanName = other891.resourcePlanName; + poolPath = other891.poolPath; + allocFraction = other891.allocFraction; + queryParallelism = other891.queryParallelism; + schedulingPolicy = other891.schedulingPolicy; + __isset = other891.__isset; return *this; } void WMPool::printTo(std::ostream& out) const { @@ -22040,16 +22098,7 @@ void swap(WMNullablePool &a, WMNullablePool &b) { swap(a.__isset, b.__isset); } -WMNullablePool::WMNullablePool(const WMNullablePool& other891) { - resourcePlanName = other891.resourcePlanName; - poolPath = other891.poolPath; - allocFraction = other891.allocFraction; - queryParallelism = other891.queryParallelism; - schedulingPolicy = other891.schedulingPolicy; - isSetSchedulingPolicy = other891.isSetSchedulingPolicy; - __isset = other891.__isset; -} -WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other892) { +WMNullablePool::WMNullablePool(const WMNullablePool& other892) { resourcePlanName = other892.resourcePlanName; poolPath = other892.poolPath; allocFraction = other892.allocFraction; @@ -22057,6 +22106,15 @@ WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other892) { schedulingPolicy = other892.schedulingPolicy; isSetSchedulingPolicy = other892.isSetSchedulingPolicy; __isset = other892.__isset; +} +WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other893) { + resourcePlanName = other893.resourcePlanName; + poolPath = other893.poolPath; + allocFraction = other893.allocFraction; + queryParallelism = other893.queryParallelism; + schedulingPolicy = other893.schedulingPolicy; + isSetSchedulingPolicy = other893.isSetSchedulingPolicy; + __isset = other893.__isset; return *this; } void WMNullablePool::printTo(std::ostream& out) const { @@ -22221,21 +22279,21 @@ void swap(WMTrigger &a, WMTrigger &b) { swap(a.__isset, b.__isset); } -WMTrigger::WMTrigger(const WMTrigger& other893) { - resourcePlanName = other893.resourcePlanName; - triggerName = other893.triggerName; - triggerExpression = other893.triggerExpression; - actionExpression = other893.actionExpression; - isInUnmanaged = other893.isInUnmanaged; - __isset = other893.__isset; -} -WMTrigger& WMTrigger::operator=(const WMTrigger& other894) { +WMTrigger::WMTrigger(const WMTrigger& other894) { resourcePlanName = other894.resourcePlanName; triggerName = other894.triggerName; triggerExpression = other894.triggerExpression; actionExpression = other894.actionExpression; isInUnmanaged = other894.isInUnmanaged; __isset = other894.__isset; +} +WMTrigger& WMTrigger::operator=(const WMTrigger& other895) { + resourcePlanName = other895.resourcePlanName; + triggerName = other895.triggerName; + triggerExpression = other895.triggerExpression; + actionExpression = other895.actionExpression; + isInUnmanaged = other895.isInUnmanaged; + __isset = other895.__isset; return *this; } void WMTrigger::printTo(std::ostream& out) const { @@ -22400,21 +22458,21 @@ void swap(WMMapping &a, WMMapping &b) { swap(a.__isset, b.__isset); } -WMMapping::WMMapping(const WMMapping& other895) { - resourcePlanName = other895.resourcePlanName; - entityType = other895.entityType; - entityName = other895.entityName; - poolPath = other895.poolPath; - ordering = other895.ordering; - __isset = other895.__isset; -} -WMMapping& WMMapping::operator=(const WMMapping& other896) { +WMMapping::WMMapping(const WMMapping& other896) { resourcePlanName = other896.resourcePlanName; entityType = other896.entityType; entityName = other896.entityName; poolPath = other896.poolPath; ordering = other896.ordering; __isset = other896.__isset; +} +WMMapping& WMMapping::operator=(const WMMapping& other897) { + resourcePlanName = other897.resourcePlanName; + entityType = other897.entityType; + entityName = other897.entityName; + poolPath = other897.poolPath; + ordering = other897.ordering; + __isset = other897.__isset; return *this; } void WMMapping::printTo(std::ostream& out) const { @@ -22520,13 +22578,13 @@ void swap(WMPoolTrigger &a, WMPoolTrigger &b) { swap(a.trigger, b.trigger); } -WMPoolTrigger::WMPoolTrigger(const WMPoolTrigger& other897) { - pool = other897.pool; - trigger = other897.trigger; -} -WMPoolTrigger& WMPoolTrigger::operator=(const WMPoolTrigger& other898) { +WMPoolTrigger::WMPoolTrigger(const WMPoolTrigger& other898) { pool = other898.pool; trigger = other898.trigger; +} +WMPoolTrigger& WMPoolTrigger::operator=(const WMPoolTrigger& other899) { + pool = other899.pool; + trigger = other899.trigger; return *this; } void WMPoolTrigger::printTo(std::ostream& out) const { @@ -22600,14 +22658,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->pools.clear(); - uint32_t _size899; - ::apache::thrift::protocol::TType _etype902; - xfer += iprot->readListBegin(_etype902, _size899); - this->pools.resize(_size899); - uint32_t _i903; - for (_i903 = 0; _i903 < _size899; ++_i903) + uint32_t _size900; + ::apache::thrift::protocol::TType _etype903; + xfer += iprot->readListBegin(_etype903, _size900); + this->pools.resize(_size900); + uint32_t _i904; + for (_i904 = 0; _i904 < _size900; ++_i904) { - xfer += this->pools[_i903].read(iprot); + xfer += this->pools[_i904].read(iprot); } xfer += iprot->readListEnd(); } @@ -22620,14 +22678,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->mappings.clear(); - uint32_t _size904; - ::apache::thrift::protocol::TType _etype907; - xfer += iprot->readListBegin(_etype907, _size904); - this->mappings.resize(_size904); - uint32_t _i908; - for (_i908 = 0; _i908 < _size904; ++_i908) + uint32_t _size905; + ::apache::thrift::protocol::TType _etype908; + xfer += iprot->readListBegin(_etype908, _size905); + this->mappings.resize(_size905); + uint32_t _i909; + for (_i909 = 0; _i909 < _size905; ++_i909) { - xfer += this->mappings[_i908].read(iprot); + xfer += this->mappings[_i909].read(iprot); } xfer += iprot->readListEnd(); } @@ -22640,14 +22698,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->triggers.clear(); - uint32_t _size909; - ::apache::thrift::protocol::TType _etype912; - xfer += iprot->readListBegin(_etype912, _size909); - this->triggers.resize(_size909); - uint32_t _i913; - for (_i913 = 0; _i913 < _size909; ++_i913) + uint32_t _size910; + ::apache::thrift::protocol::TType _etype913; + xfer += iprot->readListBegin(_etype913, _size910); + this->triggers.resize(_size910); + uint32_t _i914; + for (_i914 = 0; _i914 < _size910; ++_i914) { - xfer += this->triggers[_i913].read(iprot); + xfer += this->triggers[_i914].read(iprot); } xfer += iprot->readListEnd(); } @@ -22660,14 +22718,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->poolTriggers.clear(); - uint32_t _size914; - ::apache::thrift::protocol::TType _etype917; - xfer += iprot->readListBegin(_etype917, _size914); - this->poolTriggers.resize(_size914); - uint32_t _i918; - for (_i918 = 0; _i918 < _size914; ++_i918) + uint32_t _size915; + ::apache::thrift::protocol::TType _etype918; + xfer += iprot->readListBegin(_etype918, _size915); + this->poolTriggers.resize(_size915); + uint32_t _i919; + for (_i919 = 0; _i919 < _size915; ++_i919) { - xfer += this->poolTriggers[_i918].read(iprot); + xfer += this->poolTriggers[_i919].read(iprot); } xfer += iprot->readListEnd(); } @@ -22704,10 +22762,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("pools", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->pools.size())); - std::vector ::const_iterator _iter919; - for (_iter919 = this->pools.begin(); _iter919 != this->pools.end(); ++_iter919) + std::vector ::const_iterator _iter920; + for (_iter920 = this->pools.begin(); _iter920 != this->pools.end(); ++_iter920) { - xfer += (*_iter919).write(oprot); + xfer += (*_iter920).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22717,10 +22775,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("mappings", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->mappings.size())); - std::vector ::const_iterator _iter920; - for (_iter920 = this->mappings.begin(); _iter920 != this->mappings.end(); ++_iter920) + std::vector ::const_iterator _iter921; + for (_iter921 = this->mappings.begin(); _iter921 != this->mappings.end(); ++_iter921) { - xfer += (*_iter920).write(oprot); + xfer += (*_iter921).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22730,10 +22788,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("triggers", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->triggers.size())); - std::vector ::const_iterator _iter921; - for (_iter921 = this->triggers.begin(); _iter921 != this->triggers.end(); ++_iter921) + std::vector ::const_iterator _iter922; + for (_iter922 = this->triggers.begin(); _iter922 != this->triggers.end(); ++_iter922) { - xfer += (*_iter921).write(oprot); + xfer += (*_iter922).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22743,10 +22801,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("poolTriggers", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->poolTriggers.size())); - std::vector ::const_iterator _iter922; - for (_iter922 = this->poolTriggers.begin(); _iter922 != this->poolTriggers.end(); ++_iter922) + std::vector ::const_iterator _iter923; + for (_iter923 = this->poolTriggers.begin(); _iter923 != this->poolTriggers.end(); ++_iter923) { - xfer += (*_iter922).write(oprot); + xfer += (*_iter923).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22767,21 +22825,21 @@ void swap(WMFullResourcePlan &a, WMFullResourcePlan &b) { swap(a.__isset, b.__isset); } -WMFullResourcePlan::WMFullResourcePlan(const WMFullResourcePlan& other923) { - plan = other923.plan; - pools = other923.pools; - mappings = other923.mappings; - triggers = other923.triggers; - poolTriggers = other923.poolTriggers; - __isset = other923.__isset; -} -WMFullResourcePlan& WMFullResourcePlan::operator=(const WMFullResourcePlan& other924) { +WMFullResourcePlan::WMFullResourcePlan(const WMFullResourcePlan& other924) { plan = other924.plan; pools = other924.pools; mappings = other924.mappings; triggers = other924.triggers; poolTriggers = other924.poolTriggers; __isset = other924.__isset; +} +WMFullResourcePlan& WMFullResourcePlan::operator=(const WMFullResourcePlan& other925) { + plan = other925.plan; + pools = other925.pools; + mappings = other925.mappings; + triggers = other925.triggers; + poolTriggers = other925.poolTriggers; + __isset = other925.__isset; return *this; } void WMFullResourcePlan::printTo(std::ostream& out) const { @@ -22886,15 +22944,15 @@ void swap(WMCreateResourcePlanRequest &a, WMCreateResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMCreateResourcePlanRequest::WMCreateResourcePlanRequest(const WMCreateResourcePlanRequest& other925) { - resourcePlan = other925.resourcePlan; - copyFrom = other925.copyFrom; - __isset = other925.__isset; -} -WMCreateResourcePlanRequest& WMCreateResourcePlanRequest::operator=(const WMCreateResourcePlanRequest& other926) { +WMCreateResourcePlanRequest::WMCreateResourcePlanRequest(const WMCreateResourcePlanRequest& other926) { resourcePlan = other926.resourcePlan; copyFrom = other926.copyFrom; __isset = other926.__isset; +} +WMCreateResourcePlanRequest& WMCreateResourcePlanRequest::operator=(const WMCreateResourcePlanRequest& other927) { + resourcePlan = other927.resourcePlan; + copyFrom = other927.copyFrom; + __isset = other927.__isset; return *this; } void WMCreateResourcePlanRequest::printTo(std::ostream& out) const { @@ -22954,11 +23012,11 @@ void swap(WMCreateResourcePlanResponse &a, WMCreateResourcePlanResponse &b) { (void) b; } -WMCreateResourcePlanResponse::WMCreateResourcePlanResponse(const WMCreateResourcePlanResponse& other927) { - (void) other927; -} -WMCreateResourcePlanResponse& WMCreateResourcePlanResponse::operator=(const WMCreateResourcePlanResponse& other928) { +WMCreateResourcePlanResponse::WMCreateResourcePlanResponse(const WMCreateResourcePlanResponse& other928) { (void) other928; +} +WMCreateResourcePlanResponse& WMCreateResourcePlanResponse::operator=(const WMCreateResourcePlanResponse& other929) { + (void) other929; return *this; } void WMCreateResourcePlanResponse::printTo(std::ostream& out) const { @@ -23016,11 +23074,11 @@ void swap(WMGetActiveResourcePlanRequest &a, WMGetActiveResourcePlanRequest &b) (void) b; } -WMGetActiveResourcePlanRequest::WMGetActiveResourcePlanRequest(const WMGetActiveResourcePlanRequest& other929) { - (void) other929; -} -WMGetActiveResourcePlanRequest& WMGetActiveResourcePlanRequest::operator=(const WMGetActiveResourcePlanRequest& other930) { +WMGetActiveResourcePlanRequest::WMGetActiveResourcePlanRequest(const WMGetActiveResourcePlanRequest& other930) { (void) other930; +} +WMGetActiveResourcePlanRequest& WMGetActiveResourcePlanRequest::operator=(const WMGetActiveResourcePlanRequest& other931) { + (void) other931; return *this; } void WMGetActiveResourcePlanRequest::printTo(std::ostream& out) const { @@ -23101,13 +23159,13 @@ void swap(WMGetActiveResourcePlanResponse &a, WMGetActiveResourcePlanResponse &b swap(a.__isset, b.__isset); } -WMGetActiveResourcePlanResponse::WMGetActiveResourcePlanResponse(const WMGetActiveResourcePlanResponse& other931) { - resourcePlan = other931.resourcePlan; - __isset = other931.__isset; -} -WMGetActiveResourcePlanResponse& WMGetActiveResourcePlanResponse::operator=(const WMGetActiveResourcePlanResponse& other932) { +WMGetActiveResourcePlanResponse::WMGetActiveResourcePlanResponse(const WMGetActiveResourcePlanResponse& other932) { resourcePlan = other932.resourcePlan; __isset = other932.__isset; +} +WMGetActiveResourcePlanResponse& WMGetActiveResourcePlanResponse::operator=(const WMGetActiveResourcePlanResponse& other933) { + resourcePlan = other933.resourcePlan; + __isset = other933.__isset; return *this; } void WMGetActiveResourcePlanResponse::printTo(std::ostream& out) const { @@ -23189,13 +23247,13 @@ void swap(WMGetResourcePlanRequest &a, WMGetResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMGetResourcePlanRequest::WMGetResourcePlanRequest(const WMGetResourcePlanRequest& other933) { - resourcePlanName = other933.resourcePlanName; - __isset = other933.__isset; -} -WMGetResourcePlanRequest& WMGetResourcePlanRequest::operator=(const WMGetResourcePlanRequest& other934) { +WMGetResourcePlanRequest::WMGetResourcePlanRequest(const WMGetResourcePlanRequest& other934) { resourcePlanName = other934.resourcePlanName; __isset = other934.__isset; +} +WMGetResourcePlanRequest& WMGetResourcePlanRequest::operator=(const WMGetResourcePlanRequest& other935) { + resourcePlanName = other935.resourcePlanName; + __isset = other935.__isset; return *this; } void WMGetResourcePlanRequest::printTo(std::ostream& out) const { @@ -23277,13 +23335,13 @@ void swap(WMGetResourcePlanResponse &a, WMGetResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMGetResourcePlanResponse::WMGetResourcePlanResponse(const WMGetResourcePlanResponse& other935) { - resourcePlan = other935.resourcePlan; - __isset = other935.__isset; -} -WMGetResourcePlanResponse& WMGetResourcePlanResponse::operator=(const WMGetResourcePlanResponse& other936) { +WMGetResourcePlanResponse::WMGetResourcePlanResponse(const WMGetResourcePlanResponse& other936) { resourcePlan = other936.resourcePlan; __isset = other936.__isset; +} +WMGetResourcePlanResponse& WMGetResourcePlanResponse::operator=(const WMGetResourcePlanResponse& other937) { + resourcePlan = other937.resourcePlan; + __isset = other937.__isset; return *this; } void WMGetResourcePlanResponse::printTo(std::ostream& out) const { @@ -23342,11 +23400,11 @@ void swap(WMGetAllResourcePlanRequest &a, WMGetAllResourcePlanRequest &b) { (void) b; } -WMGetAllResourcePlanRequest::WMGetAllResourcePlanRequest(const WMGetAllResourcePlanRequest& other937) { - (void) other937; -} -WMGetAllResourcePlanRequest& WMGetAllResourcePlanRequest::operator=(const WMGetAllResourcePlanRequest& other938) { +WMGetAllResourcePlanRequest::WMGetAllResourcePlanRequest(const WMGetAllResourcePlanRequest& other938) { (void) other938; +} +WMGetAllResourcePlanRequest& WMGetAllResourcePlanRequest::operator=(const WMGetAllResourcePlanRequest& other939) { + (void) other939; return *this; } void WMGetAllResourcePlanRequest::printTo(std::ostream& out) const { @@ -23390,14 +23448,14 @@ uint32_t WMGetAllResourcePlanResponse::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->resourcePlans.clear(); - uint32_t _size939; - ::apache::thrift::protocol::TType _etype942; - xfer += iprot->readListBegin(_etype942, _size939); - this->resourcePlans.resize(_size939); - uint32_t _i943; - for (_i943 = 0; _i943 < _size939; ++_i943) + uint32_t _size940; + ::apache::thrift::protocol::TType _etype943; + xfer += iprot->readListBegin(_etype943, _size940); + this->resourcePlans.resize(_size940); + uint32_t _i944; + for (_i944 = 0; _i944 < _size940; ++_i944) { - xfer += this->resourcePlans[_i943].read(iprot); + xfer += this->resourcePlans[_i944].read(iprot); } xfer += iprot->readListEnd(); } @@ -23427,10 +23485,10 @@ uint32_t WMGetAllResourcePlanResponse::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("resourcePlans", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->resourcePlans.size())); - std::vector ::const_iterator _iter944; - for (_iter944 = this->resourcePlans.begin(); _iter944 != this->resourcePlans.end(); ++_iter944) + std::vector ::const_iterator _iter945; + for (_iter945 = this->resourcePlans.begin(); _iter945 != this->resourcePlans.end(); ++_iter945) { - xfer += (*_iter944).write(oprot); + xfer += (*_iter945).write(oprot); } xfer += oprot->writeListEnd(); } @@ -23447,13 +23505,13 @@ void swap(WMGetAllResourcePlanResponse &a, WMGetAllResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMGetAllResourcePlanResponse::WMGetAllResourcePlanResponse(const WMGetAllResourcePlanResponse& other945) { - resourcePlans = other945.resourcePlans; - __isset = other945.__isset; -} -WMGetAllResourcePlanResponse& WMGetAllResourcePlanResponse::operator=(const WMGetAllResourcePlanResponse& other946) { +WMGetAllResourcePlanResponse::WMGetAllResourcePlanResponse(const WMGetAllResourcePlanResponse& other946) { resourcePlans = other946.resourcePlans; __isset = other946.__isset; +} +WMGetAllResourcePlanResponse& WMGetAllResourcePlanResponse::operator=(const WMGetAllResourcePlanResponse& other947) { + resourcePlans = other947.resourcePlans; + __isset = other947.__isset; return *this; } void WMGetAllResourcePlanResponse::printTo(std::ostream& out) const { @@ -23611,21 +23669,21 @@ void swap(WMAlterResourcePlanRequest &a, WMAlterResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMAlterResourcePlanRequest::WMAlterResourcePlanRequest(const WMAlterResourcePlanRequest& other947) { - resourcePlanName = other947.resourcePlanName; - resourcePlan = other947.resourcePlan; - isEnableAndActivate = other947.isEnableAndActivate; - isForceDeactivate = other947.isForceDeactivate; - isReplace = other947.isReplace; - __isset = other947.__isset; -} -WMAlterResourcePlanRequest& WMAlterResourcePlanRequest::operator=(const WMAlterResourcePlanRequest& other948) { +WMAlterResourcePlanRequest::WMAlterResourcePlanRequest(const WMAlterResourcePlanRequest& other948) { resourcePlanName = other948.resourcePlanName; resourcePlan = other948.resourcePlan; isEnableAndActivate = other948.isEnableAndActivate; isForceDeactivate = other948.isForceDeactivate; isReplace = other948.isReplace; __isset = other948.__isset; +} +WMAlterResourcePlanRequest& WMAlterResourcePlanRequest::operator=(const WMAlterResourcePlanRequest& other949) { + resourcePlanName = other949.resourcePlanName; + resourcePlan = other949.resourcePlan; + isEnableAndActivate = other949.isEnableAndActivate; + isForceDeactivate = other949.isForceDeactivate; + isReplace = other949.isReplace; + __isset = other949.__isset; return *this; } void WMAlterResourcePlanRequest::printTo(std::ostream& out) const { @@ -23711,13 +23769,13 @@ void swap(WMAlterResourcePlanResponse &a, WMAlterResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMAlterResourcePlanResponse::WMAlterResourcePlanResponse(const WMAlterResourcePlanResponse& other949) { - fullResourcePlan = other949.fullResourcePlan; - __isset = other949.__isset; -} -WMAlterResourcePlanResponse& WMAlterResourcePlanResponse::operator=(const WMAlterResourcePlanResponse& other950) { +WMAlterResourcePlanResponse::WMAlterResourcePlanResponse(const WMAlterResourcePlanResponse& other950) { fullResourcePlan = other950.fullResourcePlan; __isset = other950.__isset; +} +WMAlterResourcePlanResponse& WMAlterResourcePlanResponse::operator=(const WMAlterResourcePlanResponse& other951) { + fullResourcePlan = other951.fullResourcePlan; + __isset = other951.__isset; return *this; } void WMAlterResourcePlanResponse::printTo(std::ostream& out) const { @@ -23799,13 +23857,13 @@ void swap(WMValidateResourcePlanRequest &a, WMValidateResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMValidateResourcePlanRequest::WMValidateResourcePlanRequest(const WMValidateResourcePlanRequest& other951) { - resourcePlanName = other951.resourcePlanName; - __isset = other951.__isset; -} -WMValidateResourcePlanRequest& WMValidateResourcePlanRequest::operator=(const WMValidateResourcePlanRequest& other952) { +WMValidateResourcePlanRequest::WMValidateResourcePlanRequest(const WMValidateResourcePlanRequest& other952) { resourcePlanName = other952.resourcePlanName; __isset = other952.__isset; +} +WMValidateResourcePlanRequest& WMValidateResourcePlanRequest::operator=(const WMValidateResourcePlanRequest& other953) { + resourcePlanName = other953.resourcePlanName; + __isset = other953.__isset; return *this; } void WMValidateResourcePlanRequest::printTo(std::ostream& out) const { @@ -23855,14 +23913,14 @@ uint32_t WMValidateResourcePlanResponse::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->errors.clear(); - uint32_t _size953; - ::apache::thrift::protocol::TType _etype956; - xfer += iprot->readListBegin(_etype956, _size953); - this->errors.resize(_size953); - uint32_t _i957; - for (_i957 = 0; _i957 < _size953; ++_i957) + uint32_t _size954; + ::apache::thrift::protocol::TType _etype957; + xfer += iprot->readListBegin(_etype957, _size954); + this->errors.resize(_size954); + uint32_t _i958; + for (_i958 = 0; _i958 < _size954; ++_i958) { - xfer += iprot->readString(this->errors[_i957]); + xfer += iprot->readString(this->errors[_i958]); } xfer += iprot->readListEnd(); } @@ -23875,14 +23933,14 @@ uint32_t WMValidateResourcePlanResponse::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->warnings.clear(); - uint32_t _size958; - ::apache::thrift::protocol::TType _etype961; - xfer += iprot->readListBegin(_etype961, _size958); - this->warnings.resize(_size958); - uint32_t _i962; - for (_i962 = 0; _i962 < _size958; ++_i962) + uint32_t _size959; + ::apache::thrift::protocol::TType _etype962; + xfer += iprot->readListBegin(_etype962, _size959); + this->warnings.resize(_size959); + uint32_t _i963; + for (_i963 = 0; _i963 < _size959; ++_i963) { - xfer += iprot->readString(this->warnings[_i962]); + xfer += iprot->readString(this->warnings[_i963]); } xfer += iprot->readListEnd(); } @@ -23912,10 +23970,10 @@ uint32_t WMValidateResourcePlanResponse::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("errors", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->errors.size())); - std::vector ::const_iterator _iter963; - for (_iter963 = this->errors.begin(); _iter963 != this->errors.end(); ++_iter963) + std::vector ::const_iterator _iter964; + for (_iter964 = this->errors.begin(); _iter964 != this->errors.end(); ++_iter964) { - xfer += oprot->writeString((*_iter963)); + xfer += oprot->writeString((*_iter964)); } xfer += oprot->writeListEnd(); } @@ -23925,10 +23983,10 @@ uint32_t WMValidateResourcePlanResponse::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("warnings", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->warnings.size())); - std::vector ::const_iterator _iter964; - for (_iter964 = this->warnings.begin(); _iter964 != this->warnings.end(); ++_iter964) + std::vector ::const_iterator _iter965; + for (_iter965 = this->warnings.begin(); _iter965 != this->warnings.end(); ++_iter965) { - xfer += oprot->writeString((*_iter964)); + xfer += oprot->writeString((*_iter965)); } xfer += oprot->writeListEnd(); } @@ -23946,15 +24004,15 @@ void swap(WMValidateResourcePlanResponse &a, WMValidateResourcePlanResponse &b) swap(a.__isset, b.__isset); } -WMValidateResourcePlanResponse::WMValidateResourcePlanResponse(const WMValidateResourcePlanResponse& other965) { - errors = other965.errors; - warnings = other965.warnings; - __isset = other965.__isset; -} -WMValidateResourcePlanResponse& WMValidateResourcePlanResponse::operator=(const WMValidateResourcePlanResponse& other966) { +WMValidateResourcePlanResponse::WMValidateResourcePlanResponse(const WMValidateResourcePlanResponse& other966) { errors = other966.errors; warnings = other966.warnings; __isset = other966.__isset; +} +WMValidateResourcePlanResponse& WMValidateResourcePlanResponse::operator=(const WMValidateResourcePlanResponse& other967) { + errors = other967.errors; + warnings = other967.warnings; + __isset = other967.__isset; return *this; } void WMValidateResourcePlanResponse::printTo(std::ostream& out) const { @@ -24037,13 +24095,13 @@ void swap(WMDropResourcePlanRequest &a, WMDropResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMDropResourcePlanRequest::WMDropResourcePlanRequest(const WMDropResourcePlanRequest& other967) { - resourcePlanName = other967.resourcePlanName; - __isset = other967.__isset; -} -WMDropResourcePlanRequest& WMDropResourcePlanRequest::operator=(const WMDropResourcePlanRequest& other968) { +WMDropResourcePlanRequest::WMDropResourcePlanRequest(const WMDropResourcePlanRequest& other968) { resourcePlanName = other968.resourcePlanName; __isset = other968.__isset; +} +WMDropResourcePlanRequest& WMDropResourcePlanRequest::operator=(const WMDropResourcePlanRequest& other969) { + resourcePlanName = other969.resourcePlanName; + __isset = other969.__isset; return *this; } void WMDropResourcePlanRequest::printTo(std::ostream& out) const { @@ -24102,11 +24160,11 @@ void swap(WMDropResourcePlanResponse &a, WMDropResourcePlanResponse &b) { (void) b; } -WMDropResourcePlanResponse::WMDropResourcePlanResponse(const WMDropResourcePlanResponse& other969) { - (void) other969; -} -WMDropResourcePlanResponse& WMDropResourcePlanResponse::operator=(const WMDropResourcePlanResponse& other970) { +WMDropResourcePlanResponse::WMDropResourcePlanResponse(const WMDropResourcePlanResponse& other970) { (void) other970; +} +WMDropResourcePlanResponse& WMDropResourcePlanResponse::operator=(const WMDropResourcePlanResponse& other971) { + (void) other971; return *this; } void WMDropResourcePlanResponse::printTo(std::ostream& out) const { @@ -24187,13 +24245,13 @@ void swap(WMCreateTriggerRequest &a, WMCreateTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMCreateTriggerRequest::WMCreateTriggerRequest(const WMCreateTriggerRequest& other971) { - trigger = other971.trigger; - __isset = other971.__isset; -} -WMCreateTriggerRequest& WMCreateTriggerRequest::operator=(const WMCreateTriggerRequest& other972) { +WMCreateTriggerRequest::WMCreateTriggerRequest(const WMCreateTriggerRequest& other972) { trigger = other972.trigger; __isset = other972.__isset; +} +WMCreateTriggerRequest& WMCreateTriggerRequest::operator=(const WMCreateTriggerRequest& other973) { + trigger = other973.trigger; + __isset = other973.__isset; return *this; } void WMCreateTriggerRequest::printTo(std::ostream& out) const { @@ -24252,11 +24310,11 @@ void swap(WMCreateTriggerResponse &a, WMCreateTriggerResponse &b) { (void) b; } -WMCreateTriggerResponse::WMCreateTriggerResponse(const WMCreateTriggerResponse& other973) { - (void) other973; -} -WMCreateTriggerResponse& WMCreateTriggerResponse::operator=(const WMCreateTriggerResponse& other974) { +WMCreateTriggerResponse::WMCreateTriggerResponse(const WMCreateTriggerResponse& other974) { (void) other974; +} +WMCreateTriggerResponse& WMCreateTriggerResponse::operator=(const WMCreateTriggerResponse& other975) { + (void) other975; return *this; } void WMCreateTriggerResponse::printTo(std::ostream& out) const { @@ -24337,13 +24395,13 @@ void swap(WMAlterTriggerRequest &a, WMAlterTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMAlterTriggerRequest::WMAlterTriggerRequest(const WMAlterTriggerRequest& other975) { - trigger = other975.trigger; - __isset = other975.__isset; -} -WMAlterTriggerRequest& WMAlterTriggerRequest::operator=(const WMAlterTriggerRequest& other976) { +WMAlterTriggerRequest::WMAlterTriggerRequest(const WMAlterTriggerRequest& other976) { trigger = other976.trigger; __isset = other976.__isset; +} +WMAlterTriggerRequest& WMAlterTriggerRequest::operator=(const WMAlterTriggerRequest& other977) { + trigger = other977.trigger; + __isset = other977.__isset; return *this; } void WMAlterTriggerRequest::printTo(std::ostream& out) const { @@ -24402,11 +24460,11 @@ void swap(WMAlterTriggerResponse &a, WMAlterTriggerResponse &b) { (void) b; } -WMAlterTriggerResponse::WMAlterTriggerResponse(const WMAlterTriggerResponse& other977) { - (void) other977; -} -WMAlterTriggerResponse& WMAlterTriggerResponse::operator=(const WMAlterTriggerResponse& other978) { +WMAlterTriggerResponse::WMAlterTriggerResponse(const WMAlterTriggerResponse& other978) { (void) other978; +} +WMAlterTriggerResponse& WMAlterTriggerResponse::operator=(const WMAlterTriggerResponse& other979) { + (void) other979; return *this; } void WMAlterTriggerResponse::printTo(std::ostream& out) const { @@ -24506,15 +24564,15 @@ void swap(WMDropTriggerRequest &a, WMDropTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMDropTriggerRequest::WMDropTriggerRequest(const WMDropTriggerRequest& other979) { - resourcePlanName = other979.resourcePlanName; - triggerName = other979.triggerName; - __isset = other979.__isset; -} -WMDropTriggerRequest& WMDropTriggerRequest::operator=(const WMDropTriggerRequest& other980) { +WMDropTriggerRequest::WMDropTriggerRequest(const WMDropTriggerRequest& other980) { resourcePlanName = other980.resourcePlanName; triggerName = other980.triggerName; __isset = other980.__isset; +} +WMDropTriggerRequest& WMDropTriggerRequest::operator=(const WMDropTriggerRequest& other981) { + resourcePlanName = other981.resourcePlanName; + triggerName = other981.triggerName; + __isset = other981.__isset; return *this; } void WMDropTriggerRequest::printTo(std::ostream& out) const { @@ -24574,11 +24632,11 @@ void swap(WMDropTriggerResponse &a, WMDropTriggerResponse &b) { (void) b; } -WMDropTriggerResponse::WMDropTriggerResponse(const WMDropTriggerResponse& other981) { - (void) other981; -} -WMDropTriggerResponse& WMDropTriggerResponse::operator=(const WMDropTriggerResponse& other982) { +WMDropTriggerResponse::WMDropTriggerResponse(const WMDropTriggerResponse& other982) { (void) other982; +} +WMDropTriggerResponse& WMDropTriggerResponse::operator=(const WMDropTriggerResponse& other983) { + (void) other983; return *this; } void WMDropTriggerResponse::printTo(std::ostream& out) const { @@ -24659,13 +24717,13 @@ void swap(WMGetTriggersForResourePlanRequest &a, WMGetTriggersForResourePlanRequ swap(a.__isset, b.__isset); } -WMGetTriggersForResourePlanRequest::WMGetTriggersForResourePlanRequest(const WMGetTriggersForResourePlanRequest& other983) { - resourcePlanName = other983.resourcePlanName; - __isset = other983.__isset; -} -WMGetTriggersForResourePlanRequest& WMGetTriggersForResourePlanRequest::operator=(const WMGetTriggersForResourePlanRequest& other984) { +WMGetTriggersForResourePlanRequest::WMGetTriggersForResourePlanRequest(const WMGetTriggersForResourePlanRequest& other984) { resourcePlanName = other984.resourcePlanName; __isset = other984.__isset; +} +WMGetTriggersForResourePlanRequest& WMGetTriggersForResourePlanRequest::operator=(const WMGetTriggersForResourePlanRequest& other985) { + resourcePlanName = other985.resourcePlanName; + __isset = other985.__isset; return *this; } void WMGetTriggersForResourePlanRequest::printTo(std::ostream& out) const { @@ -24710,14 +24768,14 @@ uint32_t WMGetTriggersForResourePlanResponse::read(::apache::thrift::protocol::T if (ftype == ::apache::thrift::protocol::T_LIST) { { this->triggers.clear(); - uint32_t _size985; - ::apache::thrift::protocol::TType _etype988; - xfer += iprot->readListBegin(_etype988, _size985); - this->triggers.resize(_size985); - uint32_t _i989; - for (_i989 = 0; _i989 < _size985; ++_i989) + uint32_t _size986; + ::apache::thrift::protocol::TType _etype989; + xfer += iprot->readListBegin(_etype989, _size986); + this->triggers.resize(_size986); + uint32_t _i990; + for (_i990 = 0; _i990 < _size986; ++_i990) { - xfer += this->triggers[_i989].read(iprot); + xfer += this->triggers[_i990].read(iprot); } xfer += iprot->readListEnd(); } @@ -24747,10 +24805,10 @@ uint32_t WMGetTriggersForResourePlanResponse::write(::apache::thrift::protocol:: xfer += oprot->writeFieldBegin("triggers", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->triggers.size())); - std::vector ::const_iterator _iter990; - for (_iter990 = this->triggers.begin(); _iter990 != this->triggers.end(); ++_iter990) + std::vector ::const_iterator _iter991; + for (_iter991 = this->triggers.begin(); _iter991 != this->triggers.end(); ++_iter991) { - xfer += (*_iter990).write(oprot); + xfer += (*_iter991).write(oprot); } xfer += oprot->writeListEnd(); } @@ -24767,13 +24825,13 @@ void swap(WMGetTriggersForResourePlanResponse &a, WMGetTriggersForResourePlanRes swap(a.__isset, b.__isset); } -WMGetTriggersForResourePlanResponse::WMGetTriggersForResourePlanResponse(const WMGetTriggersForResourePlanResponse& other991) { - triggers = other991.triggers; - __isset = other991.__isset; -} -WMGetTriggersForResourePlanResponse& WMGetTriggersForResourePlanResponse::operator=(const WMGetTriggersForResourePlanResponse& other992) { +WMGetTriggersForResourePlanResponse::WMGetTriggersForResourePlanResponse(const WMGetTriggersForResourePlanResponse& other992) { triggers = other992.triggers; __isset = other992.__isset; +} +WMGetTriggersForResourePlanResponse& WMGetTriggersForResourePlanResponse::operator=(const WMGetTriggersForResourePlanResponse& other993) { + triggers = other993.triggers; + __isset = other993.__isset; return *this; } void WMGetTriggersForResourePlanResponse::printTo(std::ostream& out) const { @@ -24855,13 +24913,13 @@ void swap(WMCreatePoolRequest &a, WMCreatePoolRequest &b) { swap(a.__isset, b.__isset); } -WMCreatePoolRequest::WMCreatePoolRequest(const WMCreatePoolRequest& other993) { - pool = other993.pool; - __isset = other993.__isset; -} -WMCreatePoolRequest& WMCreatePoolRequest::operator=(const WMCreatePoolRequest& other994) { +WMCreatePoolRequest::WMCreatePoolRequest(const WMCreatePoolRequest& other994) { pool = other994.pool; __isset = other994.__isset; +} +WMCreatePoolRequest& WMCreatePoolRequest::operator=(const WMCreatePoolRequest& other995) { + pool = other995.pool; + __isset = other995.__isset; return *this; } void WMCreatePoolRequest::printTo(std::ostream& out) const { @@ -24920,11 +24978,11 @@ void swap(WMCreatePoolResponse &a, WMCreatePoolResponse &b) { (void) b; } -WMCreatePoolResponse::WMCreatePoolResponse(const WMCreatePoolResponse& other995) { - (void) other995; -} -WMCreatePoolResponse& WMCreatePoolResponse::operator=(const WMCreatePoolResponse& other996) { +WMCreatePoolResponse::WMCreatePoolResponse(const WMCreatePoolResponse& other996) { (void) other996; +} +WMCreatePoolResponse& WMCreatePoolResponse::operator=(const WMCreatePoolResponse& other997) { + (void) other997; return *this; } void WMCreatePoolResponse::printTo(std::ostream& out) const { @@ -25024,15 +25082,15 @@ void swap(WMAlterPoolRequest &a, WMAlterPoolRequest &b) { swap(a.__isset, b.__isset); } -WMAlterPoolRequest::WMAlterPoolRequest(const WMAlterPoolRequest& other997) { - pool = other997.pool; - poolPath = other997.poolPath; - __isset = other997.__isset; -} -WMAlterPoolRequest& WMAlterPoolRequest::operator=(const WMAlterPoolRequest& other998) { +WMAlterPoolRequest::WMAlterPoolRequest(const WMAlterPoolRequest& other998) { pool = other998.pool; poolPath = other998.poolPath; __isset = other998.__isset; +} +WMAlterPoolRequest& WMAlterPoolRequest::operator=(const WMAlterPoolRequest& other999) { + pool = other999.pool; + poolPath = other999.poolPath; + __isset = other999.__isset; return *this; } void WMAlterPoolRequest::printTo(std::ostream& out) const { @@ -25092,11 +25150,11 @@ void swap(WMAlterPoolResponse &a, WMAlterPoolResponse &b) { (void) b; } -WMAlterPoolResponse::WMAlterPoolResponse(const WMAlterPoolResponse& other999) { - (void) other999; -} -WMAlterPoolResponse& WMAlterPoolResponse::operator=(const WMAlterPoolResponse& other1000) { +WMAlterPoolResponse::WMAlterPoolResponse(const WMAlterPoolResponse& other1000) { (void) other1000; +} +WMAlterPoolResponse& WMAlterPoolResponse::operator=(const WMAlterPoolResponse& other1001) { + (void) other1001; return *this; } void WMAlterPoolResponse::printTo(std::ostream& out) const { @@ -25196,15 +25254,15 @@ void swap(WMDropPoolRequest &a, WMDropPoolRequest &b) { swap(a.__isset, b.__isset); } -WMDropPoolRequest::WMDropPoolRequest(const WMDropPoolRequest& other1001) { - resourcePlanName = other1001.resourcePlanName; - poolPath = other1001.poolPath; - __isset = other1001.__isset; -} -WMDropPoolRequest& WMDropPoolRequest::operator=(const WMDropPoolRequest& other1002) { +WMDropPoolRequest::WMDropPoolRequest(const WMDropPoolRequest& other1002) { resourcePlanName = other1002.resourcePlanName; poolPath = other1002.poolPath; __isset = other1002.__isset; +} +WMDropPoolRequest& WMDropPoolRequest::operator=(const WMDropPoolRequest& other1003) { + resourcePlanName = other1003.resourcePlanName; + poolPath = other1003.poolPath; + __isset = other1003.__isset; return *this; } void WMDropPoolRequest::printTo(std::ostream& out) const { @@ -25264,11 +25322,11 @@ void swap(WMDropPoolResponse &a, WMDropPoolResponse &b) { (void) b; } -WMDropPoolResponse::WMDropPoolResponse(const WMDropPoolResponse& other1003) { - (void) other1003; -} -WMDropPoolResponse& WMDropPoolResponse::operator=(const WMDropPoolResponse& other1004) { +WMDropPoolResponse::WMDropPoolResponse(const WMDropPoolResponse& other1004) { (void) other1004; +} +WMDropPoolResponse& WMDropPoolResponse::operator=(const WMDropPoolResponse& other1005) { + (void) other1005; return *this; } void WMDropPoolResponse::printTo(std::ostream& out) const { @@ -25368,15 +25426,15 @@ void swap(WMCreateOrUpdateMappingRequest &a, WMCreateOrUpdateMappingRequest &b) swap(a.__isset, b.__isset); } -WMCreateOrUpdateMappingRequest::WMCreateOrUpdateMappingRequest(const WMCreateOrUpdateMappingRequest& other1005) { - mapping = other1005.mapping; - update = other1005.update; - __isset = other1005.__isset; -} -WMCreateOrUpdateMappingRequest& WMCreateOrUpdateMappingRequest::operator=(const WMCreateOrUpdateMappingRequest& other1006) { +WMCreateOrUpdateMappingRequest::WMCreateOrUpdateMappingRequest(const WMCreateOrUpdateMappingRequest& other1006) { mapping = other1006.mapping; update = other1006.update; __isset = other1006.__isset; +} +WMCreateOrUpdateMappingRequest& WMCreateOrUpdateMappingRequest::operator=(const WMCreateOrUpdateMappingRequest& other1007) { + mapping = other1007.mapping; + update = other1007.update; + __isset = other1007.__isset; return *this; } void WMCreateOrUpdateMappingRequest::printTo(std::ostream& out) const { @@ -25436,11 +25494,11 @@ void swap(WMCreateOrUpdateMappingResponse &a, WMCreateOrUpdateMappingResponse &b (void) b; } -WMCreateOrUpdateMappingResponse::WMCreateOrUpdateMappingResponse(const WMCreateOrUpdateMappingResponse& other1007) { - (void) other1007; -} -WMCreateOrUpdateMappingResponse& WMCreateOrUpdateMappingResponse::operator=(const WMCreateOrUpdateMappingResponse& other1008) { +WMCreateOrUpdateMappingResponse::WMCreateOrUpdateMappingResponse(const WMCreateOrUpdateMappingResponse& other1008) { (void) other1008; +} +WMCreateOrUpdateMappingResponse& WMCreateOrUpdateMappingResponse::operator=(const WMCreateOrUpdateMappingResponse& other1009) { + (void) other1009; return *this; } void WMCreateOrUpdateMappingResponse::printTo(std::ostream& out) const { @@ -25521,13 +25579,13 @@ void swap(WMDropMappingRequest &a, WMDropMappingRequest &b) { swap(a.__isset, b.__isset); } -WMDropMappingRequest::WMDropMappingRequest(const WMDropMappingRequest& other1009) { - mapping = other1009.mapping; - __isset = other1009.__isset; -} -WMDropMappingRequest& WMDropMappingRequest::operator=(const WMDropMappingRequest& other1010) { +WMDropMappingRequest::WMDropMappingRequest(const WMDropMappingRequest& other1010) { mapping = other1010.mapping; __isset = other1010.__isset; +} +WMDropMappingRequest& WMDropMappingRequest::operator=(const WMDropMappingRequest& other1011) { + mapping = other1011.mapping; + __isset = other1011.__isset; return *this; } void WMDropMappingRequest::printTo(std::ostream& out) const { @@ -25586,11 +25644,11 @@ void swap(WMDropMappingResponse &a, WMDropMappingResponse &b) { (void) b; } -WMDropMappingResponse::WMDropMappingResponse(const WMDropMappingResponse& other1011) { - (void) other1011; -} -WMDropMappingResponse& WMDropMappingResponse::operator=(const WMDropMappingResponse& other1012) { +WMDropMappingResponse::WMDropMappingResponse(const WMDropMappingResponse& other1012) { (void) other1012; +} +WMDropMappingResponse& WMDropMappingResponse::operator=(const WMDropMappingResponse& other1013) { + (void) other1013; return *this; } void WMDropMappingResponse::printTo(std::ostream& out) const { @@ -25728,19 +25786,19 @@ void swap(WMCreateOrDropTriggerToPoolMappingRequest &a, WMCreateOrDropTriggerToP swap(a.__isset, b.__isset); } -WMCreateOrDropTriggerToPoolMappingRequest::WMCreateOrDropTriggerToPoolMappingRequest(const WMCreateOrDropTriggerToPoolMappingRequest& other1013) { - resourcePlanName = other1013.resourcePlanName; - triggerName = other1013.triggerName; - poolPath = other1013.poolPath; - drop = other1013.drop; - __isset = other1013.__isset; -} -WMCreateOrDropTriggerToPoolMappingRequest& WMCreateOrDropTriggerToPoolMappingRequest::operator=(const WMCreateOrDropTriggerToPoolMappingRequest& other1014) { +WMCreateOrDropTriggerToPoolMappingRequest::WMCreateOrDropTriggerToPoolMappingRequest(const WMCreateOrDropTriggerToPoolMappingRequest& other1014) { resourcePlanName = other1014.resourcePlanName; triggerName = other1014.triggerName; poolPath = other1014.poolPath; drop = other1014.drop; __isset = other1014.__isset; +} +WMCreateOrDropTriggerToPoolMappingRequest& WMCreateOrDropTriggerToPoolMappingRequest::operator=(const WMCreateOrDropTriggerToPoolMappingRequest& other1015) { + resourcePlanName = other1015.resourcePlanName; + triggerName = other1015.triggerName; + poolPath = other1015.poolPath; + drop = other1015.drop; + __isset = other1015.__isset; return *this; } void WMCreateOrDropTriggerToPoolMappingRequest::printTo(std::ostream& out) const { @@ -25802,11 +25860,11 @@ void swap(WMCreateOrDropTriggerToPoolMappingResponse &a, WMCreateOrDropTriggerTo (void) b; } -WMCreateOrDropTriggerToPoolMappingResponse::WMCreateOrDropTriggerToPoolMappingResponse(const WMCreateOrDropTriggerToPoolMappingResponse& other1015) { - (void) other1015; -} -WMCreateOrDropTriggerToPoolMappingResponse& WMCreateOrDropTriggerToPoolMappingResponse::operator=(const WMCreateOrDropTriggerToPoolMappingResponse& other1016) { +WMCreateOrDropTriggerToPoolMappingResponse::WMCreateOrDropTriggerToPoolMappingResponse(const WMCreateOrDropTriggerToPoolMappingResponse& other1016) { (void) other1016; +} +WMCreateOrDropTriggerToPoolMappingResponse& WMCreateOrDropTriggerToPoolMappingResponse::operator=(const WMCreateOrDropTriggerToPoolMappingResponse& other1017) { + (void) other1017; return *this; } void WMCreateOrDropTriggerToPoolMappingResponse::printTo(std::ostream& out) const { @@ -25885,13 +25943,13 @@ void swap(MetaException &a, MetaException &b) { swap(a.__isset, b.__isset); } -MetaException::MetaException(const MetaException& other1017) : TException() { - message = other1017.message; - __isset = other1017.__isset; -} -MetaException& MetaException::operator=(const MetaException& other1018) { +MetaException::MetaException(const MetaException& other1018) : TException() { message = other1018.message; __isset = other1018.__isset; +} +MetaException& MetaException::operator=(const MetaException& other1019) { + message = other1019.message; + __isset = other1019.__isset; return *this; } void MetaException::printTo(std::ostream& out) const { @@ -25982,13 +26040,13 @@ void swap(UnknownTableException &a, UnknownTableException &b) { swap(a.__isset, b.__isset); } -UnknownTableException::UnknownTableException(const UnknownTableException& other1019) : TException() { - message = other1019.message; - __isset = other1019.__isset; -} -UnknownTableException& UnknownTableException::operator=(const UnknownTableException& other1020) { +UnknownTableException::UnknownTableException(const UnknownTableException& other1020) : TException() { message = other1020.message; __isset = other1020.__isset; +} +UnknownTableException& UnknownTableException::operator=(const UnknownTableException& other1021) { + message = other1021.message; + __isset = other1021.__isset; return *this; } void UnknownTableException::printTo(std::ostream& out) const { @@ -26079,13 +26137,13 @@ void swap(UnknownDBException &a, UnknownDBException &b) { swap(a.__isset, b.__isset); } -UnknownDBException::UnknownDBException(const UnknownDBException& other1021) : TException() { - message = other1021.message; - __isset = other1021.__isset; -} -UnknownDBException& UnknownDBException::operator=(const UnknownDBException& other1022) { +UnknownDBException::UnknownDBException(const UnknownDBException& other1022) : TException() { message = other1022.message; __isset = other1022.__isset; +} +UnknownDBException& UnknownDBException::operator=(const UnknownDBException& other1023) { + message = other1023.message; + __isset = other1023.__isset; return *this; } void UnknownDBException::printTo(std::ostream& out) const { @@ -26176,13 +26234,13 @@ void swap(AlreadyExistsException &a, AlreadyExistsException &b) { swap(a.__isset, b.__isset); } -AlreadyExistsException::AlreadyExistsException(const AlreadyExistsException& other1023) : TException() { - message = other1023.message; - __isset = other1023.__isset; -} -AlreadyExistsException& AlreadyExistsException::operator=(const AlreadyExistsException& other1024) { +AlreadyExistsException::AlreadyExistsException(const AlreadyExistsException& other1024) : TException() { message = other1024.message; __isset = other1024.__isset; +} +AlreadyExistsException& AlreadyExistsException::operator=(const AlreadyExistsException& other1025) { + message = other1025.message; + __isset = other1025.__isset; return *this; } void AlreadyExistsException::printTo(std::ostream& out) const { @@ -26273,13 +26331,13 @@ void swap(InvalidPartitionException &a, InvalidPartitionException &b) { swap(a.__isset, b.__isset); } -InvalidPartitionException::InvalidPartitionException(const InvalidPartitionException& other1025) : TException() { - message = other1025.message; - __isset = other1025.__isset; -} -InvalidPartitionException& InvalidPartitionException::operator=(const InvalidPartitionException& other1026) { +InvalidPartitionException::InvalidPartitionException(const InvalidPartitionException& other1026) : TException() { message = other1026.message; __isset = other1026.__isset; +} +InvalidPartitionException& InvalidPartitionException::operator=(const InvalidPartitionException& other1027) { + message = other1027.message; + __isset = other1027.__isset; return *this; } void InvalidPartitionException::printTo(std::ostream& out) const { @@ -26370,13 +26428,13 @@ void swap(UnknownPartitionException &a, UnknownPartitionException &b) { swap(a.__isset, b.__isset); } -UnknownPartitionException::UnknownPartitionException(const UnknownPartitionException& other1027) : TException() { - message = other1027.message; - __isset = other1027.__isset; -} -UnknownPartitionException& UnknownPartitionException::operator=(const UnknownPartitionException& other1028) { +UnknownPartitionException::UnknownPartitionException(const UnknownPartitionException& other1028) : TException() { message = other1028.message; __isset = other1028.__isset; +} +UnknownPartitionException& UnknownPartitionException::operator=(const UnknownPartitionException& other1029) { + message = other1029.message; + __isset = other1029.__isset; return *this; } void UnknownPartitionException::printTo(std::ostream& out) const { @@ -26467,13 +26525,13 @@ void swap(InvalidObjectException &a, InvalidObjectException &b) { swap(a.__isset, b.__isset); } -InvalidObjectException::InvalidObjectException(const InvalidObjectException& other1029) : TException() { - message = other1029.message; - __isset = other1029.__isset; -} -InvalidObjectException& InvalidObjectException::operator=(const InvalidObjectException& other1030) { +InvalidObjectException::InvalidObjectException(const InvalidObjectException& other1030) : TException() { message = other1030.message; __isset = other1030.__isset; +} +InvalidObjectException& InvalidObjectException::operator=(const InvalidObjectException& other1031) { + message = other1031.message; + __isset = other1031.__isset; return *this; } void InvalidObjectException::printTo(std::ostream& out) const { @@ -26564,13 +26622,13 @@ void swap(NoSuchObjectException &a, NoSuchObjectException &b) { swap(a.__isset, b.__isset); } -NoSuchObjectException::NoSuchObjectException(const NoSuchObjectException& other1031) : TException() { - message = other1031.message; - __isset = other1031.__isset; -} -NoSuchObjectException& NoSuchObjectException::operator=(const NoSuchObjectException& other1032) { +NoSuchObjectException::NoSuchObjectException(const NoSuchObjectException& other1032) : TException() { message = other1032.message; __isset = other1032.__isset; +} +NoSuchObjectException& NoSuchObjectException::operator=(const NoSuchObjectException& other1033) { + message = other1033.message; + __isset = other1033.__isset; return *this; } void NoSuchObjectException::printTo(std::ostream& out) const { @@ -26661,13 +26719,13 @@ void swap(IndexAlreadyExistsException &a, IndexAlreadyExistsException &b) { swap(a.__isset, b.__isset); } -IndexAlreadyExistsException::IndexAlreadyExistsException(const IndexAlreadyExistsException& other1033) : TException() { - message = other1033.message; - __isset = other1033.__isset; -} -IndexAlreadyExistsException& IndexAlreadyExistsException::operator=(const IndexAlreadyExistsException& other1034) { +IndexAlreadyExistsException::IndexAlreadyExistsException(const IndexAlreadyExistsException& other1034) : TException() { message = other1034.message; __isset = other1034.__isset; +} +IndexAlreadyExistsException& IndexAlreadyExistsException::operator=(const IndexAlreadyExistsException& other1035) { + message = other1035.message; + __isset = other1035.__isset; return *this; } void IndexAlreadyExistsException::printTo(std::ostream& out) const { @@ -26758,13 +26816,13 @@ void swap(InvalidOperationException &a, InvalidOperationException &b) { swap(a.__isset, b.__isset); } -InvalidOperationException::InvalidOperationException(const InvalidOperationException& other1035) : TException() { - message = other1035.message; - __isset = other1035.__isset; -} -InvalidOperationException& InvalidOperationException::operator=(const InvalidOperationException& other1036) { +InvalidOperationException::InvalidOperationException(const InvalidOperationException& other1036) : TException() { message = other1036.message; __isset = other1036.__isset; +} +InvalidOperationException& InvalidOperationException::operator=(const InvalidOperationException& other1037) { + message = other1037.message; + __isset = other1037.__isset; return *this; } void InvalidOperationException::printTo(std::ostream& out) const { @@ -26855,13 +26913,13 @@ void swap(ConfigValSecurityException &a, ConfigValSecurityException &b) { swap(a.__isset, b.__isset); } -ConfigValSecurityException::ConfigValSecurityException(const ConfigValSecurityException& other1037) : TException() { - message = other1037.message; - __isset = other1037.__isset; -} -ConfigValSecurityException& ConfigValSecurityException::operator=(const ConfigValSecurityException& other1038) { +ConfigValSecurityException::ConfigValSecurityException(const ConfigValSecurityException& other1038) : TException() { message = other1038.message; __isset = other1038.__isset; +} +ConfigValSecurityException& ConfigValSecurityException::operator=(const ConfigValSecurityException& other1039) { + message = other1039.message; + __isset = other1039.__isset; return *this; } void ConfigValSecurityException::printTo(std::ostream& out) const { @@ -26952,13 +27010,13 @@ void swap(InvalidInputException &a, InvalidInputException &b) { swap(a.__isset, b.__isset); } -InvalidInputException::InvalidInputException(const InvalidInputException& other1039) : TException() { - message = other1039.message; - __isset = other1039.__isset; -} -InvalidInputException& InvalidInputException::operator=(const InvalidInputException& other1040) { +InvalidInputException::InvalidInputException(const InvalidInputException& other1040) : TException() { message = other1040.message; __isset = other1040.__isset; +} +InvalidInputException& InvalidInputException::operator=(const InvalidInputException& other1041) { + message = other1041.message; + __isset = other1041.__isset; return *this; } void InvalidInputException::printTo(std::ostream& out) const { @@ -27049,13 +27107,13 @@ void swap(NoSuchTxnException &a, NoSuchTxnException &b) { swap(a.__isset, b.__isset); } -NoSuchTxnException::NoSuchTxnException(const NoSuchTxnException& other1041) : TException() { - message = other1041.message; - __isset = other1041.__isset; -} -NoSuchTxnException& NoSuchTxnException::operator=(const NoSuchTxnException& other1042) { +NoSuchTxnException::NoSuchTxnException(const NoSuchTxnException& other1042) : TException() { message = other1042.message; __isset = other1042.__isset; +} +NoSuchTxnException& NoSuchTxnException::operator=(const NoSuchTxnException& other1043) { + message = other1043.message; + __isset = other1043.__isset; return *this; } void NoSuchTxnException::printTo(std::ostream& out) const { @@ -27146,13 +27204,13 @@ void swap(TxnAbortedException &a, TxnAbortedException &b) { swap(a.__isset, b.__isset); } -TxnAbortedException::TxnAbortedException(const TxnAbortedException& other1043) : TException() { - message = other1043.message; - __isset = other1043.__isset; -} -TxnAbortedException& TxnAbortedException::operator=(const TxnAbortedException& other1044) { +TxnAbortedException::TxnAbortedException(const TxnAbortedException& other1044) : TException() { message = other1044.message; __isset = other1044.__isset; +} +TxnAbortedException& TxnAbortedException::operator=(const TxnAbortedException& other1045) { + message = other1045.message; + __isset = other1045.__isset; return *this; } void TxnAbortedException::printTo(std::ostream& out) const { @@ -27243,13 +27301,13 @@ void swap(TxnOpenException &a, TxnOpenException &b) { swap(a.__isset, b.__isset); } -TxnOpenException::TxnOpenException(const TxnOpenException& other1045) : TException() { - message = other1045.message; - __isset = other1045.__isset; -} -TxnOpenException& TxnOpenException::operator=(const TxnOpenException& other1046) { +TxnOpenException::TxnOpenException(const TxnOpenException& other1046) : TException() { message = other1046.message; __isset = other1046.__isset; +} +TxnOpenException& TxnOpenException::operator=(const TxnOpenException& other1047) { + message = other1047.message; + __isset = other1047.__isset; return *this; } void TxnOpenException::printTo(std::ostream& out) const { @@ -27340,13 +27398,13 @@ void swap(NoSuchLockException &a, NoSuchLockException &b) { swap(a.__isset, b.__isset); } -NoSuchLockException::NoSuchLockException(const NoSuchLockException& other1047) : TException() { - message = other1047.message; - __isset = other1047.__isset; -} -NoSuchLockException& NoSuchLockException::operator=(const NoSuchLockException& other1048) { +NoSuchLockException::NoSuchLockException(const NoSuchLockException& other1048) : TException() { message = other1048.message; __isset = other1048.__isset; +} +NoSuchLockException& NoSuchLockException::operator=(const NoSuchLockException& other1049) { + message = other1049.message; + __isset = other1049.__isset; return *this; } void NoSuchLockException::printTo(std::ostream& out) const { diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h index 4c09bc8fe6..44a2c83f39 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h +++ b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h @@ -132,6 +132,16 @@ struct EventRequestType { extern const std::map _EventRequestType_VALUES_TO_NAMES; +struct BucketingVersion { + enum type { + INVALID_BUCKETING = 0, + JAVA_BUCKETING = 1, + MURMUR_BUCKETING = 2 + }; +}; + +extern const std::map _BucketingVersion_VALUES_TO_NAMES; + struct FunctionType { enum type { JAVA = 1 @@ -2372,7 +2382,7 @@ inline std::ostream& operator<<(std::ostream& out, const StorageDescriptor& obj) } typedef struct _Table__isset { - _Table__isset() : tableName(false), dbName(false), owner(false), createTime(false), lastAccessTime(false), retention(false), sd(false), partitionKeys(false), parameters(false), viewOriginalText(false), viewExpandedText(false), tableType(false), privileges(false), temporary(true), rewriteEnabled(false), creationMetadata(false) {} + _Table__isset() : tableName(false), dbName(false), owner(false), createTime(false), lastAccessTime(false), retention(false), sd(false), partitionKeys(false), parameters(false), viewOriginalText(false), viewExpandedText(false), tableType(false), privileges(false), temporary(true), rewriteEnabled(false), creationMetadata(false), bucketingVersion(true), loadInBucketedTable(true) {} bool tableName :1; bool dbName :1; bool owner :1; @@ -2389,6 +2399,8 @@ typedef struct _Table__isset { bool temporary :1; bool rewriteEnabled :1; bool creationMetadata :1; + bool bucketingVersion :1; + bool loadInBucketedTable :1; } _Table__isset; class Table { @@ -2396,7 +2408,9 @@ class Table { Table(const Table&); Table& operator=(const Table&); - Table() : tableName(), dbName(), owner(), createTime(0), lastAccessTime(0), retention(0), viewOriginalText(), viewExpandedText(), tableType(), temporary(false), rewriteEnabled(0) { + Table() : tableName(), dbName(), owner(), createTime(0), lastAccessTime(0), retention(0), viewOriginalText(), viewExpandedText(), tableType(), temporary(false), rewriteEnabled(0), bucketingVersion((BucketingVersion::type)1), loadInBucketedTable(false) { + bucketingVersion = (BucketingVersion::type)1; + } virtual ~Table() throw(); @@ -2416,6 +2430,8 @@ class Table { bool temporary; bool rewriteEnabled; CreationMetadata creationMetadata; + BucketingVersion::type bucketingVersion; + bool loadInBucketedTable; _Table__isset __isset; @@ -2451,6 +2467,10 @@ class Table { void __set_creationMetadata(const CreationMetadata& val); + void __set_bucketingVersion(const BucketingVersion::type val); + + void __set_loadInBucketedTable(const bool val); + bool operator == (const Table & rhs) const { if (!(tableName == rhs.tableName)) @@ -2493,6 +2513,14 @@ class Table { return false; else if (__isset.creationMetadata && !(creationMetadata == rhs.creationMetadata)) return false; + if (__isset.bucketingVersion != rhs.__isset.bucketingVersion) + return false; + else if (__isset.bucketingVersion && !(bucketingVersion == rhs.bucketingVersion)) + return false; + if (__isset.loadInBucketedTable != rhs.__isset.loadInBucketedTable) + return false; + else if (__isset.loadInBucketedTable && !(loadInBucketedTable == rhs.loadInBucketedTable)) + return false; return true; } bool operator != (const Table &rhs) const { diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/Table.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/Table.java index a132e5e838..068e542ec1 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/Table.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/Table.java @@ -54,6 +54,8 @@ private static final org.apache.thrift.protocol.TField TEMPORARY_FIELD_DESC = new org.apache.thrift.protocol.TField("temporary", org.apache.thrift.protocol.TType.BOOL, (short)14); private static final org.apache.thrift.protocol.TField REWRITE_ENABLED_FIELD_DESC = new org.apache.thrift.protocol.TField("rewriteEnabled", org.apache.thrift.protocol.TType.BOOL, (short)15); private static final org.apache.thrift.protocol.TField CREATION_METADATA_FIELD_DESC = new org.apache.thrift.protocol.TField("creationMetadata", org.apache.thrift.protocol.TType.STRUCT, (short)16); + private static final org.apache.thrift.protocol.TField BUCKETING_VERSION_FIELD_DESC = new org.apache.thrift.protocol.TField("bucketingVersion", org.apache.thrift.protocol.TType.I32, (short)17); + private static final org.apache.thrift.protocol.TField LOAD_IN_BUCKETED_TABLE_FIELD_DESC = new org.apache.thrift.protocol.TField("loadInBucketedTable", org.apache.thrift.protocol.TType.BOOL, (short)18); private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); static { @@ -77,6 +79,8 @@ private boolean temporary; // optional private boolean rewriteEnabled; // optional private CreationMetadata creationMetadata; // optional + private BucketingVersion bucketingVersion; // optional + private boolean loadInBucketedTable; // optional /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { @@ -95,7 +99,13 @@ PRIVILEGES((short)13, "privileges"), TEMPORARY((short)14, "temporary"), REWRITE_ENABLED((short)15, "rewriteEnabled"), - CREATION_METADATA((short)16, "creationMetadata"); + CREATION_METADATA((short)16, "creationMetadata"), + /** + * + * @see BucketingVersion + */ + BUCKETING_VERSION((short)17, "bucketingVersion"), + LOAD_IN_BUCKETED_TABLE((short)18, "loadInBucketedTable"); private static final Map byName = new HashMap(); @@ -142,6 +152,10 @@ public static _Fields findByThriftId(int fieldId) { return REWRITE_ENABLED; case 16: // CREATION_METADATA return CREATION_METADATA; + case 17: // BUCKETING_VERSION + return BUCKETING_VERSION; + case 18: // LOAD_IN_BUCKETED_TABLE + return LOAD_IN_BUCKETED_TABLE; default: return null; } @@ -187,8 +201,9 @@ public String getFieldName() { private static final int __RETENTION_ISSET_ID = 2; private static final int __TEMPORARY_ISSET_ID = 3; private static final int __REWRITEENABLED_ISSET_ID = 4; + private static final int __LOADINBUCKETEDTABLE_ISSET_ID = 5; private byte __isset_bitfield = 0; - private static final _Fields optionals[] = {_Fields.PRIVILEGES,_Fields.TEMPORARY,_Fields.REWRITE_ENABLED,_Fields.CREATION_METADATA}; + private static final _Fields optionals[] = {_Fields.PRIVILEGES,_Fields.TEMPORARY,_Fields.REWRITE_ENABLED,_Fields.CREATION_METADATA,_Fields.BUCKETING_VERSION,_Fields.LOAD_IN_BUCKETED_TABLE}; public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); @@ -227,6 +242,10 @@ public String getFieldName() { new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); tmpMap.put(_Fields.CREATION_METADATA, new org.apache.thrift.meta_data.FieldMetaData("creationMetadata", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT , "CreationMetadata"))); + tmpMap.put(_Fields.BUCKETING_VERSION, new org.apache.thrift.meta_data.FieldMetaData("bucketingVersion", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.EnumMetaData(org.apache.thrift.protocol.TType.ENUM, BucketingVersion.class))); + tmpMap.put(_Fields.LOAD_IN_BUCKETED_TABLE, new org.apache.thrift.meta_data.FieldMetaData("loadInBucketedTable", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.BOOL))); metaDataMap = Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(Table.class, metaDataMap); } @@ -234,6 +253,10 @@ public String getFieldName() { public Table() { this.temporary = false; + this.bucketingVersion = org.apache.hadoop.hive.metastore.api.BucketingVersion.JAVA_BUCKETING; + + this.loadInBucketedTable = false; + } public Table( @@ -316,6 +339,10 @@ public Table(Table other) { if (other.isSetCreationMetadata()) { this.creationMetadata = other.creationMetadata; } + if (other.isSetBucketingVersion()) { + this.bucketingVersion = other.bucketingVersion; + } + this.loadInBucketedTable = other.loadInBucketedTable; } public Table deepCopy() { @@ -345,6 +372,10 @@ public void clear() { setRewriteEnabledIsSet(false); this.rewriteEnabled = false; this.creationMetadata = null; + this.bucketingVersion = org.apache.hadoop.hive.metastore.api.BucketingVersion.JAVA_BUCKETING; + + this.loadInBucketedTable = false; + } public String getTableName() { @@ -736,6 +767,59 @@ public void setCreationMetadataIsSet(boolean value) { } } + /** + * + * @see BucketingVersion + */ + public BucketingVersion getBucketingVersion() { + return this.bucketingVersion; + } + + /** + * + * @see BucketingVersion + */ + public void setBucketingVersion(BucketingVersion bucketingVersion) { + this.bucketingVersion = bucketingVersion; + } + + public void unsetBucketingVersion() { + this.bucketingVersion = null; + } + + /** Returns true if field bucketingVersion is set (has been assigned a value) and false otherwise */ + public boolean isSetBucketingVersion() { + return this.bucketingVersion != null; + } + + public void setBucketingVersionIsSet(boolean value) { + if (!value) { + this.bucketingVersion = null; + } + } + + public boolean isLoadInBucketedTable() { + return this.loadInBucketedTable; + } + + public void setLoadInBucketedTable(boolean loadInBucketedTable) { + this.loadInBucketedTable = loadInBucketedTable; + setLoadInBucketedTableIsSet(true); + } + + public void unsetLoadInBucketedTable() { + __isset_bitfield = EncodingUtils.clearBit(__isset_bitfield, __LOADINBUCKETEDTABLE_ISSET_ID); + } + + /** Returns true if field loadInBucketedTable is set (has been assigned a value) and false otherwise */ + public boolean isSetLoadInBucketedTable() { + return EncodingUtils.testBit(__isset_bitfield, __LOADINBUCKETEDTABLE_ISSET_ID); + } + + public void setLoadInBucketedTableIsSet(boolean value) { + __isset_bitfield = EncodingUtils.setBit(__isset_bitfield, __LOADINBUCKETEDTABLE_ISSET_ID, value); + } + public void setFieldValue(_Fields field, Object value) { switch (field) { case TABLE_NAME: @@ -866,6 +950,22 @@ public void setFieldValue(_Fields field, Object value) { } break; + case BUCKETING_VERSION: + if (value == null) { + unsetBucketingVersion(); + } else { + setBucketingVersion((BucketingVersion)value); + } + break; + + case LOAD_IN_BUCKETED_TABLE: + if (value == null) { + unsetLoadInBucketedTable(); + } else { + setLoadInBucketedTable((Boolean)value); + } + break; + } } @@ -919,6 +1019,12 @@ public Object getFieldValue(_Fields field) { case CREATION_METADATA: return getCreationMetadata(); + case BUCKETING_VERSION: + return getBucketingVersion(); + + case LOAD_IN_BUCKETED_TABLE: + return isLoadInBucketedTable(); + } throw new IllegalStateException(); } @@ -962,6 +1068,10 @@ public boolean isSet(_Fields field) { return isSetRewriteEnabled(); case CREATION_METADATA: return isSetCreationMetadata(); + case BUCKETING_VERSION: + return isSetBucketingVersion(); + case LOAD_IN_BUCKETED_TABLE: + return isSetLoadInBucketedTable(); } throw new IllegalStateException(); } @@ -1123,6 +1233,24 @@ public boolean equals(Table that) { return false; } + boolean this_present_bucketingVersion = true && this.isSetBucketingVersion(); + boolean that_present_bucketingVersion = true && that.isSetBucketingVersion(); + if (this_present_bucketingVersion || that_present_bucketingVersion) { + if (!(this_present_bucketingVersion && that_present_bucketingVersion)) + return false; + if (!this.bucketingVersion.equals(that.bucketingVersion)) + return false; + } + + boolean this_present_loadInBucketedTable = true && this.isSetLoadInBucketedTable(); + boolean that_present_loadInBucketedTable = true && that.isSetLoadInBucketedTable(); + if (this_present_loadInBucketedTable || that_present_loadInBucketedTable) { + if (!(this_present_loadInBucketedTable && that_present_loadInBucketedTable)) + return false; + if (this.loadInBucketedTable != that.loadInBucketedTable) + return false; + } + return true; } @@ -1210,6 +1338,16 @@ public int hashCode() { if (present_creationMetadata) list.add(creationMetadata); + boolean present_bucketingVersion = true && (isSetBucketingVersion()); + list.add(present_bucketingVersion); + if (present_bucketingVersion) + list.add(bucketingVersion.getValue()); + + boolean present_loadInBucketedTable = true && (isSetLoadInBucketedTable()); + list.add(present_loadInBucketedTable); + if (present_loadInBucketedTable) + list.add(loadInBucketedTable); + return list.hashCode(); } @@ -1381,6 +1519,26 @@ public int compareTo(Table other) { return lastComparison; } } + lastComparison = Boolean.valueOf(isSetBucketingVersion()).compareTo(other.isSetBucketingVersion()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetBucketingVersion()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.bucketingVersion, other.bucketingVersion); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetLoadInBucketedTable()).compareTo(other.isSetLoadInBucketedTable()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetLoadInBucketedTable()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.loadInBucketedTable, other.loadInBucketedTable); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -1516,6 +1674,22 @@ public String toString() { } first = false; } + if (isSetBucketingVersion()) { + if (!first) sb.append(", "); + sb.append("bucketingVersion:"); + if (this.bucketingVersion == null) { + sb.append("null"); + } else { + sb.append(this.bucketingVersion); + } + first = false; + } + if (isSetLoadInBucketedTable()) { + if (!first) sb.append(", "); + sb.append("loadInBucketedTable:"); + sb.append(this.loadInBucketedTable); + first = false; + } sb.append(")"); return sb.toString(); } @@ -1721,6 +1895,22 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, Table struct) throw org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 17: // BUCKETING_VERSION + if (schemeField.type == org.apache.thrift.protocol.TType.I32) { + struct.bucketingVersion = org.apache.hadoop.hive.metastore.api.BucketingVersion.findByValue(iprot.readI32()); + struct.setBucketingVersionIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 18: // LOAD_IN_BUCKETED_TABLE + if (schemeField.type == org.apache.thrift.protocol.TType.BOOL) { + struct.loadInBucketedTable = iprot.readBool(); + struct.setLoadInBucketedTableIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -1827,6 +2017,18 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, Table struct) thro oprot.writeFieldEnd(); } } + if (struct.bucketingVersion != null) { + if (struct.isSetBucketingVersion()) { + oprot.writeFieldBegin(BUCKETING_VERSION_FIELD_DESC); + oprot.writeI32(struct.bucketingVersion.getValue()); + oprot.writeFieldEnd(); + } + } + if (struct.isSetLoadInBucketedTable()) { + oprot.writeFieldBegin(LOAD_IN_BUCKETED_TABLE_FIELD_DESC); + oprot.writeBool(struct.loadInBucketedTable); + oprot.writeFieldEnd(); + } oprot.writeFieldStop(); oprot.writeStructEnd(); } @@ -1893,7 +2095,13 @@ public void write(org.apache.thrift.protocol.TProtocol prot, Table struct) throw if (struct.isSetCreationMetadata()) { optionals.set(15); } - oprot.writeBitSet(optionals, 16); + if (struct.isSetBucketingVersion()) { + optionals.set(16); + } + if (struct.isSetLoadInBucketedTable()) { + optionals.set(17); + } + oprot.writeBitSet(optionals, 18); if (struct.isSetTableName()) { oprot.writeString(struct.tableName); } @@ -1955,12 +2163,18 @@ public void write(org.apache.thrift.protocol.TProtocol prot, Table struct) throw if (struct.isSetCreationMetadata()) { struct.creationMetadata.write(oprot); } + if (struct.isSetBucketingVersion()) { + oprot.writeI32(struct.bucketingVersion.getValue()); + } + if (struct.isSetLoadInBucketedTable()) { + oprot.writeBool(struct.loadInBucketedTable); + } } @Override public void read(org.apache.thrift.protocol.TProtocol prot, Table struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; - BitSet incoming = iprot.readBitSet(16); + BitSet incoming = iprot.readBitSet(18); if (incoming.get(0)) { struct.tableName = iprot.readString(); struct.setTableNameIsSet(true); @@ -2049,6 +2263,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, Table struct) throws struct.creationMetadata.read(iprot); struct.setCreationMetadataIsSet(true); } + if (incoming.get(16)) { + struct.bucketingVersion = org.apache.hadoop.hive.metastore.api.BucketingVersion.findByValue(iprot.readI32()); + struct.setBucketingVersionIsSet(true); + } + if (incoming.get(17)) { + struct.loadInBucketedTable = iprot.readBool(); + struct.setLoadInBucketedTableIsSet(true); + } } } diff --git a/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php b/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php index a5b578ef37..687911ef60 100644 --- a/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php +++ b/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php @@ -142,6 +142,17 @@ final class EventRequestType { ); } +final class BucketingVersion { + const INVALID_BUCKETING = 0; + const JAVA_BUCKETING = 1; + const MURMUR_BUCKETING = 2; + static public $__names = array( + 0 => 'INVALID_BUCKETING', + 1 => 'JAVA_BUCKETING', + 2 => 'MURMUR_BUCKETING', + ); +} + final class FunctionType { const JAVA = 1; static public $__names = array( @@ -5042,6 +5053,14 @@ class Table { * @var \metastore\CreationMetadata */ public $creationMetadata = null; + /** + * @var int + */ + public $bucketingVersion = 1; + /** + * @var bool + */ + public $loadInBucketedTable = false; public function __construct($vals=null) { if (!isset(self::$_TSPEC)) { @@ -5126,6 +5145,14 @@ class Table { 'type' => TType::STRUCT, 'class' => '\metastore\CreationMetadata', ), + 17 => array( + 'var' => 'bucketingVersion', + 'type' => TType::I32, + ), + 18 => array( + 'var' => 'loadInBucketedTable', + 'type' => TType::BOOL, + ), ); } if (is_array($vals)) { @@ -5177,6 +5204,12 @@ class Table { if (isset($vals['creationMetadata'])) { $this->creationMetadata = $vals['creationMetadata']; } + if (isset($vals['bucketingVersion'])) { + $this->bucketingVersion = $vals['bucketingVersion']; + } + if (isset($vals['loadInBucketedTable'])) { + $this->loadInBucketedTable = $vals['loadInBucketedTable']; + } } } @@ -5338,6 +5371,20 @@ class Table { $xfer += $input->skip($ftype); } break; + case 17: + if ($ftype == TType::I32) { + $xfer += $input->readI32($this->bucketingVersion); + } else { + $xfer += $input->skip($ftype); + } + break; + case 18: + if ($ftype == TType::BOOL) { + $xfer += $input->readBool($this->loadInBucketedTable); + } else { + $xfer += $input->skip($ftype); + } + break; default: $xfer += $input->skip($ftype); break; @@ -5465,6 +5512,16 @@ class Table { $xfer += $this->creationMetadata->write($output); $xfer += $output->writeFieldEnd(); } + if ($this->bucketingVersion !== null) { + $xfer += $output->writeFieldBegin('bucketingVersion', TType::I32, 17); + $xfer += $output->writeI32($this->bucketingVersion); + $xfer += $output->writeFieldEnd(); + } + if ($this->loadInBucketedTable !== null) { + $xfer += $output->writeFieldBegin('loadInBucketedTable', TType::BOOL, 18); + $xfer += $output->writeBool($this->loadInBucketedTable); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; diff --git a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py index 5598859042..86c1937f30 100644 --- a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py +++ b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py @@ -211,6 +211,23 @@ class EventRequestType: "DELETE": 3, } +class BucketingVersion: + INVALID_BUCKETING = 0 + JAVA_BUCKETING = 1 + MURMUR_BUCKETING = 2 + + _VALUES_TO_NAMES = { + 0: "INVALID_BUCKETING", + 1: "JAVA_BUCKETING", + 2: "MURMUR_BUCKETING", + } + + _NAMES_TO_VALUES = { + "INVALID_BUCKETING": 0, + "JAVA_BUCKETING": 1, + "MURMUR_BUCKETING": 2, + } + class FunctionType: JAVA = 1 @@ -3468,6 +3485,8 @@ class Table: - temporary - rewriteEnabled - creationMetadata + - bucketingVersion + - loadInBucketedTable """ thrift_spec = ( @@ -3488,9 +3507,11 @@ class Table: (14, TType.BOOL, 'temporary', None, False, ), # 14 (15, TType.BOOL, 'rewriteEnabled', None, None, ), # 15 (16, TType.STRUCT, 'creationMetadata', (CreationMetadata, CreationMetadata.thrift_spec), None, ), # 16 + (17, TType.I32, 'bucketingVersion', None, 1, ), # 17 + (18, TType.BOOL, 'loadInBucketedTable', None, False, ), # 18 ) - def __init__(self, tableName=None, dbName=None, owner=None, createTime=None, lastAccessTime=None, retention=None, sd=None, partitionKeys=None, parameters=None, viewOriginalText=None, viewExpandedText=None, tableType=None, privileges=None, temporary=thrift_spec[14][4], rewriteEnabled=None, creationMetadata=None,): + def __init__(self, tableName=None, dbName=None, owner=None, createTime=None, lastAccessTime=None, retention=None, sd=None, partitionKeys=None, parameters=None, viewOriginalText=None, viewExpandedText=None, tableType=None, privileges=None, temporary=thrift_spec[14][4], rewriteEnabled=None, creationMetadata=None, bucketingVersion=thrift_spec[17][4], loadInBucketedTable=thrift_spec[18][4],): self.tableName = tableName self.dbName = dbName self.owner = owner @@ -3507,6 +3528,8 @@ def __init__(self, tableName=None, dbName=None, owner=None, createTime=None, las self.temporary = temporary self.rewriteEnabled = rewriteEnabled self.creationMetadata = creationMetadata + self.bucketingVersion = bucketingVersion + self.loadInBucketedTable = loadInBucketedTable def read(self, iprot): if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: @@ -3612,6 +3635,16 @@ def read(self, iprot): self.creationMetadata.read(iprot) else: iprot.skip(ftype) + elif fid == 17: + if ftype == TType.I32: + self.bucketingVersion = iprot.readI32() + else: + iprot.skip(ftype) + elif fid == 18: + if ftype == TType.BOOL: + self.loadInBucketedTable = iprot.readBool() + else: + iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() @@ -3693,6 +3726,14 @@ def write(self, oprot): oprot.writeFieldBegin('creationMetadata', TType.STRUCT, 16) self.creationMetadata.write(oprot) oprot.writeFieldEnd() + if self.bucketingVersion is not None: + oprot.writeFieldBegin('bucketingVersion', TType.I32, 17) + oprot.writeI32(self.bucketingVersion) + oprot.writeFieldEnd() + if self.loadInBucketedTable is not None: + oprot.writeFieldBegin('loadInBucketedTable', TType.BOOL, 18) + oprot.writeBool(self.loadInBucketedTable) + oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() @@ -3718,6 +3759,8 @@ def __hash__(self): value = (value * 31) ^ hash(self.temporary) value = (value * 31) ^ hash(self.rewriteEnabled) value = (value * 31) ^ hash(self.creationMetadata) + value = (value * 31) ^ hash(self.bucketingVersion) + value = (value * 31) ^ hash(self.loadInBucketedTable) return value def __repr__(self): diff --git a/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb b/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb index bc58cfe0ef..2c3edb7da8 100644 --- a/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb +++ b/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb @@ -98,6 +98,14 @@ module EventRequestType VALID_VALUES = Set.new([INSERT, UPDATE, DELETE]).freeze end +module BucketingVersion + INVALID_BUCKETING = 0 + JAVA_BUCKETING = 1 + MURMUR_BUCKETING = 2 + VALUE_MAP = {0 => "INVALID_BUCKETING", 1 => "JAVA_BUCKETING", 2 => "MURMUR_BUCKETING"} + VALID_VALUES = Set.new([INVALID_BUCKETING, JAVA_BUCKETING, MURMUR_BUCKETING]).freeze +end + module FunctionType JAVA = 1 VALUE_MAP = {1 => "JAVA"} @@ -810,6 +818,8 @@ class Table TEMPORARY = 14 REWRITEENABLED = 15 CREATIONMETADATA = 16 + BUCKETINGVERSION = 17 + LOADINBUCKETEDTABLE = 18 FIELDS = { TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'}, @@ -827,12 +837,17 @@ class Table PRIVILEGES => {:type => ::Thrift::Types::STRUCT, :name => 'privileges', :class => ::PrincipalPrivilegeSet, :optional => true}, TEMPORARY => {:type => ::Thrift::Types::BOOL, :name => 'temporary', :default => false, :optional => true}, REWRITEENABLED => {:type => ::Thrift::Types::BOOL, :name => 'rewriteEnabled', :optional => true}, - CREATIONMETADATA => {:type => ::Thrift::Types::STRUCT, :name => 'creationMetadata', :class => ::CreationMetadata, :optional => true} + CREATIONMETADATA => {:type => ::Thrift::Types::STRUCT, :name => 'creationMetadata', :class => ::CreationMetadata, :optional => true}, + BUCKETINGVERSION => {:type => ::Thrift::Types::I32, :name => 'bucketingVersion', :default => 1, :optional => true, :enum_class => ::BucketingVersion}, + LOADINBUCKETEDTABLE => {:type => ::Thrift::Types::BOOL, :name => 'loadInBucketedTable', :default => false, :optional => true} } def struct_fields; FIELDS; end def validate + unless @bucketingVersion.nil? || ::BucketingVersion::VALID_VALUES.include?(@bucketingVersion) + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Invalid value of field bucketingVersion!') + end end ::Thrift::Struct.generate_accessors self diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java index 3d1c67f97c..abfe27caae 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java @@ -1667,6 +1667,8 @@ private Table convertToTable(MTable mtbl) throws MetaException { mtbl.getViewOriginalText(), mtbl.getViewExpandedText(), tableType); t.setCreationMetadata(convertToCreationMetadata(mtbl.getCreationMetadata())); t.setRewriteEnabled(mtbl.isRewriteEnabled()); + t.setBucketingVersion(mtbl.getBucketingVersion()); + t.setLoadInBucketedTable(mtbl.isLoadInBucketedTable()); return t; } @@ -1705,7 +1707,8 @@ private MTable convertToMTable(Table tbl) throws InvalidObjectException, .getCreateTime(), tbl.getLastAccessTime(), tbl.getRetention(), convertToMFieldSchemas(tbl.getPartitionKeys()), tbl.getParameters(), tbl.getViewOriginalText(), tbl.getViewExpandedText(), tbl.isRewriteEnabled(), - convertToMCreationMetadata(tbl.getCreationMetadata()), tableType); + convertToMCreationMetadata(tbl.getCreationMetadata()), tableType, + tbl.getBucketingVersion(), tbl.isLoadInBucketedTable()); } private List convertToMFieldSchemas(List keys) { diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTable.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTable.java index aea16ade7d..72ca7d56aa 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTable.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTable.java @@ -18,6 +18,8 @@ package org.apache.hadoop.hive.metastore.model; +import org.apache.hadoop.hive.metastore.api.BucketingVersion; + import java.util.List; import java.util.Map; @@ -37,6 +39,8 @@ private boolean rewriteEnabled; private MCreationMetadata creationMetadata; private String tableType; + private BucketingVersion bucketingVersion; + private boolean loadInBucketedTable; public MTable() {} @@ -58,7 +62,7 @@ public MTable(String tableName, MDatabase database, MStorageDescriptor sd, Strin int createTime, int lastAccessTime, int retention, List partitionKeys, Map parameters, String viewOriginalText, String viewExpandedText, boolean rewriteEnabled, MCreationMetadata creationMetadata, - String tableType) { + String tableType, BucketingVersion bucketingVersion, boolean loadInBucketedTable) { this.tableName = tableName; this.database = database; this.sd = sd; @@ -73,6 +77,8 @@ public MTable(String tableName, MDatabase database, MStorageDescriptor sd, Strin this.rewriteEnabled = rewriteEnabled; this.creationMetadata = creationMetadata; this.tableType = tableType; + this.bucketingVersion = bucketingVersion; + this.loadInBucketedTable = loadInBucketedTable; } /** @@ -270,4 +276,32 @@ public void setTableType(String tableType) { public String getTableType() { return tableType; } + + /** + * @param bucketingVersion used in bucketed table + */ + public void setBucketingVersion(BucketingVersion bucketingVersion) { + this.bucketingVersion = bucketingVersion; + } + + /** + * @return the bucketingVersion + */ + public BucketingVersion getBucketingVersion() { + return bucketingVersion; + } + /** + * @param loadInBucketedTable + */ + public void setLoadInBucketedTable(boolean loadInBucketedTable) { + this.loadInBucketedTable = loadInBucketedTable; + } + + /** + * @return the expertMode + */ + public boolean isLoadInBucketedTable() { + return loadInBucketedTable; + } + } diff --git a/standalone-metastore/src/main/thrift/hive_metastore.thrift b/standalone-metastore/src/main/thrift/hive_metastore.thrift index 371b97590c..cf4f64f6ce 100644 --- a/standalone-metastore/src/main/thrift/hive_metastore.thrift +++ b/standalone-metastore/src/main/thrift/hive_metastore.thrift @@ -310,6 +310,13 @@ struct StorageDescriptor { 12: optional bool storedAsSubDirectories // stored as subdirectories or not } +// Hash version for bucketing table +enum BucketingVersion { + INVALID_BUCKETING = 0, + JAVA_BUCKETING = 1, + MURMUR_BUCKETING = 2, +} + // table information struct Table { 1: string tableName, // name of the table @@ -327,7 +334,9 @@ struct Table { 13: optional PrincipalPrivilegeSet privileges, 14: optional bool temporary=false, 15: optional bool rewriteEnabled, // rewrite enabled or not - 16: optional CreationMetadata creationMetadata // only for MVs, it stores table names used and txn list at MV creation + 16: optional CreationMetadata creationMetadata, // only for MVs, it stores table names used and txn list at MV creation + 17: optional BucketingVersion bucketingVersion = BucketingVersion.JAVA_BUCKETING, // For bucketed table only. Default : 2, for existing tables, 1, for new tables 2. + 18: optional bool loadInBucketedTable = false // For bucketed table only. Default : false. true if user loads data using “load data” command. } struct Partition { diff --git a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/cache/TestCachedStore.java b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/cache/TestCachedStore.java index bd61df654a..09bed5d050 100644 --- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/cache/TestCachedStore.java +++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/cache/TestCachedStore.java @@ -29,21 +29,7 @@ import org.apache.hadoop.hive.metastore.ObjectStore; import org.apache.hadoop.hive.metastore.TableType; import org.apache.hadoop.hive.metastore.TestObjectStore.MockPartitionExpressionProxy; -import org.apache.hadoop.hive.metastore.api.AggrStats; -import org.apache.hadoop.hive.metastore.api.BasicTxnInfo; -import org.apache.hadoop.hive.metastore.api.BooleanColumnStatsData; -import org.apache.hadoop.hive.metastore.api.ColumnStatistics; -import org.apache.hadoop.hive.metastore.api.ColumnStatisticsData; -import org.apache.hadoop.hive.metastore.api.ColumnStatisticsDesc; -import org.apache.hadoop.hive.metastore.api.ColumnStatisticsObj; -import org.apache.hadoop.hive.metastore.api.Database; -import org.apache.hadoop.hive.metastore.api.FieldSchema; -import org.apache.hadoop.hive.metastore.api.NoSuchObjectException; -import org.apache.hadoop.hive.metastore.api.Partition; -import org.apache.hadoop.hive.metastore.api.PrincipalType; -import org.apache.hadoop.hive.metastore.api.SerDeInfo; -import org.apache.hadoop.hive.metastore.api.StorageDescriptor; -import org.apache.hadoop.hive.metastore.api.Table; +import org.apache.hadoop.hive.metastore.api.*; import org.apache.hadoop.hive.metastore.columnstats.cache.LongColumnStatsDataInspector; import org.apache.hadoop.hive.metastore.columnstats.cache.StringColumnStatsDataInspector; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; @@ -225,6 +211,8 @@ public void testTableOps() throws Exception { Table tbl1 = new Table(tblName1, dbName, tblOwner, 0, 0, 0, sd, new ArrayList<>(), tblParams, null, null, TableType.MANAGED_TABLE.toString()); + tbl1.setBucketingVersion(BucketingVersion.JAVA_BUCKETING); + tbl1.setLoadInBucketedTable(false); cachedStore.createTable(tbl1); tbl1 = cachedStore.getTable(dbName, tblName1); diff --git a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestTablesCreateDropAlterTruncate.java b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestTablesCreateDropAlterTruncate.java index 00f38eeec5..f10be72044 100644 --- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestTablesCreateDropAlterTruncate.java +++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestTablesCreateDropAlterTruncate.java @@ -21,18 +21,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.metastore.IMetaStoreClient; -import org.apache.hadoop.hive.metastore.api.AlreadyExistsException; -import org.apache.hadoop.hive.metastore.api.EnvironmentContext; -import org.apache.hadoop.hive.metastore.api.FieldSchema; -import org.apache.hadoop.hive.metastore.api.InvalidObjectException; -import org.apache.hadoop.hive.metastore.api.InvalidOperationException; -import org.apache.hadoop.hive.metastore.api.MetaException; -import org.apache.hadoop.hive.metastore.api.NoSuchObjectException; -import org.apache.hadoop.hive.metastore.api.Partition; -import org.apache.hadoop.hive.metastore.api.SerDeInfo; -import org.apache.hadoop.hive.metastore.api.SkewedInfo; -import org.apache.hadoop.hive.metastore.api.StorageDescriptor; -import org.apache.hadoop.hive.metastore.api.Table; +import org.apache.hadoop.hive.metastore.api.*; import org.apache.hadoop.hive.metastore.client.builder.DatabaseBuilder; import org.apache.hadoop.hive.metastore.client.builder.PartitionBuilder; import org.apache.hadoop.hive.metastore.client.builder.TableBuilder; @@ -224,6 +213,9 @@ public void tearDown() throws Exception { public void testCreateGetDeleteTable() throws Exception { // Try to create a table with all of the parameters set Table table = getTableWithAllParametersSet(); + // Set parameters set outside + table.setBucketingVersion(BucketingVersion.MURMUR_BUCKETING); + table.setLoadInBucketedTable(false); client.createTable(table); Table createdTable = client.getTable(table.getDbName(), table.getTableName()); // The createTime will be set on the server side, so the comparison should skip it @@ -684,6 +676,9 @@ public void testAlterTable() throws Exception { // Partition keys can not be set, but getTableWithAllParametersSet is added one, so remove for // this test newTable.setPartitionKeys(originalTable.getPartitionKeys()); + // Set the optional bucketingVersion and expertMode with default values + newTable.setBucketingVersion(BucketingVersion.JAVA_BUCKETING); + newTable.setLoadInBucketedTable(false); client.alter_table(originalDatabase, originalTableName, newTable); Table alteredTable = client.getTable(originalDatabase, originalTableName);