diff --git a/hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/listener/DbNotificationListener.java b/hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/listener/DbNotificationListener.java index 6321f9bdb7..717cc8afa8 100644 --- a/hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/listener/DbNotificationListener.java +++ b/hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/listener/DbNotificationListener.java @@ -23,6 +23,7 @@ import java.sql.SQLException; import java.sql.Statement; import java.util.ArrayList; +import java.util.Collections; import java.util.Iterator; import java.util.List; import java.util.concurrent.TimeUnit; @@ -75,11 +76,14 @@ import org.apache.hadoop.hive.metastore.events.AbortTxnEvent; import org.apache.hadoop.hive.metastore.events.AllocWriteIdEvent; import org.apache.hadoop.hive.metastore.events.ListenerEvent; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; +import org.apache.hadoop.hive.metastore.messaging.AcidWriteMessage; import org.apache.hadoop.hive.metastore.messaging.EventMessage.EventType; import org.apache.hadoop.hive.metastore.messaging.MessageFactory; import org.apache.hadoop.hive.metastore.messaging.OpenTxnMessage; import org.apache.hadoop.hive.metastore.messaging.PartitionFiles; import org.apache.hadoop.hive.metastore.tools.SQLGenerator; +import org.apache.hadoop.hive.metastore.txn.TxnUtils; import org.apache.hadoop.util.StringUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -269,10 +273,16 @@ public boolean hasNext() { public PartitionFiles next() { try { Partition p = partitionIter.next(); - List files = Lists.newArrayList(new FileIterator(p.getSd().getLocation())); + Iterator fileIterator; + //For transactional tables, the actual file copy will be done by acid write event during replay of commit txn. + if (!TxnUtils.isTransactionalTable(t)) { + List files = Lists.newArrayList(new FileIterator(p.getSd().getLocation())); + fileIterator = files.iterator(); + } else { + fileIterator = Collections.emptyIterator(); + } PartitionFiles partitionFiles = - new PartitionFiles(Warehouse.makePartName(t.getPartitionKeys(), p.getValues()), - files.iterator()); + new PartitionFiles(Warehouse.makePartName(t.getPartitionKeys(), p.getValues()), fileIterator); return partitionFiles; } catch (MetaException e) { throw new RuntimeException(e); @@ -414,10 +424,15 @@ public void onDropFunction(DropFunctionEvent fnEvent) throws MetaException { class FileChksumIterator implements Iterator { private List files; private List chksums; + private List subDirs; int i = 0; FileChksumIterator(List files, List chksums) { + this(files, chksums, null); + } + FileChksumIterator(List files, List chksums, List subDirs) { this.files = files; this.chksums = chksums; + this.subDirs = subDirs; } @Override public boolean hasNext() { @@ -428,7 +443,8 @@ public boolean hasNext() { public String next() { String result; try { - result = ReplChangeManager.encodeFileUri(files.get(i), chksums != null ? chksums.get(i) : null, null); + result = ReplChangeManager.encodeFileUri(files.get(i), chksums != null ? chksums.get(i) : null, + subDirs != null ? subDirs.get(i) : null); } catch (IOException e) { // File operations failed LOG.error("Encoding file URI failed with error " + e.getMessage()); @@ -623,6 +639,23 @@ public void onAllocWriteId(AllocWriteIdEvent allocWriteIdEvent, Connection dbCon } } + @Override + public void onAcidWrite(AcidWriteEvent acidWriteEvent, Connection dbConn, SQLGenerator sqlGenerator) + throws MetaException { + AcidWriteMessage msg = msgFactory.buildAcidWriteMessage(acidWriteEvent, + new FileChksumIterator(acidWriteEvent.getFiles(), acidWriteEvent.getChecksums(), + acidWriteEvent.getSubDirs())); + NotificationEvent event = new NotificationEvent(0, now(), EventType.ACID_WRITE.toString(), msg.toString()); + event.setMessageFormat(msgFactory.getMessageFormat()); + event.setDbName(acidWriteEvent.getDatabase()); + event.setTableName(acidWriteEvent.getTable()); + try { + addWriteNotificationLog(event, acidWriteEvent, dbConn, sqlGenerator, msg); + } catch (SQLException e) { + throw new MetaException("Unable to add write notification log " + StringUtils.stringifyException(e)); + } + } + private int now() { long millis = System.currentTimeMillis(); millis /= 1000; @@ -634,12 +667,133 @@ private int now() { return (int)millis; } + /** + * Close statement instance. + * @param stmt statement instance. + */ + private static void closeStmt(Statement stmt) { + try { + if (stmt != null && !stmt.isClosed()) { + stmt.close(); + } + } catch (SQLException e) { + LOG.warn("Failed to close statement " + e.getMessage()); + } + } + + /** + * Close the ResultSet. + * @param rs may be {@code null} + */ + private static void close(ResultSet rs) { + try { + if (rs != null && !rs.isClosed()) { + rs.close(); + } + } catch(SQLException ex) { + LOG.warn("Failed to close result set " + ex.getMessage()); + } + } + + private long getNextNLId(Statement stmt, SQLGenerator sqlGenerator, String sequence) + throws SQLException, MetaException { + String s = sqlGenerator.addForUpdateClause("select \"NEXT_VAL\" from " + + "\"SEQUENCE_TABLE\" where \"SEQUENCE_NAME\" = " + quoteString(sequence)); + LOG.debug("Going to execute query <" + s + ">"); + ResultSet rs = null; + try { + rs = stmt.executeQuery(s); + if (!rs.next()) { + throw new MetaException("Transaction database not properly configured, can't find next NL id."); + } + + long nextNLId = rs.getLong(1); + long updatedNLId = nextNLId + 1; + s = "update \"SEQUENCE_TABLE\" set \"NEXT_VAL\" = " + updatedNLId + " where \"SEQUENCE_NAME\" = " + + quoteString(sequence); + LOG.debug("Going to execute update <" + s + ">"); + stmt.executeUpdate(s); + return nextNLId; + }finally { + close(rs); + } + } + + private void addWriteNotificationLog(NotificationEvent event, AcidWriteEvent acidWriteEvent, Connection dbConn, + SQLGenerator sqlGenerator, AcidWriteMessage msg) throws MetaException, SQLException { + LOG.debug("DbNotificationListener: adding write notification log for : {}", event.getMessage()); + assert ((dbConn != null) && (sqlGenerator != null)); + + Statement stmt =null; + ResultSet rs = null; + String dbName = acidWriteEvent.getDatabase(); + String tblName = acidWriteEvent.getTable(); + String partition = acidWriteEvent.getPartition(); + String tableObj = msg.getTableObjStr(); + String partitionObj = msg.getPartitionObjStr(); + String files = ReplChangeManager.joinWithSeparator(msg.getFiles()); + + try { + stmt = dbConn.createStatement(); + if (sqlGenerator.getDbProduct() == MYSQL) { + stmt.execute("SET @@session.sql_mode=ANSI_QUOTES"); + } + + String s = sqlGenerator.addForUpdateClause("select \"WNL_FILES\", \"WNL_ID\" from" + + " \"TXN_WRITE_NOTIFICATION_LOG\" " + + "where \"WNL_DATABASE\" = " + quoteString(dbName) + + "and \"WNL_TABLE\" = " + quoteString(tblName) + " and \"WNL_PARTITION\" = " + + quoteString(partition) + " and \"WNL_TXNID\" = " + Long.toString(acidWriteEvent.getTxnId())); + LOG.debug("Going to execute query <" + s + ">"); + rs = stmt.executeQuery(s); + if (!rs.next()) { + // if rs is empty then no lock is taken and thus it can not cause deadlock. + long nextNLId = getNextNLId(stmt, sqlGenerator, + "org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog"); + s = "insert into \"TXN_WRITE_NOTIFICATION_LOG\" (\"WNL_ID\", \"WNL_TXNID\", \"WNL_WRITEID\"," + + " \"WNL_DATABASE\", \"WNL_TABLE\"," + + " \"WNL_PARTITION\", \"WNL_TABLE_OBJ\", \"WNL_PARTITION_OBJ\", \"WNL_FILES\", \"WNL_EVENT_TIME\")" + + " values (" + nextNLId + + "," + acidWriteEvent.getTxnId() + "," + acidWriteEvent.getWriteId()+ "," + + quoteString(dbName)+ "," + quoteString(tblName)+ "," + quoteString(partition)+ "," + + quoteString(tableObj)+ "," + quoteString(partitionObj) + "," + quoteString(files)+ + "," + now() + ")"; + LOG.info("Going to execute insert <" + s + ">"); + stmt.execute(sqlGenerator.addEscapeCharacters(s)); + } else { + String existingFiles = rs.getString(1); + if (existingFiles.contains(sqlGenerator.addEscapeCharacters(files))) { + // If list of files are already present then no need to update it again. This scenario can come in case of + // retry done to the meta store for the same operation. + LOG.info("file list " + files + " already present"); + return; + } + long nlId = rs.getLong(2); + files = ReplChangeManager.joinWithSeparator(Lists.newArrayList(files, existingFiles)); + s = "update \"TXN_WRITE_NOTIFICATION_LOG\" set \"WNL_TABLE_OBJ\" = " + quoteString(tableObj) + "," + + " \"WNL_PARTITION_OBJ\" = " + quoteString(partitionObj) + "," + + " \"WNL_FILES\" = " + quoteString(files) + "," + + " \"WNL_EVENT_TIME\" = " + now() + + " where \"WNL_ID\" = " + nlId; + LOG.info("Going to execute update <" + s + ">"); + stmt.executeUpdate(sqlGenerator.addEscapeCharacters(s)); + } + } catch (SQLException e) { + LOG.warn("failed to add write notification log" + e.getMessage()); + throw e; + } finally { + closeStmt(stmt); + close(rs); + } + } + static String quoteString(String input) { return "'" + input + "'"; } private void addNotificationLog(NotificationEvent event, ListenerEvent listenerEvent, Connection dbConn, SQLGenerator sqlGenerator) throws MetaException, SQLException { + LOG.debug("DbNotificationListener: adding notification log for : {}", event.getMessage()); if ((dbConn == null) || (sqlGenerator == null)) { LOG.info("connection or sql generator is not set so executing sql via DN"); process(event, listenerEvent); @@ -669,22 +823,8 @@ private void addNotificationLog(NotificationEvent event, ListenerEvent listenerE LOG.debug("Going to execute update <" + s + ">"); stmt.executeUpdate(s); - s = sqlGenerator.addForUpdateClause("select \"NEXT_VAL\" from " + - "\"SEQUENCE_TABLE\" where \"SEQUENCE_NAME\" = " + - " 'org.apache.hadoop.hive.metastore.model.MNotificationLog'"); - LOG.debug("Going to execute query <" + s + ">"); - rs = stmt.executeQuery(s); - if (!rs.next()) { - throw new MetaException("failed to get next NEXT_VAL from SEQUENCE_TABLE"); - } - - long nextNLId = rs.getLong(1); - long updatedNLId = nextNLId + 1; - s = "update \"SEQUENCE_TABLE\" set \"NEXT_VAL\" = " + updatedNLId + " where \"SEQUENCE_NAME\" = " + - - " 'org.apache.hadoop.hive.metastore.model.MNotificationLog'"; - LOG.debug("Going to execute update <" + s + ">"); - stmt.executeUpdate(s); + long nextNLId = getNextNLId(stmt, sqlGenerator, + "org.apache.hadoop.hive.metastore.model.MNotificationLog"); List insert = new ArrayList<>(); @@ -712,20 +852,8 @@ private void addNotificationLog(NotificationEvent event, ListenerEvent listenerE LOG.warn("failed to add notification log" + e.getMessage()); throw e; } finally { - if (stmt != null && !stmt.isClosed()) { - try { - stmt.close(); - } catch (SQLException e) { - LOG.warn("Failed to close statement " + e.getMessage()); - } - } - if (rs != null && !rs.isClosed()) { - try { - rs.close(); - } catch (SQLException e) { - LOG.warn("Failed to close result set " + e.getMessage()); - } - } + closeStmt(stmt); + close(rs); } } @@ -742,12 +870,12 @@ private void process(NotificationEvent event, ListenerEvent listenerEvent) throw event.getMessage()); HMSHandler.getMSForConf(conf).addNotificationEvent(event); - // Set the DB_NOTIFICATION_EVENT_ID for future reference by other listeners. - if (event.isSetEventId()) { - listenerEvent.putParameter( - MetaStoreEventListenerConstants.DB_NOTIFICATION_EVENT_ID_KEY_NAME, - Long.toString(event.getEventId())); - } + // Set the DB_NOTIFICATION_EVENT_ID for future reference by other listeners. + if (event.isSetEventId()) { + listenerEvent.putParameter( + MetaStoreEventListenerConstants.DB_NOTIFICATION_EVENT_ID_KEY_NAME, + Long.toString(event.getEventId())); + } } private static class CleanerThread extends Thread { @@ -768,6 +896,7 @@ public void run() { while (true) { try { rs.cleanNotificationEvents(ttl); + rs.cleanWriteNotificationEvents(ttl); } catch (Exception ex) { //catching exceptions here makes sure that the thread doesn't die in case of unexpected //exceptions diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java index 7271c3d977..33f2ea09e5 100644 --- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java +++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/DummyRawStoreFailEvent.java @@ -89,6 +89,7 @@ import org.apache.hadoop.hive.metastore.api.WMMapping; import org.apache.hadoop.hive.metastore.api.WMPool; import org.apache.hadoop.hive.metastore.api.WMNullablePool; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.ColStatsObjWithSourceInfo; import org.apache.thrift.TException; @@ -872,6 +873,20 @@ public void cleanNotificationEvents(int olderThan) { objectStore.cleanNotificationEvents(olderThan); } + @Override + public void cleanWriteNotificationEvents(int olderThan) { + if (!shouldEventSucceed) { + //throw exception to simulate an issue with cleaner thread + throw new RuntimeException("Dummy exception while cleaning write notifications"); + } + objectStore.cleanWriteNotificationEvents(olderThan); + } + + @Override + public List getAllWriteEventInfo(long txnId, String dbName, String tableName) throws MetaException { + return objectStore.getAllWriteEventInfo(txnId, dbName, tableName); + } + @Override public CurrentNotificationEventId getCurrentNotificationEventId() { return objectStore.getCurrentNotificationEventId(); diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java index eef917e9f4..82429e36a5 100644 --- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java +++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java @@ -75,6 +75,7 @@ import org.apache.hadoop.hive.metastore.events.AbortTxnEvent; import org.apache.hadoop.hive.metastore.events.ListenerEvent; import org.apache.hadoop.hive.metastore.events.AllocWriteIdEvent; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.messaging.AddPartitionMessage; import org.apache.hadoop.hive.metastore.messaging.AlterPartitionMessage; import org.apache.hadoop.hive.metastore.messaging.AlterTableMessage; @@ -238,6 +239,10 @@ public void onAbortTxn(AbortTxnEvent abortTxnEvent) throws MetaException { public void onAllocWriteId(AllocWriteIdEvent allocWriteIdEvent) throws MetaException { pushEventId(EventType.ALLOC_WRITE_ID, allocWriteIdEvent); } + + public void onAcidWrite(AcidWriteEvent acidWriteEvent) throws MetaException { + pushEventId(EventType.ACID_WRITE, acidWriteEvent); + } } @SuppressWarnings("rawtypes") diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java index d1f0def580..1e0efe72ce 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java @@ -2832,78 +2832,6 @@ public void testRemoveStats() throws IOException { verifyRun("SELECT max(a) from " + replDbName + ".ptned2 where b=1", new String[]{"8"}, driverMirror); } - // TODO: This test should be removed once ACID tables replication is supported. - @Test - public void testSkipTables() throws Exception { - String testName = "skipTables"; - String dbName = createDB(testName, driver); - String replDbName = dbName + "_dupe"; - - // TODO: this is wrong; this test sets up dummy txn manager and so it cannot create ACID tables. - // If I change it to use proper txn manager, the setup for some tests hangs. - // This used to work by accident, now this works due a test flag. The test needs to be fixed. - // Create table - run("CREATE TABLE " + dbName + ".acid_table (key int, value int) PARTITIONED BY (load_date date) " + - "CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES ('transactional'='true')", driver); - run("CREATE TABLE " + dbName + ".mm_table (key int, value int) PARTITIONED BY (load_date date) " + - "CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES ('transactional'='true'," + - " 'transactional_properties'='insert_only')", driver); - verifyIfTableExist(dbName, "acid_table", metaStoreClient); - verifyIfTableExist(dbName, "mm_table", metaStoreClient); - - // Bootstrap test - Tuple bootstrapDump = bootstrapLoadAndVerify(dbName, replDbName); - String replDumpId = bootstrapDump.lastReplId; - verifyIfTableNotExist(replDbName, "acid_table", metaStoreClientMirror); - verifyIfTableNotExist(replDbName, "mm_table", metaStoreClientMirror); - - // Test alter table - run("ALTER TABLE " + dbName + ".acid_table RENAME TO " + dbName + ".acid_table_rename", driver); - verifyIfTableExist(dbName, "acid_table_rename", metaStoreClient); - - // Dummy create table command to mark proper last repl ID after dump - run("CREATE TABLE " + dbName + ".dummy (a int)", driver); - - // Perform REPL-DUMP/LOAD - Tuple incrementalDump = incrementalLoadAndVerify(dbName, replDumpId, replDbName); - replDumpId = incrementalDump.lastReplId; - verifyIfTableNotExist(replDbName, "acid_table_rename", metaStoreClientMirror); - - // Create another table for incremental repl verification - run("CREATE TABLE " + dbName + ".acid_table_incremental (key int, value int) PARTITIONED BY (load_date date) " + - "CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES ('transactional'='true')", driver); - run("CREATE TABLE " + dbName + ".mm_table_incremental (key int, value int) PARTITIONED BY (load_date date) " + - "CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES ('transactional'='true'," + - " 'transactional_properties'='insert_only')", driver); - verifyIfTableExist(dbName, "acid_table_incremental", metaStoreClient); - verifyIfTableExist(dbName, "mm_table_incremental", metaStoreClient); - - // Dummy insert into command to mark proper last repl ID after dump - run("INSERT INTO " + dbName + ".dummy values(1)", driver); - - // Perform REPL-DUMP/LOAD - incrementalDump = incrementalLoadAndVerify(dbName, replDumpId, replDbName); - replDumpId = incrementalDump.lastReplId; - verifyIfTableNotExist(replDbName, "acid_table_incremental", metaStoreClientMirror); - verifyIfTableNotExist(replDbName, "mm_table_incremental", metaStoreClientMirror); - - // Test adding a constraint - run("ALTER TABLE " + dbName + ".acid_table_incremental ADD CONSTRAINT key_pk PRIMARY KEY (key) DISABLE NOVALIDATE", driver); - try { - List pks = metaStoreClient.getPrimaryKeys(new PrimaryKeysRequest(dbName, "acid_table_incremental")); - assertEquals(pks.size(), 1); - } catch (TException te) { - assertNull(te); - } - - // Dummy insert into command to mark proper last repl ID after dump - run("INSERT INTO " + dbName + ".dummy values(2)", driver); - - // Perform REPL-DUMP/LOAD - incrementalLoadAndVerify(dbName, replDumpId, replDbName); - verifyIfTableNotExist(replDbName, "acid_table_incremental", metaStoreClientMirror); - } - @Test public void testDeleteStagingDir() throws IOException { String testName = "deleteStagingDir"; diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java index 3ee0747ebd..746888633e 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcidTables.java @@ -37,6 +37,7 @@ import org.apache.hadoop.hive.ql.ErrorMsg; import org.apache.hadoop.hive.ql.processors.CommandProcessorResponse; import org.junit.rules.TestName; + import org.junit.rules.TestRule; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -53,6 +54,8 @@ import java.util.HashMap; import java.util.List; import javax.annotation.Nullable; +import java.util.Collections; +import com.google.common.collect.Lists; /** * TestReplicationScenariosAcidTables - test replication for ACID tables @@ -66,8 +69,13 @@ protected static final Logger LOG = LoggerFactory.getLogger(TestReplicationScenarios.class); private static WarehouseInstance primary, replica, replicaNonAcid; - private String primaryDbName, replicatedDbName; private static HiveConf conf; + private String primaryDbName, replicatedDbName, primaryDbNameExtra; + private enum OperationType { + REPL_TEST_ACID_INSERT, REPL_TEST_ACID_INSERT_SELECT, REPL_TEST_ACID_CTAS, + REPL_TEST_ACID_INSERT_OVERWRITE, REPL_TEST_ACID_INSERT_IMPORT, REPL_TEST_ACID_INSERT_LOADLOCAL, + REPL_TEST_ACID_INSERT_UNION + } @BeforeClass public static void classLevelSetup() throws Exception { @@ -80,9 +88,13 @@ public static void classLevelSetup() throws Exception { put("fs.defaultFS", miniDFSCluster.getFileSystem().getUri().toString()); put("hive.support.concurrency", "true"); put("hive.txn.manager", "org.apache.hadoop.hive.ql.lockmgr.DbTxnManager"); - put("hive.repl.dump.include.acid.tables", "true"); put("hive.metastore.client.capability.check", "false"); put("hive.repl.bootstrap.dump.open.txn.timeout", "1s"); + put("hive.exec.dynamic.partition.mode", "nonstrict"); + put("hive.strict.checks.bucketing", "false"); + put("hive.mapred.mode", "nonstrict"); + put("mapred.input.dir.recursive", "true"); + put("hive.metastore.disallow.incompatible.col.type.changes", "false"); }}; primary = new WarehouseInstance(LOG, miniDFSCluster, overridesForHiveConf); replica = new WarehouseInstance(LOG, miniDFSCluster, overridesForHiveConf); @@ -90,7 +102,6 @@ public static void classLevelSetup() throws Exception { put("fs.defaultFS", miniDFSCluster.getFileSystem().getUri().toString()); put("hive.support.concurrency", "false"); put("hive.txn.manager", "org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager"); - put("hive.repl.dump.include.acid.tables", "true"); put("hive.metastore.client.capability.check", "false"); }}; replicaNonAcid = new WarehouseInstance(LOG, miniDFSCluster, overridesForHiveConf1); @@ -109,6 +120,9 @@ public void setup() throws Throwable { replicatedDbName = "replicated_" + primaryDbName; primary.run("create database " + primaryDbName + " WITH DBPROPERTIES ( '" + SOURCE_OF_REPLICATION + "' = '1,2,3')"); + primaryDbNameExtra = primaryDbName+"_extra"; + primary.run("create database " + primaryDbNameExtra + " WITH DBPROPERTIES ( '" + + SOURCE_OF_REPLICATION + "' = '1,2,3')"); } @After @@ -116,6 +130,7 @@ public void tearDown() throws Throwable { primary.run("drop database if exists " + primaryDbName + " cascade"); replica.run("drop database if exists " + replicatedDbName + " cascade"); replicaNonAcid.run("drop database if exists " + replicatedDbName + " cascade"); + primary.run("drop database if exists " + primaryDbName + "_extra cascade"); } @Test @@ -488,4 +503,585 @@ public void testDumpAcidTableWithTableDirMissing() throws Throwable { primary.run("DROP TABLE " + dbName + ".normal"); primary.run("drop database " + dbName); } + + @Test + public void testAcidTableIncrementalReplication() throws Throwable { + WarehouseInstance.Tuple bootStrapDump = primary.dump(primaryDbName, null); + replica.load(replicatedDbName, bootStrapDump.dumpLocation) + .run("REPL STATUS " + replicatedDbName) + .verifyResult(bootStrapDump.lastReplicationId); + List selectStmtList = new ArrayList<>(); + List expectedValues = new ArrayList<>(); + + appendInsert(selectStmtList, expectedValues); + appendDelete(selectStmtList, expectedValues); + appendUpdate(selectStmtList, expectedValues); + appendTruncate(selectStmtList, expectedValues); + appendInsertIntoFromSelect(selectStmtList, expectedValues); + appendMerge(selectStmtList, expectedValues); + appendCreateAsSelect(selectStmtList, expectedValues); + appendImport(selectStmtList, expectedValues); + appendInsertOverwrite(selectStmtList, expectedValues); + //appendLoadLocal(selectStmtList, expectedValues); + appendInsertUnion(selectStmtList, expectedValues); + appendMultiStatementTxn(selectStmtList, expectedValues); + appendMultiStatementTxnUpdateDelete(selectStmtList, expectedValues); + + verifyIncrementalLoad(selectStmtList, expectedValues, bootStrapDump.lastReplicationId); + } + + private void appendInsert(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testInsert"; + String tableNameMM = tableName + "_MM"; + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + private void appendDelete(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testDelete"; + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + deleteRecords(tableName); + selectStmtList.add("select count(*) from " + tableName); + expectedValues.add(new String[] {"0"}); + } + + private void appendUpdate(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testUpdate"; + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + updateRecords(tableName); + selectStmtList.add("select value from " + tableName + " order by value"); + expectedValues.add(new String[] {"1", "100", "100", "100", "100"}); + } + + private void appendTruncate(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testTruncate"; + String tableNameMM = tableName + "_MM"; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + truncateTable(primaryDbName, tableName); + selectStmtList.add("select count(*) from " + tableName); + expectedValues.add(new String[] {"0"}); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + truncateTable(primaryDbName, tableNameMM); + selectStmtList.add("select count(*) from " + tableNameMM); + expectedValues.add(new String[] {"0"}); + } + + private void appendInsertIntoFromSelect(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testInsertIntoFromSelect"; + String tableNameMM =tableName + "_MM"; + String tableNameSelect = testName.getMethodName() + "_Select"; + String tableNameSelectMM = testName.getMethodName() + "_SelectMM"; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableName, tableNameSelect, false, OperationType.REPL_TEST_ACID_INSERT_SELECT); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameSelect + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableNameMM, tableNameSelectMM, true, OperationType.REPL_TEST_ACID_INSERT_SELECT); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameSelectMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + private void appendMerge(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testMerge"; + String tableNameMerge = testName.getMethodName() + "_Merge"; + + insertForMerge(tableName, tableNameMerge, false); + selectStmtList.add("select last_update_user from " + tableName + " order by last_update_user"); + expectedValues.add(new String[] {"creation", "creation", "creation", "creation", "creation", + "creation", "creation", "merge_update", "merge_insert", "merge_insert"}); + selectStmtList.add("select ID from " + tableNameMerge + " order by ID"); + expectedValues.add(new String[] {"1", "4", "7", "8", "8", "11"}); + } + + private void appendCreateAsSelect(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testCreateAsSelect"; + String tableNameMM = tableName + "_MM"; + String tableNameCTAS = testName.getMethodName() + "_CTAS"; + String tableNameCTASMM = testName.getMethodName() + "_CTASMM"; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableName, tableNameCTAS, false, OperationType.REPL_TEST_ACID_CTAS); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameCTAS + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableNameMM, tableNameCTASMM, true, OperationType.REPL_TEST_ACID_CTAS); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameCTASMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + private void appendImport(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testImport"; + String tableNameMM = tableName + "_MM"; + String tableNameImport = testName.getMethodName() + "_Import"; + String tableNameImportMM = testName.getMethodName() + "_ImportMM"; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableName, tableNameImport, false, OperationType.REPL_TEST_ACID_INSERT_IMPORT); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameImport + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableNameMM, tableNameImportMM, true, OperationType.REPL_TEST_ACID_INSERT_IMPORT); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameImportMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + private void appendInsertOverwrite(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testInsertOverwrite"; + String tableNameOW = testName.getMethodName() +"_OW"; + String tableNameMM = tableName + "_MM"; + String tableNameOWMM = testName.getMethodName() +"_OWMM"; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableName, tableNameOW, false, OperationType.REPL_TEST_ACID_INSERT_OVERWRITE); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameOW + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableNameMM, tableNameOWMM, true, OperationType.REPL_TEST_ACID_INSERT_OVERWRITE); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameOWMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + //TODO: need to check why its failing. Loading to acid table from local path is failing. + private void appendLoadLocal(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testLoadLocal"; + String tableNameLL = testName.getMethodName() +"_LL"; + String tableNameMM = tableName + "_MM"; + String tableNameLLMM = testName.getMethodName() +"_LLMM"; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableName, tableNameLL, false, OperationType.REPL_TEST_ACID_INSERT_LOADLOCAL); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameLL + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableNameMM, tableNameLLMM, true, OperationType.REPL_TEST_ACID_INSERT_LOADLOCAL); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + selectStmtList.add("select key from " + tableNameLLMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + private void appendInsertUnion(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testInsertUnion"; + String tableNameUnion = testName.getMethodName() +"_UNION"; + String tableNameMM = tableName + "_MM"; + String tableNameUnionMM = testName.getMethodName() +"_UNIONMM"; + String[] resultArray = new String[]{"1", "2", "3", "4", "5"}; + String[] resultArrayUnion = new String[]{"1", "1", "2", "2", "3", "3", "4", "4", "5", "5"}; + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableName, tableNameUnion, false, OperationType.REPL_TEST_ACID_INSERT_UNION); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(resultArray); + selectStmtList.add( "select key from " + tableNameUnion + " order by key"); + expectedValues.add(resultArrayUnion); + selectStmtList.add("select key from " + tableName + "_nopart" + " order by key"); + expectedValues.add(resultArray); + selectStmtList.add("select key from " + tableNameUnion + "_nopart" + " order by key"); + expectedValues.add(resultArrayUnion); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + insertRecords(tableNameMM, tableNameUnionMM, true, OperationType.REPL_TEST_ACID_INSERT_UNION); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(resultArray); + selectStmtList.add( "select key from " + tableNameUnionMM + " order by key"); + expectedValues.add(resultArrayUnion); + selectStmtList.add("select key from " + tableNameMM + "_nopart" + " order by key"); + expectedValues.add(resultArray); + selectStmtList.add("select key from " + tableNameUnionMM + "_nopart" + " order by key"); + expectedValues.add(resultArrayUnion); + } + + private void appendMultiStatementTxn(List selectStmtList, List expectedValues) throws Throwable { + String tableName = testName.getMethodName() + "testMultiStatementTxn"; + String[] resultArray = new String[]{"1", "2", "3", "4", "5"}; + String tableNameMM = tableName + "_MM"; + String tableProperty = "'transactional'='true'"; + + insertIntoDB(primaryDbName, tableName, tableProperty, resultArray, true); + selectStmtList.add("select key from " + tableName + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + + tableProperty = setMMtableProperty(tableProperty); + insertIntoDB(primaryDbName, tableNameMM, tableProperty, resultArray, true); + selectStmtList.add("select key from " + tableNameMM + " order by key"); + expectedValues.add(new String[]{"1", "2", "3", "4", "5"}); + } + + private void appendMultiStatementTxnUpdateDelete(List selectStmtList, List expectedValues) + throws Throwable { + String tableName = testName.getMethodName() + "testMultiStatementTxnUpdate"; + String tableNameDelete = testName.getMethodName() + "testMultiStatementTxnDelete"; + String[] resultArray = new String[]{"1", "2", "3", "4", "5"}; + String tableProperty = "'transactional'='true'"; + + insertIntoDB(primaryDbName, tableName, tableProperty, resultArray, true); + updateRecords(tableName); + selectStmtList.add("select value from " + tableName + " order by value"); + expectedValues.add(new String[] {"1", "100", "100", "100", "100"}); + + insertIntoDB(primaryDbName, tableNameDelete, tableProperty, resultArray, true); + deleteRecords(tableNameDelete); + selectStmtList.add("select count(*) from " + tableNameDelete); + expectedValues.add(new String[] {"0"}); + } + + @Test + public void testReplCM() throws Throwable { + String tableName = testName.getMethodName(); + String tableNameMM = testName.getMethodName() + "_MM"; + String[] result = new String[]{"5"}; + + WarehouseInstance.Tuple incrementalDump; + WarehouseInstance.Tuple bootStrapDump = primary.dump(primaryDbName, null); + replica.load(replicatedDbName, bootStrapDump.dumpLocation) + .run("REPL STATUS " + replicatedDbName) + .verifyResult(bootStrapDump.lastReplicationId); + + insertRecords(tableName, null, false, OperationType.REPL_TEST_ACID_INSERT); + incrementalDump = primary.dump(primaryDbName, bootStrapDump.lastReplicationId); + truncateTable(primaryDbName, tableName); + replica.loadWithoutExplain(replicatedDbName, incrementalDump.dumpLocation) + .run("REPL STATUS " + replicatedDbName).verifyResult(incrementalDump.lastReplicationId); + verifyResultsInReplica(Lists.newArrayList("select count(*) from " + tableName, + "select count(*) from " + tableName + "_nopart"), + Lists.newArrayList(result, result)); + + insertRecords(tableNameMM, null, true, OperationType.REPL_TEST_ACID_INSERT); + incrementalDump = primary.dump(primaryDbName, bootStrapDump.lastReplicationId); + truncateTable(primaryDbName, tableNameMM); + replica.loadWithoutExplain(replicatedDbName, incrementalDump.dumpLocation) + .run("REPL STATUS " + replicatedDbName).verifyResult(incrementalDump.lastReplicationId); + verifyResultsInReplica(Lists.newArrayList("select count(*) from " + tableNameMM, + "select count(*) from " + tableNameMM + "_nopart"), + Lists.newArrayList(result, result)); + } + + @Test + public void testMultiDBTxn() throws Throwable { + String tableName = testName.getMethodName(); + String dbName1 = tableName + "_db1"; + String dbName2 = tableName + "_db2"; + String[] resultArray = new String[]{"1", "2", "3", "4", "5"}; + String tableProperty = "'transactional'='true'"; + String txnStrStart = "START TRANSACTION"; + String txnStrCommit = "COMMIT"; + + WarehouseInstance.Tuple incrementalDump; + primary.run("alter database default set dbproperties ('repl.source.for' = '1, 2, 3')"); + WarehouseInstance.Tuple bootStrapDump = primary.dump("`*`", null); + + primary.run("use " + primaryDbName) + .run("create database " + dbName1 + " WITH DBPROPERTIES ( '" + SOURCE_OF_REPLICATION + "' = '1,2,3')") + .run("create database " + dbName2 + " WITH DBPROPERTIES ( '" + SOURCE_OF_REPLICATION + "' = '1,2,3')") + .run("CREATE TABLE " + dbName1 + "." + tableName + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("use " + dbName1) + .run("SHOW TABLES LIKE '" + tableName + "'") + .verifyResult(tableName) + .run("CREATE TABLE " + dbName2 + "." + tableName + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("use " + dbName2) + .run("SHOW TABLES LIKE '" + tableName + "'") + .verifyResult(tableName) + .run(txnStrStart) + .run("INSERT INTO " + dbName2 + "." + tableName + " partition (load_date='2016-03-02') VALUES (5, 5)") + .run("INSERT INTO " + dbName1 + "." + tableName + " partition (load_date='2016-03-01') VALUES (1, 1)") + .run("INSERT INTO " + dbName1 + "." + tableName + " partition (load_date='2016-03-01') VALUES (2, 2)") + .run("INSERT INTO " + dbName2 + "." + tableName + " partition (load_date='2016-03-01') VALUES (2, 2)") + .run("INSERT INTO " + dbName2 + "." + tableName + " partition (load_date='2016-03-02') VALUES (3, 3)") + .run("INSERT INTO " + dbName1 + "." + tableName + " partition (load_date='2016-03-02') VALUES (3, 3)") + .run("INSERT INTO " + dbName1 + "." + tableName + " partition (load_date='2016-03-03') VALUES (4, 4)") + .run("INSERT INTO " + dbName1 + "." + tableName + " partition (load_date='2016-03-02') VALUES (5, 5)") + .run("INSERT INTO " + dbName2 + "." + tableName + " partition (load_date='2016-03-01') VALUES (1, 1)") + .run("INSERT INTO " + dbName2 + "." + tableName + " partition (load_date='2016-03-03') VALUES (4, 4)") + .run("select key from " + dbName2 + "." + tableName + " order by key") + .verifyResults(resultArray) + .run("select key from " + dbName1 + "." + tableName + " order by key") + .verifyResults(resultArray) + .run(txnStrCommit); + + incrementalDump = primary.dump("`*`", bootStrapDump.lastReplicationId); + + // Due to the limitation that we can only have one instance of Persistence Manager Factory in a JVM + // we are not able to create multiple embedded derby instances for two different MetaStore instances. + primary.run("drop database " + primaryDbName + " cascade"); + primary.run("drop database " + dbName1 + " cascade"); + primary.run("drop database " + dbName2 + " cascade"); + //End of additional steps + + replica.loadWithoutExplain("", bootStrapDump.dumpLocation) + .run("REPL STATUS default") + .verifyResult(bootStrapDump.lastReplicationId); + + replica.loadWithoutExplain("", incrementalDump.dumpLocation) + .run("REPL STATUS " + dbName1) + .run("select key from " + dbName1 + "." + tableName + " order by key") + .verifyResults(resultArray) + .run("select key from " + dbName2 + "." + tableName + " order by key") + .verifyResults(resultArray); + + replica.run("drop database " + primaryDbName + " cascade"); + replica.run("drop database " + dbName1 + " cascade"); + replica.run("drop database " + dbName2 + " cascade"); + } + + private void verifyResultsInReplica(List selectStmtList, List expectedValues) throws Throwable { + for (int idx = 0; idx < selectStmtList.size(); idx++) { + replica.run("use " + replicatedDbName) + .run(selectStmtList.get(idx)) + .verifyResults(expectedValues.get(idx)); + } + } + + private WarehouseInstance.Tuple verifyIncrementalLoad(List selectStmtList, + List expectedValues, String lastReplId) throws Throwable { + WarehouseInstance.Tuple incrementalDump = primary.dump(primaryDbName, lastReplId); + replica.loadWithoutExplain(replicatedDbName, incrementalDump.dumpLocation) + .run("REPL STATUS " + replicatedDbName).verifyResult(incrementalDump.lastReplicationId); + verifyResultsInReplica(selectStmtList, expectedValues); + + replica.loadWithoutExplain(replicatedDbName, incrementalDump.dumpLocation) + .run("REPL STATUS " + replicatedDbName).verifyResult(incrementalDump.lastReplicationId); + verifyResultsInReplica(selectStmtList, expectedValues); + return incrementalDump; + } + + private void deleteRecords(String tableName) throws Throwable { + primary.run("use " + primaryDbName) + .run("delete from " + tableName) + .run("select count(*) from " + tableName) + .verifyResult("0"); + } + + private void updateRecords(String tableName) throws Throwable { + primary.run("use " + primaryDbName) + .run("update " + tableName + " set value = 100 where key >= 2") + .run("select value from " + tableName + " order by value") + .verifyResults(new String[] {"1", "100", "100", "100", "100"}); + } + + private void truncateTable(String dbName, String tableName) throws Throwable { + primary.run("use " + dbName) + .run("truncate table " + tableName) + .run("select count(*) from " + tableName) + .verifyResult("0") + .run("truncate table " + tableName + "_nopart") + .run("select count(*) from " + tableName + "_nopart") + .verifyResult("0"); + } + + private WarehouseInstance.Tuple verifyLoad(String tableName, String tableNameOp, String lastReplId) throws Throwable { + String[] resultArray = new String[]{"1", "2", "3", "4", "5"}; + if (tableNameOp == null) { + return verifyIncrementalLoad(Lists.newArrayList("select key from " + tableName + " order by key", + "select key from " + tableName + "_nopart order by key"), + Lists.newArrayList(resultArray, resultArray), lastReplId); + } + return verifyIncrementalLoad(Lists.newArrayList("select key from " + tableName + " order by key", + "select key from " + tableNameOp + " order by key", + "select key from " + tableName + "_nopart" + " order by key", + "select key from " + tableNameOp + "_nopart" + " order by key"), + Lists.newArrayList(resultArray, resultArray, resultArray, resultArray), lastReplId); + } + + private void insertIntoDB(String dbName, String tableName, String tableProperty, String[] resultArray, boolean isTxn) + throws Throwable { + String txnStrStart = "START TRANSACTION"; + String txnStrCommit = "COMMIT"; + if (!isTxn) { + txnStrStart = "use " + dbName; //dummy + txnStrCommit = "use " + dbName; //dummy + } + primary.run("use " + dbName); + primary.run("CREATE TABLE " + tableName + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("SHOW TABLES LIKE '" + tableName + "'") + .verifyResult(tableName) + .run("CREATE TABLE " + tableName + "_nopart (key int, value int) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("SHOW TABLES LIKE '" + tableName + "_nopart'") + .run("ALTER TABLE " + tableName + " ADD PARTITION (load_date='2016-03-03')") + .run(txnStrStart) + .run("INSERT INTO " + tableName + " partition (load_date='2016-03-01') VALUES (1, 1)") + .run("INSERT INTO " + tableName + " partition (load_date='2016-03-01') VALUES (2, 2)") + .run("INSERT INTO " + tableName + " partition (load_date='2016-03-02') VALUES (3, 3)") + .run("INSERT INTO " + tableName + " partition (load_date='2016-03-03') VALUES (4, 4)") + .run("INSERT INTO " + tableName + " partition (load_date='2016-03-02') VALUES (5, 5)") + .run("select key from " + tableName + " order by key") + .verifyResults(resultArray) + .run("INSERT INTO " + tableName + "_nopart (key, value) select key, value from " + tableName) + .run("select key from " + tableName + "_nopart" + " order by key") + .verifyResults(resultArray) + .run(txnStrCommit); + } + + private void insertIntoDB(String dbName, String tableName, String tableProperty, String[] resultArray) + throws Throwable { + insertIntoDB(dbName, tableName, tableProperty, resultArray, false); + } + + private void insertRecords(String tableName, String tableNameOp, boolean isMMTable, + OperationType opType) throws Throwable { + insertRecordsIntoDB(primaryDbName, tableName, tableNameOp, isMMTable, opType); + } + + private void insertRecordsIntoDB(String DbName, String tableName, String tableNameOp, boolean isMMTable, + OperationType opType) throws Throwable { + String[] resultArray = new String[]{"1", "2", "3", "4", "5"}; + String tableProperty = "'transactional'='true'"; + if (isMMTable) { + tableProperty = setMMtableProperty(tableProperty); + } + primary.run("use " + DbName); + + switch (opType) { + case REPL_TEST_ACID_INSERT: + insertIntoDB(DbName, tableName, tableProperty, resultArray); + insertIntoDB(primaryDbNameExtra, tableName, tableProperty, resultArray); + return; + case REPL_TEST_ACID_INSERT_OVERWRITE: + primary.run("CREATE TABLE " + tableNameOp + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( "+ tableProperty + " )") + .run("INSERT INTO " + tableNameOp + " partition (load_date='2016-03-01') VALUES (2, 2)") + .run("INSERT INTO " + tableNameOp + " partition (load_date='2016-03-01') VALUES (10, 12)") + .run("INSERT INTO " + tableNameOp + " partition (load_date='2016-03-02') VALUES (11, 1)") + .run("select key from " + tableNameOp + " order by key") + .verifyResults(new String[]{"2", "10", "11"}) + .run("insert overwrite table " + tableNameOp + " select * from " + tableName) + .run("CREATE TABLE " + tableNameOp + "_nopart (key int, value int) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( "+ tableProperty + " )") + .run("INSERT INTO " + tableNameOp + "_nopart VALUES (2, 2)") + .run("INSERT INTO " + tableNameOp + "_nopart VALUES (10, 12)") + .run("INSERT INTO " + tableNameOp + "_nopart VALUES (11, 1)") + .run("select key from " + tableNameOp + "_nopart" + " order by key") + .verifyResults(new String[]{"2", "10", "11"}) + .run("insert overwrite table " + tableNameOp + "_nopart select * from " + tableName + "_nopart") + .run("select key from " + tableNameOp + "_nopart" + " order by key"); + break; + case REPL_TEST_ACID_INSERT_SELECT: + primary.run("CREATE TABLE " + tableNameOp + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + " )") + .run("insert into " + tableNameOp + " partition (load_date) select * from " + tableName) + .run("CREATE TABLE " + tableNameOp + "_nopart (key int, value int) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + " )") + .run("insert into " + tableNameOp + "_nopart select * from " + tableName + "_nopart"); + break; + case REPL_TEST_ACID_INSERT_IMPORT: + String path = "hdfs:///tmp/" + DbName + "/"; + String exportPath = "'" + path + tableName + "/'"; + String exportPathNoPart = "'" + path + tableName + "_nopart/'"; + primary.run("export table " + tableName + " to " + exportPath) + .run("import table " + tableNameOp + " from " + exportPath) + .run("export table " + tableName + "_nopart to " + exportPathNoPart) + .run("import table " + tableNameOp + "_nopart from " + exportPathNoPart); + break; + case REPL_TEST_ACID_CTAS: + primary.run("create table " + tableNameOp + " as select * from " + tableName) + .run("create table " + tableNameOp + "_nopart as select * from " + tableName + "_nopart"); + break; + case REPL_TEST_ACID_INSERT_LOADLOCAL: + primary.run("CREATE TABLE " + tableNameOp + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("SHOW TABLES LIKE '" + tableNameOp + "'") + .verifyResult(tableNameOp) + .run("INSERT OVERWRITE LOCAL DIRECTORY './test.dat' SELECT a.* FROM " + tableName + " a") + .run("LOAD DATA LOCAL INPATH './test.dat' OVERWRITE INTO TABLE " + tableNameOp + + " PARTITION (load_date='2008-08-15')") + .run("CREATE TABLE " + tableNameOp + "_nopart (key int, value int) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("SHOW TABLES LIKE '" + tableNameOp + "_nopart'") + .verifyResult(tableNameOp + "_nopart") + .run("LOAD DATA LOCAL INPATH './test.dat' OVERWRITE INTO TABLE " + tableNameOp + "_nopart"); + break; + case REPL_TEST_ACID_INSERT_UNION: + primary.run("CREATE TABLE " + tableNameOp + " (key int, value int) PARTITIONED BY (load_date date) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("SHOW TABLES LIKE '" + tableNameOp + "'") + .verifyResult(tableNameOp) + .run("insert overwrite table " + tableNameOp + " partition (load_date) select * from " + tableName + + " union all select * from " + tableName) + .run("CREATE TABLE " + tableNameOp + "_nopart (key int, value int) " + + "CLUSTERED BY(key) INTO 3 BUCKETS STORED AS ORC TBLPROPERTIES ( " + tableProperty + ")") + .run("insert overwrite table " + tableNameOp + "_nopart select * from " + tableName + + "_nopart union all select * from " + tableName + "_nopart"); + resultArray = new String[]{"1", "2", "3", "4", "5", "1", "2", "3", "4", "5"}; + break; + default: + return; + } + primary.run("select key from " + tableNameOp + " order by key").verifyResults(resultArray); + primary.run("select key from " + tableNameOp + "_nopart" + " order by key").verifyResults(resultArray); + } + + private String setMMtableProperty(String tableProperty) throws Throwable { + return tableProperty.concat(", 'transactional_properties' = 'insert_only'"); + } + + private void insertForMerge(String tableName, String tableNameMerge, boolean isMMTable) throws Throwable { + String tableProperty = "'transactional'='true'"; + if (isMMTable) { + tableProperty = setMMtableProperty(tableProperty); + } + primary.run("use " + primaryDbName) + .run("CREATE TABLE " + tableName + "( ID int, TranValue string, last_update_user string) PARTITIONED BY " + + "(tran_date string) CLUSTERED BY (ID) into 5 buckets STORED AS ORC TBLPROPERTIES " + + " ( "+ tableProperty + " )") + .run("SHOW TABLES LIKE '" + tableName + "'") + .verifyResult(tableName) + .run("CREATE TABLE " + tableNameMerge + " ( ID int, TranValue string, tran_date string) STORED AS ORC ") + .run("SHOW TABLES LIKE '" + tableNameMerge + "'") + .verifyResult(tableNameMerge) + .run("INSERT INTO " + tableName + " PARTITION (tran_date) VALUES (1, 'value_01', 'creation', '20170410')," + + " (2, 'value_02', 'creation', '20170410'), (3, 'value_03', 'creation', '20170410'), " + + " (4, 'value_04', 'creation', '20170410'), (5, 'value_05', 'creation', '20170413'), " + + " (6, 'value_06', 'creation', '20170413'), (7, 'value_07', 'creation', '20170413'), " + + " (8, 'value_08', 'creation', '20170413'), (9, 'value_09', 'creation', '20170413'), " + + " (10, 'value_10','creation', '20170413')") + .run("select ID from " + tableName + " order by ID") + .verifyResults(new String[] {"1", "2", "3", "4", "5", "6", "7", "8", "9", "10"}) + .run("INSERT INTO " + tableNameMerge + " VALUES (1, 'value_01', '20170410'), " + + " (4, NULL, '20170410'), (7, 'value_77777', '20170413'), " + + " (8, NULL, '20170413'), (8, 'value_08', '20170415'), " + + "(11, 'value_11', '20170415')") + .run("select ID from " + tableNameMerge + " order by ID") + .verifyResults(new String[] {"1", "4", "7", "8", "8", "11"}) + .run("MERGE INTO " + tableName + " AS T USING " + tableNameMerge + " AS S ON T.ID = S.ID and" + + " T.tran_date = S.tran_date WHEN MATCHED AND (T.TranValue != S.TranValue AND S.TranValue " + + " IS NOT NULL) THEN UPDATE SET TranValue = S.TranValue, last_update_user = " + + " 'merge_update' WHEN MATCHED AND S.TranValue IS NULL THEN DELETE WHEN NOT MATCHED " + + " THEN INSERT VALUES (S.ID, S.TranValue,'merge_insert', S.tran_date)") + .run("select last_update_user from " + tableName + " order by last_update_user") + .verifyResults(new String[] {"creation", "creation", "creation", "creation", "creation", + "creation", "creation", "merge_update", "merge_insert", "merge_insert"}); + } } diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java index fbd893b017..3df27a715a 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java @@ -318,8 +318,7 @@ public void testMetadataBootstrapDump() throws Throwable { "clustered by(key) into 2 buckets stored as orc tblproperties ('transactional'='true')") .run("create table table1 (i int, j int)") .run("insert into table1 values (1,2)") - .dump(primaryDbName, null, Arrays.asList("'hive.repl.dump.metadata.only'='true'", - "'hive.repl.dump.include.acid.tables'='true'")); + .dump(primaryDbName, null, Arrays.asList("'hive.repl.dump.metadata.only'='true'")); replica.load(replicatedDbName, tuple.dumpLocation) .run("use " + replicatedDbName) @@ -338,8 +337,7 @@ public void testIncrementalMetadataReplication() throws Throwable { .run("create table table2 (a int, city string) partitioned by (country string)") .run("create table table3 (i int, j int)") .run("insert into table1 values (1,2)") - .dump(primaryDbName, null, Arrays.asList("'hive.repl.dump.metadata.only'='true'", - "'hive.repl.dump.include.acid.tables'='true'")); + .dump(primaryDbName, null, Arrays.asList("'hive.repl.dump.metadata.only'='true'")); replica.load(replicatedDbName, bootstrapTuple.dumpLocation) .run("use " + replicatedDbName) @@ -464,8 +462,7 @@ public void testBootStrapDumpOfWarehouse() throws Throwable { SOURCE_OF_REPLICATION + "' = '1,2,3')") .run("use " + dbTwo) .run("create table t1 (i int, j int)") - .dump("`*`", null, Arrays.asList("'hive.repl.dump.metadata.only'='true'", - "'hive.repl.dump.include.acid.tables'='true'")); + .dump("`*`", null, Arrays.asList("'hive.repl.dump.metadata.only'='true'")); /* Due to the limitation that we can only have one instance of Persistence Manager Factory in a JVM @@ -524,8 +521,7 @@ public void testIncrementalDumpOfWarehouse() throws Throwable { .run("use " + dbOne) .run("create table t1 (i int, j int) partitioned by (load_date date) " + "clustered by(i) into 2 buckets stored as orc tblproperties ('transactional'='true') ") - .dump("`*`", null, Arrays.asList("'hive.repl.dump.metadata.only'='true'", - "'hive.repl.dump.include.acid.tables'='true'")); + .dump("`*`", null, Arrays.asList("'hive.repl.dump.metadata.only'='true'")); String dbTwo = primaryDbName + randomTwo; WarehouseInstance.Tuple incrementalTuple = primary @@ -536,8 +532,7 @@ public void testIncrementalDumpOfWarehouse() throws Throwable { .run("use " + dbOne) .run("create table t2 (a int, b int)") .dump("`*`", bootstrapTuple.lastReplicationId, - Arrays.asList("'hive.repl.dump.metadata.only'='true'", - "'hive.repl.dump.include.acid.tables'='true'")); + Arrays.asList("'hive.repl.dump.metadata.only'='true'")); /* Due to the limitation that we can only have one instance of Persistence Manager Factory in a JVM diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java index e9875b4bdd..d84b8c9327 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java @@ -250,6 +250,11 @@ WarehouseInstance load(String replicatedDbName, String dumpLocation) throws Thro return this; } + WarehouseInstance loadWithoutExplain(String replicatedDbName, String dumpLocation) throws Throwable { + run("REPL LOAD " + replicatedDbName + " FROM '" + dumpLocation + "'"); + return this; + } + WarehouseInstance load(String replicatedDbName, String dumpLocation, List withClauseOptions) throws Throwable { String replLoadCmd = "REPL LOAD " + replicatedDbName + " FROM '" + dumpLocation + "'"; diff --git a/ql/if/queryplan.thrift b/ql/if/queryplan.thrift index ad778e3245..d0ba7041bd 100644 --- a/ql/if/queryplan.thrift +++ b/ql/if/queryplan.thrift @@ -61,6 +61,7 @@ enum OperatorType { RCFILEMERGE, MERGEJOIN, SPARKPRUNINGSINK, + TOPNKEY } struct Operator { diff --git a/ql/src/gen/thrift/gen-cpp/queryplan_types.cpp b/ql/src/gen/thrift/gen-cpp/queryplan_types.cpp index b6eb12ab13..e28ac4c013 100644 --- a/ql/src/gen/thrift/gen-cpp/queryplan_types.cpp +++ b/ql/src/gen/thrift/gen-cpp/queryplan_types.cpp @@ -59,7 +59,8 @@ int _kOperatorTypeValues[] = { OperatorType::ORCFILEMERGE, OperatorType::RCFILEMERGE, OperatorType::MERGEJOIN, - OperatorType::SPARKPRUNINGSINK + OperatorType::SPARKPRUNINGSINK, + OperatorType::TOPNKEY }; const char* _kOperatorTypeNames[] = { "JOIN", @@ -87,9 +88,10 @@ const char* _kOperatorTypeNames[] = { "ORCFILEMERGE", "RCFILEMERGE", "MERGEJOIN", - "SPARKPRUNINGSINK" + "SPARKPRUNINGSINK", + "TOPNKEY" }; -const std::map _OperatorType_VALUES_TO_NAMES(::apache::thrift::TEnumIterator(26, _kOperatorTypeValues, _kOperatorTypeNames), ::apache::thrift::TEnumIterator(-1, NULL, NULL)); +const std::map _OperatorType_VALUES_TO_NAMES(::apache::thrift::TEnumIterator(27, _kOperatorTypeValues, _kOperatorTypeNames), ::apache::thrift::TEnumIterator(-1, NULL, NULL)); int _kTaskTypeValues[] = { TaskType::MAP, diff --git a/ql/src/gen/thrift/gen-cpp/queryplan_types.h b/ql/src/gen/thrift/gen-cpp/queryplan_types.h index eb02107e64..6bdea4b56a 100644 --- a/ql/src/gen/thrift/gen-cpp/queryplan_types.h +++ b/ql/src/gen/thrift/gen-cpp/queryplan_types.h @@ -64,7 +64,8 @@ struct OperatorType { ORCFILEMERGE = 22, RCFILEMERGE = 23, MERGEJOIN = 24, - SPARKPRUNINGSINK = 25 + SPARKPRUNINGSINK = 25, + TOPNKEY = 26 }; }; diff --git a/ql/src/gen/thrift/gen-php/Types.php b/ql/src/gen/thrift/gen-php/Types.php index df4e41db93..4ceec886b4 100644 --- a/ql/src/gen/thrift/gen-php/Types.php +++ b/ql/src/gen/thrift/gen-php/Types.php @@ -60,6 +60,7 @@ final class OperatorType { const RCFILEMERGE = 23; const MERGEJOIN = 24; const SPARKPRUNINGSINK = 25; + const TOPNKEY = 26; static public $__names = array( 0 => 'JOIN', 1 => 'MAPJOIN', @@ -87,6 +88,7 @@ final class OperatorType { 23 => 'RCFILEMERGE', 24 => 'MERGEJOIN', 25 => 'SPARKPRUNINGSINK', + 26 => 'TOPNKEY', ); } diff --git a/ql/src/gen/thrift/gen-py/queryplan/ttypes.py b/ql/src/gen/thrift/gen-py/queryplan/ttypes.py index 85d39fdbe1..5638d35326 100644 --- a/ql/src/gen/thrift/gen-py/queryplan/ttypes.py +++ b/ql/src/gen/thrift/gen-py/queryplan/ttypes.py @@ -71,6 +71,7 @@ class OperatorType: RCFILEMERGE = 23 MERGEJOIN = 24 SPARKPRUNINGSINK = 25 + TOPNKEY = 26 _VALUES_TO_NAMES = { 0: "JOIN", @@ -99,6 +100,7 @@ class OperatorType: 23: "RCFILEMERGE", 24: "MERGEJOIN", 25: "SPARKPRUNINGSINK", + 26: "TOPNKEY", } _NAMES_TO_VALUES = { @@ -128,6 +130,7 @@ class OperatorType: "RCFILEMERGE": 23, "MERGEJOIN": 24, "SPARKPRUNINGSINK": 25, + "TOPNKEY": 26, } class TaskType: diff --git a/ql/src/gen/thrift/gen-rb/queryplan_types.rb b/ql/src/gen/thrift/gen-rb/queryplan_types.rb index 6010f3d21e..04af9754cd 100644 --- a/ql/src/gen/thrift/gen-rb/queryplan_types.rb +++ b/ql/src/gen/thrift/gen-rb/queryplan_types.rb @@ -47,8 +47,9 @@ module OperatorType RCFILEMERGE = 23 MERGEJOIN = 24 SPARKPRUNINGSINK = 25 - VALUE_MAP = {0 => "JOIN", 1 => "MAPJOIN", 2 => "EXTRACT", 3 => "FILTER", 4 => "FORWARD", 5 => "GROUPBY", 6 => "LIMIT", 7 => "SCRIPT", 8 => "SELECT", 9 => "TABLESCAN", 10 => "FILESINK", 11 => "REDUCESINK", 12 => "UNION", 13 => "UDTF", 14 => "LATERALVIEWJOIN", 15 => "LATERALVIEWFORWARD", 16 => "HASHTABLESINK", 17 => "HASHTABLEDUMMY", 18 => "PTF", 19 => "MUX", 20 => "DEMUX", 21 => "EVENT", 22 => "ORCFILEMERGE", 23 => "RCFILEMERGE", 24 => "MERGEJOIN", 25 => "SPARKPRUNINGSINK"} - VALID_VALUES = Set.new([JOIN, MAPJOIN, EXTRACT, FILTER, FORWARD, GROUPBY, LIMIT, SCRIPT, SELECT, TABLESCAN, FILESINK, REDUCESINK, UNION, UDTF, LATERALVIEWJOIN, LATERALVIEWFORWARD, HASHTABLESINK, HASHTABLEDUMMY, PTF, MUX, DEMUX, EVENT, ORCFILEMERGE, RCFILEMERGE, MERGEJOIN, SPARKPRUNINGSINK]).freeze + TOPNKEY = 26 + VALUE_MAP = {0 => "JOIN", 1 => "MAPJOIN", 2 => "EXTRACT", 3 => "FILTER", 4 => "FORWARD", 5 => "GROUPBY", 6 => "LIMIT", 7 => "SCRIPT", 8 => "SELECT", 9 => "TABLESCAN", 10 => "FILESINK", 11 => "REDUCESINK", 12 => "UNION", 13 => "UDTF", 14 => "LATERALVIEWJOIN", 15 => "LATERALVIEWFORWARD", 16 => "HASHTABLESINK", 17 => "HASHTABLEDUMMY", 18 => "PTF", 19 => "MUX", 20 => "DEMUX", 21 => "EVENT", 22 => "ORCFILEMERGE", 23 => "RCFILEMERGE", 24 => "MERGEJOIN", 25 => "SPARKPRUNINGSINK", 26 => "TOPNKEY"} + VALID_VALUES = Set.new([JOIN, MAPJOIN, EXTRACT, FILTER, FORWARD, GROUPBY, LIMIT, SCRIPT, SELECT, TABLESCAN, FILESINK, REDUCESINK, UNION, UDTF, LATERALVIEWJOIN, LATERALVIEWFORWARD, HASHTABLESINK, HASHTABLEDUMMY, PTF, MUX, DEMUX, EVENT, ORCFILEMERGE, RCFILEMERGE, MERGEJOIN, SPARKPRUNINGSINK, TOPNKEY]).freeze end module TaskType diff --git a/ql/src/java/org/apache/hadoop/hive/metastore/SynchronizedMetaStoreClient.java b/ql/src/java/org/apache/hadoop/hive/metastore/SynchronizedMetaStoreClient.java index f87a6aa426..2ba6d0796c 100644 --- a/ql/src/java/org/apache/hadoop/hive/metastore/SynchronizedMetaStoreClient.java +++ b/ql/src/java/org/apache/hadoop/hive/metastore/SynchronizedMetaStoreClient.java @@ -35,6 +35,7 @@ import org.apache.hadoop.hive.metastore.api.ShowLocksRequest; import org.apache.hadoop.hive.metastore.api.ShowLocksResponse; import org.apache.hadoop.hive.metastore.api.UnknownTableException; +import org.apache.hadoop.hive.metastore.api.WriteNotificationLogRequest; import org.apache.thrift.TException; @@ -109,6 +110,10 @@ public synchronized FireEventResponse fireListenerEvent(FireEventRequest rqst) t return client.fireListenerEvent(rqst); } + public synchronized void addWriteNotificationLog(WriteNotificationLogRequest rqst) throws TException { + client.addWriteNotificationLog(rqst); + } + public synchronized void close() { client.close(); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java index 2bb3ec4a5a..e2f88cfea3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java @@ -132,7 +132,11 @@ private void moveFileInDfs (Path sourcePath, Path targetPath, HiveConf conf) if (HiveConf.getBoolVar(conf, HiveConf.ConfVars.HIVE_INSERT_INTO_MULTILEVEL_DIRS)) { deletePath = createTargetPath(targetPath, tgtFs); } - Hive.clearDestForSubDirSrc(conf, targetPath, sourcePath, false); + //For acid table incremental replication, just copy the content of staging directory to destination. + //No need to clean it. + if (work.isNeedCleanTarget()) { + Hive.clearDestForSubDirSrc(conf, targetPath, sourcePath, false); + } // Set isManaged to false as this is not load data operation for which it is needed. if (!Hive.moveFile(conf, sourcePath, targetPath, true, false, false)) { try { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplCopyTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplCopyTask.java index 4ace861165..4a9db0a858 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplCopyTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplCopyTask.java @@ -151,10 +151,11 @@ protected int execute(DriverContext driverContext) { continue; } String destFileName = srcFile.getCmPath().getName(); - Path destFile = new Path(toPath, destFileName); + Path destRoot = CopyUtils.getCopyDestination(srcFile, toPath); + Path destFile = new Path(destRoot, destFileName); if (dstFs.exists(destFile)) { String destFileWithSourceName = srcFile.getSourcePath().getName(); - Path newDestFile = new Path(toPath, destFileWithSourceName); + Path newDestFile = new Path(destRoot, destFileWithSourceName); boolean result = dstFs.rename(destFile, newDestFile); if (!result) { throw new IllegalStateException( diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplTxnTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplTxnTask.java index 5bbc25ab2a..c2953c5ce0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplTxnTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/ReplTxnTask.java @@ -18,6 +18,8 @@ package org.apache.hadoop.hive.ql.exec; +import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; +import org.apache.hadoop.hive.metastore.api.Database; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; import org.apache.hadoop.hive.ql.DriverContext; import org.apache.hadoop.hive.ql.lockmgr.HiveTxnManager; @@ -60,8 +62,19 @@ public int execute(DriverContext driverContext) { return 0; } } catch (InvalidTableException e) { - LOG.info("Table does not exist so, ignoring the operation as it might be a retry(idempotent) case."); - return 0; + // In scenarios like import to mm tables, the alloc write id event is generated before create table event. + try { + Database database = Hive.get().getDatabase(work.getDbName()); + if (!replicationSpec.allowReplacementInto(database.getParameters())) { + // if the event is already replayed, then no need to replay it again. + LOG.debug("ReplTxnTask: Event is skipped as it is already replayed. Event Id: " + + replicationSpec.getReplicationState() + "Event Type: " + work.getOperationType()); + return 0; + } + } catch (HiveException e1) { + LOG.error("Get database failed with exception " + e1.getMessage()); + return 1; + } } catch (HiveException e) { LOG.error("Get table failed with exception " + e.getMessage()); return 1; @@ -85,10 +98,16 @@ public int execute(DriverContext driverContext) { } return 0; case REPL_COMMIT_TXN: - for (long txnId : work.getTxnIds()) { - txnManager.replCommitTxn(replPolicy, txnId); - LOG.info("Replayed CommitTxn Event for policy " + replPolicy + " with srcTxn " + txnId); - } + // Currently only one commit txn per event is supported. + assert (work.getTxnIds().size() == 1); + + long txnId = work.getTxnIds().get(0); + CommitTxnRequest commitTxnRequest = new CommitTxnRequest(txnId); + commitTxnRequest.setReplPolicy(work.getReplPolicy()); + commitTxnRequest.setWriteEventInfos(work.getWriteEventInfos()); + txnManager.replCommitTxn(commitTxnRequest); + LOG.info("Replayed CommitTxn Event for replPolicy: " + replPolicy + " with srcTxn: " + txnId + + "WriteEventInfos: " + work.getWriteEventInfos()); return 0; case REPL_ALLOC_WRITE_ID: assert work.getTxnToWriteIdList() != null; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java index a14802fe41..940e381f49 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java @@ -199,7 +199,9 @@ private void dumpEvent(NotificationEvent ev, Path evRoot, Path cmRoot) throws Ex cmRoot, getHive(), conf, - getNewEventOnlyReplicationSpec(ev.getEventId()) + getNewEventOnlyReplicationSpec(ev.getEventId()), + work.dbNameOrPattern, + work.tableNameOrPattern ); EventHandler eventHandler = EventHandlerFactory.handlerFor(ev); eventHandler.handle(context); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java index 7fce67fc3e..16ba82ef6d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java @@ -22,7 +22,6 @@ import java.io.IOException; import java.io.Serializable; -import java.nio.ByteBuffer; import java.util.ArrayList; import java.util.Arrays; import java.util.Collections; @@ -71,21 +70,7 @@ import org.slf4j.LoggerFactory; import com.google.common.annotations.VisibleForTesting; - -import java.io.IOException; -import java.io.Serializable; import java.nio.charset.Charset; -import java.util.ArrayList; -import java.util.Collections; -import java.util.HashMap; -import java.util.List; -import java.util.Map; -import java.util.Properties; -import java.util.Set; -import java.util.regex.Pattern; - -import static org.apache.hadoop.hive.ql.exec.Utilities.COPY_KEYWORD; - /** * Utilities that are shared by all of the ACID input and output formats. They @@ -1907,6 +1892,28 @@ public static String getAcidSubDir(Path dataPath) { return null; } + //Get the first level acid directory (if any) from a given path + public static String getFirstLevelAcidDirPath(Path dataPath, FileSystem fileSystem) throws IOException { + if (dataPath == null) { + return null; + } + String firstLevelAcidDir = getAcidSubDir(dataPath); + if (firstLevelAcidDir != null) { + return firstLevelAcidDir; + } + + String acidDirPath = getFirstLevelAcidDirPath(dataPath.getParent(), fileSystem); + if (acidDirPath == null) { + return null; + } + + // We need the path for directory so no need to append file name + if (fileSystem.isDirectory(dataPath)) { + return acidDirPath + Path.SEPARATOR + dataPath.getName(); + } + return acidDirPath; + } + public static boolean isAcidEnabled(HiveConf hiveConf) { String txnMgr = hiveConf.getVar(HiveConf.ConfVars.HIVE_TXN_MANAGER); boolean concurrency = hiveConf.getBoolVar(HiveConf.ConfVars.HIVE_SUPPORT_CONCURRENCY); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java index bcc05081aa..ec8527ef27 100755 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java @@ -26,30 +26,32 @@ import java.util.Comparator; import java.util.HashSet; import java.util.Iterator; -import java.util.LinkedList; import java.util.List; import java.util.Map; import java.util.Map.Entry; import java.util.Set; import java.util.concurrent.ConcurrentHashMap; +import org.apache.hadoop.hive.common.FileUtils; +import org.apache.hadoop.hive.common.StringInternUtils; +import org.apache.hadoop.hive.common.ValidTxnWriteIdList; +import org.apache.hadoop.hive.common.ValidWriteIdList; +import org.apache.hadoop.hive.ql.exec.SerializationUtilities; +import org.apache.hive.common.util.HiveStringUtils; +import org.apache.hive.common.util.Ref; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hive.common.FileUtils; -import org.apache.hadoop.hive.common.JavaUtils; -import org.apache.hadoop.hive.common.StringInternUtils; -import org.apache.hadoop.hive.common.ValidTxnWriteIdList; -import org.apache.hadoop.hive.common.ValidWriteIdList; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil; import org.apache.hadoop.hive.llap.io.api.LlapIo; import org.apache.hadoop.hive.llap.io.api.LlapProxy; import org.apache.hadoop.hive.ql.exec.Operator; -import org.apache.hadoop.hive.ql.exec.SerializationUtilities; import org.apache.hadoop.hive.ql.exec.TableScanOperator; import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.exec.spark.SparkDynamicPartitionPruner; @@ -62,8 +64,6 @@ import org.apache.hadoop.hive.ql.plan.PartitionDesc; import org.apache.hadoop.hive.ql.plan.TableDesc; import org.apache.hadoop.hive.ql.plan.TableScanDesc; -import org.apache.hadoop.hive.ql.plan.VectorPartitionDesc; -import org.apache.hadoop.hive.ql.plan.VectorPartitionDesc.VectorMapOperatorReadType; import org.apache.hadoop.hive.ql.session.SessionState; import org.apache.hadoop.hive.serde2.ColumnProjectionUtils; import org.apache.hadoop.hive.serde2.Deserializer; @@ -78,10 +78,7 @@ import org.apache.hadoop.mapred.RecordReader; import org.apache.hadoop.mapred.Reporter; import org.apache.hadoop.util.StringUtils; -import org.apache.hive.common.util.Ref; import org.apache.hive.common.util.ReflectionUtil; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; /** * HiveInputFormat is a parameterized InputFormat which looks at the path name @@ -460,8 +457,9 @@ private void addSplitsForGroup(List dirs, TableScanOperator tableScan, Job InputFormat inputFormat, Class inputFormatClass, int splits, TableDesc table, List result) throws IOException { + String tableName = table.getTableName(); ValidWriteIdList validWriteIdList = AcidUtils.getTableValidWriteIdList( - conf, table.getTableName()); + conf, tableName == null ? null : HiveStringUtils.normalizeIdentifier(tableName)); ValidWriteIdList validMmWriteIdList = getMmValidWriteIds(conf, table, validWriteIdList); try { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java index 4fd1d4ec54..78980fad93 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java @@ -37,6 +37,7 @@ Licensed to the Apache Software Foundation (ASF) under one import org.apache.hadoop.hive.metastore.api.NoSuchTxnException; import org.apache.hadoop.hive.metastore.api.TxnAbortedException; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; +import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; import org.apache.hadoop.hive.metastore.txn.TxnUtils; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; import org.apache.hadoop.hive.ql.Context; @@ -638,14 +639,15 @@ public void releaseLocks(List hiveLocks) throws LockException { } @Override - public void replCommitTxn(String replPolicy, long srcTxnId) throws LockException { + public void replCommitTxn(CommitTxnRequest rqst) throws LockException { try { - getMS().replCommitTxn(srcTxnId, replPolicy); + getMS().replCommitTxn(rqst); } catch (NoSuchTxnException e) { - LOG.error("Metastore could not find " + JavaUtils.txnIdToString(srcTxnId)); - throw new LockException(e, ErrorMsg.TXN_NO_SUCH_TRANSACTION, JavaUtils.txnIdToString(srcTxnId)); + LOG.error("Metastore could not find " + JavaUtils.txnIdToString(rqst.getTxnid())); + throw new LockException(e, ErrorMsg.TXN_NO_SUCH_TRANSACTION, JavaUtils.txnIdToString(rqst.getTxnid())); } catch (TxnAbortedException e) { - LockException le = new LockException(e, ErrorMsg.TXN_ABORTED, JavaUtils.txnIdToString(srcTxnId), e.getMessage()); + LockException le = new LockException(e, ErrorMsg.TXN_ABORTED, + JavaUtils.txnIdToString(rqst.getTxnid()), e.getMessage()); LOG.error(le.getMessage()); throw le; } catch (TException e) { @@ -1013,7 +1015,11 @@ public int getStmtIdAndIncrement() { assert isTxnOpen(); return stmtId++; } - + @Override + public int getCurrentStmtId() { + assert isTxnOpen(); + return stmtId; + } @Override public long getTableWriteId(String dbName, String tableName) throws LockException { assert isTxnOpen(); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java index ab9d67e441..1feddebf2b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java @@ -18,6 +18,7 @@ package org.apache.hadoop.hive.ql.lockmgr; import org.apache.hadoop.hive.common.ValidTxnWriteIdList; +import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -74,6 +75,10 @@ public int getStmtIdAndIncrement() { return 0; } @Override + public int getCurrentStmtId() { + return 0; + } + @Override public long getTableWriteId(String dbName, String tableName) throws LockException { return 0L; } @@ -220,7 +225,7 @@ public void commitTxn() throws LockException { } @Override - public void replCommitTxn(String replPolicy, long srcTxnId) throws LockException { + public void replCommitTxn(CommitTxnRequest rqst) throws LockException { // No-op } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java index 5f68e085a0..9575552ff0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveTxnManager.java @@ -19,6 +19,7 @@ import org.apache.hadoop.hive.common.ValidTxnList; import org.apache.hadoop.hive.common.ValidTxnWriteIdList; +import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; import org.apache.hadoop.hive.metastore.api.LockResponse; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; import org.apache.hadoop.hive.ql.Context; @@ -61,11 +62,11 @@ /** * Commit the transaction in target cluster. - * @param replPolicy Replication policy to uniquely identify the source cluster. - * @param srcTxnId The id of the transaction at the source cluster + * + * @param rqst Commit transaction request having information related to commit txn and write events. * @throws LockException in case of failure to commit the transaction. */ - void replCommitTxn(String replPolicy, long srcTxnId) throws LockException; + void replCommitTxn(CommitTxnRequest rqst) throws LockException; /** * Abort the transaction in target cluster. @@ -295,6 +296,9 @@ void replAllocateTableWriteIdsBatch(String dbName, String tableName, String repl */ int getStmtIdAndIncrement(); + // Can be used by operation to set the stmt id when allocation is done somewhere else. + int getCurrentStmtId(); + /** * Acquire the materialization rebuild lock for a given view. We need to specify the fully * qualified name of the materialized view and the open transaction ID so we can identify diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java index 2e05e15616..953cd1d43a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java @@ -161,6 +161,7 @@ import org.apache.hadoop.hive.metastore.api.WMTrigger; import org.apache.hadoop.hive.metastore.api.WMValidateResourcePlanResponse; import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; +import org.apache.hadoop.hive.metastore.api.WriteNotificationLogRequest; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; import org.apache.hadoop.hive.ql.ErrorMsg; import org.apache.hadoop.hive.ql.exec.AbstractFileMergeOperator; @@ -1762,8 +1763,14 @@ public Partition loadPartition(Path loadPath, Table tbl, Map par PerfLogger perfLogger = SessionState.getPerfLogger(); perfLogger.PerfLogBegin("MoveTask", "FileMoves"); - List newFiles = Collections.synchronizedList(new ArrayList()); + List newFiles = null; + // If config is set, table is not temporary and partition being inserted exists, capture + // the list of files added. For not yet existing partitions (insert overwrite to new partition + // or dynamic partition inserts), the add partition event will capture the list of files added. + if (areEventsForDmlNeeded(tbl, oldPart)) { + newFiles = Collections.synchronizedList(new ArrayList()); + } // Note: the stats for ACID tables do not have any coordination with either Hive ACID logic // like txn commits, time outs, etc.; nor the lower level sync in metastore pertaining @@ -1776,8 +1783,8 @@ public Partition loadPartition(Path loadPath, Table tbl, Map par Utilities.FILE_OP_LOGGER.trace("not moving " + loadPath + " to " + newPartPath + " (MM)"); } assert !isAcidIUDoperation; - if (areEventsForDmlNeeded(tbl, oldPart)) { - newFiles = listFilesCreatedByQuery(loadPath, writeId, stmtId); + if (newFiles != null) { + listFilesCreatedByQuery(loadPath, writeId, stmtId, isMmTableWrite ? isInsertOverwrite : false, newFiles); } if (Utilities.FILE_OP_LOGGER.isTraceEnabled()) { Utilities.FILE_OP_LOGGER.trace("maybe deleting stuff from " + oldPartPath @@ -1826,8 +1833,15 @@ public Partition loadPartition(Path loadPath, Table tbl, Map par // or dynamic partition inserts), the add partition event will capture the list of files added. // Generate an insert event only if inserting into an existing partition // When inserting into a new partition, the add partition event takes care of insert event - if (conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML) && !tbl.isTemporary() && (null != oldPart)) { - fireInsertEvent(tbl, partSpec, (loadFileType == LoadFileType.REPLACE_ALL), newFiles); + if ((null != oldPart) && (null != newFiles)) { + if (isTxnTable) { + addWriteNotificationLog(tbl, partSpec, newFiles, writeId); + } else { + fireInsertEvent(tbl, partSpec, (loadFileType == LoadFileType.REPLACE_ALL), newFiles); + } + } else { + LOG.debug("No new files were created, and is not a replace, or we're inserting into a " + + "partition that does not exist yet. Skipping generating INSERT event."); } // column stats will be inaccurate @@ -1897,6 +1911,12 @@ public Partition loadPartition(Path loadPath, Table tbl, Map par } throw e; } + + // For acid table, add the acid_write event with file list at the time of load itself. But + // it should be done after partition is created. + if (isTxnTable && (null != newFiles)) { + addWriteNotificationLog(tbl, partSpec, newFiles, writeId); + } } else { setStatsPropAndAlterPartition(hasFollowingStatsTask, tbl, newTPart); } @@ -1949,50 +1969,47 @@ private Path fixFullAcidPathForLoadData(LoadFileType loadFileType, Path destPath } private boolean areEventsForDmlNeeded(Table tbl, Partition oldPart) { - return conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML) && !tbl.isTemporary() && oldPart != null; + // For Acid IUD, add partition is a meta data only operation. So need to add the new files added + // information into the TXN_WRITE_NOTIFICATION_LOG table. + return conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML) && !tbl.isTemporary() && + ((null != oldPart) || AcidUtils.isTransactionalTable(tbl)); + } + + private void listFilesInsideAcidDirectory(Path acidDir, FileSystem srcFs, List newFiles) throws IOException { + // list out all the files/directory in the path + FileStatus[] acidFiles; + acidFiles = srcFs.listStatus(acidDir); + if (acidFiles == null) { + LOG.debug("No files added by this query in: " + acidDir); + return; + } + for (FileStatus acidFile : acidFiles) { + // need to list out only files, ignore folders. + if (!acidFile.isDirectory()) { + newFiles.add(acidFile.getPath()); + } else { + listFilesInsideAcidDirectory(acidFile.getPath(), srcFs, newFiles); + } + } } - private List listFilesCreatedByQuery(Path loadPath, long writeId, int stmtId) throws HiveException { - List newFiles = new ArrayList(); - final String filePrefix = AcidUtils.deltaSubdir(writeId, writeId, stmtId); - FileStatus[] srcs; - FileSystem srcFs; + private void listFilesCreatedByQuery(Path loadPath, long writeId, int stmtId, + boolean isInsertOverwrite, List newFiles) throws HiveException { + Path acidDir = new Path(loadPath, AcidUtils.baseOrDeltaSubdir(isInsertOverwrite, writeId, writeId, stmtId)); try { - srcFs = loadPath.getFileSystem(conf); - srcs = srcFs.listStatus(loadPath); + FileSystem srcFs = loadPath.getFileSystem(conf); + if (srcFs.exists(acidDir) && srcFs.isDirectory(acidDir)){ + // list out all the files in the path + listFilesInsideAcidDirectory(acidDir, srcFs, newFiles); + } else { + LOG.info("directory does not exist: " + acidDir); + return; + } } catch (IOException e) { LOG.error("Error listing files", e); throw new HiveException(e); } - if (srcs == null) { - LOG.info("No sources specified: " + loadPath); - return newFiles; - } - PathFilter subdirFilter = null; - - // Note: just like the move path, we only do one level of recursion. - for (FileStatus src : srcs) { - if (src.isDirectory()) { - if (subdirFilter == null) { - subdirFilter = new PathFilter() { - @Override - public boolean accept(Path path) { - return path.getName().startsWith(filePrefix); - } - }; - } - try { - for (FileStatus srcFile : srcFs.listStatus(src.getPath(), subdirFilter)) { - newFiles.add(srcFile.getPath()); - } - } catch (IOException e) { - throw new HiveException(e); - } - } else if (src.getPath().getName().startsWith(filePrefix)) { - newFiles.add(src.getPath()); - } - } - return newFiles; + return; } private void setStatsPropAndAlterPartition(boolean hasFollowingStatsTask, Table tbl, @@ -2335,13 +2352,17 @@ public void loadTable(Path loadPath, String tableName, LoadFileType loadFileType boolean isSkewedStoreAsSubdir, boolean isAcidIUDoperation, boolean hasFollowingStatsTask, Long writeId, int stmtId, boolean isInsertOverwrite) throws HiveException { - List newFiles = Collections.synchronizedList(new ArrayList()); + List newFiles = null; Table tbl = getTable(tableName); assert tbl.getPath() != null : "null==getPath() for " + tbl.getTableName(); boolean isTxnTable = AcidUtils.isTransactionalTable(tbl); boolean isMmTable = AcidUtils.isInsertOnlyTable(tbl); boolean isFullAcidTable = AcidUtils.isFullAcidTable(tbl); + if (conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML) && !tbl.isTemporary()) { + newFiles = Collections.synchronizedList(new ArrayList()); + } + // Note: this assumes both paths are qualified; which they are, currently. if ((isMmTable || isFullAcidTable) && loadPath.equals(tbl.getPath())) { /** @@ -2353,7 +2374,11 @@ public void loadTable(Path loadPath, String tableName, LoadFileType loadFileType Utilities.FILE_OP_LOGGER.debug( "not moving " + loadPath + " to " + tbl.getPath() + " (MM)"); } - newFiles = listFilesCreatedByQuery(loadPath, writeId, stmtId); + + //new files list is required only for event notification. + if (newFiles != null) { + listFilesCreatedByQuery(loadPath, writeId, stmtId, isMmTable ? isInsertOverwrite : false, newFiles); + } } else { // Either a non-MM query, or a load into MM table from an external source. Path tblPath = tbl.getPath(); @@ -2421,11 +2446,10 @@ public void loadTable(Path loadPath, String tableName, LoadFileType loadFileType alterTable(tbl, environmentContext); - if (conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML) && !tbl.isTemporary()) { - fireInsertEvent(tbl, null, (loadFileType == LoadFileType.REPLACE_ALL), newFiles); + if (AcidUtils.isTransactionalTable(tbl)) { + addWriteNotificationLog(tbl, null, newFiles, writeId); } else { - fireInsertEvent(tbl, null, (loadFileType == LoadFileType.REPLACE_ALL), null); - + fireInsertEvent(tbl, null, (loadFileType == LoadFileType.REPLACE_ALL), newFiles); } } @@ -2677,6 +2701,48 @@ private void alterPartitionSpecInMemory(Table tbl, tpart.getSd().setLocation(partPath); } + private void addWriteNotificationLog(Table tbl, Map partitionSpec, + List newFiles, Long writeId) throws HiveException { + if (!conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML)) { + LOG.debug("write notification log is ignored as dml event logging is disabled"); + return; + } + + if (tbl.isTemporary()) { + LOG.debug("write notification log is ignored as " + tbl.getTableName() + " is temporary : " + writeId); + return; + } + + if (newFiles == null || newFiles.isEmpty()) { + LOG.debug("write notification log is ignored as file list is empty"); + return; + } + + LOG.debug("adding write notification log for operation " + writeId + " table " + tbl.getCompleteName() + + "partition " + partitionSpec + " list of files " + newFiles); + + try { + FileSystem fileSystem = tbl.getDataLocation().getFileSystem(conf); + Long txnId = SessionState.get().getTxnMgr().getCurrentTxnId(); + + InsertEventRequestData insertData = new InsertEventRequestData(); + insertData.setReplace(true); + + WriteNotificationLogRequest rqst = new WriteNotificationLogRequest(txnId, writeId, + tbl.getDbName(), tbl.getTableName(), insertData); + addInsertFileInformation(newFiles, fileSystem, insertData); + + if (partitionSpec != null && !partitionSpec.isEmpty()) { + for (FieldSchema fs : tbl.getPartitionKeys()) { + rqst.addToPartitionVals(partitionSpec.get(fs.getName())); + } + } + getSynchronizedMSC().addWriteNotificationLog(rqst); + } catch (IOException | TException e) { + throw new HiveException(e); + } + } + private void fireInsertEvent(Table tbl, Map partitionSpec, boolean replace, List newFiles) throws HiveException { if (conf.getBoolVar(ConfVars.FIRE_EVENTS_FOR_DML)) { @@ -2753,6 +2819,7 @@ private static void addInsertNonDirectoryInformation(Path p, FileSystem fileSyst InsertEventRequestData insertData) throws IOException { insertData.addToFilesAdded(p.toString()); FileChecksum cksum = fileSystem.getFileChecksum(p); + String acidDirPath = AcidUtils.getFirstLevelAcidDirPath(p.getParent(), fileSystem); // File checksum is not implemented for local filesystem (RawLocalFileSystem) if (cksum != null) { String checksumString = @@ -2762,6 +2829,11 @@ private static void addInsertNonDirectoryInformation(Path p, FileSystem fileSyst // Add an empty checksum string for filesystems that don't generate one insertData.addToFilesAddedChecksum(""); } + + // acid dir will be present only for acid write operations. + if (acidDirPath != null) { + insertData.addToSubDirectoryList(acidDirPath); + } } public boolean dropPartition(String tblName, List part_vals, boolean deleteData) @@ -3718,7 +3790,6 @@ public static boolean moveFile(final HiveConf conf, Path srcf, final Path destf, @Override public Void call() throws HiveException { SessionState.setCurrentSessionState(parentSession); - final String group = srcStatus.getGroup(); try { boolean success = false; if (destFs instanceof DistributedFileSystem) { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveUtils.java index f1c4d9827b..e04a0f3dce 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveUtils.java @@ -21,6 +21,7 @@ import java.util.ArrayList; import java.util.List; +import org.apache.hadoop.fs.Path; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.conf.Configuration; @@ -431,11 +432,19 @@ public static String getLocalDirList(Configuration conf) { public static String getReplPolicy(String dbName, String tableName) { if ((dbName == null) || (dbName.isEmpty())) { - return null; + return "*.*"; } else if ((tableName == null) || (tableName.isEmpty())) { return dbName.toLowerCase() + ".*"; } else { return dbName.toLowerCase() + "." + tableName.toLowerCase(); } } + + public static Path getDumpPath(Path root, String dbName, String tableName) { + assert (dbName != null); + if ((tableName != null) && (!tableName.isEmpty())) { + return new Path(root, dbName + "." + tableName); + } + return new Path(root, dbName); + } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java index d34de61842..eb594f825d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java @@ -56,6 +56,7 @@ import org.apache.hadoop.hive.ql.plan.DDLWork; import org.apache.hadoop.hive.ql.plan.DropTableDesc; import org.apache.hadoop.hive.ql.plan.LoadTableDesc; +import org.apache.hadoop.hive.ql.plan.LoadMultiFilesDesc; import org.apache.hadoop.hive.ql.plan.LoadTableDesc.LoadFileType; import org.apache.hadoop.hive.ql.plan.MoveWork; import org.apache.hadoop.hive.ql.session.SessionState; @@ -249,9 +250,11 @@ public static boolean prepareImport(boolean isImportCmd, throw new HiveException(e); } + boolean inReplicationScope = false; if ((replicationSpec != null) && replicationSpec.isInReplicationScope()){ tblDesc.setReplicationSpec(replicationSpec); StatsSetupConst.setBasicStatsState(tblDesc.getTblProps(), StatsSetupConst.FALSE); + inReplicationScope = true; } if (isExternalSet) { @@ -275,7 +278,7 @@ public static boolean prepareImport(boolean isImportCmd, for (Partition partition : partitions) { // TODO: this should ideally not create AddPartitionDesc per partition AddPartitionDesc partsDesc = getBaseAddPartitionDescFromPartition(fromPath, dbname, tblDesc, partition); - if ((replicationSpec != null) && replicationSpec.isInReplicationScope()){ + if (inReplicationScope){ StatsSetupConst.setBasicStatsState(partsDesc.getPartition(0).getPartParams(), StatsSetupConst.FALSE); } partitionDescs.add(partsDesc); @@ -335,13 +338,14 @@ public static boolean prepareImport(boolean isImportCmd, //if importing into existing transactional table or will create a new transactional table //(because Export was done from transactional table), need a writeId // Explain plan doesn't open a txn and hence no need to allocate write id. - if (x.getCtx().getExplainConfig() == null) { + // In replication flow, no need to allocate write id. It will be allocated using the alloc write id event. + if (x.getCtx().getExplainConfig() == null && !inReplicationScope) { writeId = txnMgr.getTableWriteId(tblDesc.getDatabaseName(), tblDesc.getTableName()); stmtId = txnMgr.getStmtIdAndIncrement(); } } - if (!replicationSpec.isInReplicationScope()) { + if (!inReplicationScope) { createRegularImportTasks( tblDesc, partitionDescs, isPartSpecSet, replicationSpec, table, @@ -390,7 +394,7 @@ private static ImportTableDesc getBaseCreateTableDescFromTable(String dbName, Path dataPath = new Path(fromURI.toString(), EximUtil.DATA_PATH_NAME); Path destPath = null, loadPath = null; LoadFileType lft; - if (AcidUtils.isTransactionalTable(table)) { + if (AcidUtils.isTransactionalTable(table) && !replicationSpec.isInReplicationScope()) { String mmSubdir = replace ? AcidUtils.baseDir(writeId) : AcidUtils.deltaSubdir(writeId, writeId, stmtId); destPath = new Path(tgtPath, mmSubdir); @@ -428,13 +432,26 @@ private static ImportTableDesc getBaseCreateTableDescFromTable(String dbName, copyTask = TaskFactory.get(new CopyWork(dataPath, destPath, false)); } - LoadTableDesc loadTableWork = new LoadTableDesc( - loadPath, Utilities.getTableDesc(table), new TreeMap<>(), lft, writeId); - loadTableWork.setStmtId(stmtId); + MoveWork moveWork = new MoveWork(x.getInputs(), x.getOutputs(), null, null, false); + + + if (replicationSpec.isInReplicationScope() && AcidUtils.isTransactionalTable(table)) { + LoadMultiFilesDesc loadFilesWork = new LoadMultiFilesDesc( + Collections.singletonList(destPath), + Collections.singletonList(tgtPath), + true, null, null); + moveWork.setMultiFilesDesc(loadFilesWork); + moveWork.setNeedCleanTarget(false); + } else { + LoadTableDesc loadTableWork = new LoadTableDesc( + loadPath, Utilities.getTableDesc(table), new TreeMap<>(), lft, writeId); + loadTableWork.setStmtId(stmtId); + moveWork.setLoadTableWork(loadTableWork); + } + //if Importing into existing table, FileFormat is checked by // ImportSemanticAnalzyer.checked checkTable() - MoveWork mv = new MoveWork(x.getInputs(), x.getOutputs(), loadTableWork, null, false); - Task loadTableTask = TaskFactory.get(mv, x.getConf()); + Task loadTableTask = TaskFactory.get(moveWork, x.getConf()); copyTask.addDependentTask(loadTableTask); x.getTasks().add(copyTask); return loadTableTask; @@ -498,8 +515,10 @@ private static ImportTableDesc getBaseCreateTableDescFromTable(String dbName, + partSpecToString(partSpec.getPartSpec()) + " with source location: " + srcLocation); Path tgtLocation = new Path(partSpec.getLocation()); - Path destPath = !AcidUtils.isTransactionalTable(table.getParameters()) ? - x.getCtx().getExternalTmpPath(tgtLocation) + //Replication scope the write id will be invalid + Boolean useStagingDirectory = !AcidUtils.isTransactionalTable(table.getParameters()) || + replicationSpec.isInReplicationScope(); + Path destPath = useStagingDirectory ? x.getCtx().getExternalTmpPath(tgtLocation) : new Path(tgtLocation, AcidUtils.deltaSubdir(writeId, writeId, stmtId)); Path moveTaskSrc = !AcidUtils.isTransactionalTable(table.getParameters()) ? destPath : tgtLocation; if (Utilities.FILE_OP_LOGGER.isTraceEnabled()) { @@ -523,17 +542,29 @@ private static ImportTableDesc getBaseCreateTableDescFromTable(String dbName, Task addPartTask = TaskFactory.get( new DDLWork(x.getInputs(), x.getOutputs(), addPartitionDesc), x.getConf()); + MoveWork moveWork = new MoveWork(x.getInputs(), x.getOutputs(), + null, null, false); + // Note: this sets LoadFileType incorrectly for ACID; is that relevant for import? // See setLoadFileType and setIsAcidIow calls elsewhere for an example. - LoadTableDesc loadTableWork = new LoadTableDesc(moveTaskSrc, Utilities.getTableDesc(table), - partSpec.getPartSpec(), - replicationSpec.isReplace() ? LoadFileType.REPLACE_ALL : LoadFileType.OVERWRITE_EXISTING, - writeId); - loadTableWork.setStmtId(stmtId); - loadTableWork.setInheritTableSpecs(false); - Task loadPartTask = TaskFactory.get( - new MoveWork(x.getInputs(), x.getOutputs(), loadTableWork, null, false), - x.getConf()); + if (replicationSpec.isInReplicationScope() && AcidUtils.isTransactionalTable(tblDesc.getTblProps())) { + LoadMultiFilesDesc loadFilesWork = new LoadMultiFilesDesc( + Collections.singletonList(destPath), + Collections.singletonList(tgtLocation), + true, null, null); + moveWork.setMultiFilesDesc(loadFilesWork); + moveWork.setNeedCleanTarget(false); + } else { + LoadTableDesc loadTableWork = new LoadTableDesc(moveTaskSrc, Utilities.getTableDesc(table), + partSpec.getPartSpec(), + replicationSpec.isReplace() ? LoadFileType.REPLACE_ALL : LoadFileType.OVERWRITE_EXISTING, + writeId); + loadTableWork.setStmtId(stmtId); + loadTableWork.setInheritTableSpecs(false); + moveWork.setLoadTableWork(loadTableWork); + } + + Task loadPartTask = TaskFactory.get(moveWork, x.getConf()); copyTask.addDependentTask(loadPartTask); addPartTask.addDependentTask(loadPartTask); x.getTasks().add(copyTask); @@ -1005,7 +1036,8 @@ private static void createReplImportTasks( t.addDependentTask( addSinglePartition(fromURI, fs, tblDesc, table, wh, addPartitionDesc, replicationSpec, x, writeId, stmtId)); if (updatedMetadata != null) { - updatedMetadata.addPartition(addPartitionDesc.getPartition(0).getPartSpec()); + updatedMetadata.addPartition(table.getDbName(), table.getTableName(), + addPartitionDesc.getPartition(0).getPartSpec()); } } } else { @@ -1057,13 +1089,15 @@ private static void createReplImportTasks( x.getTasks().add(addSinglePartition( fromURI, fs, tblDesc, table, wh, addPartitionDesc, replicationSpec, x, writeId, stmtId)); if (updatedMetadata != null) { - updatedMetadata.addPartition(addPartitionDesc.getPartition(0).getPartSpec()); + updatedMetadata.addPartition(table.getDbName(), table.getTableName(), + addPartitionDesc.getPartition(0).getPartSpec()); } } else { x.getTasks().add(alterSinglePartition( fromURI, fs, tblDesc, table, wh, addPartitionDesc, replicationSpec, null, x)); if (updatedMetadata != null) { - updatedMetadata.addPartition(addPartitionDesc.getPartition(0).getPartSpec()); + updatedMetadata.addPartition(table.getDbName(), table.getTableName(), + addPartitionDesc.getPartition(0).getPartSpec()); } } } else { @@ -1078,7 +1112,8 @@ private static void createReplImportTasks( fromURI, fs, tblDesc, table, wh, addPartitionDesc, replicationSpec, ptn, x)); } if (updatedMetadata != null) { - updatedMetadata.addPartition(addPartitionDesc.getPartition(0).getPartSpec()); + updatedMetadata.addPartition(table.getDbName(), table.getTableName(), + addPartitionDesc.getPartition(0).getPartSpec()); } if (lockType == WriteEntity.WriteType.DDL_NO_LOCK){ lockType = WriteEntity.WriteType.DDL_SHARED; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java index f37de3e808..c5714a560d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ReplicationSemanticAnalyzer.java @@ -588,49 +588,55 @@ private void analyzeReplLoad(ASTNode ast) throws SemanticException { private List> addUpdateReplStateTasks( boolean isDatabaseLoad, - UpdatedMetaDataTracker updatedMetadata, + UpdatedMetaDataTracker updatedMetaDataTracker, List> importTasks) throws SemanticException { - String replState = updatedMetadata.getReplicationState(); - String dbName = updatedMetadata.getDatabase(); - String tableName = updatedMetadata.getTable(); - - // If no import tasks generated by the event or no table updated for table level load, then no - // need to update the repl state to any object. - if (importTasks.isEmpty() || (!isDatabaseLoad && (tableName == null))) { - LOG.debug("No objects need update of repl state: Either 0 import tasks or table level load"); + // If no import tasks generated by the event then no need to update the repl state to any object. + if (importTasks.isEmpty()) { + LOG.debug("No objects need update of repl state: 0 import tasks"); return importTasks; } // Create a barrier task for dependency collection of import tasks - Task barrierTask = TaskFactory.get(new DependencyCollectionWork()); - - // Link import tasks to the barrier task which will in-turn linked with repl state update tasks - for (Task t : importTasks){ - t.addDependentTask(barrierTask); - LOG.debug("Added {}:{} as a precursor of barrier task {}:{}", - t.getClass(), t.getId(), barrierTask.getClass(), barrierTask.getId()); - } + Task barrierTask = TaskFactory.get(new DependencyCollectionWork(), conf); List> tasks = new ArrayList<>(); Task updateReplIdTask; // If any partition is updated, then update repl state in partition object - for (final Map partSpec : updatedMetadata.getPartitions()) { - updateReplIdTask = tableUpdateReplStateTask(dbName, tableName, partSpec, replState, barrierTask); - tasks.add(updateReplIdTask); + for (UpdatedMetaDataTracker.UpdateMetaData updateMetaData : updatedMetaDataTracker.getUpdateMetaDataList()) { + String replState = updateMetaData.getReplState(); + String dbName = updateMetaData.getDbName(); + String tableName = updateMetaData.getTableName(); + // If any partition is updated, then update repl state in partition object + for (final Map partSpec : updateMetaData.getPartitionsList()) { + updateReplIdTask = tableUpdateReplStateTask(dbName, tableName, partSpec, replState, barrierTask); + tasks.add(updateReplIdTask); + } + + if (tableName != null) { + // If any table/partition is updated, then update repl state in table object + updateReplIdTask = tableUpdateReplStateTask(dbName, tableName, null, replState, barrierTask); + tasks.add(updateReplIdTask); + } + + // For table level load, need not update replication state for the database + if (isDatabaseLoad) { + // If any table/partition is updated, then update repl state in db object + updateReplIdTask = dbUpdateReplStateTask(dbName, replState, barrierTask); + tasks.add(updateReplIdTask); + } } - if (tableName != null) { - // If any table/partition is updated, then update repl state in table object - updateReplIdTask = tableUpdateReplStateTask(dbName, tableName, null, replState, barrierTask); - tasks.add(updateReplIdTask); + if (tasks.isEmpty()) { + LOG.debug("No objects need update of repl state: 0 update tracker tasks"); + return importTasks; } - // For table level load, need not update replication state for the database - if (isDatabaseLoad) { - // If any table/partition is updated, then update repl state in db object - updateReplIdTask = dbUpdateReplStateTask(dbName, replState, barrierTask); - tasks.add(updateReplIdTask); + // Link import tasks to the barrier task which will in-turn linked with repl state update tasks + for (Task t : importTasks){ + t.addDependentTask(barrierTask); + LOG.debug("Added {}:{} as a precursor of barrier task {}:{}", + t.getClass(), t.getId(), barrierTask.getClass(), barrierTask.getId()); } // At least one task would have been added to update the repl state diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index e10729e649..26db245e68 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -7316,7 +7316,7 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) } try { if (ctx.getExplainConfig() != null) { - writeId = 0L; // For explain plan, txn won't be opened and doesn't make sense to allocate write id + writeId = null; // For explain plan, txn won't be opened and doesn't make sense to allocate write id } else { if (isMmTable) { writeId = txnMgr.getTableWriteId(dest_tab.getDbName(), dest_tab.getTableName()); @@ -7331,6 +7331,9 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) boolean isReplace = !qb.getParseInfo().isInsertIntoTable( dest_tab.getDbName(), dest_tab.getTableName()); ltd = new LoadTableDesc(queryTmpdir, table_desc, dpCtx, acidOp, isReplace, writeId); + if (writeId != null) { + ltd.setStmtId(txnMgr.getCurrentStmtId()); + } // For Acid table, Insert Overwrite shouldn't replace the table content. We keep the old // deltas and base and leave them up to the cleaner to clean up boolean isInsertInto = qb.getParseInfo().isInsertIntoTable( @@ -7426,6 +7429,9 @@ protected Operator genFileSinkPlan(String dest, QB qb, Operator input) throw new SemanticException("Failed to allocate write Id", ex); } ltd = new LoadTableDesc(queryTmpdir, table_desc, dest_part.getSpec(), acidOp, writeId); + if (writeId != null) { + ltd.setStmtId(txnMgr.getCurrentStmtId()); + } // For the current context for generating File Sink Operator, it is either INSERT INTO or INSERT OVERWRITE. // So the next line works. boolean isInsertInto = !qb.getParseInfo().isDestToOpTypeInsertOverwrite(dest); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java index d9483f8205..97eb4b45f3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java @@ -179,9 +179,23 @@ private void analyzeAcidExport(ASTNode ast) throws SemanticException { String newTableName = getTmptTableNameForExport(exportTable); //this is db.table Map tblProps = new HashMap<>(); tblProps.put(hive_metastoreConstants.TABLE_IS_TRANSACTIONAL, Boolean.FALSE.toString()); + String location; + + // for temporary tables we set the location to something in the session's scratch dir + // it has the same life cycle as the tmp table + try { + // Generate a unique ID for temp table path. + // This path will be fixed for the life of the temp table. + Path path = new Path(SessionState.getTempTableSpace(conf), UUID.randomUUID().toString()); + path = Warehouse.getDnsPath(path, conf); + location = path.toString(); + } catch (MetaException err) { + throw new SemanticException("Error while generating temp table path:", err); + } + CreateTableLikeDesc ctlt = new CreateTableLikeDesc(newTableName, false, true, null, - null, null, null, null, + null, location, null, null, tblProps, true, //important so we get an exception on name collision Warehouse.getQualifiedName(exportTable.getTTable()), false); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/CopyUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/CopyUtils.java index 7e8d520d93..9a54c1e59b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/CopyUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/CopyUtils.java @@ -400,7 +400,7 @@ private boolean isLocal(FileSystem fs) { return result; } - private Path getCopyDestination(ReplChangeManager.FileInfo fileInfo, Path destRoot) { + public static Path getCopyDestination(ReplChangeManager.FileInfo fileInfo, Path destRoot) { if (fileInfo.getSubDir() == null) { return destRoot; } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/Utils.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/Utils.java index c0701c5b87..62d699f706 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/Utils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/Utils.java @@ -186,10 +186,6 @@ public static Boolean shouldReplicate(ReplicationSpec replicationSpec, Table tab return false; } - boolean isAcidTable = AcidUtils.isTransactionalTable(tableHandle); - if (isAcidTable) { - return hiveConf.getBoolVar(HiveConf.ConfVars.REPL_DUMP_INCLUDE_ACID_TABLES); - } return !tableHandle.isTemporary(); } return true; diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CommitTxnHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CommitTxnHandler.java index db97d7c945..f04cd93069 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CommitTxnHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CommitTxnHandler.java @@ -18,9 +18,27 @@ */ package org.apache.hadoop.hive.ql.parse.repl.dump.events; +import com.google.common.collect.Lists; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hive.conf.HiveConf; +import org.apache.hadoop.hive.metastore.HiveMetaStore; +import org.apache.hadoop.hive.metastore.ReplChangeManager; import org.apache.hadoop.hive.metastore.api.NotificationEvent; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; +import org.apache.hadoop.hive.metastore.messaging.CommitTxnMessage; +import org.apache.hadoop.hive.metastore.utils.StringUtils; +import org.apache.hadoop.hive.ql.metadata.HiveUtils; +import org.apache.hadoop.hive.ql.metadata.Partition; +import org.apache.hadoop.hive.ql.parse.EximUtil; +import org.apache.hadoop.hive.ql.parse.SemanticException; import org.apache.hadoop.hive.ql.parse.repl.DumpType; import org.apache.hadoop.hive.ql.parse.repl.load.DumpMetaData; +import org.apache.hadoop.fs.FileSystem; +import java.io.BufferedWriter; +import java.io.IOException; +import java.io.OutputStreamWriter; +import java.util.ArrayList; +import java.util.List; class CommitTxnHandler extends AbstractEventHandler { @@ -28,11 +46,116 @@ super(event); } + private BufferedWriter writer(Context withinContext, Path dataPath) throws IOException { + Path filesPath = new Path(dataPath, EximUtil.FILES_NAME); + FileSystem fs = dataPath.getFileSystem(withinContext.hiveConf); + return new BufferedWriter(new OutputStreamWriter(fs.create(filesPath))); + } + + private void writeDumpFiles(Context withinContext, Iterable files, Path dataPath) throws IOException { + // encoded filename/checksum of files, write into _files + try (BufferedWriter fileListWriter = writer(withinContext, dataPath)) { + for (String file : files) { + fileListWriter.write(file + "\n"); + } + } + } + + private void createDumpFile(Context withinContext, org.apache.hadoop.hive.ql.metadata.Table qlMdTable, + List qlPtns, List> fileListArray) throws IOException, SemanticException { + if (fileListArray == null || fileListArray.isEmpty()) { + return; + } + + Path metaDataPath = new Path(withinContext.eventRoot, EximUtil.METADATA_NAME); + withinContext.replicationSpec.setIsReplace(true); + EximUtil.createExportDump(metaDataPath.getFileSystem(withinContext.hiveConf), metaDataPath, + qlMdTable, qlPtns, + withinContext.replicationSpec, + withinContext.hiveConf); + + if ((null == qlPtns) || qlPtns.isEmpty()) { + Path dataPath = new Path(withinContext.eventRoot, EximUtil.DATA_PATH_NAME); + writeDumpFiles(withinContext, fileListArray.get(0), dataPath); + } else { + for (int idx = 0; idx < qlPtns.size(); idx++) { + Path dataPath = new Path(withinContext.eventRoot, qlPtns.get(idx).getName()); + writeDumpFiles(withinContext, fileListArray.get(idx), dataPath); + } + } + } + + private void createDumpFileForTable(Context withinContext, org.apache.hadoop.hive.ql.metadata.Table qlMdTable, + List qlPtns, List> fileListArray) throws IOException, SemanticException { + Path newPath = HiveUtils.getDumpPath(withinContext.eventRoot, qlMdTable.getDbName(), qlMdTable.getTableName()); + Context context = new Context(withinContext); + context.setEventRoot(newPath); + createDumpFile(context, qlMdTable, qlPtns, fileListArray); + } + @Override public void handle(Context withinContext) throws Exception { LOG.info("Processing#{} COMMIT_TXN message : {}", fromEventId(), event.getMessage()); + String payload = event.getMessage(); + + if (!withinContext.hiveConf.getBoolVar(HiveConf.ConfVars.REPL_DUMP_METADATA_ONLY)) { + CommitTxnMessage commitTxnMessage = deserializer.getCommitTxnMessage(event.getMessage()); + + String contextDbName = withinContext.dbName == null ? null : + StringUtils.normalizeIdentifier(withinContext.dbName); + String contextTableName = withinContext.tableName == null ? null : + StringUtils.normalizeIdentifier(withinContext.tableName); + List writeEventInfoList = HiveMetaStore.HMSHandler.getMSForConf(withinContext.hiveConf). + getAllWriteEventInfo(commitTxnMessage.getTxnId(), contextDbName, contextTableName); + int numEntry = (writeEventInfoList != null ? writeEventInfoList.size() : 0); + if (numEntry != 0) { + commitTxnMessage.addWriteEventInfo(writeEventInfoList); + payload = commitTxnMessage.toString(); + LOG.debug("payload for commit txn event : " + payload); + } + + org.apache.hadoop.hive.ql.metadata.Table qlMdTablePrev = null; + org.apache.hadoop.hive.ql.metadata.Table qlMdTable = null; + List qlPtns = new ArrayList<>(); + List> filesTobeAdded = new ArrayList<>(); + + // The below loop creates dump directory for each table. It reads through the list of write notification events, + // groups the entries per table and creates the lists of files to be replicated. The event directory in the dump + // path will have subdirectory for each table. This folder will have metadata for the table and the list of files + // to be replicated. The entries are added in the table with txn id, db name,table name, partition name + // combination as primary key, so the entries with same table will come together. Only basic table metadata is + // used during import, so we need not dump the latest table metadata. + for (int idx = 0; idx < numEntry; idx++) { + qlMdTable = new org.apache.hadoop.hive.ql.metadata.Table(commitTxnMessage.getTableObj(idx)); + if (qlMdTablePrev == null) { + qlMdTablePrev = qlMdTable; + } + + // one dump directory per table + if (!qlMdTablePrev.getCompleteName().equals(qlMdTable.getCompleteName())) { + createDumpFileForTable(withinContext, qlMdTablePrev, qlPtns, filesTobeAdded); + qlPtns = new ArrayList<>(); + filesTobeAdded = new ArrayList<>(); + qlMdTablePrev = qlMdTable; + } + + if (qlMdTable.isPartitioned() && (null != commitTxnMessage.getPartitionObj(idx))) { + qlPtns.add(new org.apache.hadoop.hive.ql.metadata.Partition(qlMdTable, + commitTxnMessage.getPartitionObj(idx))); + } + + filesTobeAdded.add(Lists.newArrayList( + ReplChangeManager.getListFromSeparatedString(commitTxnMessage.getFiles(idx)))); + } + + //Dump last table in the list + if (qlMdTablePrev != null) { + createDumpFileForTable(withinContext, qlMdTablePrev, qlPtns, filesTobeAdded); + } + } + DumpMetaData dmd = withinContext.createDmd(this); - dmd.setPayload(event.getMessage()); + dmd.setPayload(payload); dmd.write(); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/EventHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/EventHandler.java index c0fa7b2d4e..ec35f4e94d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/EventHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/EventHandler.java @@ -35,18 +35,37 @@ DumpType dumpType(); class Context { - final Path eventRoot, cmRoot; + Path eventRoot; + final Path cmRoot; final Hive db; final HiveConf hiveConf; final ReplicationSpec replicationSpec; + final String dbName; + final String tableName; public Context(Path eventRoot, Path cmRoot, Hive db, HiveConf hiveConf, - ReplicationSpec replicationSpec) { + ReplicationSpec replicationSpec, String dbName, String tableName) { this.eventRoot = eventRoot; this.cmRoot = cmRoot; this.db = db; this.hiveConf = hiveConf; this.replicationSpec = replicationSpec; + this.dbName = dbName; + this.tableName = tableName; + } + + public Context(Context other) { + this.eventRoot = other.eventRoot; + this.cmRoot = other.cmRoot; + this.db = other.db; + this.hiveConf = other.hiveConf; + this.replicationSpec = other.replicationSpec; + this.dbName = other.dbName; + this.tableName = other.tableName; + } + + public void setEventRoot(Path eventRoot) { + this.eventRoot = eventRoot; } DumpMetaData createDmd(EventHandler eventHandler) { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/InsertHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/InsertHandler.java index 5ac3af0c30..cf3822a3fe 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/InsertHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/InsertHandler.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.NotificationEvent; import org.apache.hadoop.hive.metastore.messaging.InsertMessage; +import org.apache.hadoop.hive.ql.io.AcidUtils; import org.apache.hadoop.hive.ql.metadata.Partition; import org.apache.hadoop.hive.ql.parse.EximUtil; import org.apache.hadoop.hive.ql.parse.repl.DumpType; @@ -53,6 +54,9 @@ public void handle(Context withinContext) throws Exception { return; } + // In case of ACID tables, insert event should not have fired. + assert(!AcidUtils.isTransactionalTable(qlMdTable)); + List qlPtns = null; if (qlMdTable.isPartitioned() && (null != insertMsg.getPtnObj())) { qlPtns = Collections.singletonList(partitionObject(qlMdTable, insertMsg)); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/UpdatedMetaDataTracker.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/UpdatedMetaDataTracker.java index d76f4193e0..614e0713ae 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/UpdatedMetaDataTracker.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/UpdatedMetaDataTracker.java @@ -17,7 +17,10 @@ */ package org.apache.hadoop.hive.ql.parse.repl.load; +import org.apache.hadoop.hive.ql.parse.SemanticException; +import org.apache.hive.common.util.HiveStringUtils; import java.util.ArrayList; +import java.util.HashMap; import java.util.Map; import java.util.List; @@ -25,52 +28,113 @@ * Utility class to help track and return the metadata which are updated by repl load */ public class UpdatedMetaDataTracker { - private String replState; - private String dbName; - private String tableName; - private List> partitionsList; - public UpdatedMetaDataTracker() { - this.replState = null; - this.dbName = null; - this.tableName = null; - this.partitionsList = new ArrayList<>(); + /** + * Utility class to store replication state of a table. + */ + public static class UpdateMetaData { + private String replState; + private String dbName; + private String tableName; + private List> partitionsList; + + UpdateMetaData(String replState, String dbName, String tableName, Map partSpec) { + this.replState = replState; + this.dbName = dbName; + this.tableName = tableName; + this.partitionsList = new ArrayList<>(); + if (partSpec != null) { + this.partitionsList.add(partSpec); + } + } + + public String getReplState() { + return replState; + } + + public String getDbName() { + return dbName; + } + + public String getTableName() { + return tableName; + } + + public List> getPartitionsList() { + return partitionsList; + } + + public void addPartition(Map partSpec) { + this.partitionsList.add(partSpec); + } } - public void copyUpdatedMetadata(UpdatedMetaDataTracker other) { - this.replState = other.replState; - this.dbName = other.dbName; - this.tableName = other.tableName; - this.partitionsList = other.getPartitions(); + private List updateMetaDataList; + private Map updateMetaDataMap; + + public UpdatedMetaDataTracker() { + updateMetaDataList = new ArrayList<>(); + updateMetaDataMap = new HashMap<>(); } - public void set(String replState, String dbName, String tableName, Map partSpec) { - this.replState = replState; - this.dbName = dbName; - this.tableName = tableName; - if (partSpec != null) { - addPartition(partSpec); + public void copyUpdatedMetadata(UpdatedMetaDataTracker other) { + int size = updateMetaDataList.size(); + for (UpdateMetaData updateMetaDataOther : other.updateMetaDataList) { + String key = getKey(normalizeIdentifier(updateMetaDataOther.getDbName()), + normalizeIdentifier(updateMetaDataOther.getTableName())); + Integer idx = updateMetaDataMap.get(key); + if (idx == null) { + updateMetaDataList.add(updateMetaDataOther); + updateMetaDataMap.put(key, size++); + } else if (updateMetaDataOther.partitionsList != null && updateMetaDataOther.partitionsList.size() != 0) { + UpdateMetaData updateMetaData = updateMetaDataList.get(idx); + for (Map partSpec : updateMetaDataOther.partitionsList) { + updateMetaData.addPartition(partSpec); + } + } } } - public void addPartition(Map partSpec) { - partitionsList.add(partSpec); + public void set(String replState, String dbName, String tableName, Map partSpec) + throws SemanticException { + if (dbName == null) { + throw new SemanticException("db name can not be null"); + } + String key = getKey(normalizeIdentifier(dbName), normalizeIdentifier(tableName)); + Integer idx = updateMetaDataMap.get(key); + if (idx == null) { + updateMetaDataList.add(new UpdateMetaData(replState, dbName, tableName, partSpec)); + updateMetaDataMap.put(key, updateMetaDataList.size() - 1); + } else { + updateMetaDataList.get(idx).addPartition(partSpec); + } } - public String getReplicationState() { - return replState; + public void addPartition(String dbName, String tableName, Map partSpec) throws SemanticException { + if (dbName == null) { + throw new SemanticException("db name can not be null"); + } + String key = getKey(normalizeIdentifier(dbName), normalizeIdentifier(tableName)); + Integer idx = updateMetaDataMap.get(key); + if (idx == null) { + throw new SemanticException("add partition to metadata map failed as list is not yet set for table : " + key); + } + updateMetaDataList.get(idx).addPartition(partSpec); } - public String getDatabase() { - return dbName; + public List getUpdateMetaDataList() { + return updateMetaDataList; } - public String getTable() { - return tableName; + private String getKey(String dbName, String tableName) { + if (tableName == null) { + return dbName + ".*"; + } + return dbName + "." + tableName; } - public List> getPartitions() { - return partitionsList; + private String normalizeIdentifier(String name) { + return name == null ? null : HiveStringUtils.normalizeIdentifier(name); } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AbortTxnHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AbortTxnHandler.java index afc7426217..d3f3306366 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AbortTxnHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AbortTxnHandler.java @@ -48,7 +48,12 @@ msg.getTxnId(), ReplTxnWork.OperationType.REPL_ABORT_TXN, context.eventOnlyReplicationSpec()), context.hiveConf ); - updatedMetadata.set(context.dmd.getEventTo().toString(), context.dbName, context.tableName, null); + + // For warehouse level dump, don't update the metadata of database as we don't know this txn is for which database. + // Anyways, if this event gets executed again, it is taken care of. + if (!context.isDbNameEmpty()) { + updatedMetadata.set(context.dmd.getEventTo().toString(), context.dbName, context.tableName, null); + } context.log.debug("Added Abort txn task : {}", abortTxnTask.getId()); return Collections.singletonList(abortTxnTask); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AllocWriteIdHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AllocWriteIdHandler.java index 9bdbf649a6..63f2577bb5 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AllocWriteIdHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/AllocWriteIdHandler.java @@ -52,7 +52,7 @@ .getTableName()); // Repl policy should be created based on the table name in context. - ReplTxnWork work = new ReplTxnWork(HiveUtils.getReplPolicy(dbName, context.tableName), dbName, tableName, + ReplTxnWork work = new ReplTxnWork(HiveUtils.getReplPolicy(context.dbName, context.tableName), dbName, tableName, ReplTxnWork.OperationType.REPL_ALLOC_WRITE_ID, msg.getTxnToWriteIdList(), context.eventOnlyReplicationSpec()); Task allocWriteIdTask = TaskFactory.get(work, context.hiveConf); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CommitTxnHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CommitTxnHandler.java index d25102ef1d..87a6ff6d4e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CommitTxnHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/CommitTxnHandler.java @@ -17,7 +17,12 @@ */ package org.apache.hadoop.hive.ql.parse.repl.load.message; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.messaging.CommitTxnMessage; +import org.apache.hadoop.hive.ql.exec.repl.bootstrap.AddDependencyToLeaves; +import org.apache.hadoop.hive.ql.exec.util.DAGTraversal; +import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.plan.ReplTxnWork; import org.apache.hadoop.hive.ql.exec.Task; import org.apache.hadoop.hive.ql.exec.TaskFactory; @@ -25,7 +30,7 @@ import org.apache.hadoop.hive.ql.metadata.HiveUtils; import org.apache.hadoop.hive.ql.parse.SemanticException; import java.io.Serializable; -import java.util.Collections; +import java.util.ArrayList; import java.util.List; /** @@ -35,20 +40,75 @@ public class CommitTxnHandler extends AbstractMessageHandler { @Override public List> handle(Context context) - throws SemanticException { + throws SemanticException { if (!AcidUtils.isAcidEnabled(context.hiveConf)) { context.log.error("Cannot load transaction events as acid is not enabled"); throw new SemanticException("Cannot load transaction events as acid is not enabled"); } CommitTxnMessage msg = deserializer.getCommitTxnMessage(context.dmd.getPayload()); - Task commitTxnTask = TaskFactory.get( - new ReplTxnWork(HiveUtils.getReplPolicy(context.dbName, context.tableName), context.dbName, context.tableName, - msg.getTxnId(), ReplTxnWork.OperationType.REPL_COMMIT_TXN, context.eventOnlyReplicationSpec()), - context.hiveConf - ); - updatedMetadata.set(context.dmd.getEventTo().toString(), context.dbName, context.tableName, null); + int numEntry = (msg.getTables() == null ? 0 : msg.getTables().size()); + List> tasks = new ArrayList<>(); + String dbName = context.dbName; + String tableNamePrev = null; + String tblName = context.tableName; + + ReplTxnWork work = new ReplTxnWork(HiveUtils.getReplPolicy(context.dbName, context.tableName), context.dbName, + context.tableName, msg.getTxnId(), ReplTxnWork.OperationType.REPL_COMMIT_TXN, context.eventOnlyReplicationSpec()); + + if (numEntry > 0) { + context.log.debug("Commit txn handler for txnid " + msg.getTxnId() + " databases : " + msg.getDatabases() + + " tables : " + msg.getTables() + " partitions : " + msg.getPartitions() + " files : " + + msg.getFilesList() + " write ids : " + msg.getWriteIds()); + } + + for (int idx = 0; idx < numEntry; idx++) { + String actualTblName = msg.getTables().get(idx); + String actualDBName = msg.getDatabases().get(idx); + String completeName = Table.getCompleteName(actualDBName, actualTblName); + + // One import task per table. Events for same table are kept together in one dump directory during dump and are + // grouped together in commit txn message. + if (tableNamePrev == null || !(completeName.equals(tableNamePrev))) { + // The data location is created by source, so the location should be formed based on the table name in msg. + Path location = HiveUtils.getDumpPath(new Path(context.location), actualDBName, actualTblName); + tblName = context.isTableNameEmpty() ? actualTblName : context.tableName; + // for warehouse level dump, use db name from write event + dbName = (context.isDbNameEmpty() ? actualDBName : context.dbName); + Context currentContext = new Context(context, dbName, tblName); + currentContext.setLocation(location.toUri().toString()); + + // Piggybacking in Import logic for now + TableHandler tableHandler = new TableHandler(); + tasks.addAll((tableHandler.handle(currentContext))); + readEntitySet.addAll(tableHandler.readEntities()); + writeEntitySet.addAll(tableHandler.writeEntities()); + getUpdatedMetadata().copyUpdatedMetadata(tableHandler.getUpdatedMetadata()); + tableNamePrev = completeName; + } + + try { + WriteEventInfo writeEventInfo = new WriteEventInfo(msg.getWriteIds().get(idx), + dbName, tblName, msg.getFiles(idx)); + if (msg.getPartitions().get(idx) != null && !msg.getPartitions().get(idx).isEmpty()) { + writeEventInfo.setPartition(msg.getPartitions().get(idx)); + } + work.addWriteEventInfo(writeEventInfo); + } catch (Exception e) { + throw new SemanticException("Failed to extract write event info from commit txn message : " + e.getMessage()); + } + } + + Task commitTxnTask = TaskFactory.get(work, context.hiveConf); + + // For warehouse level dump, don't update the metadata of database as we don't know this txn is for which database. + // Anyways, if this event gets executed again, it is taken care of. + if (!context.isDbNameEmpty()) { + updatedMetadata.set(context.dmd.getEventTo().toString(), context.dbName, context.tableName, null); + } context.log.debug("Added Commit txn task : {}", commitTxnTask.getId()); - return Collections.singletonList(commitTxnTask); + DAGTraversal.traverse(tasks, new AddDependencyToLeaves(commitTxnTask)); + return tasks; } } + diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/MessageHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/MessageHandler.java index ef4a901b8f..cdf51dd12c 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/MessageHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/MessageHandler.java @@ -46,8 +46,8 @@ UpdatedMetaDataTracker getUpdatedMetadata(); class Context { - public String dbName; - public final String tableName, location; + public String location; + public final String tableName, dbName; public final Task precursor; public DumpMetaData dmd; final HiveConf hiveConf; @@ -101,5 +101,9 @@ ReplicationSpec eventOnlyReplicationSpec() throws SemanticException { public HiveTxnManager getTxnMgr() { return nestedContext.getHiveTxnManager(); } + + public void setLocation(String location) { + this.location = location; + } } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/OpenTxnHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/OpenTxnHandler.java index 190e02172f..5dcc44e56b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/OpenTxnHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/load/message/OpenTxnHandler.java @@ -47,7 +47,12 @@ msg.getTxnIds(), ReplTxnWork.OperationType.REPL_OPEN_TXN, context.eventOnlyReplicationSpec()), context.hiveConf ); - updatedMetadata.set(context.dmd.getEventTo().toString(), context.dbName, context.tableName, null); + + // For warehouse level dump, don't update the metadata of database as we don't know this txn is for which database. + // Anyways, if this event gets executed again, it is taken care of. + if (!context.isDbNameEmpty()) { + updatedMetadata.set(context.dmd.getEventTo().toString(), context.dbName, context.tableName, null); + } context.log.debug("Added Open txn task : {}", openTxnTask.getId()); return Collections.singletonList(openTxnTask); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/MoveWork.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/MoveWork.java index 9a1e3a1af5..47a56d5ade 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/MoveWork.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/MoveWork.java @@ -40,6 +40,7 @@ private LoadMultiFilesDesc loadMultiFilesWork; private boolean checkFileFormat; private boolean srcLocal; + private boolean needCleanTarget; /** * ReadEntitites that are passed to the hooks. @@ -63,6 +64,7 @@ public MoveWork() { private MoveWork(HashSet inputs, HashSet outputs) { this.inputs = inputs; this.outputs = outputs; + this.needCleanTarget = true; } public MoveWork(HashSet inputs, HashSet outputs, @@ -93,6 +95,7 @@ public MoveWork(final MoveWork o) { srcLocal = o.isSrcLocal(); inputs = o.getInputs(); outputs = o.getOutputs(); + needCleanTarget = o.needCleanTarget; } @Explain(displayName = "tables", explainLevels = { Level.USER, Level.DEFAULT, Level.EXTENDED }) @@ -153,5 +156,12 @@ public boolean isSrcLocal() { public void setSrcLocal(boolean srcLocal) { this.srcLocal = srcLocal; } - + + public boolean isNeedCleanTarget() { + return needCleanTarget; + } + + public void setNeedCleanTarget(boolean needCleanTarget) { + this.needCleanTarget = needCleanTarget; + } } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/ReplTxnWork.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/ReplTxnWork.java index 3c853c9ccd..a6ab83659d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/ReplTxnWork.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/ReplTxnWork.java @@ -20,8 +20,10 @@ import java.io.Serializable; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.ql.parse.ReplicationSpec; import org.apache.hadoop.hive.ql.plan.Explain.Level; +import java.util.ArrayList; import java.util.Collections; import java.util.List; @@ -40,6 +42,7 @@ private List txnIds; private List txnToWriteIdList; private ReplicationSpec replicationSpec; + private List writeEventInfos; /** * OperationType. @@ -60,6 +63,7 @@ public ReplTxnWork(String replPolicy, String dbName, String tableName, List txnIds, OperationType type, @@ -86,6 +90,13 @@ public ReplTxnWork(String dbName, String tableName, List partNames, this.operation = type; } + public void addWriteEventInfo(WriteEventInfo writeEventInfo) { + if (this.writeEventInfos == null) { + this.writeEventInfos = new ArrayList<>(); + } + this.writeEventInfos.add(writeEventInfo); + } + public List getTxnIds() { return txnIds; } @@ -121,4 +132,8 @@ public OperationType getOperationType() { public ReplicationSpec getReplicationSpec() { return replicationSpec; } + + public List getWriteEventInfos() { + return writeEventInfos; + } } diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp index cd356e1e0f..9d57d4cffa 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp +++ b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp @@ -2334,14 +2334,14 @@ uint32_t ThriftHiveMetastore_get_databases_result::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1195; - ::apache::thrift::protocol::TType _etype1198; - xfer += iprot->readListBegin(_etype1198, _size1195); - this->success.resize(_size1195); - uint32_t _i1199; - for (_i1199 = 0; _i1199 < _size1195; ++_i1199) + uint32_t _size1219; + ::apache::thrift::protocol::TType _etype1222; + xfer += iprot->readListBegin(_etype1222, _size1219); + this->success.resize(_size1219); + uint32_t _i1223; + for (_i1223 = 0; _i1223 < _size1219; ++_i1223) { - xfer += iprot->readString(this->success[_i1199]); + xfer += iprot->readString(this->success[_i1223]); } xfer += iprot->readListEnd(); } @@ -2380,10 +2380,10 @@ uint32_t ThriftHiveMetastore_get_databases_result::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1200; - for (_iter1200 = this->success.begin(); _iter1200 != this->success.end(); ++_iter1200) + std::vector ::const_iterator _iter1224; + for (_iter1224 = this->success.begin(); _iter1224 != this->success.end(); ++_iter1224) { - xfer += oprot->writeString((*_iter1200)); + xfer += oprot->writeString((*_iter1224)); } xfer += oprot->writeListEnd(); } @@ -2428,14 +2428,14 @@ uint32_t ThriftHiveMetastore_get_databases_presult::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1201; - ::apache::thrift::protocol::TType _etype1204; - xfer += iprot->readListBegin(_etype1204, _size1201); - (*(this->success)).resize(_size1201); - uint32_t _i1205; - for (_i1205 = 0; _i1205 < _size1201; ++_i1205) + uint32_t _size1225; + ::apache::thrift::protocol::TType _etype1228; + xfer += iprot->readListBegin(_etype1228, _size1225); + (*(this->success)).resize(_size1225); + uint32_t _i1229; + for (_i1229 = 0; _i1229 < _size1225; ++_i1229) { - xfer += iprot->readString((*(this->success))[_i1205]); + xfer += iprot->readString((*(this->success))[_i1229]); } xfer += iprot->readListEnd(); } @@ -2552,14 +2552,14 @@ uint32_t ThriftHiveMetastore_get_all_databases_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1206; - ::apache::thrift::protocol::TType _etype1209; - xfer += iprot->readListBegin(_etype1209, _size1206); - this->success.resize(_size1206); - uint32_t _i1210; - for (_i1210 = 0; _i1210 < _size1206; ++_i1210) + uint32_t _size1230; + ::apache::thrift::protocol::TType _etype1233; + xfer += iprot->readListBegin(_etype1233, _size1230); + this->success.resize(_size1230); + uint32_t _i1234; + for (_i1234 = 0; _i1234 < _size1230; ++_i1234) { - xfer += iprot->readString(this->success[_i1210]); + xfer += iprot->readString(this->success[_i1234]); } xfer += iprot->readListEnd(); } @@ -2598,10 +2598,10 @@ uint32_t ThriftHiveMetastore_get_all_databases_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1211; - for (_iter1211 = this->success.begin(); _iter1211 != this->success.end(); ++_iter1211) + std::vector ::const_iterator _iter1235; + for (_iter1235 = this->success.begin(); _iter1235 != this->success.end(); ++_iter1235) { - xfer += oprot->writeString((*_iter1211)); + xfer += oprot->writeString((*_iter1235)); } xfer += oprot->writeListEnd(); } @@ -2646,14 +2646,14 @@ uint32_t ThriftHiveMetastore_get_all_databases_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1212; - ::apache::thrift::protocol::TType _etype1215; - xfer += iprot->readListBegin(_etype1215, _size1212); - (*(this->success)).resize(_size1212); - uint32_t _i1216; - for (_i1216 = 0; _i1216 < _size1212; ++_i1216) + uint32_t _size1236; + ::apache::thrift::protocol::TType _etype1239; + xfer += iprot->readListBegin(_etype1239, _size1236); + (*(this->success)).resize(_size1236); + uint32_t _i1240; + for (_i1240 = 0; _i1240 < _size1236; ++_i1240) { - xfer += iprot->readString((*(this->success))[_i1216]); + xfer += iprot->readString((*(this->success))[_i1240]); } xfer += iprot->readListEnd(); } @@ -3715,17 +3715,17 @@ uint32_t ThriftHiveMetastore_get_type_all_result::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size1217; - ::apache::thrift::protocol::TType _ktype1218; - ::apache::thrift::protocol::TType _vtype1219; - xfer += iprot->readMapBegin(_ktype1218, _vtype1219, _size1217); - uint32_t _i1221; - for (_i1221 = 0; _i1221 < _size1217; ++_i1221) + uint32_t _size1241; + ::apache::thrift::protocol::TType _ktype1242; + ::apache::thrift::protocol::TType _vtype1243; + xfer += iprot->readMapBegin(_ktype1242, _vtype1243, _size1241); + uint32_t _i1245; + for (_i1245 = 0; _i1245 < _size1241; ++_i1245) { - std::string _key1222; - xfer += iprot->readString(_key1222); - Type& _val1223 = this->success[_key1222]; - xfer += _val1223.read(iprot); + std::string _key1246; + xfer += iprot->readString(_key1246); + Type& _val1247 = this->success[_key1246]; + xfer += _val1247.read(iprot); } xfer += iprot->readMapEnd(); } @@ -3764,11 +3764,11 @@ uint32_t ThriftHiveMetastore_get_type_all_result::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::map ::const_iterator _iter1224; - for (_iter1224 = this->success.begin(); _iter1224 != this->success.end(); ++_iter1224) + std::map ::const_iterator _iter1248; + for (_iter1248 = this->success.begin(); _iter1248 != this->success.end(); ++_iter1248) { - xfer += oprot->writeString(_iter1224->first); - xfer += _iter1224->second.write(oprot); + xfer += oprot->writeString(_iter1248->first); + xfer += _iter1248->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -3813,17 +3813,17 @@ uint32_t ThriftHiveMetastore_get_type_all_presult::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size1225; - ::apache::thrift::protocol::TType _ktype1226; - ::apache::thrift::protocol::TType _vtype1227; - xfer += iprot->readMapBegin(_ktype1226, _vtype1227, _size1225); - uint32_t _i1229; - for (_i1229 = 0; _i1229 < _size1225; ++_i1229) + uint32_t _size1249; + ::apache::thrift::protocol::TType _ktype1250; + ::apache::thrift::protocol::TType _vtype1251; + xfer += iprot->readMapBegin(_ktype1250, _vtype1251, _size1249); + uint32_t _i1253; + for (_i1253 = 0; _i1253 < _size1249; ++_i1253) { - std::string _key1230; - xfer += iprot->readString(_key1230); - Type& _val1231 = (*(this->success))[_key1230]; - xfer += _val1231.read(iprot); + std::string _key1254; + xfer += iprot->readString(_key1254); + Type& _val1255 = (*(this->success))[_key1254]; + xfer += _val1255.read(iprot); } xfer += iprot->readMapEnd(); } @@ -3977,14 +3977,14 @@ uint32_t ThriftHiveMetastore_get_fields_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1232; - ::apache::thrift::protocol::TType _etype1235; - xfer += iprot->readListBegin(_etype1235, _size1232); - this->success.resize(_size1232); - uint32_t _i1236; - for (_i1236 = 0; _i1236 < _size1232; ++_i1236) + uint32_t _size1256; + ::apache::thrift::protocol::TType _etype1259; + xfer += iprot->readListBegin(_etype1259, _size1256); + this->success.resize(_size1256); + uint32_t _i1260; + for (_i1260 = 0; _i1260 < _size1256; ++_i1260) { - xfer += this->success[_i1236].read(iprot); + xfer += this->success[_i1260].read(iprot); } xfer += iprot->readListEnd(); } @@ -4039,10 +4039,10 @@ uint32_t ThriftHiveMetastore_get_fields_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1237; - for (_iter1237 = this->success.begin(); _iter1237 != this->success.end(); ++_iter1237) + std::vector ::const_iterator _iter1261; + for (_iter1261 = this->success.begin(); _iter1261 != this->success.end(); ++_iter1261) { - xfer += (*_iter1237).write(oprot); + xfer += (*_iter1261).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4095,14 +4095,14 @@ uint32_t ThriftHiveMetastore_get_fields_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1238; - ::apache::thrift::protocol::TType _etype1241; - xfer += iprot->readListBegin(_etype1241, _size1238); - (*(this->success)).resize(_size1238); - uint32_t _i1242; - for (_i1242 = 0; _i1242 < _size1238; ++_i1242) + uint32_t _size1262; + ::apache::thrift::protocol::TType _etype1265; + xfer += iprot->readListBegin(_etype1265, _size1262); + (*(this->success)).resize(_size1262); + uint32_t _i1266; + for (_i1266 = 0; _i1266 < _size1262; ++_i1266) { - xfer += (*(this->success))[_i1242].read(iprot); + xfer += (*(this->success))[_i1266].read(iprot); } xfer += iprot->readListEnd(); } @@ -4288,14 +4288,14 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_result::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1243; - ::apache::thrift::protocol::TType _etype1246; - xfer += iprot->readListBegin(_etype1246, _size1243); - this->success.resize(_size1243); - uint32_t _i1247; - for (_i1247 = 0; _i1247 < _size1243; ++_i1247) + uint32_t _size1267; + ::apache::thrift::protocol::TType _etype1270; + xfer += iprot->readListBegin(_etype1270, _size1267); + this->success.resize(_size1267); + uint32_t _i1271; + for (_i1271 = 0; _i1271 < _size1267; ++_i1271) { - xfer += this->success[_i1247].read(iprot); + xfer += this->success[_i1271].read(iprot); } xfer += iprot->readListEnd(); } @@ -4350,10 +4350,10 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_result::write(: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1248; - for (_iter1248 = this->success.begin(); _iter1248 != this->success.end(); ++_iter1248) + std::vector ::const_iterator _iter1272; + for (_iter1272 = this->success.begin(); _iter1272 != this->success.end(); ++_iter1272) { - xfer += (*_iter1248).write(oprot); + xfer += (*_iter1272).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4406,14 +4406,14 @@ uint32_t ThriftHiveMetastore_get_fields_with_environment_context_presult::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1249; - ::apache::thrift::protocol::TType _etype1252; - xfer += iprot->readListBegin(_etype1252, _size1249); - (*(this->success)).resize(_size1249); - uint32_t _i1253; - for (_i1253 = 0; _i1253 < _size1249; ++_i1253) + uint32_t _size1273; + ::apache::thrift::protocol::TType _etype1276; + xfer += iprot->readListBegin(_etype1276, _size1273); + (*(this->success)).resize(_size1273); + uint32_t _i1277; + for (_i1277 = 0; _i1277 < _size1273; ++_i1277) { - xfer += (*(this->success))[_i1253].read(iprot); + xfer += (*(this->success))[_i1277].read(iprot); } xfer += iprot->readListEnd(); } @@ -4583,14 +4583,14 @@ uint32_t ThriftHiveMetastore_get_schema_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1254; - ::apache::thrift::protocol::TType _etype1257; - xfer += iprot->readListBegin(_etype1257, _size1254); - this->success.resize(_size1254); - uint32_t _i1258; - for (_i1258 = 0; _i1258 < _size1254; ++_i1258) + uint32_t _size1278; + ::apache::thrift::protocol::TType _etype1281; + xfer += iprot->readListBegin(_etype1281, _size1278); + this->success.resize(_size1278); + uint32_t _i1282; + for (_i1282 = 0; _i1282 < _size1278; ++_i1282) { - xfer += this->success[_i1258].read(iprot); + xfer += this->success[_i1282].read(iprot); } xfer += iprot->readListEnd(); } @@ -4645,10 +4645,10 @@ uint32_t ThriftHiveMetastore_get_schema_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1259; - for (_iter1259 = this->success.begin(); _iter1259 != this->success.end(); ++_iter1259) + std::vector ::const_iterator _iter1283; + for (_iter1283 = this->success.begin(); _iter1283 != this->success.end(); ++_iter1283) { - xfer += (*_iter1259).write(oprot); + xfer += (*_iter1283).write(oprot); } xfer += oprot->writeListEnd(); } @@ -4701,14 +4701,14 @@ uint32_t ThriftHiveMetastore_get_schema_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1260; - ::apache::thrift::protocol::TType _etype1263; - xfer += iprot->readListBegin(_etype1263, _size1260); - (*(this->success)).resize(_size1260); - uint32_t _i1264; - for (_i1264 = 0; _i1264 < _size1260; ++_i1264) + uint32_t _size1284; + ::apache::thrift::protocol::TType _etype1287; + xfer += iprot->readListBegin(_etype1287, _size1284); + (*(this->success)).resize(_size1284); + uint32_t _i1288; + for (_i1288 = 0; _i1288 < _size1284; ++_i1288) { - xfer += (*(this->success))[_i1264].read(iprot); + xfer += (*(this->success))[_i1288].read(iprot); } xfer += iprot->readListEnd(); } @@ -4894,14 +4894,14 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_result::read(:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1265; - ::apache::thrift::protocol::TType _etype1268; - xfer += iprot->readListBegin(_etype1268, _size1265); - this->success.resize(_size1265); - uint32_t _i1269; - for (_i1269 = 0; _i1269 < _size1265; ++_i1269) + uint32_t _size1289; + ::apache::thrift::protocol::TType _etype1292; + xfer += iprot->readListBegin(_etype1292, _size1289); + this->success.resize(_size1289); + uint32_t _i1293; + for (_i1293 = 0; _i1293 < _size1289; ++_i1293) { - xfer += this->success[_i1269].read(iprot); + xfer += this->success[_i1293].read(iprot); } xfer += iprot->readListEnd(); } @@ -4956,10 +4956,10 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_result::write(: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1270; - for (_iter1270 = this->success.begin(); _iter1270 != this->success.end(); ++_iter1270) + std::vector ::const_iterator _iter1294; + for (_iter1294 = this->success.begin(); _iter1294 != this->success.end(); ++_iter1294) { - xfer += (*_iter1270).write(oprot); + xfer += (*_iter1294).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5012,14 +5012,14 @@ uint32_t ThriftHiveMetastore_get_schema_with_environment_context_presult::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1271; - ::apache::thrift::protocol::TType _etype1274; - xfer += iprot->readListBegin(_etype1274, _size1271); - (*(this->success)).resize(_size1271); - uint32_t _i1275; - for (_i1275 = 0; _i1275 < _size1271; ++_i1275) + uint32_t _size1295; + ::apache::thrift::protocol::TType _etype1298; + xfer += iprot->readListBegin(_etype1298, _size1295); + (*(this->success)).resize(_size1295); + uint32_t _i1299; + for (_i1299 = 0; _i1299 < _size1295; ++_i1299) { - xfer += (*(this->success))[_i1275].read(iprot); + xfer += (*(this->success))[_i1299].read(iprot); } xfer += iprot->readListEnd(); } @@ -5612,14 +5612,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->primaryKeys.clear(); - uint32_t _size1276; - ::apache::thrift::protocol::TType _etype1279; - xfer += iprot->readListBegin(_etype1279, _size1276); - this->primaryKeys.resize(_size1276); - uint32_t _i1280; - for (_i1280 = 0; _i1280 < _size1276; ++_i1280) + uint32_t _size1300; + ::apache::thrift::protocol::TType _etype1303; + xfer += iprot->readListBegin(_etype1303, _size1300); + this->primaryKeys.resize(_size1300); + uint32_t _i1304; + for (_i1304 = 0; _i1304 < _size1300; ++_i1304) { - xfer += this->primaryKeys[_i1280].read(iprot); + xfer += this->primaryKeys[_i1304].read(iprot); } xfer += iprot->readListEnd(); } @@ -5632,14 +5632,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->foreignKeys.clear(); - uint32_t _size1281; - ::apache::thrift::protocol::TType _etype1284; - xfer += iprot->readListBegin(_etype1284, _size1281); - this->foreignKeys.resize(_size1281); - uint32_t _i1285; - for (_i1285 = 0; _i1285 < _size1281; ++_i1285) + uint32_t _size1305; + ::apache::thrift::protocol::TType _etype1308; + xfer += iprot->readListBegin(_etype1308, _size1305); + this->foreignKeys.resize(_size1305); + uint32_t _i1309; + for (_i1309 = 0; _i1309 < _size1305; ++_i1309) { - xfer += this->foreignKeys[_i1285].read(iprot); + xfer += this->foreignKeys[_i1309].read(iprot); } xfer += iprot->readListEnd(); } @@ -5652,14 +5652,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->uniqueConstraints.clear(); - uint32_t _size1286; - ::apache::thrift::protocol::TType _etype1289; - xfer += iprot->readListBegin(_etype1289, _size1286); - this->uniqueConstraints.resize(_size1286); - uint32_t _i1290; - for (_i1290 = 0; _i1290 < _size1286; ++_i1290) + uint32_t _size1310; + ::apache::thrift::protocol::TType _etype1313; + xfer += iprot->readListBegin(_etype1313, _size1310); + this->uniqueConstraints.resize(_size1310); + uint32_t _i1314; + for (_i1314 = 0; _i1314 < _size1310; ++_i1314) { - xfer += this->uniqueConstraints[_i1290].read(iprot); + xfer += this->uniqueConstraints[_i1314].read(iprot); } xfer += iprot->readListEnd(); } @@ -5672,14 +5672,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->notNullConstraints.clear(); - uint32_t _size1291; - ::apache::thrift::protocol::TType _etype1294; - xfer += iprot->readListBegin(_etype1294, _size1291); - this->notNullConstraints.resize(_size1291); - uint32_t _i1295; - for (_i1295 = 0; _i1295 < _size1291; ++_i1295) + uint32_t _size1315; + ::apache::thrift::protocol::TType _etype1318; + xfer += iprot->readListBegin(_etype1318, _size1315); + this->notNullConstraints.resize(_size1315); + uint32_t _i1319; + for (_i1319 = 0; _i1319 < _size1315; ++_i1319) { - xfer += this->notNullConstraints[_i1295].read(iprot); + xfer += this->notNullConstraints[_i1319].read(iprot); } xfer += iprot->readListEnd(); } @@ -5692,14 +5692,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->defaultConstraints.clear(); - uint32_t _size1296; - ::apache::thrift::protocol::TType _etype1299; - xfer += iprot->readListBegin(_etype1299, _size1296); - this->defaultConstraints.resize(_size1296); - uint32_t _i1300; - for (_i1300 = 0; _i1300 < _size1296; ++_i1300) + uint32_t _size1320; + ::apache::thrift::protocol::TType _etype1323; + xfer += iprot->readListBegin(_etype1323, _size1320); + this->defaultConstraints.resize(_size1320); + uint32_t _i1324; + for (_i1324 = 0; _i1324 < _size1320; ++_i1324) { - xfer += this->defaultConstraints[_i1300].read(iprot); + xfer += this->defaultConstraints[_i1324].read(iprot); } xfer += iprot->readListEnd(); } @@ -5712,14 +5712,14 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->checkConstraints.clear(); - uint32_t _size1301; - ::apache::thrift::protocol::TType _etype1304; - xfer += iprot->readListBegin(_etype1304, _size1301); - this->checkConstraints.resize(_size1301); - uint32_t _i1305; - for (_i1305 = 0; _i1305 < _size1301; ++_i1305) + uint32_t _size1325; + ::apache::thrift::protocol::TType _etype1328; + xfer += iprot->readListBegin(_etype1328, _size1325); + this->checkConstraints.resize(_size1325); + uint32_t _i1329; + for (_i1329 = 0; _i1329 < _size1325; ++_i1329) { - xfer += this->checkConstraints[_i1305].read(iprot); + xfer += this->checkConstraints[_i1329].read(iprot); } xfer += iprot->readListEnd(); } @@ -5752,10 +5752,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->primaryKeys.size())); - std::vector ::const_iterator _iter1306; - for (_iter1306 = this->primaryKeys.begin(); _iter1306 != this->primaryKeys.end(); ++_iter1306) + std::vector ::const_iterator _iter1330; + for (_iter1330 = this->primaryKeys.begin(); _iter1330 != this->primaryKeys.end(); ++_iter1330) { - xfer += (*_iter1306).write(oprot); + xfer += (*_iter1330).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5764,10 +5764,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->foreignKeys.size())); - std::vector ::const_iterator _iter1307; - for (_iter1307 = this->foreignKeys.begin(); _iter1307 != this->foreignKeys.end(); ++_iter1307) + std::vector ::const_iterator _iter1331; + for (_iter1331 = this->foreignKeys.begin(); _iter1331 != this->foreignKeys.end(); ++_iter1331) { - xfer += (*_iter1307).write(oprot); + xfer += (*_iter1331).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5776,10 +5776,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->uniqueConstraints.size())); - std::vector ::const_iterator _iter1308; - for (_iter1308 = this->uniqueConstraints.begin(); _iter1308 != this->uniqueConstraints.end(); ++_iter1308) + std::vector ::const_iterator _iter1332; + for (_iter1332 = this->uniqueConstraints.begin(); _iter1332 != this->uniqueConstraints.end(); ++_iter1332) { - xfer += (*_iter1308).write(oprot); + xfer += (*_iter1332).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5788,10 +5788,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->notNullConstraints.size())); - std::vector ::const_iterator _iter1309; - for (_iter1309 = this->notNullConstraints.begin(); _iter1309 != this->notNullConstraints.end(); ++_iter1309) + std::vector ::const_iterator _iter1333; + for (_iter1333 = this->notNullConstraints.begin(); _iter1333 != this->notNullConstraints.end(); ++_iter1333) { - xfer += (*_iter1309).write(oprot); + xfer += (*_iter1333).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5800,10 +5800,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("defaultConstraints", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->defaultConstraints.size())); - std::vector ::const_iterator _iter1310; - for (_iter1310 = this->defaultConstraints.begin(); _iter1310 != this->defaultConstraints.end(); ++_iter1310) + std::vector ::const_iterator _iter1334; + for (_iter1334 = this->defaultConstraints.begin(); _iter1334 != this->defaultConstraints.end(); ++_iter1334) { - xfer += (*_iter1310).write(oprot); + xfer += (*_iter1334).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5812,10 +5812,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_args::write(::apache: xfer += oprot->writeFieldBegin("checkConstraints", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->checkConstraints.size())); - std::vector ::const_iterator _iter1311; - for (_iter1311 = this->checkConstraints.begin(); _iter1311 != this->checkConstraints.end(); ++_iter1311) + std::vector ::const_iterator _iter1335; + for (_iter1335 = this->checkConstraints.begin(); _iter1335 != this->checkConstraints.end(); ++_iter1335) { - xfer += (*_iter1311).write(oprot); + xfer += (*_iter1335).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5843,10 +5843,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("primaryKeys", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->primaryKeys)).size())); - std::vector ::const_iterator _iter1312; - for (_iter1312 = (*(this->primaryKeys)).begin(); _iter1312 != (*(this->primaryKeys)).end(); ++_iter1312) + std::vector ::const_iterator _iter1336; + for (_iter1336 = (*(this->primaryKeys)).begin(); _iter1336 != (*(this->primaryKeys)).end(); ++_iter1336) { - xfer += (*_iter1312).write(oprot); + xfer += (*_iter1336).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5855,10 +5855,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("foreignKeys", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->foreignKeys)).size())); - std::vector ::const_iterator _iter1313; - for (_iter1313 = (*(this->foreignKeys)).begin(); _iter1313 != (*(this->foreignKeys)).end(); ++_iter1313) + std::vector ::const_iterator _iter1337; + for (_iter1337 = (*(this->foreignKeys)).begin(); _iter1337 != (*(this->foreignKeys)).end(); ++_iter1337) { - xfer += (*_iter1313).write(oprot); + xfer += (*_iter1337).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5867,10 +5867,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("uniqueConstraints", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->uniqueConstraints)).size())); - std::vector ::const_iterator _iter1314; - for (_iter1314 = (*(this->uniqueConstraints)).begin(); _iter1314 != (*(this->uniqueConstraints)).end(); ++_iter1314) + std::vector ::const_iterator _iter1338; + for (_iter1338 = (*(this->uniqueConstraints)).begin(); _iter1338 != (*(this->uniqueConstraints)).end(); ++_iter1338) { - xfer += (*_iter1314).write(oprot); + xfer += (*_iter1338).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5879,10 +5879,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("notNullConstraints", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->notNullConstraints)).size())); - std::vector ::const_iterator _iter1315; - for (_iter1315 = (*(this->notNullConstraints)).begin(); _iter1315 != (*(this->notNullConstraints)).end(); ++_iter1315) + std::vector ::const_iterator _iter1339; + for (_iter1339 = (*(this->notNullConstraints)).begin(); _iter1339 != (*(this->notNullConstraints)).end(); ++_iter1339) { - xfer += (*_iter1315).write(oprot); + xfer += (*_iter1339).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5891,10 +5891,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("defaultConstraints", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->defaultConstraints)).size())); - std::vector ::const_iterator _iter1316; - for (_iter1316 = (*(this->defaultConstraints)).begin(); _iter1316 != (*(this->defaultConstraints)).end(); ++_iter1316) + std::vector ::const_iterator _iter1340; + for (_iter1340 = (*(this->defaultConstraints)).begin(); _iter1340 != (*(this->defaultConstraints)).end(); ++_iter1340) { - xfer += (*_iter1316).write(oprot); + xfer += (*_iter1340).write(oprot); } xfer += oprot->writeListEnd(); } @@ -5903,10 +5903,10 @@ uint32_t ThriftHiveMetastore_create_table_with_constraints_pargs::write(::apache xfer += oprot->writeFieldBegin("checkConstraints", ::apache::thrift::protocol::T_LIST, 7); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->checkConstraints)).size())); - std::vector ::const_iterator _iter1317; - for (_iter1317 = (*(this->checkConstraints)).begin(); _iter1317 != (*(this->checkConstraints)).end(); ++_iter1317) + std::vector ::const_iterator _iter1341; + for (_iter1341 = (*(this->checkConstraints)).begin(); _iter1341 != (*(this->checkConstraints)).end(); ++_iter1341) { - xfer += (*_iter1317).write(oprot); + xfer += (*_iter1341).write(oprot); } xfer += oprot->writeListEnd(); } @@ -8074,14 +8074,14 @@ uint32_t ThriftHiveMetastore_truncate_table_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partNames.clear(); - uint32_t _size1318; - ::apache::thrift::protocol::TType _etype1321; - xfer += iprot->readListBegin(_etype1321, _size1318); - this->partNames.resize(_size1318); - uint32_t _i1322; - for (_i1322 = 0; _i1322 < _size1318; ++_i1322) + uint32_t _size1342; + ::apache::thrift::protocol::TType _etype1345; + xfer += iprot->readListBegin(_etype1345, _size1342); + this->partNames.resize(_size1342); + uint32_t _i1346; + for (_i1346 = 0; _i1346 < _size1342; ++_i1346) { - xfer += iprot->readString(this->partNames[_i1322]); + xfer += iprot->readString(this->partNames[_i1346]); } xfer += iprot->readListEnd(); } @@ -8118,10 +8118,10 @@ uint32_t ThriftHiveMetastore_truncate_table_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partNames.size())); - std::vector ::const_iterator _iter1323; - for (_iter1323 = this->partNames.begin(); _iter1323 != this->partNames.end(); ++_iter1323) + std::vector ::const_iterator _iter1347; + for (_iter1347 = this->partNames.begin(); _iter1347 != this->partNames.end(); ++_iter1347) { - xfer += oprot->writeString((*_iter1323)); + xfer += oprot->writeString((*_iter1347)); } xfer += oprot->writeListEnd(); } @@ -8153,10 +8153,10 @@ uint32_t ThriftHiveMetastore_truncate_table_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->partNames)).size())); - std::vector ::const_iterator _iter1324; - for (_iter1324 = (*(this->partNames)).begin(); _iter1324 != (*(this->partNames)).end(); ++_iter1324) + std::vector ::const_iterator _iter1348; + for (_iter1348 = (*(this->partNames)).begin(); _iter1348 != (*(this->partNames)).end(); ++_iter1348) { - xfer += oprot->writeString((*_iter1324)); + xfer += oprot->writeString((*_iter1348)); } xfer += oprot->writeListEnd(); } @@ -8400,14 +8400,14 @@ uint32_t ThriftHiveMetastore_get_tables_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1325; - ::apache::thrift::protocol::TType _etype1328; - xfer += iprot->readListBegin(_etype1328, _size1325); - this->success.resize(_size1325); - uint32_t _i1329; - for (_i1329 = 0; _i1329 < _size1325; ++_i1329) + uint32_t _size1349; + ::apache::thrift::protocol::TType _etype1352; + xfer += iprot->readListBegin(_etype1352, _size1349); + this->success.resize(_size1349); + uint32_t _i1353; + for (_i1353 = 0; _i1353 < _size1349; ++_i1353) { - xfer += iprot->readString(this->success[_i1329]); + xfer += iprot->readString(this->success[_i1353]); } xfer += iprot->readListEnd(); } @@ -8446,10 +8446,10 @@ uint32_t ThriftHiveMetastore_get_tables_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1330; - for (_iter1330 = this->success.begin(); _iter1330 != this->success.end(); ++_iter1330) + std::vector ::const_iterator _iter1354; + for (_iter1354 = this->success.begin(); _iter1354 != this->success.end(); ++_iter1354) { - xfer += oprot->writeString((*_iter1330)); + xfer += oprot->writeString((*_iter1354)); } xfer += oprot->writeListEnd(); } @@ -8494,14 +8494,14 @@ uint32_t ThriftHiveMetastore_get_tables_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1331; - ::apache::thrift::protocol::TType _etype1334; - xfer += iprot->readListBegin(_etype1334, _size1331); - (*(this->success)).resize(_size1331); - uint32_t _i1335; - for (_i1335 = 0; _i1335 < _size1331; ++_i1335) + uint32_t _size1355; + ::apache::thrift::protocol::TType _etype1358; + xfer += iprot->readListBegin(_etype1358, _size1355); + (*(this->success)).resize(_size1355); + uint32_t _i1359; + for (_i1359 = 0; _i1359 < _size1355; ++_i1359) { - xfer += iprot->readString((*(this->success))[_i1335]); + xfer += iprot->readString((*(this->success))[_i1359]); } xfer += iprot->readListEnd(); } @@ -8671,14 +8671,14 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_result::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1336; - ::apache::thrift::protocol::TType _etype1339; - xfer += iprot->readListBegin(_etype1339, _size1336); - this->success.resize(_size1336); - uint32_t _i1340; - for (_i1340 = 0; _i1340 < _size1336; ++_i1340) + uint32_t _size1360; + ::apache::thrift::protocol::TType _etype1363; + xfer += iprot->readListBegin(_etype1363, _size1360); + this->success.resize(_size1360); + uint32_t _i1364; + for (_i1364 = 0; _i1364 < _size1360; ++_i1364) { - xfer += iprot->readString(this->success[_i1340]); + xfer += iprot->readString(this->success[_i1364]); } xfer += iprot->readListEnd(); } @@ -8717,10 +8717,10 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_result::write(::apache::thrift:: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1341; - for (_iter1341 = this->success.begin(); _iter1341 != this->success.end(); ++_iter1341) + std::vector ::const_iterator _iter1365; + for (_iter1365 = this->success.begin(); _iter1365 != this->success.end(); ++_iter1365) { - xfer += oprot->writeString((*_iter1341)); + xfer += oprot->writeString((*_iter1365)); } xfer += oprot->writeListEnd(); } @@ -8765,14 +8765,14 @@ uint32_t ThriftHiveMetastore_get_tables_by_type_presult::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1342; - ::apache::thrift::protocol::TType _etype1345; - xfer += iprot->readListBegin(_etype1345, _size1342); - (*(this->success)).resize(_size1342); - uint32_t _i1346; - for (_i1346 = 0; _i1346 < _size1342; ++_i1346) + uint32_t _size1366; + ::apache::thrift::protocol::TType _etype1369; + xfer += iprot->readListBegin(_etype1369, _size1366); + (*(this->success)).resize(_size1366); + uint32_t _i1370; + for (_i1370 = 0; _i1370 < _size1366; ++_i1370) { - xfer += iprot->readString((*(this->success))[_i1346]); + xfer += iprot->readString((*(this->success))[_i1370]); } xfer += iprot->readListEnd(); } @@ -8910,14 +8910,14 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_result::read(: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1347; - ::apache::thrift::protocol::TType _etype1350; - xfer += iprot->readListBegin(_etype1350, _size1347); - this->success.resize(_size1347); - uint32_t _i1351; - for (_i1351 = 0; _i1351 < _size1347; ++_i1351) + uint32_t _size1371; + ::apache::thrift::protocol::TType _etype1374; + xfer += iprot->readListBegin(_etype1374, _size1371); + this->success.resize(_size1371); + uint32_t _i1375; + for (_i1375 = 0; _i1375 < _size1371; ++_i1375) { - xfer += iprot->readString(this->success[_i1351]); + xfer += iprot->readString(this->success[_i1375]); } xfer += iprot->readListEnd(); } @@ -8956,10 +8956,10 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_result::write( xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1352; - for (_iter1352 = this->success.begin(); _iter1352 != this->success.end(); ++_iter1352) + std::vector ::const_iterator _iter1376; + for (_iter1376 = this->success.begin(); _iter1376 != this->success.end(); ++_iter1376) { - xfer += oprot->writeString((*_iter1352)); + xfer += oprot->writeString((*_iter1376)); } xfer += oprot->writeListEnd(); } @@ -9004,14 +9004,14 @@ uint32_t ThriftHiveMetastore_get_materialized_views_for_rewriting_presult::read( if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1353; - ::apache::thrift::protocol::TType _etype1356; - xfer += iprot->readListBegin(_etype1356, _size1353); - (*(this->success)).resize(_size1353); - uint32_t _i1357; - for (_i1357 = 0; _i1357 < _size1353; ++_i1357) + uint32_t _size1377; + ::apache::thrift::protocol::TType _etype1380; + xfer += iprot->readListBegin(_etype1380, _size1377); + (*(this->success)).resize(_size1377); + uint32_t _i1381; + for (_i1381 = 0; _i1381 < _size1377; ++_i1381) { - xfer += iprot->readString((*(this->success))[_i1357]); + xfer += iprot->readString((*(this->success))[_i1381]); } xfer += iprot->readListEnd(); } @@ -9086,14 +9086,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tbl_types.clear(); - uint32_t _size1358; - ::apache::thrift::protocol::TType _etype1361; - xfer += iprot->readListBegin(_etype1361, _size1358); - this->tbl_types.resize(_size1358); - uint32_t _i1362; - for (_i1362 = 0; _i1362 < _size1358; ++_i1362) + uint32_t _size1382; + ::apache::thrift::protocol::TType _etype1385; + xfer += iprot->readListBegin(_etype1385, _size1382); + this->tbl_types.resize(_size1382); + uint32_t _i1386; + for (_i1386 = 0; _i1386 < _size1382; ++_i1386) { - xfer += iprot->readString(this->tbl_types[_i1362]); + xfer += iprot->readString(this->tbl_types[_i1386]); } xfer += iprot->readListEnd(); } @@ -9130,10 +9130,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("tbl_types", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tbl_types.size())); - std::vector ::const_iterator _iter1363; - for (_iter1363 = this->tbl_types.begin(); _iter1363 != this->tbl_types.end(); ++_iter1363) + std::vector ::const_iterator _iter1387; + for (_iter1387 = this->tbl_types.begin(); _iter1387 != this->tbl_types.end(); ++_iter1387) { - xfer += oprot->writeString((*_iter1363)); + xfer += oprot->writeString((*_iter1387)); } xfer += oprot->writeListEnd(); } @@ -9165,10 +9165,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("tbl_types", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->tbl_types)).size())); - std::vector ::const_iterator _iter1364; - for (_iter1364 = (*(this->tbl_types)).begin(); _iter1364 != (*(this->tbl_types)).end(); ++_iter1364) + std::vector ::const_iterator _iter1388; + for (_iter1388 = (*(this->tbl_types)).begin(); _iter1388 != (*(this->tbl_types)).end(); ++_iter1388) { - xfer += oprot->writeString((*_iter1364)); + xfer += oprot->writeString((*_iter1388)); } xfer += oprot->writeListEnd(); } @@ -9209,14 +9209,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1365; - ::apache::thrift::protocol::TType _etype1368; - xfer += iprot->readListBegin(_etype1368, _size1365); - this->success.resize(_size1365); - uint32_t _i1369; - for (_i1369 = 0; _i1369 < _size1365; ++_i1369) + uint32_t _size1389; + ::apache::thrift::protocol::TType _etype1392; + xfer += iprot->readListBegin(_etype1392, _size1389); + this->success.resize(_size1389); + uint32_t _i1393; + for (_i1393 = 0; _i1393 < _size1389; ++_i1393) { - xfer += this->success[_i1369].read(iprot); + xfer += this->success[_i1393].read(iprot); } xfer += iprot->readListEnd(); } @@ -9255,10 +9255,10 @@ uint32_t ThriftHiveMetastore_get_table_meta_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1370; - for (_iter1370 = this->success.begin(); _iter1370 != this->success.end(); ++_iter1370) + std::vector ::const_iterator _iter1394; + for (_iter1394 = this->success.begin(); _iter1394 != this->success.end(); ++_iter1394) { - xfer += (*_iter1370).write(oprot); + xfer += (*_iter1394).write(oprot); } xfer += oprot->writeListEnd(); } @@ -9303,14 +9303,14 @@ uint32_t ThriftHiveMetastore_get_table_meta_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1371; - ::apache::thrift::protocol::TType _etype1374; - xfer += iprot->readListBegin(_etype1374, _size1371); - (*(this->success)).resize(_size1371); - uint32_t _i1375; - for (_i1375 = 0; _i1375 < _size1371; ++_i1375) + uint32_t _size1395; + ::apache::thrift::protocol::TType _etype1398; + xfer += iprot->readListBegin(_etype1398, _size1395); + (*(this->success)).resize(_size1395); + uint32_t _i1399; + for (_i1399 = 0; _i1399 < _size1395; ++_i1399) { - xfer += (*(this->success))[_i1375].read(iprot); + xfer += (*(this->success))[_i1399].read(iprot); } xfer += iprot->readListEnd(); } @@ -9448,14 +9448,14 @@ uint32_t ThriftHiveMetastore_get_all_tables_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1376; - ::apache::thrift::protocol::TType _etype1379; - xfer += iprot->readListBegin(_etype1379, _size1376); - this->success.resize(_size1376); - uint32_t _i1380; - for (_i1380 = 0; _i1380 < _size1376; ++_i1380) + uint32_t _size1400; + ::apache::thrift::protocol::TType _etype1403; + xfer += iprot->readListBegin(_etype1403, _size1400); + this->success.resize(_size1400); + uint32_t _i1404; + for (_i1404 = 0; _i1404 < _size1400; ++_i1404) { - xfer += iprot->readString(this->success[_i1380]); + xfer += iprot->readString(this->success[_i1404]); } xfer += iprot->readListEnd(); } @@ -9494,10 +9494,10 @@ uint32_t ThriftHiveMetastore_get_all_tables_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1381; - for (_iter1381 = this->success.begin(); _iter1381 != this->success.end(); ++_iter1381) + std::vector ::const_iterator _iter1405; + for (_iter1405 = this->success.begin(); _iter1405 != this->success.end(); ++_iter1405) { - xfer += oprot->writeString((*_iter1381)); + xfer += oprot->writeString((*_iter1405)); } xfer += oprot->writeListEnd(); } @@ -9542,14 +9542,14 @@ uint32_t ThriftHiveMetastore_get_all_tables_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1382; - ::apache::thrift::protocol::TType _etype1385; - xfer += iprot->readListBegin(_etype1385, _size1382); - (*(this->success)).resize(_size1382); - uint32_t _i1386; - for (_i1386 = 0; _i1386 < _size1382; ++_i1386) + uint32_t _size1406; + ::apache::thrift::protocol::TType _etype1409; + xfer += iprot->readListBegin(_etype1409, _size1406); + (*(this->success)).resize(_size1406); + uint32_t _i1410; + for (_i1410 = 0; _i1410 < _size1406; ++_i1410) { - xfer += iprot->readString((*(this->success))[_i1386]); + xfer += iprot->readString((*(this->success))[_i1410]); } xfer += iprot->readListEnd(); } @@ -9859,14 +9859,14 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_args::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tbl_names.clear(); - uint32_t _size1387; - ::apache::thrift::protocol::TType _etype1390; - xfer += iprot->readListBegin(_etype1390, _size1387); - this->tbl_names.resize(_size1387); - uint32_t _i1391; - for (_i1391 = 0; _i1391 < _size1387; ++_i1391) + uint32_t _size1411; + ::apache::thrift::protocol::TType _etype1414; + xfer += iprot->readListBegin(_etype1414, _size1411); + this->tbl_names.resize(_size1411); + uint32_t _i1415; + for (_i1415 = 0; _i1415 < _size1411; ++_i1415) { - xfer += iprot->readString(this->tbl_names[_i1391]); + xfer += iprot->readString(this->tbl_names[_i1415]); } xfer += iprot->readListEnd(); } @@ -9899,10 +9899,10 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_args::write(::apache::thr xfer += oprot->writeFieldBegin("tbl_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tbl_names.size())); - std::vector ::const_iterator _iter1392; - for (_iter1392 = this->tbl_names.begin(); _iter1392 != this->tbl_names.end(); ++_iter1392) + std::vector ::const_iterator _iter1416; + for (_iter1416 = this->tbl_names.begin(); _iter1416 != this->tbl_names.end(); ++_iter1416) { - xfer += oprot->writeString((*_iter1392)); + xfer += oprot->writeString((*_iter1416)); } xfer += oprot->writeListEnd(); } @@ -9930,10 +9930,10 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_pargs::write(::apache::th xfer += oprot->writeFieldBegin("tbl_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->tbl_names)).size())); - std::vector ::const_iterator _iter1393; - for (_iter1393 = (*(this->tbl_names)).begin(); _iter1393 != (*(this->tbl_names)).end(); ++_iter1393) + std::vector ::const_iterator _iter1417; + for (_iter1417 = (*(this->tbl_names)).begin(); _iter1417 != (*(this->tbl_names)).end(); ++_iter1417) { - xfer += oprot->writeString((*_iter1393)); + xfer += oprot->writeString((*_iter1417)); } xfer += oprot->writeListEnd(); } @@ -9974,14 +9974,14 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1394; - ::apache::thrift::protocol::TType _etype1397; - xfer += iprot->readListBegin(_etype1397, _size1394); - this->success.resize(_size1394); - uint32_t _i1398; - for (_i1398 = 0; _i1398 < _size1394; ++_i1398) + uint32_t _size1418; + ::apache::thrift::protocol::TType _etype1421; + xfer += iprot->readListBegin(_etype1421, _size1418); + this->success.resize(_size1418); + uint32_t _i1422; + for (_i1422 = 0; _i1422 < _size1418; ++_i1422) { - xfer += this->success[_i1398].read(iprot); + xfer += this->success[_i1422].read(iprot); } xfer += iprot->readListEnd(); } @@ -10012,10 +10012,10 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1399; - for (_iter1399 = this->success.begin(); _iter1399 != this->success.end(); ++_iter1399) + std::vector
::const_iterator _iter1423; + for (_iter1423 = this->success.begin(); _iter1423 != this->success.end(); ++_iter1423) { - xfer += (*_iter1399).write(oprot); + xfer += (*_iter1423).write(oprot); } xfer += oprot->writeListEnd(); } @@ -10056,14 +10056,14 @@ uint32_t ThriftHiveMetastore_get_table_objects_by_name_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1400; - ::apache::thrift::protocol::TType _etype1403; - xfer += iprot->readListBegin(_etype1403, _size1400); - (*(this->success)).resize(_size1400); - uint32_t _i1404; - for (_i1404 = 0; _i1404 < _size1400; ++_i1404) + uint32_t _size1424; + ::apache::thrift::protocol::TType _etype1427; + xfer += iprot->readListBegin(_etype1427, _size1424); + (*(this->success)).resize(_size1424); + uint32_t _i1428; + for (_i1428 = 0; _i1428 < _size1424; ++_i1428) { - xfer += (*(this->success))[_i1404].read(iprot); + xfer += (*(this->success))[_i1428].read(iprot); } xfer += iprot->readListEnd(); } @@ -11237,14 +11237,14 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1405; - ::apache::thrift::protocol::TType _etype1408; - xfer += iprot->readListBegin(_etype1408, _size1405); - this->success.resize(_size1405); - uint32_t _i1409; - for (_i1409 = 0; _i1409 < _size1405; ++_i1409) + uint32_t _size1429; + ::apache::thrift::protocol::TType _etype1432; + xfer += iprot->readListBegin(_etype1432, _size1429); + this->success.resize(_size1429); + uint32_t _i1433; + for (_i1433 = 0; _i1433 < _size1429; ++_i1433) { - xfer += iprot->readString(this->success[_i1409]); + xfer += iprot->readString(this->success[_i1433]); } xfer += iprot->readListEnd(); } @@ -11299,10 +11299,10 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1410; - for (_iter1410 = this->success.begin(); _iter1410 != this->success.end(); ++_iter1410) + std::vector ::const_iterator _iter1434; + for (_iter1434 = this->success.begin(); _iter1434 != this->success.end(); ++_iter1434) { - xfer += oprot->writeString((*_iter1410)); + xfer += oprot->writeString((*_iter1434)); } xfer += oprot->writeListEnd(); } @@ -11355,14 +11355,14 @@ uint32_t ThriftHiveMetastore_get_table_names_by_filter_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1411; - ::apache::thrift::protocol::TType _etype1414; - xfer += iprot->readListBegin(_etype1414, _size1411); - (*(this->success)).resize(_size1411); - uint32_t _i1415; - for (_i1415 = 0; _i1415 < _size1411; ++_i1415) + uint32_t _size1435; + ::apache::thrift::protocol::TType _etype1438; + xfer += iprot->readListBegin(_etype1438, _size1435); + (*(this->success)).resize(_size1435); + uint32_t _i1439; + for (_i1439 = 0; _i1439 < _size1435; ++_i1439) { - xfer += iprot->readString((*(this->success))[_i1415]); + xfer += iprot->readString((*(this->success))[_i1439]); } xfer += iprot->readListEnd(); } @@ -12696,14 +12696,14 @@ uint32_t ThriftHiveMetastore_add_partitions_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1416; - ::apache::thrift::protocol::TType _etype1419; - xfer += iprot->readListBegin(_etype1419, _size1416); - this->new_parts.resize(_size1416); - uint32_t _i1420; - for (_i1420 = 0; _i1420 < _size1416; ++_i1420) + uint32_t _size1440; + ::apache::thrift::protocol::TType _etype1443; + xfer += iprot->readListBegin(_etype1443, _size1440); + this->new_parts.resize(_size1440); + uint32_t _i1444; + for (_i1444 = 0; _i1444 < _size1440; ++_i1444) { - xfer += this->new_parts[_i1420].read(iprot); + xfer += this->new_parts[_i1444].read(iprot); } xfer += iprot->readListEnd(); } @@ -12732,10 +12732,10 @@ uint32_t ThriftHiveMetastore_add_partitions_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1421; - for (_iter1421 = this->new_parts.begin(); _iter1421 != this->new_parts.end(); ++_iter1421) + std::vector ::const_iterator _iter1445; + for (_iter1445 = this->new_parts.begin(); _iter1445 != this->new_parts.end(); ++_iter1445) { - xfer += (*_iter1421).write(oprot); + xfer += (*_iter1445).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12759,10 +12759,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1422; - for (_iter1422 = (*(this->new_parts)).begin(); _iter1422 != (*(this->new_parts)).end(); ++_iter1422) + std::vector ::const_iterator _iter1446; + for (_iter1446 = (*(this->new_parts)).begin(); _iter1446 != (*(this->new_parts)).end(); ++_iter1446) { - xfer += (*_iter1422).write(oprot); + xfer += (*_iter1446).write(oprot); } xfer += oprot->writeListEnd(); } @@ -12971,14 +12971,14 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_args::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1423; - ::apache::thrift::protocol::TType _etype1426; - xfer += iprot->readListBegin(_etype1426, _size1423); - this->new_parts.resize(_size1423); - uint32_t _i1427; - for (_i1427 = 0; _i1427 < _size1423; ++_i1427) + uint32_t _size1447; + ::apache::thrift::protocol::TType _etype1450; + xfer += iprot->readListBegin(_etype1450, _size1447); + this->new_parts.resize(_size1447); + uint32_t _i1451; + for (_i1451 = 0; _i1451 < _size1447; ++_i1451) { - xfer += this->new_parts[_i1427].read(iprot); + xfer += this->new_parts[_i1451].read(iprot); } xfer += iprot->readListEnd(); } @@ -13007,10 +13007,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_args::write(::apache::thrift:: xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1428; - for (_iter1428 = this->new_parts.begin(); _iter1428 != this->new_parts.end(); ++_iter1428) + std::vector ::const_iterator _iter1452; + for (_iter1452 = this->new_parts.begin(); _iter1452 != this->new_parts.end(); ++_iter1452) { - xfer += (*_iter1428).write(oprot); + xfer += (*_iter1452).write(oprot); } xfer += oprot->writeListEnd(); } @@ -13034,10 +13034,10 @@ uint32_t ThriftHiveMetastore_add_partitions_pspec_pargs::write(::apache::thrift: xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1429; - for (_iter1429 = (*(this->new_parts)).begin(); _iter1429 != (*(this->new_parts)).end(); ++_iter1429) + std::vector ::const_iterator _iter1453; + for (_iter1453 = (*(this->new_parts)).begin(); _iter1453 != (*(this->new_parts)).end(); ++_iter1453) { - xfer += (*_iter1429).write(oprot); + xfer += (*_iter1453).write(oprot); } xfer += oprot->writeListEnd(); } @@ -13262,14 +13262,14 @@ uint32_t ThriftHiveMetastore_append_partition_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1430; - ::apache::thrift::protocol::TType _etype1433; - xfer += iprot->readListBegin(_etype1433, _size1430); - this->part_vals.resize(_size1430); - uint32_t _i1434; - for (_i1434 = 0; _i1434 < _size1430; ++_i1434) + uint32_t _size1454; + ::apache::thrift::protocol::TType _etype1457; + xfer += iprot->readListBegin(_etype1457, _size1454); + this->part_vals.resize(_size1454); + uint32_t _i1458; + for (_i1458 = 0; _i1458 < _size1454; ++_i1458) { - xfer += iprot->readString(this->part_vals[_i1434]); + xfer += iprot->readString(this->part_vals[_i1458]); } xfer += iprot->readListEnd(); } @@ -13306,10 +13306,10 @@ uint32_t ThriftHiveMetastore_append_partition_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1435; - for (_iter1435 = this->part_vals.begin(); _iter1435 != this->part_vals.end(); ++_iter1435) + std::vector ::const_iterator _iter1459; + for (_iter1459 = this->part_vals.begin(); _iter1459 != this->part_vals.end(); ++_iter1459) { - xfer += oprot->writeString((*_iter1435)); + xfer += oprot->writeString((*_iter1459)); } xfer += oprot->writeListEnd(); } @@ -13341,10 +13341,10 @@ uint32_t ThriftHiveMetastore_append_partition_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1436; - for (_iter1436 = (*(this->part_vals)).begin(); _iter1436 != (*(this->part_vals)).end(); ++_iter1436) + std::vector ::const_iterator _iter1460; + for (_iter1460 = (*(this->part_vals)).begin(); _iter1460 != (*(this->part_vals)).end(); ++_iter1460) { - xfer += oprot->writeString((*_iter1436)); + xfer += oprot->writeString((*_iter1460)); } xfer += oprot->writeListEnd(); } @@ -13816,14 +13816,14 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_args::rea if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1437; - ::apache::thrift::protocol::TType _etype1440; - xfer += iprot->readListBegin(_etype1440, _size1437); - this->part_vals.resize(_size1437); - uint32_t _i1441; - for (_i1441 = 0; _i1441 < _size1437; ++_i1441) + uint32_t _size1461; + ::apache::thrift::protocol::TType _etype1464; + xfer += iprot->readListBegin(_etype1464, _size1461); + this->part_vals.resize(_size1461); + uint32_t _i1465; + for (_i1465 = 0; _i1465 < _size1461; ++_i1465) { - xfer += iprot->readString(this->part_vals[_i1441]); + xfer += iprot->readString(this->part_vals[_i1465]); } xfer += iprot->readListEnd(); } @@ -13868,10 +13868,10 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_args::wri xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1442; - for (_iter1442 = this->part_vals.begin(); _iter1442 != this->part_vals.end(); ++_iter1442) + std::vector ::const_iterator _iter1466; + for (_iter1466 = this->part_vals.begin(); _iter1466 != this->part_vals.end(); ++_iter1466) { - xfer += oprot->writeString((*_iter1442)); + xfer += oprot->writeString((*_iter1466)); } xfer += oprot->writeListEnd(); } @@ -13907,10 +13907,10 @@ uint32_t ThriftHiveMetastore_append_partition_with_environment_context_pargs::wr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1443; - for (_iter1443 = (*(this->part_vals)).begin(); _iter1443 != (*(this->part_vals)).end(); ++_iter1443) + std::vector ::const_iterator _iter1467; + for (_iter1467 = (*(this->part_vals)).begin(); _iter1467 != (*(this->part_vals)).end(); ++_iter1467) { - xfer += oprot->writeString((*_iter1443)); + xfer += oprot->writeString((*_iter1467)); } xfer += oprot->writeListEnd(); } @@ -14713,14 +14713,14 @@ uint32_t ThriftHiveMetastore_drop_partition_args::read(::apache::thrift::protoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1444; - ::apache::thrift::protocol::TType _etype1447; - xfer += iprot->readListBegin(_etype1447, _size1444); - this->part_vals.resize(_size1444); - uint32_t _i1448; - for (_i1448 = 0; _i1448 < _size1444; ++_i1448) + uint32_t _size1468; + ::apache::thrift::protocol::TType _etype1471; + xfer += iprot->readListBegin(_etype1471, _size1468); + this->part_vals.resize(_size1468); + uint32_t _i1472; + for (_i1472 = 0; _i1472 < _size1468; ++_i1472) { - xfer += iprot->readString(this->part_vals[_i1448]); + xfer += iprot->readString(this->part_vals[_i1472]); } xfer += iprot->readListEnd(); } @@ -14765,10 +14765,10 @@ uint32_t ThriftHiveMetastore_drop_partition_args::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1449; - for (_iter1449 = this->part_vals.begin(); _iter1449 != this->part_vals.end(); ++_iter1449) + std::vector ::const_iterator _iter1473; + for (_iter1473 = this->part_vals.begin(); _iter1473 != this->part_vals.end(); ++_iter1473) { - xfer += oprot->writeString((*_iter1449)); + xfer += oprot->writeString((*_iter1473)); } xfer += oprot->writeListEnd(); } @@ -14804,10 +14804,10 @@ uint32_t ThriftHiveMetastore_drop_partition_pargs::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1450; - for (_iter1450 = (*(this->part_vals)).begin(); _iter1450 != (*(this->part_vals)).end(); ++_iter1450) + std::vector ::const_iterator _iter1474; + for (_iter1474 = (*(this->part_vals)).begin(); _iter1474 != (*(this->part_vals)).end(); ++_iter1474) { - xfer += oprot->writeString((*_iter1450)); + xfer += oprot->writeString((*_iter1474)); } xfer += oprot->writeListEnd(); } @@ -15016,14 +15016,14 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_args::read( if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1451; - ::apache::thrift::protocol::TType _etype1454; - xfer += iprot->readListBegin(_etype1454, _size1451); - this->part_vals.resize(_size1451); - uint32_t _i1455; - for (_i1455 = 0; _i1455 < _size1451; ++_i1455) + uint32_t _size1475; + ::apache::thrift::protocol::TType _etype1478; + xfer += iprot->readListBegin(_etype1478, _size1475); + this->part_vals.resize(_size1475); + uint32_t _i1479; + for (_i1479 = 0; _i1479 < _size1475; ++_i1479) { - xfer += iprot->readString(this->part_vals[_i1455]); + xfer += iprot->readString(this->part_vals[_i1479]); } xfer += iprot->readListEnd(); } @@ -15076,10 +15076,10 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_args::write xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1456; - for (_iter1456 = this->part_vals.begin(); _iter1456 != this->part_vals.end(); ++_iter1456) + std::vector ::const_iterator _iter1480; + for (_iter1480 = this->part_vals.begin(); _iter1480 != this->part_vals.end(); ++_iter1480) { - xfer += oprot->writeString((*_iter1456)); + xfer += oprot->writeString((*_iter1480)); } xfer += oprot->writeListEnd(); } @@ -15119,10 +15119,10 @@ uint32_t ThriftHiveMetastore_drop_partition_with_environment_context_pargs::writ xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1457; - for (_iter1457 = (*(this->part_vals)).begin(); _iter1457 != (*(this->part_vals)).end(); ++_iter1457) + std::vector ::const_iterator _iter1481; + for (_iter1481 = (*(this->part_vals)).begin(); _iter1481 != (*(this->part_vals)).end(); ++_iter1481) { - xfer += oprot->writeString((*_iter1457)); + xfer += oprot->writeString((*_iter1481)); } xfer += oprot->writeListEnd(); } @@ -16128,14 +16128,14 @@ uint32_t ThriftHiveMetastore_get_partition_args::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1458; - ::apache::thrift::protocol::TType _etype1461; - xfer += iprot->readListBegin(_etype1461, _size1458); - this->part_vals.resize(_size1458); - uint32_t _i1462; - for (_i1462 = 0; _i1462 < _size1458; ++_i1462) + uint32_t _size1482; + ::apache::thrift::protocol::TType _etype1485; + xfer += iprot->readListBegin(_etype1485, _size1482); + this->part_vals.resize(_size1482); + uint32_t _i1486; + for (_i1486 = 0; _i1486 < _size1482; ++_i1486) { - xfer += iprot->readString(this->part_vals[_i1462]); + xfer += iprot->readString(this->part_vals[_i1486]); } xfer += iprot->readListEnd(); } @@ -16172,10 +16172,10 @@ uint32_t ThriftHiveMetastore_get_partition_args::write(::apache::thrift::protoco xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1463; - for (_iter1463 = this->part_vals.begin(); _iter1463 != this->part_vals.end(); ++_iter1463) + std::vector ::const_iterator _iter1487; + for (_iter1487 = this->part_vals.begin(); _iter1487 != this->part_vals.end(); ++_iter1487) { - xfer += oprot->writeString((*_iter1463)); + xfer += oprot->writeString((*_iter1487)); } xfer += oprot->writeListEnd(); } @@ -16207,10 +16207,10 @@ uint32_t ThriftHiveMetastore_get_partition_pargs::write(::apache::thrift::protoc xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1464; - for (_iter1464 = (*(this->part_vals)).begin(); _iter1464 != (*(this->part_vals)).end(); ++_iter1464) + std::vector ::const_iterator _iter1488; + for (_iter1488 = (*(this->part_vals)).begin(); _iter1488 != (*(this->part_vals)).end(); ++_iter1488) { - xfer += oprot->writeString((*_iter1464)); + xfer += oprot->writeString((*_iter1488)); } xfer += oprot->writeListEnd(); } @@ -16399,17 +16399,17 @@ uint32_t ThriftHiveMetastore_exchange_partition_args::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partitionSpecs.clear(); - uint32_t _size1465; - ::apache::thrift::protocol::TType _ktype1466; - ::apache::thrift::protocol::TType _vtype1467; - xfer += iprot->readMapBegin(_ktype1466, _vtype1467, _size1465); - uint32_t _i1469; - for (_i1469 = 0; _i1469 < _size1465; ++_i1469) + uint32_t _size1489; + ::apache::thrift::protocol::TType _ktype1490; + ::apache::thrift::protocol::TType _vtype1491; + xfer += iprot->readMapBegin(_ktype1490, _vtype1491, _size1489); + uint32_t _i1493; + for (_i1493 = 0; _i1493 < _size1489; ++_i1493) { - std::string _key1470; - xfer += iprot->readString(_key1470); - std::string& _val1471 = this->partitionSpecs[_key1470]; - xfer += iprot->readString(_val1471); + std::string _key1494; + xfer += iprot->readString(_key1494); + std::string& _val1495 = this->partitionSpecs[_key1494]; + xfer += iprot->readString(_val1495); } xfer += iprot->readMapEnd(); } @@ -16470,11 +16470,11 @@ uint32_t ThriftHiveMetastore_exchange_partition_args::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->partitionSpecs.size())); - std::map ::const_iterator _iter1472; - for (_iter1472 = this->partitionSpecs.begin(); _iter1472 != this->partitionSpecs.end(); ++_iter1472) + std::map ::const_iterator _iter1496; + for (_iter1496 = this->partitionSpecs.begin(); _iter1496 != this->partitionSpecs.end(); ++_iter1496) { - xfer += oprot->writeString(_iter1472->first); - xfer += oprot->writeString(_iter1472->second); + xfer += oprot->writeString(_iter1496->first); + xfer += oprot->writeString(_iter1496->second); } xfer += oprot->writeMapEnd(); } @@ -16514,11 +16514,11 @@ uint32_t ThriftHiveMetastore_exchange_partition_pargs::write(::apache::thrift::p xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->partitionSpecs)).size())); - std::map ::const_iterator _iter1473; - for (_iter1473 = (*(this->partitionSpecs)).begin(); _iter1473 != (*(this->partitionSpecs)).end(); ++_iter1473) + std::map ::const_iterator _iter1497; + for (_iter1497 = (*(this->partitionSpecs)).begin(); _iter1497 != (*(this->partitionSpecs)).end(); ++_iter1497) { - xfer += oprot->writeString(_iter1473->first); - xfer += oprot->writeString(_iter1473->second); + xfer += oprot->writeString(_iter1497->first); + xfer += oprot->writeString(_iter1497->second); } xfer += oprot->writeMapEnd(); } @@ -16763,17 +16763,17 @@ uint32_t ThriftHiveMetastore_exchange_partitions_args::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_MAP) { { this->partitionSpecs.clear(); - uint32_t _size1474; - ::apache::thrift::protocol::TType _ktype1475; - ::apache::thrift::protocol::TType _vtype1476; - xfer += iprot->readMapBegin(_ktype1475, _vtype1476, _size1474); - uint32_t _i1478; - for (_i1478 = 0; _i1478 < _size1474; ++_i1478) + uint32_t _size1498; + ::apache::thrift::protocol::TType _ktype1499; + ::apache::thrift::protocol::TType _vtype1500; + xfer += iprot->readMapBegin(_ktype1499, _vtype1500, _size1498); + uint32_t _i1502; + for (_i1502 = 0; _i1502 < _size1498; ++_i1502) { - std::string _key1479; - xfer += iprot->readString(_key1479); - std::string& _val1480 = this->partitionSpecs[_key1479]; - xfer += iprot->readString(_val1480); + std::string _key1503; + xfer += iprot->readString(_key1503); + std::string& _val1504 = this->partitionSpecs[_key1503]; + xfer += iprot->readString(_val1504); } xfer += iprot->readMapEnd(); } @@ -16834,11 +16834,11 @@ uint32_t ThriftHiveMetastore_exchange_partitions_args::write(::apache::thrift::p xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->partitionSpecs.size())); - std::map ::const_iterator _iter1481; - for (_iter1481 = this->partitionSpecs.begin(); _iter1481 != this->partitionSpecs.end(); ++_iter1481) + std::map ::const_iterator _iter1505; + for (_iter1505 = this->partitionSpecs.begin(); _iter1505 != this->partitionSpecs.end(); ++_iter1505) { - xfer += oprot->writeString(_iter1481->first); - xfer += oprot->writeString(_iter1481->second); + xfer += oprot->writeString(_iter1505->first); + xfer += oprot->writeString(_iter1505->second); } xfer += oprot->writeMapEnd(); } @@ -16878,11 +16878,11 @@ uint32_t ThriftHiveMetastore_exchange_partitions_pargs::write(::apache::thrift:: xfer += oprot->writeFieldBegin("partitionSpecs", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->partitionSpecs)).size())); - std::map ::const_iterator _iter1482; - for (_iter1482 = (*(this->partitionSpecs)).begin(); _iter1482 != (*(this->partitionSpecs)).end(); ++_iter1482) + std::map ::const_iterator _iter1506; + for (_iter1506 = (*(this->partitionSpecs)).begin(); _iter1506 != (*(this->partitionSpecs)).end(); ++_iter1506) { - xfer += oprot->writeString(_iter1482->first); - xfer += oprot->writeString(_iter1482->second); + xfer += oprot->writeString(_iter1506->first); + xfer += oprot->writeString(_iter1506->second); } xfer += oprot->writeMapEnd(); } @@ -16939,14 +16939,14 @@ uint32_t ThriftHiveMetastore_exchange_partitions_result::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1483; - ::apache::thrift::protocol::TType _etype1486; - xfer += iprot->readListBegin(_etype1486, _size1483); - this->success.resize(_size1483); - uint32_t _i1487; - for (_i1487 = 0; _i1487 < _size1483; ++_i1487) + uint32_t _size1507; + ::apache::thrift::protocol::TType _etype1510; + xfer += iprot->readListBegin(_etype1510, _size1507); + this->success.resize(_size1507); + uint32_t _i1511; + for (_i1511 = 0; _i1511 < _size1507; ++_i1511) { - xfer += this->success[_i1487].read(iprot); + xfer += this->success[_i1511].read(iprot); } xfer += iprot->readListEnd(); } @@ -17009,10 +17009,10 @@ uint32_t ThriftHiveMetastore_exchange_partitions_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1488; - for (_iter1488 = this->success.begin(); _iter1488 != this->success.end(); ++_iter1488) + std::vector ::const_iterator _iter1512; + for (_iter1512 = this->success.begin(); _iter1512 != this->success.end(); ++_iter1512) { - xfer += (*_iter1488).write(oprot); + xfer += (*_iter1512).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17069,14 +17069,14 @@ uint32_t ThriftHiveMetastore_exchange_partitions_presult::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1489; - ::apache::thrift::protocol::TType _etype1492; - xfer += iprot->readListBegin(_etype1492, _size1489); - (*(this->success)).resize(_size1489); - uint32_t _i1493; - for (_i1493 = 0; _i1493 < _size1489; ++_i1493) + uint32_t _size1513; + ::apache::thrift::protocol::TType _etype1516; + xfer += iprot->readListBegin(_etype1516, _size1513); + (*(this->success)).resize(_size1513); + uint32_t _i1517; + for (_i1517 = 0; _i1517 < _size1513; ++_i1517) { - xfer += (*(this->success))[_i1493].read(iprot); + xfer += (*(this->success))[_i1517].read(iprot); } xfer += iprot->readListEnd(); } @@ -17175,14 +17175,14 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1494; - ::apache::thrift::protocol::TType _etype1497; - xfer += iprot->readListBegin(_etype1497, _size1494); - this->part_vals.resize(_size1494); - uint32_t _i1498; - for (_i1498 = 0; _i1498 < _size1494; ++_i1498) + uint32_t _size1518; + ::apache::thrift::protocol::TType _etype1521; + xfer += iprot->readListBegin(_etype1521, _size1518); + this->part_vals.resize(_size1518); + uint32_t _i1522; + for (_i1522 = 0; _i1522 < _size1518; ++_i1522) { - xfer += iprot->readString(this->part_vals[_i1498]); + xfer += iprot->readString(this->part_vals[_i1522]); } xfer += iprot->readListEnd(); } @@ -17203,14 +17203,14 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1499; - ::apache::thrift::protocol::TType _etype1502; - xfer += iprot->readListBegin(_etype1502, _size1499); - this->group_names.resize(_size1499); - uint32_t _i1503; - for (_i1503 = 0; _i1503 < _size1499; ++_i1503) + uint32_t _size1523; + ::apache::thrift::protocol::TType _etype1526; + xfer += iprot->readListBegin(_etype1526, _size1523); + this->group_names.resize(_size1523); + uint32_t _i1527; + for (_i1527 = 0; _i1527 < _size1523; ++_i1527) { - xfer += iprot->readString(this->group_names[_i1503]); + xfer += iprot->readString(this->group_names[_i1527]); } xfer += iprot->readListEnd(); } @@ -17247,10 +17247,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::write(::apache::thrif xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1504; - for (_iter1504 = this->part_vals.begin(); _iter1504 != this->part_vals.end(); ++_iter1504) + std::vector ::const_iterator _iter1528; + for (_iter1528 = this->part_vals.begin(); _iter1528 != this->part_vals.end(); ++_iter1528) { - xfer += oprot->writeString((*_iter1504)); + xfer += oprot->writeString((*_iter1528)); } xfer += oprot->writeListEnd(); } @@ -17263,10 +17263,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_args::write(::apache::thrif xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1505; - for (_iter1505 = this->group_names.begin(); _iter1505 != this->group_names.end(); ++_iter1505) + std::vector ::const_iterator _iter1529; + for (_iter1529 = this->group_names.begin(); _iter1529 != this->group_names.end(); ++_iter1529) { - xfer += oprot->writeString((*_iter1505)); + xfer += oprot->writeString((*_iter1529)); } xfer += oprot->writeListEnd(); } @@ -17298,10 +17298,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1506; - for (_iter1506 = (*(this->part_vals)).begin(); _iter1506 != (*(this->part_vals)).end(); ++_iter1506) + std::vector ::const_iterator _iter1530; + for (_iter1530 = (*(this->part_vals)).begin(); _iter1530 != (*(this->part_vals)).end(); ++_iter1530) { - xfer += oprot->writeString((*_iter1506)); + xfer += oprot->writeString((*_iter1530)); } xfer += oprot->writeListEnd(); } @@ -17314,10 +17314,10 @@ uint32_t ThriftHiveMetastore_get_partition_with_auth_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1507; - for (_iter1507 = (*(this->group_names)).begin(); _iter1507 != (*(this->group_names)).end(); ++_iter1507) + std::vector ::const_iterator _iter1531; + for (_iter1531 = (*(this->group_names)).begin(); _iter1531 != (*(this->group_names)).end(); ++_iter1531) { - xfer += oprot->writeString((*_iter1507)); + xfer += oprot->writeString((*_iter1531)); } xfer += oprot->writeListEnd(); } @@ -17876,14 +17876,14 @@ uint32_t ThriftHiveMetastore_get_partitions_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1508; - ::apache::thrift::protocol::TType _etype1511; - xfer += iprot->readListBegin(_etype1511, _size1508); - this->success.resize(_size1508); - uint32_t _i1512; - for (_i1512 = 0; _i1512 < _size1508; ++_i1512) + uint32_t _size1532; + ::apache::thrift::protocol::TType _etype1535; + xfer += iprot->readListBegin(_etype1535, _size1532); + this->success.resize(_size1532); + uint32_t _i1536; + for (_i1536 = 0; _i1536 < _size1532; ++_i1536) { - xfer += this->success[_i1512].read(iprot); + xfer += this->success[_i1536].read(iprot); } xfer += iprot->readListEnd(); } @@ -17930,10 +17930,10 @@ uint32_t ThriftHiveMetastore_get_partitions_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1513; - for (_iter1513 = this->success.begin(); _iter1513 != this->success.end(); ++_iter1513) + std::vector ::const_iterator _iter1537; + for (_iter1537 = this->success.begin(); _iter1537 != this->success.end(); ++_iter1537) { - xfer += (*_iter1513).write(oprot); + xfer += (*_iter1537).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17982,14 +17982,14 @@ uint32_t ThriftHiveMetastore_get_partitions_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1514; - ::apache::thrift::protocol::TType _etype1517; - xfer += iprot->readListBegin(_etype1517, _size1514); - (*(this->success)).resize(_size1514); - uint32_t _i1518; - for (_i1518 = 0; _i1518 < _size1514; ++_i1518) + uint32_t _size1538; + ::apache::thrift::protocol::TType _etype1541; + xfer += iprot->readListBegin(_etype1541, _size1538); + (*(this->success)).resize(_size1538); + uint32_t _i1542; + for (_i1542 = 0; _i1542 < _size1538; ++_i1542) { - xfer += (*(this->success))[_i1518].read(iprot); + xfer += (*(this->success))[_i1542].read(iprot); } xfer += iprot->readListEnd(); } @@ -18088,14 +18088,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_args::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1519; - ::apache::thrift::protocol::TType _etype1522; - xfer += iprot->readListBegin(_etype1522, _size1519); - this->group_names.resize(_size1519); - uint32_t _i1523; - for (_i1523 = 0; _i1523 < _size1519; ++_i1523) + uint32_t _size1543; + ::apache::thrift::protocol::TType _etype1546; + xfer += iprot->readListBegin(_etype1546, _size1543); + this->group_names.resize(_size1543); + uint32_t _i1547; + for (_i1547 = 0; _i1547 < _size1543; ++_i1547) { - xfer += iprot->readString(this->group_names[_i1523]); + xfer += iprot->readString(this->group_names[_i1547]); } xfer += iprot->readListEnd(); } @@ -18140,10 +18140,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_args::write(::apache::thri xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1524; - for (_iter1524 = this->group_names.begin(); _iter1524 != this->group_names.end(); ++_iter1524) + std::vector ::const_iterator _iter1548; + for (_iter1548 = this->group_names.begin(); _iter1548 != this->group_names.end(); ++_iter1548) { - xfer += oprot->writeString((*_iter1524)); + xfer += oprot->writeString((*_iter1548)); } xfer += oprot->writeListEnd(); } @@ -18183,10 +18183,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_pargs::write(::apache::thr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1525; - for (_iter1525 = (*(this->group_names)).begin(); _iter1525 != (*(this->group_names)).end(); ++_iter1525) + std::vector ::const_iterator _iter1549; + for (_iter1549 = (*(this->group_names)).begin(); _iter1549 != (*(this->group_names)).end(); ++_iter1549) { - xfer += oprot->writeString((*_iter1525)); + xfer += oprot->writeString((*_iter1549)); } xfer += oprot->writeListEnd(); } @@ -18227,14 +18227,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1526; - ::apache::thrift::protocol::TType _etype1529; - xfer += iprot->readListBegin(_etype1529, _size1526); - this->success.resize(_size1526); - uint32_t _i1530; - for (_i1530 = 0; _i1530 < _size1526; ++_i1530) + uint32_t _size1550; + ::apache::thrift::protocol::TType _etype1553; + xfer += iprot->readListBegin(_etype1553, _size1550); + this->success.resize(_size1550); + uint32_t _i1554; + for (_i1554 = 0; _i1554 < _size1550; ++_i1554) { - xfer += this->success[_i1530].read(iprot); + xfer += this->success[_i1554].read(iprot); } xfer += iprot->readListEnd(); } @@ -18281,10 +18281,10 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1531; - for (_iter1531 = this->success.begin(); _iter1531 != this->success.end(); ++_iter1531) + std::vector ::const_iterator _iter1555; + for (_iter1555 = this->success.begin(); _iter1555 != this->success.end(); ++_iter1555) { - xfer += (*_iter1531).write(oprot); + xfer += (*_iter1555).write(oprot); } xfer += oprot->writeListEnd(); } @@ -18333,14 +18333,14 @@ uint32_t ThriftHiveMetastore_get_partitions_with_auth_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1532; - ::apache::thrift::protocol::TType _etype1535; - xfer += iprot->readListBegin(_etype1535, _size1532); - (*(this->success)).resize(_size1532); - uint32_t _i1536; - for (_i1536 = 0; _i1536 < _size1532; ++_i1536) + uint32_t _size1556; + ::apache::thrift::protocol::TType _etype1559; + xfer += iprot->readListBegin(_etype1559, _size1556); + (*(this->success)).resize(_size1556); + uint32_t _i1560; + for (_i1560 = 0; _i1560 < _size1556; ++_i1560) { - xfer += (*(this->success))[_i1536].read(iprot); + xfer += (*(this->success))[_i1560].read(iprot); } xfer += iprot->readListEnd(); } @@ -18518,14 +18518,14 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_result::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1537; - ::apache::thrift::protocol::TType _etype1540; - xfer += iprot->readListBegin(_etype1540, _size1537); - this->success.resize(_size1537); - uint32_t _i1541; - for (_i1541 = 0; _i1541 < _size1537; ++_i1541) + uint32_t _size1561; + ::apache::thrift::protocol::TType _etype1564; + xfer += iprot->readListBegin(_etype1564, _size1561); + this->success.resize(_size1561); + uint32_t _i1565; + for (_i1565 = 0; _i1565 < _size1561; ++_i1565) { - xfer += this->success[_i1541].read(iprot); + xfer += this->success[_i1565].read(iprot); } xfer += iprot->readListEnd(); } @@ -18572,10 +18572,10 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_result::write(::apache::thrift xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1542; - for (_iter1542 = this->success.begin(); _iter1542 != this->success.end(); ++_iter1542) + std::vector ::const_iterator _iter1566; + for (_iter1566 = this->success.begin(); _iter1566 != this->success.end(); ++_iter1566) { - xfer += (*_iter1542).write(oprot); + xfer += (*_iter1566).write(oprot); } xfer += oprot->writeListEnd(); } @@ -18624,14 +18624,14 @@ uint32_t ThriftHiveMetastore_get_partitions_pspec_presult::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1543; - ::apache::thrift::protocol::TType _etype1546; - xfer += iprot->readListBegin(_etype1546, _size1543); - (*(this->success)).resize(_size1543); - uint32_t _i1547; - for (_i1547 = 0; _i1547 < _size1543; ++_i1547) + uint32_t _size1567; + ::apache::thrift::protocol::TType _etype1570; + xfer += iprot->readListBegin(_etype1570, _size1567); + (*(this->success)).resize(_size1567); + uint32_t _i1571; + for (_i1571 = 0; _i1571 < _size1567; ++_i1571) { - xfer += (*(this->success))[_i1547].read(iprot); + xfer += (*(this->success))[_i1571].read(iprot); } xfer += iprot->readListEnd(); } @@ -18809,14 +18809,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_result::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1548; - ::apache::thrift::protocol::TType _etype1551; - xfer += iprot->readListBegin(_etype1551, _size1548); - this->success.resize(_size1548); - uint32_t _i1552; - for (_i1552 = 0; _i1552 < _size1548; ++_i1552) + uint32_t _size1572; + ::apache::thrift::protocol::TType _etype1575; + xfer += iprot->readListBegin(_etype1575, _size1572); + this->success.resize(_size1572); + uint32_t _i1576; + for (_i1576 = 0; _i1576 < _size1572; ++_i1576) { - xfer += iprot->readString(this->success[_i1552]); + xfer += iprot->readString(this->success[_i1576]); } xfer += iprot->readListEnd(); } @@ -18863,10 +18863,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_result::write(::apache::thrift: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1553; - for (_iter1553 = this->success.begin(); _iter1553 != this->success.end(); ++_iter1553) + std::vector ::const_iterator _iter1577; + for (_iter1577 = this->success.begin(); _iter1577 != this->success.end(); ++_iter1577) { - xfer += oprot->writeString((*_iter1553)); + xfer += oprot->writeString((*_iter1577)); } xfer += oprot->writeListEnd(); } @@ -18915,14 +18915,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_presult::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1554; - ::apache::thrift::protocol::TType _etype1557; - xfer += iprot->readListBegin(_etype1557, _size1554); - (*(this->success)).resize(_size1554); - uint32_t _i1558; - for (_i1558 = 0; _i1558 < _size1554; ++_i1558) + uint32_t _size1578; + ::apache::thrift::protocol::TType _etype1581; + xfer += iprot->readListBegin(_etype1581, _size1578); + (*(this->success)).resize(_size1578); + uint32_t _i1582; + for (_i1582 = 0; _i1582 < _size1578; ++_i1582) { - xfer += iprot->readString((*(this->success))[_i1558]); + xfer += iprot->readString((*(this->success))[_i1582]); } xfer += iprot->readListEnd(); } @@ -19232,14 +19232,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_args::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1559; - ::apache::thrift::protocol::TType _etype1562; - xfer += iprot->readListBegin(_etype1562, _size1559); - this->part_vals.resize(_size1559); - uint32_t _i1563; - for (_i1563 = 0; _i1563 < _size1559; ++_i1563) + uint32_t _size1583; + ::apache::thrift::protocol::TType _etype1586; + xfer += iprot->readListBegin(_etype1586, _size1583); + this->part_vals.resize(_size1583); + uint32_t _i1587; + for (_i1587 = 0; _i1587 < _size1583; ++_i1587) { - xfer += iprot->readString(this->part_vals[_i1563]); + xfer += iprot->readString(this->part_vals[_i1587]); } xfer += iprot->readListEnd(); } @@ -19284,10 +19284,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_args::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1564; - for (_iter1564 = this->part_vals.begin(); _iter1564 != this->part_vals.end(); ++_iter1564) + std::vector ::const_iterator _iter1588; + for (_iter1588 = this->part_vals.begin(); _iter1588 != this->part_vals.end(); ++_iter1588) { - xfer += oprot->writeString((*_iter1564)); + xfer += oprot->writeString((*_iter1588)); } xfer += oprot->writeListEnd(); } @@ -19323,10 +19323,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_pargs::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1565; - for (_iter1565 = (*(this->part_vals)).begin(); _iter1565 != (*(this->part_vals)).end(); ++_iter1565) + std::vector ::const_iterator _iter1589; + for (_iter1589 = (*(this->part_vals)).begin(); _iter1589 != (*(this->part_vals)).end(); ++_iter1589) { - xfer += oprot->writeString((*_iter1565)); + xfer += oprot->writeString((*_iter1589)); } xfer += oprot->writeListEnd(); } @@ -19371,14 +19371,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1566; - ::apache::thrift::protocol::TType _etype1569; - xfer += iprot->readListBegin(_etype1569, _size1566); - this->success.resize(_size1566); - uint32_t _i1570; - for (_i1570 = 0; _i1570 < _size1566; ++_i1570) + uint32_t _size1590; + ::apache::thrift::protocol::TType _etype1593; + xfer += iprot->readListBegin(_etype1593, _size1590); + this->success.resize(_size1590); + uint32_t _i1594; + for (_i1594 = 0; _i1594 < _size1590; ++_i1594) { - xfer += this->success[_i1570].read(iprot); + xfer += this->success[_i1594].read(iprot); } xfer += iprot->readListEnd(); } @@ -19425,10 +19425,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1571; - for (_iter1571 = this->success.begin(); _iter1571 != this->success.end(); ++_iter1571) + std::vector ::const_iterator _iter1595; + for (_iter1595 = this->success.begin(); _iter1595 != this->success.end(); ++_iter1595) { - xfer += (*_iter1571).write(oprot); + xfer += (*_iter1595).write(oprot); } xfer += oprot->writeListEnd(); } @@ -19477,14 +19477,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1572; - ::apache::thrift::protocol::TType _etype1575; - xfer += iprot->readListBegin(_etype1575, _size1572); - (*(this->success)).resize(_size1572); - uint32_t _i1576; - for (_i1576 = 0; _i1576 < _size1572; ++_i1576) + uint32_t _size1596; + ::apache::thrift::protocol::TType _etype1599; + xfer += iprot->readListBegin(_etype1599, _size1596); + (*(this->success)).resize(_size1596); + uint32_t _i1600; + for (_i1600 = 0; _i1600 < _size1596; ++_i1600) { - xfer += (*(this->success))[_i1576].read(iprot); + xfer += (*(this->success))[_i1600].read(iprot); } xfer += iprot->readListEnd(); } @@ -19567,14 +19567,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1577; - ::apache::thrift::protocol::TType _etype1580; - xfer += iprot->readListBegin(_etype1580, _size1577); - this->part_vals.resize(_size1577); - uint32_t _i1581; - for (_i1581 = 0; _i1581 < _size1577; ++_i1581) + uint32_t _size1601; + ::apache::thrift::protocol::TType _etype1604; + xfer += iprot->readListBegin(_etype1604, _size1601); + this->part_vals.resize(_size1601); + uint32_t _i1605; + for (_i1605 = 0; _i1605 < _size1601; ++_i1605) { - xfer += iprot->readString(this->part_vals[_i1581]); + xfer += iprot->readString(this->part_vals[_i1605]); } xfer += iprot->readListEnd(); } @@ -19603,14 +19603,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1582; - ::apache::thrift::protocol::TType _etype1585; - xfer += iprot->readListBegin(_etype1585, _size1582); - this->group_names.resize(_size1582); - uint32_t _i1586; - for (_i1586 = 0; _i1586 < _size1582; ++_i1586) + uint32_t _size1606; + ::apache::thrift::protocol::TType _etype1609; + xfer += iprot->readListBegin(_etype1609, _size1606); + this->group_names.resize(_size1606); + uint32_t _i1610; + for (_i1610 = 0; _i1610 < _size1606; ++_i1610) { - xfer += iprot->readString(this->group_names[_i1586]); + xfer += iprot->readString(this->group_names[_i1610]); } xfer += iprot->readListEnd(); } @@ -19647,10 +19647,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::write(::apache::t xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1587; - for (_iter1587 = this->part_vals.begin(); _iter1587 != this->part_vals.end(); ++_iter1587) + std::vector ::const_iterator _iter1611; + for (_iter1611 = this->part_vals.begin(); _iter1611 != this->part_vals.end(); ++_iter1611) { - xfer += oprot->writeString((*_iter1587)); + xfer += oprot->writeString((*_iter1611)); } xfer += oprot->writeListEnd(); } @@ -19667,10 +19667,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_args::write(::apache::t xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1588; - for (_iter1588 = this->group_names.begin(); _iter1588 != this->group_names.end(); ++_iter1588) + std::vector ::const_iterator _iter1612; + for (_iter1612 = this->group_names.begin(); _iter1612 != this->group_names.end(); ++_iter1612) { - xfer += oprot->writeString((*_iter1588)); + xfer += oprot->writeString((*_iter1612)); } xfer += oprot->writeListEnd(); } @@ -19702,10 +19702,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_pargs::write(::apache:: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1589; - for (_iter1589 = (*(this->part_vals)).begin(); _iter1589 != (*(this->part_vals)).end(); ++_iter1589) + std::vector ::const_iterator _iter1613; + for (_iter1613 = (*(this->part_vals)).begin(); _iter1613 != (*(this->part_vals)).end(); ++_iter1613) { - xfer += oprot->writeString((*_iter1589)); + xfer += oprot->writeString((*_iter1613)); } xfer += oprot->writeListEnd(); } @@ -19722,10 +19722,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_pargs::write(::apache:: xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1590; - for (_iter1590 = (*(this->group_names)).begin(); _iter1590 != (*(this->group_names)).end(); ++_iter1590) + std::vector ::const_iterator _iter1614; + for (_iter1614 = (*(this->group_names)).begin(); _iter1614 != (*(this->group_names)).end(); ++_iter1614) { - xfer += oprot->writeString((*_iter1590)); + xfer += oprot->writeString((*_iter1614)); } xfer += oprot->writeListEnd(); } @@ -19766,14 +19766,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_result::read(::apache:: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1591; - ::apache::thrift::protocol::TType _etype1594; - xfer += iprot->readListBegin(_etype1594, _size1591); - this->success.resize(_size1591); - uint32_t _i1595; - for (_i1595 = 0; _i1595 < _size1591; ++_i1595) + uint32_t _size1615; + ::apache::thrift::protocol::TType _etype1618; + xfer += iprot->readListBegin(_etype1618, _size1615); + this->success.resize(_size1615); + uint32_t _i1619; + for (_i1619 = 0; _i1619 < _size1615; ++_i1619) { - xfer += this->success[_i1595].read(iprot); + xfer += this->success[_i1619].read(iprot); } xfer += iprot->readListEnd(); } @@ -19820,10 +19820,10 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_result::write(::apache: xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1596; - for (_iter1596 = this->success.begin(); _iter1596 != this->success.end(); ++_iter1596) + std::vector ::const_iterator _iter1620; + for (_iter1620 = this->success.begin(); _iter1620 != this->success.end(); ++_iter1620) { - xfer += (*_iter1596).write(oprot); + xfer += (*_iter1620).write(oprot); } xfer += oprot->writeListEnd(); } @@ -19872,14 +19872,14 @@ uint32_t ThriftHiveMetastore_get_partitions_ps_with_auth_presult::read(::apache: if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1597; - ::apache::thrift::protocol::TType _etype1600; - xfer += iprot->readListBegin(_etype1600, _size1597); - (*(this->success)).resize(_size1597); - uint32_t _i1601; - for (_i1601 = 0; _i1601 < _size1597; ++_i1601) + uint32_t _size1621; + ::apache::thrift::protocol::TType _etype1624; + xfer += iprot->readListBegin(_etype1624, _size1621); + (*(this->success)).resize(_size1621); + uint32_t _i1625; + for (_i1625 = 0; _i1625 < _size1621; ++_i1625) { - xfer += (*(this->success))[_i1601].read(iprot); + xfer += (*(this->success))[_i1625].read(iprot); } xfer += iprot->readListEnd(); } @@ -19962,14 +19962,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_args::read(::apache::thrift: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1602; - ::apache::thrift::protocol::TType _etype1605; - xfer += iprot->readListBegin(_etype1605, _size1602); - this->part_vals.resize(_size1602); - uint32_t _i1606; - for (_i1606 = 0; _i1606 < _size1602; ++_i1606) + uint32_t _size1626; + ::apache::thrift::protocol::TType _etype1629; + xfer += iprot->readListBegin(_etype1629, _size1626); + this->part_vals.resize(_size1626); + uint32_t _i1630; + for (_i1630 = 0; _i1630 < _size1626; ++_i1630) { - xfer += iprot->readString(this->part_vals[_i1606]); + xfer += iprot->readString(this->part_vals[_i1630]); } xfer += iprot->readListEnd(); } @@ -20014,10 +20014,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_args::write(::apache::thrift xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1607; - for (_iter1607 = this->part_vals.begin(); _iter1607 != this->part_vals.end(); ++_iter1607) + std::vector ::const_iterator _iter1631; + for (_iter1631 = this->part_vals.begin(); _iter1631 != this->part_vals.end(); ++_iter1631) { - xfer += oprot->writeString((*_iter1607)); + xfer += oprot->writeString((*_iter1631)); } xfer += oprot->writeListEnd(); } @@ -20053,10 +20053,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_pargs::write(::apache::thrif xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1608; - for (_iter1608 = (*(this->part_vals)).begin(); _iter1608 != (*(this->part_vals)).end(); ++_iter1608) + std::vector ::const_iterator _iter1632; + for (_iter1632 = (*(this->part_vals)).begin(); _iter1632 != (*(this->part_vals)).end(); ++_iter1632) { - xfer += oprot->writeString((*_iter1608)); + xfer += oprot->writeString((*_iter1632)); } xfer += oprot->writeListEnd(); } @@ -20101,14 +20101,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1609; - ::apache::thrift::protocol::TType _etype1612; - xfer += iprot->readListBegin(_etype1612, _size1609); - this->success.resize(_size1609); - uint32_t _i1613; - for (_i1613 = 0; _i1613 < _size1609; ++_i1613) + uint32_t _size1633; + ::apache::thrift::protocol::TType _etype1636; + xfer += iprot->readListBegin(_etype1636, _size1633); + this->success.resize(_size1633); + uint32_t _i1637; + for (_i1637 = 0; _i1637 < _size1633; ++_i1637) { - xfer += iprot->readString(this->success[_i1613]); + xfer += iprot->readString(this->success[_i1637]); } xfer += iprot->readListEnd(); } @@ -20155,10 +20155,10 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1614; - for (_iter1614 = this->success.begin(); _iter1614 != this->success.end(); ++_iter1614) + std::vector ::const_iterator _iter1638; + for (_iter1638 = this->success.begin(); _iter1638 != this->success.end(); ++_iter1638) { - xfer += oprot->writeString((*_iter1614)); + xfer += oprot->writeString((*_iter1638)); } xfer += oprot->writeListEnd(); } @@ -20207,14 +20207,14 @@ uint32_t ThriftHiveMetastore_get_partition_names_ps_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1615; - ::apache::thrift::protocol::TType _etype1618; - xfer += iprot->readListBegin(_etype1618, _size1615); - (*(this->success)).resize(_size1615); - uint32_t _i1619; - for (_i1619 = 0; _i1619 < _size1615; ++_i1619) + uint32_t _size1639; + ::apache::thrift::protocol::TType _etype1642; + xfer += iprot->readListBegin(_etype1642, _size1639); + (*(this->success)).resize(_size1639); + uint32_t _i1643; + for (_i1643 = 0; _i1643 < _size1639; ++_i1643) { - xfer += iprot->readString((*(this->success))[_i1619]); + xfer += iprot->readString((*(this->success))[_i1643]); } xfer += iprot->readListEnd(); } @@ -20408,14 +20408,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1620; - ::apache::thrift::protocol::TType _etype1623; - xfer += iprot->readListBegin(_etype1623, _size1620); - this->success.resize(_size1620); - uint32_t _i1624; - for (_i1624 = 0; _i1624 < _size1620; ++_i1624) + uint32_t _size1644; + ::apache::thrift::protocol::TType _etype1647; + xfer += iprot->readListBegin(_etype1647, _size1644); + this->success.resize(_size1644); + uint32_t _i1648; + for (_i1648 = 0; _i1648 < _size1644; ++_i1648) { - xfer += this->success[_i1624].read(iprot); + xfer += this->success[_i1648].read(iprot); } xfer += iprot->readListEnd(); } @@ -20462,10 +20462,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1625; - for (_iter1625 = this->success.begin(); _iter1625 != this->success.end(); ++_iter1625) + std::vector ::const_iterator _iter1649; + for (_iter1649 = this->success.begin(); _iter1649 != this->success.end(); ++_iter1649) { - xfer += (*_iter1625).write(oprot); + xfer += (*_iter1649).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20514,14 +20514,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_filter_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1626; - ::apache::thrift::protocol::TType _etype1629; - xfer += iprot->readListBegin(_etype1629, _size1626); - (*(this->success)).resize(_size1626); - uint32_t _i1630; - for (_i1630 = 0; _i1630 < _size1626; ++_i1630) + uint32_t _size1650; + ::apache::thrift::protocol::TType _etype1653; + xfer += iprot->readListBegin(_etype1653, _size1650); + (*(this->success)).resize(_size1650); + uint32_t _i1654; + for (_i1654 = 0; _i1654 < _size1650; ++_i1654) { - xfer += (*(this->success))[_i1630].read(iprot); + xfer += (*(this->success))[_i1654].read(iprot); } xfer += iprot->readListEnd(); } @@ -20715,14 +20715,14 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_result::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1631; - ::apache::thrift::protocol::TType _etype1634; - xfer += iprot->readListBegin(_etype1634, _size1631); - this->success.resize(_size1631); - uint32_t _i1635; - for (_i1635 = 0; _i1635 < _size1631; ++_i1635) + uint32_t _size1655; + ::apache::thrift::protocol::TType _etype1658; + xfer += iprot->readListBegin(_etype1658, _size1655); + this->success.resize(_size1655); + uint32_t _i1659; + for (_i1659 = 0; _i1659 < _size1655; ++_i1659) { - xfer += this->success[_i1635].read(iprot); + xfer += this->success[_i1659].read(iprot); } xfer += iprot->readListEnd(); } @@ -20769,10 +20769,10 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_result::write(::apache::th xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1636; - for (_iter1636 = this->success.begin(); _iter1636 != this->success.end(); ++_iter1636) + std::vector ::const_iterator _iter1660; + for (_iter1660 = this->success.begin(); _iter1660 != this->success.end(); ++_iter1660) { - xfer += (*_iter1636).write(oprot); + xfer += (*_iter1660).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20821,14 +20821,14 @@ uint32_t ThriftHiveMetastore_get_part_specs_by_filter_presult::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1637; - ::apache::thrift::protocol::TType _etype1640; - xfer += iprot->readListBegin(_etype1640, _size1637); - (*(this->success)).resize(_size1637); - uint32_t _i1641; - for (_i1641 = 0; _i1641 < _size1637; ++_i1641) + uint32_t _size1661; + ::apache::thrift::protocol::TType _etype1664; + xfer += iprot->readListBegin(_etype1664, _size1661); + (*(this->success)).resize(_size1661); + uint32_t _i1665; + for (_i1665 = 0; _i1665 < _size1661; ++_i1665) { - xfer += (*(this->success))[_i1641].read(iprot); + xfer += (*(this->success))[_i1665].read(iprot); } xfer += iprot->readListEnd(); } @@ -21397,14 +21397,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_args::read(::apache::thrift if (ftype == ::apache::thrift::protocol::T_LIST) { { this->names.clear(); - uint32_t _size1642; - ::apache::thrift::protocol::TType _etype1645; - xfer += iprot->readListBegin(_etype1645, _size1642); - this->names.resize(_size1642); - uint32_t _i1646; - for (_i1646 = 0; _i1646 < _size1642; ++_i1646) + uint32_t _size1666; + ::apache::thrift::protocol::TType _etype1669; + xfer += iprot->readListBegin(_etype1669, _size1666); + this->names.resize(_size1666); + uint32_t _i1670; + for (_i1670 = 0; _i1670 < _size1666; ++_i1670) { - xfer += iprot->readString(this->names[_i1646]); + xfer += iprot->readString(this->names[_i1670]); } xfer += iprot->readListEnd(); } @@ -21441,10 +21441,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_args::write(::apache::thrif xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->names.size())); - std::vector ::const_iterator _iter1647; - for (_iter1647 = this->names.begin(); _iter1647 != this->names.end(); ++_iter1647) + std::vector ::const_iterator _iter1671; + for (_iter1671 = this->names.begin(); _iter1671 != this->names.end(); ++_iter1671) { - xfer += oprot->writeString((*_iter1647)); + xfer += oprot->writeString((*_iter1671)); } xfer += oprot->writeListEnd(); } @@ -21476,10 +21476,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_pargs::write(::apache::thri xfer += oprot->writeFieldBegin("names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->names)).size())); - std::vector ::const_iterator _iter1648; - for (_iter1648 = (*(this->names)).begin(); _iter1648 != (*(this->names)).end(); ++_iter1648) + std::vector ::const_iterator _iter1672; + for (_iter1672 = (*(this->names)).begin(); _iter1672 != (*(this->names)).end(); ++_iter1672) { - xfer += oprot->writeString((*_iter1648)); + xfer += oprot->writeString((*_iter1672)); } xfer += oprot->writeListEnd(); } @@ -21520,14 +21520,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1649; - ::apache::thrift::protocol::TType _etype1652; - xfer += iprot->readListBegin(_etype1652, _size1649); - this->success.resize(_size1649); - uint32_t _i1653; - for (_i1653 = 0; _i1653 < _size1649; ++_i1653) + uint32_t _size1673; + ::apache::thrift::protocol::TType _etype1676; + xfer += iprot->readListBegin(_etype1676, _size1673); + this->success.resize(_size1673); + uint32_t _i1677; + for (_i1677 = 0; _i1677 < _size1673; ++_i1677) { - xfer += this->success[_i1653].read(iprot); + xfer += this->success[_i1677].read(iprot); } xfer += iprot->readListEnd(); } @@ -21574,10 +21574,10 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1654; - for (_iter1654 = this->success.begin(); _iter1654 != this->success.end(); ++_iter1654) + std::vector ::const_iterator _iter1678; + for (_iter1678 = this->success.begin(); _iter1678 != this->success.end(); ++_iter1678) { - xfer += (*_iter1654).write(oprot); + xfer += (*_iter1678).write(oprot); } xfer += oprot->writeListEnd(); } @@ -21626,14 +21626,14 @@ uint32_t ThriftHiveMetastore_get_partitions_by_names_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1655; - ::apache::thrift::protocol::TType _etype1658; - xfer += iprot->readListBegin(_etype1658, _size1655); - (*(this->success)).resize(_size1655); - uint32_t _i1659; - for (_i1659 = 0; _i1659 < _size1655; ++_i1659) + uint32_t _size1679; + ::apache::thrift::protocol::TType _etype1682; + xfer += iprot->readListBegin(_etype1682, _size1679); + (*(this->success)).resize(_size1679); + uint32_t _i1683; + for (_i1683 = 0; _i1683 < _size1679; ++_i1683) { - xfer += (*(this->success))[_i1659].read(iprot); + xfer += (*(this->success))[_i1683].read(iprot); } xfer += iprot->readListEnd(); } @@ -21955,14 +21955,14 @@ uint32_t ThriftHiveMetastore_alter_partitions_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1660; - ::apache::thrift::protocol::TType _etype1663; - xfer += iprot->readListBegin(_etype1663, _size1660); - this->new_parts.resize(_size1660); - uint32_t _i1664; - for (_i1664 = 0; _i1664 < _size1660; ++_i1664) + uint32_t _size1684; + ::apache::thrift::protocol::TType _etype1687; + xfer += iprot->readListBegin(_etype1687, _size1684); + this->new_parts.resize(_size1684); + uint32_t _i1688; + for (_i1688 = 0; _i1688 < _size1684; ++_i1688) { - xfer += this->new_parts[_i1664].read(iprot); + xfer += this->new_parts[_i1688].read(iprot); } xfer += iprot->readListEnd(); } @@ -21999,10 +21999,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1665; - for (_iter1665 = this->new_parts.begin(); _iter1665 != this->new_parts.end(); ++_iter1665) + std::vector ::const_iterator _iter1689; + for (_iter1689 = this->new_parts.begin(); _iter1689 != this->new_parts.end(); ++_iter1689) { - xfer += (*_iter1665).write(oprot); + xfer += (*_iter1689).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22034,10 +22034,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1666; - for (_iter1666 = (*(this->new_parts)).begin(); _iter1666 != (*(this->new_parts)).end(); ++_iter1666) + std::vector ::const_iterator _iter1690; + for (_iter1690 = (*(this->new_parts)).begin(); _iter1690 != (*(this->new_parts)).end(); ++_iter1690) { - xfer += (*_iter1666).write(oprot); + xfer += (*_iter1690).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22222,14 +22222,14 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_args::rea if (ftype == ::apache::thrift::protocol::T_LIST) { { this->new_parts.clear(); - uint32_t _size1667; - ::apache::thrift::protocol::TType _etype1670; - xfer += iprot->readListBegin(_etype1670, _size1667); - this->new_parts.resize(_size1667); - uint32_t _i1671; - for (_i1671 = 0; _i1671 < _size1667; ++_i1671) + uint32_t _size1691; + ::apache::thrift::protocol::TType _etype1694; + xfer += iprot->readListBegin(_etype1694, _size1691); + this->new_parts.resize(_size1691); + uint32_t _i1695; + for (_i1695 = 0; _i1695 < _size1691; ++_i1695) { - xfer += this->new_parts[_i1671].read(iprot); + xfer += this->new_parts[_i1695].read(iprot); } xfer += iprot->readListEnd(); } @@ -22274,10 +22274,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_args::wri xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->new_parts.size())); - std::vector ::const_iterator _iter1672; - for (_iter1672 = this->new_parts.begin(); _iter1672 != this->new_parts.end(); ++_iter1672) + std::vector ::const_iterator _iter1696; + for (_iter1696 = this->new_parts.begin(); _iter1696 != this->new_parts.end(); ++_iter1696) { - xfer += (*_iter1672).write(oprot); + xfer += (*_iter1696).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22313,10 +22313,10 @@ uint32_t ThriftHiveMetastore_alter_partitions_with_environment_context_pargs::wr xfer += oprot->writeFieldBegin("new_parts", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast((*(this->new_parts)).size())); - std::vector ::const_iterator _iter1673; - for (_iter1673 = (*(this->new_parts)).begin(); _iter1673 != (*(this->new_parts)).end(); ++_iter1673) + std::vector ::const_iterator _iter1697; + for (_iter1697 = (*(this->new_parts)).begin(); _iter1697 != (*(this->new_parts)).end(); ++_iter1697) { - xfer += (*_iter1673).write(oprot); + xfer += (*_iter1697).write(oprot); } xfer += oprot->writeListEnd(); } @@ -22760,14 +22760,14 @@ uint32_t ThriftHiveMetastore_rename_partition_args::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1674; - ::apache::thrift::protocol::TType _etype1677; - xfer += iprot->readListBegin(_etype1677, _size1674); - this->part_vals.resize(_size1674); - uint32_t _i1678; - for (_i1678 = 0; _i1678 < _size1674; ++_i1678) + uint32_t _size1698; + ::apache::thrift::protocol::TType _etype1701; + xfer += iprot->readListBegin(_etype1701, _size1698); + this->part_vals.resize(_size1698); + uint32_t _i1702; + for (_i1702 = 0; _i1702 < _size1698; ++_i1702) { - xfer += iprot->readString(this->part_vals[_i1678]); + xfer += iprot->readString(this->part_vals[_i1702]); } xfer += iprot->readListEnd(); } @@ -22812,10 +22812,10 @@ uint32_t ThriftHiveMetastore_rename_partition_args::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1679; - for (_iter1679 = this->part_vals.begin(); _iter1679 != this->part_vals.end(); ++_iter1679) + std::vector ::const_iterator _iter1703; + for (_iter1703 = this->part_vals.begin(); _iter1703 != this->part_vals.end(); ++_iter1703) { - xfer += oprot->writeString((*_iter1679)); + xfer += oprot->writeString((*_iter1703)); } xfer += oprot->writeListEnd(); } @@ -22851,10 +22851,10 @@ uint32_t ThriftHiveMetastore_rename_partition_pargs::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1680; - for (_iter1680 = (*(this->part_vals)).begin(); _iter1680 != (*(this->part_vals)).end(); ++_iter1680) + std::vector ::const_iterator _iter1704; + for (_iter1704 = (*(this->part_vals)).begin(); _iter1704 != (*(this->part_vals)).end(); ++_iter1704) { - xfer += oprot->writeString((*_iter1680)); + xfer += oprot->writeString((*_iter1704)); } xfer += oprot->writeListEnd(); } @@ -23027,14 +23027,14 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_args::read(::ap if (ftype == ::apache::thrift::protocol::T_LIST) { { this->part_vals.clear(); - uint32_t _size1681; - ::apache::thrift::protocol::TType _etype1684; - xfer += iprot->readListBegin(_etype1684, _size1681); - this->part_vals.resize(_size1681); - uint32_t _i1685; - for (_i1685 = 0; _i1685 < _size1681; ++_i1685) + uint32_t _size1705; + ::apache::thrift::protocol::TType _etype1708; + xfer += iprot->readListBegin(_etype1708, _size1705); + this->part_vals.resize(_size1705); + uint32_t _i1709; + for (_i1709 = 0; _i1709 < _size1705; ++_i1709) { - xfer += iprot->readString(this->part_vals[_i1685]); + xfer += iprot->readString(this->part_vals[_i1709]); } xfer += iprot->readListEnd(); } @@ -23071,10 +23071,10 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_args::write(::a xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::vector ::const_iterator _iter1686; - for (_iter1686 = this->part_vals.begin(); _iter1686 != this->part_vals.end(); ++_iter1686) + std::vector ::const_iterator _iter1710; + for (_iter1710 = this->part_vals.begin(); _iter1710 != this->part_vals.end(); ++_iter1710) { - xfer += oprot->writeString((*_iter1686)); + xfer += oprot->writeString((*_iter1710)); } xfer += oprot->writeListEnd(); } @@ -23102,10 +23102,10 @@ uint32_t ThriftHiveMetastore_partition_name_has_valid_characters_pargs::write(:: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::vector ::const_iterator _iter1687; - for (_iter1687 = (*(this->part_vals)).begin(); _iter1687 != (*(this->part_vals)).end(); ++_iter1687) + std::vector ::const_iterator _iter1711; + for (_iter1711 = (*(this->part_vals)).begin(); _iter1711 != (*(this->part_vals)).end(); ++_iter1711) { - xfer += oprot->writeString((*_iter1687)); + xfer += oprot->writeString((*_iter1711)); } xfer += oprot->writeListEnd(); } @@ -23580,14 +23580,14 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1688; - ::apache::thrift::protocol::TType _etype1691; - xfer += iprot->readListBegin(_etype1691, _size1688); - this->success.resize(_size1688); - uint32_t _i1692; - for (_i1692 = 0; _i1692 < _size1688; ++_i1692) + uint32_t _size1712; + ::apache::thrift::protocol::TType _etype1715; + xfer += iprot->readListBegin(_etype1715, _size1712); + this->success.resize(_size1712); + uint32_t _i1716; + for (_i1716 = 0; _i1716 < _size1712; ++_i1716) { - xfer += iprot->readString(this->success[_i1692]); + xfer += iprot->readString(this->success[_i1716]); } xfer += iprot->readListEnd(); } @@ -23626,10 +23626,10 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1693; - for (_iter1693 = this->success.begin(); _iter1693 != this->success.end(); ++_iter1693) + std::vector ::const_iterator _iter1717; + for (_iter1717 = this->success.begin(); _iter1717 != this->success.end(); ++_iter1717) { - xfer += oprot->writeString((*_iter1693)); + xfer += oprot->writeString((*_iter1717)); } xfer += oprot->writeListEnd(); } @@ -23674,14 +23674,14 @@ uint32_t ThriftHiveMetastore_partition_name_to_vals_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1694; - ::apache::thrift::protocol::TType _etype1697; - xfer += iprot->readListBegin(_etype1697, _size1694); - (*(this->success)).resize(_size1694); - uint32_t _i1698; - for (_i1698 = 0; _i1698 < _size1694; ++_i1698) + uint32_t _size1718; + ::apache::thrift::protocol::TType _etype1721; + xfer += iprot->readListBegin(_etype1721, _size1718); + (*(this->success)).resize(_size1718); + uint32_t _i1722; + for (_i1722 = 0; _i1722 < _size1718; ++_i1722) { - xfer += iprot->readString((*(this->success))[_i1698]); + xfer += iprot->readString((*(this->success))[_i1722]); } xfer += iprot->readListEnd(); } @@ -23819,17 +23819,17 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_result::read(::apache::thrif if (ftype == ::apache::thrift::protocol::T_MAP) { { this->success.clear(); - uint32_t _size1699; - ::apache::thrift::protocol::TType _ktype1700; - ::apache::thrift::protocol::TType _vtype1701; - xfer += iprot->readMapBegin(_ktype1700, _vtype1701, _size1699); - uint32_t _i1703; - for (_i1703 = 0; _i1703 < _size1699; ++_i1703) + uint32_t _size1723; + ::apache::thrift::protocol::TType _ktype1724; + ::apache::thrift::protocol::TType _vtype1725; + xfer += iprot->readMapBegin(_ktype1724, _vtype1725, _size1723); + uint32_t _i1727; + for (_i1727 = 0; _i1727 < _size1723; ++_i1727) { - std::string _key1704; - xfer += iprot->readString(_key1704); - std::string& _val1705 = this->success[_key1704]; - xfer += iprot->readString(_val1705); + std::string _key1728; + xfer += iprot->readString(_key1728); + std::string& _val1729 = this->success[_key1728]; + xfer += iprot->readString(_val1729); } xfer += iprot->readMapEnd(); } @@ -23868,11 +23868,11 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_result::write(::apache::thri xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_MAP, 0); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::map ::const_iterator _iter1706; - for (_iter1706 = this->success.begin(); _iter1706 != this->success.end(); ++_iter1706) + std::map ::const_iterator _iter1730; + for (_iter1730 = this->success.begin(); _iter1730 != this->success.end(); ++_iter1730) { - xfer += oprot->writeString(_iter1706->first); - xfer += oprot->writeString(_iter1706->second); + xfer += oprot->writeString(_iter1730->first); + xfer += oprot->writeString(_iter1730->second); } xfer += oprot->writeMapEnd(); } @@ -23917,17 +23917,17 @@ uint32_t ThriftHiveMetastore_partition_name_to_spec_presult::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_MAP) { { (*(this->success)).clear(); - uint32_t _size1707; - ::apache::thrift::protocol::TType _ktype1708; - ::apache::thrift::protocol::TType _vtype1709; - xfer += iprot->readMapBegin(_ktype1708, _vtype1709, _size1707); - uint32_t _i1711; - for (_i1711 = 0; _i1711 < _size1707; ++_i1711) + uint32_t _size1731; + ::apache::thrift::protocol::TType _ktype1732; + ::apache::thrift::protocol::TType _vtype1733; + xfer += iprot->readMapBegin(_ktype1732, _vtype1733, _size1731); + uint32_t _i1735; + for (_i1735 = 0; _i1735 < _size1731; ++_i1735) { - std::string _key1712; - xfer += iprot->readString(_key1712); - std::string& _val1713 = (*(this->success))[_key1712]; - xfer += iprot->readString(_val1713); + std::string _key1736; + xfer += iprot->readString(_key1736); + std::string& _val1737 = (*(this->success))[_key1736]; + xfer += iprot->readString(_val1737); } xfer += iprot->readMapEnd(); } @@ -24002,17 +24002,17 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::read(::apache::thrift:: if (ftype == ::apache::thrift::protocol::T_MAP) { { this->part_vals.clear(); - uint32_t _size1714; - ::apache::thrift::protocol::TType _ktype1715; - ::apache::thrift::protocol::TType _vtype1716; - xfer += iprot->readMapBegin(_ktype1715, _vtype1716, _size1714); - uint32_t _i1718; - for (_i1718 = 0; _i1718 < _size1714; ++_i1718) + uint32_t _size1738; + ::apache::thrift::protocol::TType _ktype1739; + ::apache::thrift::protocol::TType _vtype1740; + xfer += iprot->readMapBegin(_ktype1739, _vtype1740, _size1738); + uint32_t _i1742; + for (_i1742 = 0; _i1742 < _size1738; ++_i1742) { - std::string _key1719; - xfer += iprot->readString(_key1719); - std::string& _val1720 = this->part_vals[_key1719]; - xfer += iprot->readString(_val1720); + std::string _key1743; + xfer += iprot->readString(_key1743); + std::string& _val1744 = this->part_vals[_key1743]; + xfer += iprot->readString(_val1744); } xfer += iprot->readMapEnd(); } @@ -24023,9 +24023,9 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::read(::apache::thrift:: break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1721; - xfer += iprot->readI32(ecast1721); - this->eventType = (PartitionEventType::type)ecast1721; + int32_t ecast1745; + xfer += iprot->readI32(ecast1745); + this->eventType = (PartitionEventType::type)ecast1745; this->__isset.eventType = true; } else { xfer += iprot->skip(ftype); @@ -24059,11 +24059,11 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_args::write(::apache::thrift: xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::map ::const_iterator _iter1722; - for (_iter1722 = this->part_vals.begin(); _iter1722 != this->part_vals.end(); ++_iter1722) + std::map ::const_iterator _iter1746; + for (_iter1746 = this->part_vals.begin(); _iter1746 != this->part_vals.end(); ++_iter1746) { - xfer += oprot->writeString(_iter1722->first); - xfer += oprot->writeString(_iter1722->second); + xfer += oprot->writeString(_iter1746->first); + xfer += oprot->writeString(_iter1746->second); } xfer += oprot->writeMapEnd(); } @@ -24099,11 +24099,11 @@ uint32_t ThriftHiveMetastore_markPartitionForEvent_pargs::write(::apache::thrift xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::map ::const_iterator _iter1723; - for (_iter1723 = (*(this->part_vals)).begin(); _iter1723 != (*(this->part_vals)).end(); ++_iter1723) + std::map ::const_iterator _iter1747; + for (_iter1747 = (*(this->part_vals)).begin(); _iter1747 != (*(this->part_vals)).end(); ++_iter1747) { - xfer += oprot->writeString(_iter1723->first); - xfer += oprot->writeString(_iter1723->second); + xfer += oprot->writeString(_iter1747->first); + xfer += oprot->writeString(_iter1747->second); } xfer += oprot->writeMapEnd(); } @@ -24372,17 +24372,17 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_MAP) { { this->part_vals.clear(); - uint32_t _size1724; - ::apache::thrift::protocol::TType _ktype1725; - ::apache::thrift::protocol::TType _vtype1726; - xfer += iprot->readMapBegin(_ktype1725, _vtype1726, _size1724); - uint32_t _i1728; - for (_i1728 = 0; _i1728 < _size1724; ++_i1728) + uint32_t _size1748; + ::apache::thrift::protocol::TType _ktype1749; + ::apache::thrift::protocol::TType _vtype1750; + xfer += iprot->readMapBegin(_ktype1749, _vtype1750, _size1748); + uint32_t _i1752; + for (_i1752 = 0; _i1752 < _size1748; ++_i1752) { - std::string _key1729; - xfer += iprot->readString(_key1729); - std::string& _val1730 = this->part_vals[_key1729]; - xfer += iprot->readString(_val1730); + std::string _key1753; + xfer += iprot->readString(_key1753); + std::string& _val1754 = this->part_vals[_key1753]; + xfer += iprot->readString(_val1754); } xfer += iprot->readMapEnd(); } @@ -24393,9 +24393,9 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::read(::apache::thri break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1731; - xfer += iprot->readI32(ecast1731); - this->eventType = (PartitionEventType::type)ecast1731; + int32_t ecast1755; + xfer += iprot->readI32(ecast1755); + this->eventType = (PartitionEventType::type)ecast1755; this->__isset.eventType = true; } else { xfer += iprot->skip(ftype); @@ -24429,11 +24429,11 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_args::write(::apache::thr xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->part_vals.size())); - std::map ::const_iterator _iter1732; - for (_iter1732 = this->part_vals.begin(); _iter1732 != this->part_vals.end(); ++_iter1732) + std::map ::const_iterator _iter1756; + for (_iter1756 = this->part_vals.begin(); _iter1756 != this->part_vals.end(); ++_iter1756) { - xfer += oprot->writeString(_iter1732->first); - xfer += oprot->writeString(_iter1732->second); + xfer += oprot->writeString(_iter1756->first); + xfer += oprot->writeString(_iter1756->second); } xfer += oprot->writeMapEnd(); } @@ -24469,11 +24469,11 @@ uint32_t ThriftHiveMetastore_isPartitionMarkedForEvent_pargs::write(::apache::th xfer += oprot->writeFieldBegin("part_vals", ::apache::thrift::protocol::T_MAP, 3); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast((*(this->part_vals)).size())); - std::map ::const_iterator _iter1733; - for (_iter1733 = (*(this->part_vals)).begin(); _iter1733 != (*(this->part_vals)).end(); ++_iter1733) + std::map ::const_iterator _iter1757; + for (_iter1757 = (*(this->part_vals)).begin(); _iter1757 != (*(this->part_vals)).end(); ++_iter1757) { - xfer += oprot->writeString(_iter1733->first); - xfer += oprot->writeString(_iter1733->second); + xfer += oprot->writeString(_iter1757->first); + xfer += oprot->writeString(_iter1757->second); } xfer += oprot->writeMapEnd(); } @@ -29622,14 +29622,14 @@ uint32_t ThriftHiveMetastore_get_functions_result::read(::apache::thrift::protoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1734; - ::apache::thrift::protocol::TType _etype1737; - xfer += iprot->readListBegin(_etype1737, _size1734); - this->success.resize(_size1734); - uint32_t _i1738; - for (_i1738 = 0; _i1738 < _size1734; ++_i1738) + uint32_t _size1758; + ::apache::thrift::protocol::TType _etype1761; + xfer += iprot->readListBegin(_etype1761, _size1758); + this->success.resize(_size1758); + uint32_t _i1762; + for (_i1762 = 0; _i1762 < _size1758; ++_i1762) { - xfer += iprot->readString(this->success[_i1738]); + xfer += iprot->readString(this->success[_i1762]); } xfer += iprot->readListEnd(); } @@ -29668,10 +29668,10 @@ uint32_t ThriftHiveMetastore_get_functions_result::write(::apache::thrift::proto xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1739; - for (_iter1739 = this->success.begin(); _iter1739 != this->success.end(); ++_iter1739) + std::vector ::const_iterator _iter1763; + for (_iter1763 = this->success.begin(); _iter1763 != this->success.end(); ++_iter1763) { - xfer += oprot->writeString((*_iter1739)); + xfer += oprot->writeString((*_iter1763)); } xfer += oprot->writeListEnd(); } @@ -29716,14 +29716,14 @@ uint32_t ThriftHiveMetastore_get_functions_presult::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1740; - ::apache::thrift::protocol::TType _etype1743; - xfer += iprot->readListBegin(_etype1743, _size1740); - (*(this->success)).resize(_size1740); - uint32_t _i1744; - for (_i1744 = 0; _i1744 < _size1740; ++_i1744) + uint32_t _size1764; + ::apache::thrift::protocol::TType _etype1767; + xfer += iprot->readListBegin(_etype1767, _size1764); + (*(this->success)).resize(_size1764); + uint32_t _i1768; + for (_i1768 = 0; _i1768 < _size1764; ++_i1768) { - xfer += iprot->readString((*(this->success))[_i1744]); + xfer += iprot->readString((*(this->success))[_i1768]); } xfer += iprot->readListEnd(); } @@ -30683,14 +30683,14 @@ uint32_t ThriftHiveMetastore_get_role_names_result::read(::apache::thrift::proto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1745; - ::apache::thrift::protocol::TType _etype1748; - xfer += iprot->readListBegin(_etype1748, _size1745); - this->success.resize(_size1745); - uint32_t _i1749; - for (_i1749 = 0; _i1749 < _size1745; ++_i1749) + uint32_t _size1769; + ::apache::thrift::protocol::TType _etype1772; + xfer += iprot->readListBegin(_etype1772, _size1769); + this->success.resize(_size1769); + uint32_t _i1773; + for (_i1773 = 0; _i1773 < _size1769; ++_i1773) { - xfer += iprot->readString(this->success[_i1749]); + xfer += iprot->readString(this->success[_i1773]); } xfer += iprot->readListEnd(); } @@ -30729,10 +30729,10 @@ uint32_t ThriftHiveMetastore_get_role_names_result::write(::apache::thrift::prot xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1750; - for (_iter1750 = this->success.begin(); _iter1750 != this->success.end(); ++_iter1750) + std::vector ::const_iterator _iter1774; + for (_iter1774 = this->success.begin(); _iter1774 != this->success.end(); ++_iter1774) { - xfer += oprot->writeString((*_iter1750)); + xfer += oprot->writeString((*_iter1774)); } xfer += oprot->writeListEnd(); } @@ -30777,14 +30777,14 @@ uint32_t ThriftHiveMetastore_get_role_names_presult::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1751; - ::apache::thrift::protocol::TType _etype1754; - xfer += iprot->readListBegin(_etype1754, _size1751); - (*(this->success)).resize(_size1751); - uint32_t _i1755; - for (_i1755 = 0; _i1755 < _size1751; ++_i1755) + uint32_t _size1775; + ::apache::thrift::protocol::TType _etype1778; + xfer += iprot->readListBegin(_etype1778, _size1775); + (*(this->success)).resize(_size1775); + uint32_t _i1779; + for (_i1779 = 0; _i1779 < _size1775; ++_i1779) { - xfer += iprot->readString((*(this->success))[_i1755]); + xfer += iprot->readString((*(this->success))[_i1779]); } xfer += iprot->readListEnd(); } @@ -30857,9 +30857,9 @@ uint32_t ThriftHiveMetastore_grant_role_args::read(::apache::thrift::protocol::T break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1756; - xfer += iprot->readI32(ecast1756); - this->principal_type = (PrincipalType::type)ecast1756; + int32_t ecast1780; + xfer += iprot->readI32(ecast1780); + this->principal_type = (PrincipalType::type)ecast1780; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -30875,9 +30875,9 @@ uint32_t ThriftHiveMetastore_grant_role_args::read(::apache::thrift::protocol::T break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1757; - xfer += iprot->readI32(ecast1757); - this->grantorType = (PrincipalType::type)ecast1757; + int32_t ecast1781; + xfer += iprot->readI32(ecast1781); + this->grantorType = (PrincipalType::type)ecast1781; this->__isset.grantorType = true; } else { xfer += iprot->skip(ftype); @@ -31148,9 +31148,9 @@ uint32_t ThriftHiveMetastore_revoke_role_args::read(::apache::thrift::protocol:: break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1758; - xfer += iprot->readI32(ecast1758); - this->principal_type = (PrincipalType::type)ecast1758; + int32_t ecast1782; + xfer += iprot->readI32(ecast1782); + this->principal_type = (PrincipalType::type)ecast1782; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -31381,9 +31381,9 @@ uint32_t ThriftHiveMetastore_list_roles_args::read(::apache::thrift::protocol::T break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1759; - xfer += iprot->readI32(ecast1759); - this->principal_type = (PrincipalType::type)ecast1759; + int32_t ecast1783; + xfer += iprot->readI32(ecast1783); + this->principal_type = (PrincipalType::type)ecast1783; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -31472,14 +31472,14 @@ uint32_t ThriftHiveMetastore_list_roles_result::read(::apache::thrift::protocol: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1760; - ::apache::thrift::protocol::TType _etype1763; - xfer += iprot->readListBegin(_etype1763, _size1760); - this->success.resize(_size1760); - uint32_t _i1764; - for (_i1764 = 0; _i1764 < _size1760; ++_i1764) + uint32_t _size1784; + ::apache::thrift::protocol::TType _etype1787; + xfer += iprot->readListBegin(_etype1787, _size1784); + this->success.resize(_size1784); + uint32_t _i1788; + for (_i1788 = 0; _i1788 < _size1784; ++_i1788) { - xfer += this->success[_i1764].read(iprot); + xfer += this->success[_i1788].read(iprot); } xfer += iprot->readListEnd(); } @@ -31518,10 +31518,10 @@ uint32_t ThriftHiveMetastore_list_roles_result::write(::apache::thrift::protocol xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1765; - for (_iter1765 = this->success.begin(); _iter1765 != this->success.end(); ++_iter1765) + std::vector ::const_iterator _iter1789; + for (_iter1789 = this->success.begin(); _iter1789 != this->success.end(); ++_iter1789) { - xfer += (*_iter1765).write(oprot); + xfer += (*_iter1789).write(oprot); } xfer += oprot->writeListEnd(); } @@ -31566,14 +31566,14 @@ uint32_t ThriftHiveMetastore_list_roles_presult::read(::apache::thrift::protocol if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1766; - ::apache::thrift::protocol::TType _etype1769; - xfer += iprot->readListBegin(_etype1769, _size1766); - (*(this->success)).resize(_size1766); - uint32_t _i1770; - for (_i1770 = 0; _i1770 < _size1766; ++_i1770) + uint32_t _size1790; + ::apache::thrift::protocol::TType _etype1793; + xfer += iprot->readListBegin(_etype1793, _size1790); + (*(this->success)).resize(_size1790); + uint32_t _i1794; + for (_i1794 = 0; _i1794 < _size1790; ++_i1794) { - xfer += (*(this->success))[_i1770].read(iprot); + xfer += (*(this->success))[_i1794].read(iprot); } xfer += iprot->readListEnd(); } @@ -32269,14 +32269,14 @@ uint32_t ThriftHiveMetastore_get_privilege_set_args::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1771; - ::apache::thrift::protocol::TType _etype1774; - xfer += iprot->readListBegin(_etype1774, _size1771); - this->group_names.resize(_size1771); - uint32_t _i1775; - for (_i1775 = 0; _i1775 < _size1771; ++_i1775) + uint32_t _size1795; + ::apache::thrift::protocol::TType _etype1798; + xfer += iprot->readListBegin(_etype1798, _size1795); + this->group_names.resize(_size1795); + uint32_t _i1799; + for (_i1799 = 0; _i1799 < _size1795; ++_i1799) { - xfer += iprot->readString(this->group_names[_i1775]); + xfer += iprot->readString(this->group_names[_i1799]); } xfer += iprot->readListEnd(); } @@ -32313,10 +32313,10 @@ uint32_t ThriftHiveMetastore_get_privilege_set_args::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1776; - for (_iter1776 = this->group_names.begin(); _iter1776 != this->group_names.end(); ++_iter1776) + std::vector ::const_iterator _iter1800; + for (_iter1800 = this->group_names.begin(); _iter1800 != this->group_names.end(); ++_iter1800) { - xfer += oprot->writeString((*_iter1776)); + xfer += oprot->writeString((*_iter1800)); } xfer += oprot->writeListEnd(); } @@ -32348,10 +32348,10 @@ uint32_t ThriftHiveMetastore_get_privilege_set_pargs::write(::apache::thrift::pr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1777; - for (_iter1777 = (*(this->group_names)).begin(); _iter1777 != (*(this->group_names)).end(); ++_iter1777) + std::vector ::const_iterator _iter1801; + for (_iter1801 = (*(this->group_names)).begin(); _iter1801 != (*(this->group_names)).end(); ++_iter1801) { - xfer += oprot->writeString((*_iter1777)); + xfer += oprot->writeString((*_iter1801)); } xfer += oprot->writeListEnd(); } @@ -32526,9 +32526,9 @@ uint32_t ThriftHiveMetastore_list_privileges_args::read(::apache::thrift::protoc break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1778; - xfer += iprot->readI32(ecast1778); - this->principal_type = (PrincipalType::type)ecast1778; + int32_t ecast1802; + xfer += iprot->readI32(ecast1802); + this->principal_type = (PrincipalType::type)ecast1802; this->__isset.principal_type = true; } else { xfer += iprot->skip(ftype); @@ -32633,14 +32633,14 @@ uint32_t ThriftHiveMetastore_list_privileges_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1779; - ::apache::thrift::protocol::TType _etype1782; - xfer += iprot->readListBegin(_etype1782, _size1779); - this->success.resize(_size1779); - uint32_t _i1783; - for (_i1783 = 0; _i1783 < _size1779; ++_i1783) + uint32_t _size1803; + ::apache::thrift::protocol::TType _etype1806; + xfer += iprot->readListBegin(_etype1806, _size1803); + this->success.resize(_size1803); + uint32_t _i1807; + for (_i1807 = 0; _i1807 < _size1803; ++_i1807) { - xfer += this->success[_i1783].read(iprot); + xfer += this->success[_i1807].read(iprot); } xfer += iprot->readListEnd(); } @@ -32679,10 +32679,10 @@ uint32_t ThriftHiveMetastore_list_privileges_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1784; - for (_iter1784 = this->success.begin(); _iter1784 != this->success.end(); ++_iter1784) + std::vector ::const_iterator _iter1808; + for (_iter1808 = this->success.begin(); _iter1808 != this->success.end(); ++_iter1808) { - xfer += (*_iter1784).write(oprot); + xfer += (*_iter1808).write(oprot); } xfer += oprot->writeListEnd(); } @@ -32727,14 +32727,14 @@ uint32_t ThriftHiveMetastore_list_privileges_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1785; - ::apache::thrift::protocol::TType _etype1788; - xfer += iprot->readListBegin(_etype1788, _size1785); - (*(this->success)).resize(_size1785); - uint32_t _i1789; - for (_i1789 = 0; _i1789 < _size1785; ++_i1789) + uint32_t _size1809; + ::apache::thrift::protocol::TType _etype1812; + xfer += iprot->readListBegin(_etype1812, _size1809); + (*(this->success)).resize(_size1809); + uint32_t _i1813; + for (_i1813 = 0; _i1813 < _size1809; ++_i1813) { - xfer += (*(this->success))[_i1789].read(iprot); + xfer += (*(this->success))[_i1813].read(iprot); } xfer += iprot->readListEnd(); } @@ -33661,14 +33661,14 @@ uint32_t ThriftHiveMetastore_set_ugi_args::read(::apache::thrift::protocol::TPro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->group_names.clear(); - uint32_t _size1790; - ::apache::thrift::protocol::TType _etype1793; - xfer += iprot->readListBegin(_etype1793, _size1790); - this->group_names.resize(_size1790); - uint32_t _i1794; - for (_i1794 = 0; _i1794 < _size1790; ++_i1794) + uint32_t _size1814; + ::apache::thrift::protocol::TType _etype1817; + xfer += iprot->readListBegin(_etype1817, _size1814); + this->group_names.resize(_size1814); + uint32_t _i1818; + for (_i1818 = 0; _i1818 < _size1814; ++_i1818) { - xfer += iprot->readString(this->group_names[_i1794]); + xfer += iprot->readString(this->group_names[_i1818]); } xfer += iprot->readListEnd(); } @@ -33701,10 +33701,10 @@ uint32_t ThriftHiveMetastore_set_ugi_args::write(::apache::thrift::protocol::TPr xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->group_names.size())); - std::vector ::const_iterator _iter1795; - for (_iter1795 = this->group_names.begin(); _iter1795 != this->group_names.end(); ++_iter1795) + std::vector ::const_iterator _iter1819; + for (_iter1819 = this->group_names.begin(); _iter1819 != this->group_names.end(); ++_iter1819) { - xfer += oprot->writeString((*_iter1795)); + xfer += oprot->writeString((*_iter1819)); } xfer += oprot->writeListEnd(); } @@ -33732,10 +33732,10 @@ uint32_t ThriftHiveMetastore_set_ugi_pargs::write(::apache::thrift::protocol::TP xfer += oprot->writeFieldBegin("group_names", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast((*(this->group_names)).size())); - std::vector ::const_iterator _iter1796; - for (_iter1796 = (*(this->group_names)).begin(); _iter1796 != (*(this->group_names)).end(); ++_iter1796) + std::vector ::const_iterator _iter1820; + for (_iter1820 = (*(this->group_names)).begin(); _iter1820 != (*(this->group_names)).end(); ++_iter1820) { - xfer += oprot->writeString((*_iter1796)); + xfer += oprot->writeString((*_iter1820)); } xfer += oprot->writeListEnd(); } @@ -33776,14 +33776,14 @@ uint32_t ThriftHiveMetastore_set_ugi_result::read(::apache::thrift::protocol::TP if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1797; - ::apache::thrift::protocol::TType _etype1800; - xfer += iprot->readListBegin(_etype1800, _size1797); - this->success.resize(_size1797); - uint32_t _i1801; - for (_i1801 = 0; _i1801 < _size1797; ++_i1801) + uint32_t _size1821; + ::apache::thrift::protocol::TType _etype1824; + xfer += iprot->readListBegin(_etype1824, _size1821); + this->success.resize(_size1821); + uint32_t _i1825; + for (_i1825 = 0; _i1825 < _size1821; ++_i1825) { - xfer += iprot->readString(this->success[_i1801]); + xfer += iprot->readString(this->success[_i1825]); } xfer += iprot->readListEnd(); } @@ -33822,10 +33822,10 @@ uint32_t ThriftHiveMetastore_set_ugi_result::write(::apache::thrift::protocol::T xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1802; - for (_iter1802 = this->success.begin(); _iter1802 != this->success.end(); ++_iter1802) + std::vector ::const_iterator _iter1826; + for (_iter1826 = this->success.begin(); _iter1826 != this->success.end(); ++_iter1826) { - xfer += oprot->writeString((*_iter1802)); + xfer += oprot->writeString((*_iter1826)); } xfer += oprot->writeListEnd(); } @@ -33870,14 +33870,14 @@ uint32_t ThriftHiveMetastore_set_ugi_presult::read(::apache::thrift::protocol::T if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1803; - ::apache::thrift::protocol::TType _etype1806; - xfer += iprot->readListBegin(_etype1806, _size1803); - (*(this->success)).resize(_size1803); - uint32_t _i1807; - for (_i1807 = 0; _i1807 < _size1803; ++_i1807) + uint32_t _size1827; + ::apache::thrift::protocol::TType _etype1830; + xfer += iprot->readListBegin(_etype1830, _size1827); + (*(this->success)).resize(_size1827); + uint32_t _i1831; + for (_i1831 = 0; _i1831 < _size1827; ++_i1831) { - xfer += iprot->readString((*(this->success))[_i1807]); + xfer += iprot->readString((*(this->success))[_i1831]); } xfer += iprot->readListEnd(); } @@ -35188,14 +35188,14 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_result::read(::apache::th if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1808; - ::apache::thrift::protocol::TType _etype1811; - xfer += iprot->readListBegin(_etype1811, _size1808); - this->success.resize(_size1808); - uint32_t _i1812; - for (_i1812 = 0; _i1812 < _size1808; ++_i1812) + uint32_t _size1832; + ::apache::thrift::protocol::TType _etype1835; + xfer += iprot->readListBegin(_etype1835, _size1832); + this->success.resize(_size1832); + uint32_t _i1836; + for (_i1836 = 0; _i1836 < _size1832; ++_i1836) { - xfer += iprot->readString(this->success[_i1812]); + xfer += iprot->readString(this->success[_i1836]); } xfer += iprot->readListEnd(); } @@ -35226,10 +35226,10 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_result::write(::apache::t xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1813; - for (_iter1813 = this->success.begin(); _iter1813 != this->success.end(); ++_iter1813) + std::vector ::const_iterator _iter1837; + for (_iter1837 = this->success.begin(); _iter1837 != this->success.end(); ++_iter1837) { - xfer += oprot->writeString((*_iter1813)); + xfer += oprot->writeString((*_iter1837)); } xfer += oprot->writeListEnd(); } @@ -35270,14 +35270,14 @@ uint32_t ThriftHiveMetastore_get_all_token_identifiers_presult::read(::apache::t if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1814; - ::apache::thrift::protocol::TType _etype1817; - xfer += iprot->readListBegin(_etype1817, _size1814); - (*(this->success)).resize(_size1814); - uint32_t _i1818; - for (_i1818 = 0; _i1818 < _size1814; ++_i1818) + uint32_t _size1838; + ::apache::thrift::protocol::TType _etype1841; + xfer += iprot->readListBegin(_etype1841, _size1838); + (*(this->success)).resize(_size1838); + uint32_t _i1842; + for (_i1842 = 0; _i1842 < _size1838; ++_i1842) { - xfer += iprot->readString((*(this->success))[_i1818]); + xfer += iprot->readString((*(this->success))[_i1842]); } xfer += iprot->readListEnd(); } @@ -36003,14 +36003,14 @@ uint32_t ThriftHiveMetastore_get_master_keys_result::read(::apache::thrift::prot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1819; - ::apache::thrift::protocol::TType _etype1822; - xfer += iprot->readListBegin(_etype1822, _size1819); - this->success.resize(_size1819); - uint32_t _i1823; - for (_i1823 = 0; _i1823 < _size1819; ++_i1823) + uint32_t _size1843; + ::apache::thrift::protocol::TType _etype1846; + xfer += iprot->readListBegin(_etype1846, _size1843); + this->success.resize(_size1843); + uint32_t _i1847; + for (_i1847 = 0; _i1847 < _size1843; ++_i1847) { - xfer += iprot->readString(this->success[_i1823]); + xfer += iprot->readString(this->success[_i1847]); } xfer += iprot->readListEnd(); } @@ -36041,10 +36041,10 @@ uint32_t ThriftHiveMetastore_get_master_keys_result::write(::apache::thrift::pro xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->success.size())); - std::vector ::const_iterator _iter1824; - for (_iter1824 = this->success.begin(); _iter1824 != this->success.end(); ++_iter1824) + std::vector ::const_iterator _iter1848; + for (_iter1848 = this->success.begin(); _iter1848 != this->success.end(); ++_iter1848) { - xfer += oprot->writeString((*_iter1824)); + xfer += oprot->writeString((*_iter1848)); } xfer += oprot->writeListEnd(); } @@ -36085,14 +36085,14 @@ uint32_t ThriftHiveMetastore_get_master_keys_presult::read(::apache::thrift::pro if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1825; - ::apache::thrift::protocol::TType _etype1828; - xfer += iprot->readListBegin(_etype1828, _size1825); - (*(this->success)).resize(_size1825); - uint32_t _i1829; - for (_i1829 = 0; _i1829 < _size1825; ++_i1829) + uint32_t _size1849; + ::apache::thrift::protocol::TType _etype1852; + xfer += iprot->readListBegin(_etype1852, _size1849); + (*(this->success)).resize(_size1849); + uint32_t _i1853; + for (_i1853 = 0; _i1853 < _size1849; ++_i1853) { - xfer += iprot->readString((*(this->success))[_i1829]); + xfer += iprot->readString((*(this->success))[_i1853]); } xfer += iprot->readListEnd(); } @@ -40725,11 +40725,11 @@ uint32_t ThriftHiveMetastore_flushCache_presult::read(::apache::thrift::protocol } -ThriftHiveMetastore_cm_recycle_args::~ThriftHiveMetastore_cm_recycle_args() throw() { +ThriftHiveMetastore_add_write_notification_log_args::~ThriftHiveMetastore_add_write_notification_log_args() throw() { } -uint32_t ThriftHiveMetastore_cm_recycle_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_write_notification_log_args::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -40750,10 +40750,10 @@ uint32_t ThriftHiveMetastore_cm_recycle_args::read(::apache::thrift::protocol::T } switch (fid) { - case 1: + case -1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->request.read(iprot); - this->__isset.request = true; + xfer += this->rqst.read(iprot); + this->__isset.rqst = true; } else { xfer += iprot->skip(ftype); } @@ -40770,13 +40770,13 @@ uint32_t ThriftHiveMetastore_cm_recycle_args::read(::apache::thrift::protocol::T return xfer; } -uint32_t ThriftHiveMetastore_cm_recycle_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_write_notification_log_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_cm_recycle_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_write_notification_log_args"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->request.write(oprot); + xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, -1); + xfer += this->rqst.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -40785,17 +40785,17 @@ uint32_t ThriftHiveMetastore_cm_recycle_args::write(::apache::thrift::protocol:: } -ThriftHiveMetastore_cm_recycle_pargs::~ThriftHiveMetastore_cm_recycle_pargs() throw() { +ThriftHiveMetastore_add_write_notification_log_pargs::~ThriftHiveMetastore_add_write_notification_log_pargs() throw() { } -uint32_t ThriftHiveMetastore_cm_recycle_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_write_notification_log_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_cm_recycle_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_write_notification_log_pargs"); - xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->request)).write(oprot); + xfer += oprot->writeFieldBegin("rqst", ::apache::thrift::protocol::T_STRUCT, -1); + xfer += (*(this->rqst)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -40804,11 +40804,11 @@ uint32_t ThriftHiveMetastore_cm_recycle_pargs::write(::apache::thrift::protocol: } -ThriftHiveMetastore_cm_recycle_result::~ThriftHiveMetastore_cm_recycle_result() throw() { +ThriftHiveMetastore_add_write_notification_log_result::~ThriftHiveMetastore_add_write_notification_log_result() throw() { } -uint32_t ThriftHiveMetastore_cm_recycle_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_write_notification_log_result::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -40837,14 +40837,6 @@ uint32_t ThriftHiveMetastore_cm_recycle_result::read(::apache::thrift::protocol: xfer += iprot->skip(ftype); } break; - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -40857,20 +40849,16 @@ uint32_t ThriftHiveMetastore_cm_recycle_result::read(::apache::thrift::protocol: return xfer; } -uint32_t ThriftHiveMetastore_cm_recycle_result::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_add_write_notification_log_result::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; - xfer += oprot->writeStructBegin("ThriftHiveMetastore_cm_recycle_result"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_add_write_notification_log_result"); if (this->__isset.success) { xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); xfer += this->success.write(oprot); xfer += oprot->writeFieldEnd(); - } else if (this->__isset.o1) { - xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->o1.write(oprot); - xfer += oprot->writeFieldEnd(); } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); @@ -40878,11 +40866,11 @@ uint32_t ThriftHiveMetastore_cm_recycle_result::write(::apache::thrift::protocol } -ThriftHiveMetastore_cm_recycle_presult::~ThriftHiveMetastore_cm_recycle_presult() throw() { +ThriftHiveMetastore_add_write_notification_log_presult::~ThriftHiveMetastore_add_write_notification_log_presult() throw() { } -uint32_t ThriftHiveMetastore_cm_recycle_presult::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_add_write_notification_log_presult::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -40911,14 +40899,6 @@ uint32_t ThriftHiveMetastore_cm_recycle_presult::read(::apache::thrift::protocol xfer += iprot->skip(ftype); } break; - case 1: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->o1.read(iprot); - this->__isset.o1 = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -40932,11 +40912,11 @@ uint32_t ThriftHiveMetastore_cm_recycle_presult::read(::apache::thrift::protocol } -ThriftHiveMetastore_get_file_metadata_by_expr_args::~ThriftHiveMetastore_get_file_metadata_by_expr_args() throw() { +ThriftHiveMetastore_cm_recycle_args::~ThriftHiveMetastore_cm_recycle_args() throw() { } -uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_cm_recycle_args::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -40959,8 +40939,8 @@ uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::read(::apache::thri { case 1: if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->req.read(iprot); - this->__isset.req = true; + xfer += this->request.read(iprot); + this->__isset.request = true; } else { xfer += iprot->skip(ftype); } @@ -40977,13 +40957,13 @@ uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::read(::apache::thri return xfer; } -uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_cm_recycle_args::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_file_metadata_by_expr_args"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_cm_recycle_args"); - xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += this->req.write(oprot); + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->request.write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -40992,17 +40972,17 @@ uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::write(::apache::thr } -ThriftHiveMetastore_get_file_metadata_by_expr_pargs::~ThriftHiveMetastore_get_file_metadata_by_expr_pargs() throw() { +ThriftHiveMetastore_cm_recycle_pargs::~ThriftHiveMetastore_cm_recycle_pargs() throw() { } -uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t ThriftHiveMetastore_cm_recycle_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_file_metadata_by_expr_pargs"); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_cm_recycle_pargs"); - xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); - xfer += (*(this->req)).write(oprot); + xfer += oprot->writeFieldBegin("request", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->request)).write(oprot); xfer += oprot->writeFieldEnd(); xfer += oprot->writeFieldStop(); @@ -41011,11 +40991,218 @@ uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_pargs::write(::apache::th } -ThriftHiveMetastore_get_file_metadata_by_expr_result::~ThriftHiveMetastore_get_file_metadata_by_expr_result() throw() { +ThriftHiveMetastore_cm_recycle_result::~ThriftHiveMetastore_cm_recycle_result() throw() { } -uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_result::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t ThriftHiveMetastore_cm_recycle_result::read(::apache::thrift::protocol::TProtocol* iprot) { + + apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 0: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->success.read(iprot); + this->__isset.success = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o1.read(iprot); + this->__isset.o1 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t ThriftHiveMetastore_cm_recycle_result::write(::apache::thrift::protocol::TProtocol* oprot) const { + + uint32_t xfer = 0; + + xfer += oprot->writeStructBegin("ThriftHiveMetastore_cm_recycle_result"); + + if (this->__isset.success) { + xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_STRUCT, 0); + xfer += this->success.write(oprot); + xfer += oprot->writeFieldEnd(); + } else if (this->__isset.o1) { + xfer += oprot->writeFieldBegin("o1", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->o1.write(oprot); + xfer += oprot->writeFieldEnd(); + } + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_cm_recycle_presult::~ThriftHiveMetastore_cm_recycle_presult() throw() { +} + + +uint32_t ThriftHiveMetastore_cm_recycle_presult::read(::apache::thrift::protocol::TProtocol* iprot) { + + apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 0: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += (*(this->success)).read(iprot); + this->__isset.success = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->o1.read(iprot); + this->__isset.o1 = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + + +ThriftHiveMetastore_get_file_metadata_by_expr_args::~ThriftHiveMetastore_get_file_metadata_by_expr_args() throw() { +} + + +uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::read(::apache::thrift::protocol::TProtocol* iprot) { + + apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 1: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->req.read(iprot); + this->__isset.req = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_args::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_file_metadata_by_expr_args"); + + xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += this->req.write(oprot); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_get_file_metadata_by_expr_pargs::~ThriftHiveMetastore_get_file_metadata_by_expr_pargs() throw() { +} + + +uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_pargs::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("ThriftHiveMetastore_get_file_metadata_by_expr_pargs"); + + xfer += oprot->writeFieldBegin("req", ::apache::thrift::protocol::T_STRUCT, 1); + xfer += (*(this->req)).write(oprot); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + + +ThriftHiveMetastore_get_file_metadata_by_expr_result::~ThriftHiveMetastore_get_file_metadata_by_expr_result() throw() { +} + + +uint32_t ThriftHiveMetastore_get_file_metadata_by_expr_result::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -47889,14 +48076,14 @@ uint32_t ThriftHiveMetastore_get_schema_all_versions_result::read(::apache::thri if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1830; - ::apache::thrift::protocol::TType _etype1833; - xfer += iprot->readListBegin(_etype1833, _size1830); - this->success.resize(_size1830); - uint32_t _i1834; - for (_i1834 = 0; _i1834 < _size1830; ++_i1834) + uint32_t _size1854; + ::apache::thrift::protocol::TType _etype1857; + xfer += iprot->readListBegin(_etype1857, _size1854); + this->success.resize(_size1854); + uint32_t _i1858; + for (_i1858 = 0; _i1858 < _size1854; ++_i1858) { - xfer += this->success[_i1834].read(iprot); + xfer += this->success[_i1858].read(iprot); } xfer += iprot->readListEnd(); } @@ -47943,10 +48130,10 @@ uint32_t ThriftHiveMetastore_get_schema_all_versions_result::write(::apache::thr xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1835; - for (_iter1835 = this->success.begin(); _iter1835 != this->success.end(); ++_iter1835) + std::vector ::const_iterator _iter1859; + for (_iter1859 = this->success.begin(); _iter1859 != this->success.end(); ++_iter1859) { - xfer += (*_iter1835).write(oprot); + xfer += (*_iter1859).write(oprot); } xfer += oprot->writeListEnd(); } @@ -47995,14 +48182,14 @@ uint32_t ThriftHiveMetastore_get_schema_all_versions_presult::read(::apache::thr if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1836; - ::apache::thrift::protocol::TType _etype1839; - xfer += iprot->readListBegin(_etype1839, _size1836); - (*(this->success)).resize(_size1836); - uint32_t _i1840; - for (_i1840 = 0; _i1840 < _size1836; ++_i1840) + uint32_t _size1860; + ::apache::thrift::protocol::TType _etype1863; + xfer += iprot->readListBegin(_etype1863, _size1860); + (*(this->success)).resize(_size1860); + uint32_t _i1864; + for (_i1864 = 0; _i1864 < _size1860; ++_i1864) { - xfer += (*(this->success))[_i1840].read(iprot); + xfer += (*(this->success))[_i1864].read(iprot); } xfer += iprot->readListEnd(); } @@ -50055,14 +50242,14 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_result::read(::apache::thrift::pr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->success.clear(); - uint32_t _size1841; - ::apache::thrift::protocol::TType _etype1844; - xfer += iprot->readListBegin(_etype1844, _size1841); - this->success.resize(_size1841); - uint32_t _i1845; - for (_i1845 = 0; _i1845 < _size1841; ++_i1845) + uint32_t _size1865; + ::apache::thrift::protocol::TType _etype1868; + xfer += iprot->readListBegin(_etype1868, _size1865); + this->success.resize(_size1865); + uint32_t _i1869; + for (_i1869 = 0; _i1869 < _size1865; ++_i1869) { - xfer += this->success[_i1845].read(iprot); + xfer += this->success[_i1869].read(iprot); } xfer += iprot->readListEnd(); } @@ -50101,10 +50288,10 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_result::write(::apache::thrift::p xfer += oprot->writeFieldBegin("success", ::apache::thrift::protocol::T_LIST, 0); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->success.size())); - std::vector ::const_iterator _iter1846; - for (_iter1846 = this->success.begin(); _iter1846 != this->success.end(); ++_iter1846) + std::vector ::const_iterator _iter1870; + for (_iter1870 = this->success.begin(); _iter1870 != this->success.end(); ++_iter1870) { - xfer += (*_iter1846).write(oprot); + xfer += (*_iter1870).write(oprot); } xfer += oprot->writeListEnd(); } @@ -50149,14 +50336,14 @@ uint32_t ThriftHiveMetastore_get_runtime_stats_presult::read(::apache::thrift::p if (ftype == ::apache::thrift::protocol::T_LIST) { { (*(this->success)).clear(); - uint32_t _size1847; - ::apache::thrift::protocol::TType _etype1850; - xfer += iprot->readListBegin(_etype1850, _size1847); - (*(this->success)).resize(_size1847); - uint32_t _i1851; - for (_i1851 = 0; _i1851 < _size1847; ++_i1851) + uint32_t _size1871; + ::apache::thrift::protocol::TType _etype1874; + xfer += iprot->readListBegin(_etype1874, _size1871); + (*(this->success)).resize(_size1871); + uint32_t _i1875; + for (_i1875 = 0; _i1875 < _size1871; ++_i1875) { - xfer += (*(this->success))[_i1851].read(iprot); + xfer += (*(this->success))[_i1875].read(iprot); } xfer += iprot->readListEnd(); } @@ -60620,6 +60807,64 @@ void ThriftHiveMetastoreClient::recv_flushCache() return; } +void ThriftHiveMetastoreClient::add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) +{ + send_add_write_notification_log(rqst); + recv_add_write_notification_log(_return); +} + +void ThriftHiveMetastoreClient::send_add_write_notification_log(const WriteNotificationLogRequest& rqst) +{ + int32_t cseqid = 0; + oprot_->writeMessageBegin("add_write_notification_log", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_add_write_notification_log_pargs args; + args.rqst = &rqst; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); +} + +void ThriftHiveMetastoreClient::recv_add_write_notification_log(WriteNotificationLogResponse& _return) +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + iprot_->readMessageBegin(fname, mtype, rseqid); + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + throw x; + } + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("add_write_notification_log") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + ThriftHiveMetastore_add_write_notification_log_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + // _return pointer has now been filled + return; + } + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "add_write_notification_log failed: unknown result"); +} + void ThriftHiveMetastoreClient::cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) { send_cm_recycle(request); @@ -73066,6 +73311,60 @@ void ThriftHiveMetastoreProcessor::process_flushCache(int32_t seqid, ::apache::t } } +void ThriftHiveMetastoreProcessor::process_add_write_notification_log(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext) +{ + void* ctx = NULL; + if (this->eventHandler_.get() != NULL) { + ctx = this->eventHandler_->getContext("ThriftHiveMetastore.add_write_notification_log", callContext); + } + ::apache::thrift::TProcessorContextFreer freer(this->eventHandler_.get(), ctx, "ThriftHiveMetastore.add_write_notification_log"); + + if (this->eventHandler_.get() != NULL) { + this->eventHandler_->preRead(ctx, "ThriftHiveMetastore.add_write_notification_log"); + } + + ThriftHiveMetastore_add_write_notification_log_args args; + args.read(iprot); + iprot->readMessageEnd(); + uint32_t bytes = iprot->getTransport()->readEnd(); + + if (this->eventHandler_.get() != NULL) { + this->eventHandler_->postRead(ctx, "ThriftHiveMetastore.add_write_notification_log", bytes); + } + + ThriftHiveMetastore_add_write_notification_log_result result; + try { + iface_->add_write_notification_log(result.success, args.rqst); + result.__isset.success = true; + } catch (const std::exception& e) { + if (this->eventHandler_.get() != NULL) { + this->eventHandler_->handlerError(ctx, "ThriftHiveMetastore.add_write_notification_log"); + } + + ::apache::thrift::TApplicationException x(e.what()); + oprot->writeMessageBegin("add_write_notification_log", ::apache::thrift::protocol::T_EXCEPTION, seqid); + x.write(oprot); + oprot->writeMessageEnd(); + oprot->getTransport()->writeEnd(); + oprot->getTransport()->flush(); + return; + } + + if (this->eventHandler_.get() != NULL) { + this->eventHandler_->preWrite(ctx, "ThriftHiveMetastore.add_write_notification_log"); + } + + oprot->writeMessageBegin("add_write_notification_log", ::apache::thrift::protocol::T_REPLY, seqid); + result.write(oprot); + oprot->writeMessageEnd(); + bytes = oprot->getTransport()->writeEnd(); + oprot->getTransport()->flush(); + + if (this->eventHandler_.get() != NULL) { + this->eventHandler_->postWrite(ctx, "ThriftHiveMetastore.add_write_notification_log", bytes); + } +} + void ThriftHiveMetastoreProcessor::process_cm_recycle(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext) { void* ctx = NULL; @@ -90573,6 +90872,90 @@ void ThriftHiveMetastoreConcurrentClient::recv_flushCache(const int32_t seqid) } // end while(true) } +void ThriftHiveMetastoreConcurrentClient::add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) +{ + int32_t seqid = send_add_write_notification_log(rqst); + recv_add_write_notification_log(_return, seqid); +} + +int32_t ThriftHiveMetastoreConcurrentClient::send_add_write_notification_log(const WriteNotificationLogRequest& rqst) +{ + int32_t cseqid = this->sync_.generateSeqId(); + ::apache::thrift::async::TConcurrentSendSentry sentry(&this->sync_); + oprot_->writeMessageBegin("add_write_notification_log", ::apache::thrift::protocol::T_CALL, cseqid); + + ThriftHiveMetastore_add_write_notification_log_pargs args; + args.rqst = &rqst; + args.write(oprot_); + + oprot_->writeMessageEnd(); + oprot_->getTransport()->writeEnd(); + oprot_->getTransport()->flush(); + + sentry.commit(); + return cseqid; +} + +void ThriftHiveMetastoreConcurrentClient::recv_add_write_notification_log(WriteNotificationLogResponse& _return, const int32_t seqid) +{ + + int32_t rseqid = 0; + std::string fname; + ::apache::thrift::protocol::TMessageType mtype; + + // the read mutex gets dropped and reacquired as part of waitForWork() + // The destructor of this sentry wakes up other clients + ::apache::thrift::async::TConcurrentRecvSentry sentry(&this->sync_, seqid); + + while(true) { + if(!this->sync_.getPending(fname, mtype, rseqid)) { + iprot_->readMessageBegin(fname, mtype, rseqid); + } + if(seqid == rseqid) { + if (mtype == ::apache::thrift::protocol::T_EXCEPTION) { + ::apache::thrift::TApplicationException x; + x.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + sentry.commit(); + throw x; + } + if (mtype != ::apache::thrift::protocol::T_REPLY) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + } + if (fname.compare("add_write_notification_log") != 0) { + iprot_->skip(::apache::thrift::protocol::T_STRUCT); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + // in a bad state, don't commit + using ::apache::thrift::protocol::TProtocolException; + throw TProtocolException(TProtocolException::INVALID_DATA); + } + ThriftHiveMetastore_add_write_notification_log_presult result; + result.success = &_return; + result.read(iprot_); + iprot_->readMessageEnd(); + iprot_->getTransport()->readEnd(); + + if (result.__isset.success) { + // _return pointer has now been filled + sentry.commit(); + return; + } + // in a bad state, don't commit + throw ::apache::thrift::TApplicationException(::apache::thrift::TApplicationException::MISSING_RESULT, "add_write_notification_log failed: unknown result"); + } + // seqid != rseqid + this->sync_.updatePending(fname, mtype, rseqid); + + // this will temporarily unlock the readMutex, and let other clients get work done + this->sync_.waitForWork(seqid); + } // end while(true) +} + void ThriftHiveMetastoreConcurrentClient::cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) { int32_t seqid = send_cm_recycle(request); diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h index 59b2876908..a547de1126 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h +++ b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h @@ -187,6 +187,7 @@ class ThriftHiveMetastoreIf : virtual public ::facebook::fb303::FacebookService virtual void get_notification_events_count(NotificationEventsCountResponse& _return, const NotificationEventsCountRequest& rqst) = 0; virtual void fire_listener_event(FireEventResponse& _return, const FireEventRequest& rqst) = 0; virtual void flushCache() = 0; + virtual void add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) = 0; virtual void cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) = 0; virtual void get_file_metadata_by_expr(GetFileMetadataByExprResult& _return, const GetFileMetadataByExprRequest& req) = 0; virtual void get_file_metadata(GetFileMetadataResult& _return, const GetFileMetadataRequest& req) = 0; @@ -780,6 +781,9 @@ class ThriftHiveMetastoreNull : virtual public ThriftHiveMetastoreIf , virtual p void flushCache() { return; } + void add_write_notification_log(WriteNotificationLogResponse& /* _return */, const WriteNotificationLogRequest& /* rqst */) { + return; + } void cm_recycle(CmRecycleResponse& /* _return */, const CmRecycleRequest& /* request */) { return; } @@ -21228,6 +21232,110 @@ class ThriftHiveMetastore_flushCache_presult { }; +typedef struct _ThriftHiveMetastore_add_write_notification_log_args__isset { + _ThriftHiveMetastore_add_write_notification_log_args__isset() : rqst(false) {} + bool rqst :1; +} _ThriftHiveMetastore_add_write_notification_log_args__isset; + +class ThriftHiveMetastore_add_write_notification_log_args { + public: + + ThriftHiveMetastore_add_write_notification_log_args(const ThriftHiveMetastore_add_write_notification_log_args&); + ThriftHiveMetastore_add_write_notification_log_args& operator=(const ThriftHiveMetastore_add_write_notification_log_args&); + ThriftHiveMetastore_add_write_notification_log_args() { + } + + virtual ~ThriftHiveMetastore_add_write_notification_log_args() throw(); + WriteNotificationLogRequest rqst; + + _ThriftHiveMetastore_add_write_notification_log_args__isset __isset; + + void __set_rqst(const WriteNotificationLogRequest& val); + + bool operator == (const ThriftHiveMetastore_add_write_notification_log_args & rhs) const + { + if (!(rqst == rhs.rqst)) + return false; + return true; + } + bool operator != (const ThriftHiveMetastore_add_write_notification_log_args &rhs) const { + return !(*this == rhs); + } + + bool operator < (const ThriftHiveMetastore_add_write_notification_log_args & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + + +class ThriftHiveMetastore_add_write_notification_log_pargs { + public: + + + virtual ~ThriftHiveMetastore_add_write_notification_log_pargs() throw(); + const WriteNotificationLogRequest* rqst; + + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + +typedef struct _ThriftHiveMetastore_add_write_notification_log_result__isset { + _ThriftHiveMetastore_add_write_notification_log_result__isset() : success(false) {} + bool success :1; +} _ThriftHiveMetastore_add_write_notification_log_result__isset; + +class ThriftHiveMetastore_add_write_notification_log_result { + public: + + ThriftHiveMetastore_add_write_notification_log_result(const ThriftHiveMetastore_add_write_notification_log_result&); + ThriftHiveMetastore_add_write_notification_log_result& operator=(const ThriftHiveMetastore_add_write_notification_log_result&); + ThriftHiveMetastore_add_write_notification_log_result() { + } + + virtual ~ThriftHiveMetastore_add_write_notification_log_result() throw(); + WriteNotificationLogResponse success; + + _ThriftHiveMetastore_add_write_notification_log_result__isset __isset; + + void __set_success(const WriteNotificationLogResponse& val); + + bool operator == (const ThriftHiveMetastore_add_write_notification_log_result & rhs) const + { + if (!(success == rhs.success)) + return false; + return true; + } + bool operator != (const ThriftHiveMetastore_add_write_notification_log_result &rhs) const { + return !(*this == rhs); + } + + bool operator < (const ThriftHiveMetastore_add_write_notification_log_result & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + +}; + +typedef struct _ThriftHiveMetastore_add_write_notification_log_presult__isset { + _ThriftHiveMetastore_add_write_notification_log_presult__isset() : success(false) {} + bool success :1; +} _ThriftHiveMetastore_add_write_notification_log_presult__isset; + +class ThriftHiveMetastore_add_write_notification_log_presult { + public: + + + virtual ~ThriftHiveMetastore_add_write_notification_log_presult() throw(); + WriteNotificationLogResponse* success; + + _ThriftHiveMetastore_add_write_notification_log_presult__isset __isset; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + +}; + typedef struct _ThriftHiveMetastore_cm_recycle_args__isset { _ThriftHiveMetastore_cm_recycle_args__isset() : request(false) {} bool request :1; @@ -26718,6 +26826,9 @@ class ThriftHiveMetastoreClient : virtual public ThriftHiveMetastoreIf, public void flushCache(); void send_flushCache(); void recv_flushCache(); + void add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst); + void send_add_write_notification_log(const WriteNotificationLogRequest& rqst); + void recv_add_write_notification_log(WriteNotificationLogResponse& _return); void cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request); void send_cm_recycle(const CmRecycleRequest& request); void recv_cm_recycle(CmRecycleResponse& _return); @@ -27019,6 +27130,7 @@ class ThriftHiveMetastoreProcessor : public ::facebook::fb303::FacebookServiceP void process_get_notification_events_count(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_fire_listener_event(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_flushCache(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); + void process_add_write_notification_log(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_cm_recycle(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_get_file_metadata_by_expr(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); void process_get_file_metadata(int32_t seqid, ::apache::thrift::protocol::TProtocol* iprot, ::apache::thrift::protocol::TProtocol* oprot, void* callContext); @@ -27230,6 +27342,7 @@ class ThriftHiveMetastoreProcessor : public ::facebook::fb303::FacebookServiceP processMap_["get_notification_events_count"] = &ThriftHiveMetastoreProcessor::process_get_notification_events_count; processMap_["fire_listener_event"] = &ThriftHiveMetastoreProcessor::process_fire_listener_event; processMap_["flushCache"] = &ThriftHiveMetastoreProcessor::process_flushCache; + processMap_["add_write_notification_log"] = &ThriftHiveMetastoreProcessor::process_add_write_notification_log; processMap_["cm_recycle"] = &ThriftHiveMetastoreProcessor::process_cm_recycle; processMap_["get_file_metadata_by_expr"] = &ThriftHiveMetastoreProcessor::process_get_file_metadata_by_expr; processMap_["get_file_metadata"] = &ThriftHiveMetastoreProcessor::process_get_file_metadata; @@ -28884,6 +28997,16 @@ class ThriftHiveMetastoreMultiface : virtual public ThriftHiveMetastoreIf, publi ifaces_[i]->flushCache(); } + void add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) { + size_t sz = ifaces_.size(); + size_t i = 0; + for (; i < (sz - 1); ++i) { + ifaces_[i]->add_write_notification_log(_return, rqst); + } + ifaces_[i]->add_write_notification_log(_return, rqst); + return; + } + void cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) { size_t sz = ifaces_.size(); size_t i = 0; @@ -29805,6 +29928,9 @@ class ThriftHiveMetastoreConcurrentClient : virtual public ThriftHiveMetastoreIf void flushCache(); int32_t send_flushCache(); void recv_flushCache(const int32_t seqid); + void add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst); + int32_t send_add_write_notification_log(const WriteNotificationLogRequest& rqst); + void recv_add_write_notification_log(WriteNotificationLogResponse& _return, const int32_t seqid); void cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request); int32_t send_cm_recycle(const CmRecycleRequest& request); void recv_cm_recycle(CmRecycleResponse& _return, const int32_t seqid); diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp index d01b63922c..5819b1742f 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp +++ b/standalone-metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp @@ -847,6 +847,11 @@ class ThriftHiveMetastoreHandler : virtual public ThriftHiveMetastoreIf { printf("flushCache\n"); } + void add_write_notification_log(WriteNotificationLogResponse& _return, const WriteNotificationLogRequest& rqst) { + // Your implementation goes here + printf("add_write_notification_log\n"); + } + void cm_recycle(CmRecycleResponse& _return, const CmRecycleRequest& request) { // Your implementation goes here printf("cm_recycle\n"); diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp index 898b911750..26420ddf4c 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp +++ b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp @@ -16592,6 +16592,11 @@ void CommitTxnRequest::__set_replPolicy(const std::string& val) { __isset.replPolicy = true; } +void CommitTxnRequest::__set_writeEventInfos(const std::vector & val) { + this->writeEventInfos = val; +__isset.writeEventInfos = true; +} + uint32_t CommitTxnRequest::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); @@ -16630,6 +16635,26 @@ uint32_t CommitTxnRequest::read(::apache::thrift::protocol::TProtocol* iprot) { xfer += iprot->skip(ftype); } break; + case 3: + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + this->writeEventInfos.clear(); + uint32_t _size673; + ::apache::thrift::protocol::TType _etype676; + xfer += iprot->readListBegin(_etype676, _size673); + this->writeEventInfos.resize(_size673); + uint32_t _i677; + for (_i677 = 0; _i677 < _size673; ++_i677) + { + xfer += this->writeEventInfos[_i677].read(iprot); + } + xfer += iprot->readListEnd(); + } + this->__isset.writeEventInfos = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -16658,6 +16683,19 @@ uint32_t CommitTxnRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeString(this->replPolicy); xfer += oprot->writeFieldEnd(); } + if (this->__isset.writeEventInfos) { + xfer += oprot->writeFieldBegin("writeEventInfos", ::apache::thrift::protocol::T_LIST, 3); + { + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->writeEventInfos.size())); + std::vector ::const_iterator _iter678; + for (_iter678 = this->writeEventInfos.begin(); _iter678 != this->writeEventInfos.end(); ++_iter678) + { + xfer += (*_iter678).write(oprot); + } + xfer += oprot->writeListEnd(); + } + xfer += oprot->writeFieldEnd(); + } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; @@ -16667,18 +16705,21 @@ void swap(CommitTxnRequest &a, CommitTxnRequest &b) { using ::std::swap; swap(a.txnid, b.txnid); swap(a.replPolicy, b.replPolicy); + swap(a.writeEventInfos, b.writeEventInfos); swap(a.__isset, b.__isset); } -CommitTxnRequest::CommitTxnRequest(const CommitTxnRequest& other673) { - txnid = other673.txnid; - replPolicy = other673.replPolicy; - __isset = other673.__isset; +CommitTxnRequest::CommitTxnRequest(const CommitTxnRequest& other679) { + txnid = other679.txnid; + replPolicy = other679.replPolicy; + writeEventInfos = other679.writeEventInfos; + __isset = other679.__isset; } -CommitTxnRequest& CommitTxnRequest::operator=(const CommitTxnRequest& other674) { - txnid = other674.txnid; - replPolicy = other674.replPolicy; - __isset = other674.__isset; +CommitTxnRequest& CommitTxnRequest::operator=(const CommitTxnRequest& other680) { + txnid = other680.txnid; + replPolicy = other680.replPolicy; + writeEventInfos = other680.writeEventInfos; + __isset = other680.__isset; return *this; } void CommitTxnRequest::printTo(std::ostream& out) const { @@ -16686,6 +16727,231 @@ void CommitTxnRequest::printTo(std::ostream& out) const { out << "CommitTxnRequest("; out << "txnid=" << to_string(txnid); out << ", " << "replPolicy="; (__isset.replPolicy ? (out << to_string(replPolicy)) : (out << "")); + out << ", " << "writeEventInfos="; (__isset.writeEventInfos ? (out << to_string(writeEventInfos)) : (out << "")); + out << ")"; +} + + +WriteEventInfo::~WriteEventInfo() throw() { +} + + +void WriteEventInfo::__set_writeId(const int64_t val) { + this->writeId = val; +} + +void WriteEventInfo::__set_database(const std::string& val) { + this->database = val; +} + +void WriteEventInfo::__set_table(const std::string& val) { + this->table = val; +} + +void WriteEventInfo::__set_files(const std::string& val) { + this->files = val; +} + +void WriteEventInfo::__set_partition(const std::string& val) { + this->partition = val; +__isset.partition = true; +} + +void WriteEventInfo::__set_tableObj(const std::string& val) { + this->tableObj = val; +__isset.tableObj = true; +} + +void WriteEventInfo::__set_partitionObj(const std::string& val) { + this->partitionObj = val; +__isset.partitionObj = true; +} + +uint32_t WriteEventInfo::read(::apache::thrift::protocol::TProtocol* iprot) { + + apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + bool isset_writeId = false; + bool isset_database = false; + bool isset_table = false; + bool isset_files = false; + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 1: + if (ftype == ::apache::thrift::protocol::T_I64) { + xfer += iprot->readI64(this->writeId); + isset_writeId = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->database); + isset_database = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 3: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->table); + isset_table = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 4: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->files); + isset_files = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 5: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->partition); + this->__isset.partition = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 6: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->tableObj); + this->__isset.tableObj = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 7: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->partitionObj); + this->__isset.partitionObj = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + if (!isset_writeId) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_database) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_table) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_files) + throw TProtocolException(TProtocolException::INVALID_DATA); + return xfer; +} + +uint32_t WriteEventInfo::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("WriteEventInfo"); + + xfer += oprot->writeFieldBegin("writeId", ::apache::thrift::protocol::T_I64, 1); + xfer += oprot->writeI64(this->writeId); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("database", ::apache::thrift::protocol::T_STRING, 2); + xfer += oprot->writeString(this->database); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("table", ::apache::thrift::protocol::T_STRING, 3); + xfer += oprot->writeString(this->table); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("files", ::apache::thrift::protocol::T_STRING, 4); + xfer += oprot->writeString(this->files); + xfer += oprot->writeFieldEnd(); + + if (this->__isset.partition) { + xfer += oprot->writeFieldBegin("partition", ::apache::thrift::protocol::T_STRING, 5); + xfer += oprot->writeString(this->partition); + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.tableObj) { + xfer += oprot->writeFieldBegin("tableObj", ::apache::thrift::protocol::T_STRING, 6); + xfer += oprot->writeString(this->tableObj); + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.partitionObj) { + xfer += oprot->writeFieldBegin("partitionObj", ::apache::thrift::protocol::T_STRING, 7); + xfer += oprot->writeString(this->partitionObj); + xfer += oprot->writeFieldEnd(); + } + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + +void swap(WriteEventInfo &a, WriteEventInfo &b) { + using ::std::swap; + swap(a.writeId, b.writeId); + swap(a.database, b.database); + swap(a.table, b.table); + swap(a.files, b.files); + swap(a.partition, b.partition); + swap(a.tableObj, b.tableObj); + swap(a.partitionObj, b.partitionObj); + swap(a.__isset, b.__isset); +} + +WriteEventInfo::WriteEventInfo(const WriteEventInfo& other681) { + writeId = other681.writeId; + database = other681.database; + table = other681.table; + files = other681.files; + partition = other681.partition; + tableObj = other681.tableObj; + partitionObj = other681.partitionObj; + __isset = other681.__isset; +} +WriteEventInfo& WriteEventInfo::operator=(const WriteEventInfo& other682) { + writeId = other682.writeId; + database = other682.database; + table = other682.table; + files = other682.files; + partition = other682.partition; + tableObj = other682.tableObj; + partitionObj = other682.partitionObj; + __isset = other682.__isset; + return *this; +} +void WriteEventInfo::printTo(std::ostream& out) const { + using ::apache::thrift::to_string; + out << "WriteEventInfo("; + out << "writeId=" << to_string(writeId); + out << ", " << "database=" << to_string(database); + out << ", " << "table=" << to_string(table); + out << ", " << "files=" << to_string(files); + out << ", " << "partition="; (__isset.partition ? (out << to_string(partition)) : (out << "")); + out << ", " << "tableObj="; (__isset.tableObj ? (out << to_string(tableObj)) : (out << "")); + out << ", " << "partitionObj="; (__isset.partitionObj ? (out << to_string(partitionObj)) : (out << "")); out << ")"; } @@ -16789,14 +17055,14 @@ uint32_t ReplTblWriteIdStateRequest::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partNames.clear(); - uint32_t _size675; - ::apache::thrift::protocol::TType _etype678; - xfer += iprot->readListBegin(_etype678, _size675); - this->partNames.resize(_size675); - uint32_t _i679; - for (_i679 = 0; _i679 < _size675; ++_i679) + uint32_t _size683; + ::apache::thrift::protocol::TType _etype686; + xfer += iprot->readListBegin(_etype686, _size683); + this->partNames.resize(_size683); + uint32_t _i687; + for (_i687 = 0; _i687 < _size683; ++_i687) { - xfer += iprot->readString(this->partNames[_i679]); + xfer += iprot->readString(this->partNames[_i687]); } xfer += iprot->readListEnd(); } @@ -16856,10 +17122,10 @@ uint32_t ReplTblWriteIdStateRequest::write(::apache::thrift::protocol::TProtocol xfer += oprot->writeFieldBegin("partNames", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partNames.size())); - std::vector ::const_iterator _iter680; - for (_iter680 = this->partNames.begin(); _iter680 != this->partNames.end(); ++_iter680) + std::vector ::const_iterator _iter688; + for (_iter688 = this->partNames.begin(); _iter688 != this->partNames.end(); ++_iter688) { - xfer += oprot->writeString((*_iter680)); + xfer += oprot->writeString((*_iter688)); } xfer += oprot->writeListEnd(); } @@ -16881,23 +17147,23 @@ void swap(ReplTblWriteIdStateRequest &a, ReplTblWriteIdStateRequest &b) { swap(a.__isset, b.__isset); } -ReplTblWriteIdStateRequest::ReplTblWriteIdStateRequest(const ReplTblWriteIdStateRequest& other681) { - validWriteIdlist = other681.validWriteIdlist; - user = other681.user; - hostName = other681.hostName; - dbName = other681.dbName; - tableName = other681.tableName; - partNames = other681.partNames; - __isset = other681.__isset; -} -ReplTblWriteIdStateRequest& ReplTblWriteIdStateRequest::operator=(const ReplTblWriteIdStateRequest& other682) { - validWriteIdlist = other682.validWriteIdlist; - user = other682.user; - hostName = other682.hostName; - dbName = other682.dbName; - tableName = other682.tableName; - partNames = other682.partNames; - __isset = other682.__isset; +ReplTblWriteIdStateRequest::ReplTblWriteIdStateRequest(const ReplTblWriteIdStateRequest& other689) { + validWriteIdlist = other689.validWriteIdlist; + user = other689.user; + hostName = other689.hostName; + dbName = other689.dbName; + tableName = other689.tableName; + partNames = other689.partNames; + __isset = other689.__isset; +} +ReplTblWriteIdStateRequest& ReplTblWriteIdStateRequest::operator=(const ReplTblWriteIdStateRequest& other690) { + validWriteIdlist = other690.validWriteIdlist; + user = other690.user; + hostName = other690.hostName; + dbName = other690.dbName; + tableName = other690.tableName; + partNames = other690.partNames; + __isset = other690.__isset; return *this; } void ReplTblWriteIdStateRequest::printTo(std::ostream& out) const { @@ -16952,14 +17218,14 @@ uint32_t GetValidWriteIdsRequest::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fullTableNames.clear(); - uint32_t _size683; - ::apache::thrift::protocol::TType _etype686; - xfer += iprot->readListBegin(_etype686, _size683); - this->fullTableNames.resize(_size683); - uint32_t _i687; - for (_i687 = 0; _i687 < _size683; ++_i687) + uint32_t _size691; + ::apache::thrift::protocol::TType _etype694; + xfer += iprot->readListBegin(_etype694, _size691); + this->fullTableNames.resize(_size691); + uint32_t _i695; + for (_i695 = 0; _i695 < _size691; ++_i695) { - xfer += iprot->readString(this->fullTableNames[_i687]); + xfer += iprot->readString(this->fullTableNames[_i695]); } xfer += iprot->readListEnd(); } @@ -17000,10 +17266,10 @@ uint32_t GetValidWriteIdsRequest::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("fullTableNames", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->fullTableNames.size())); - std::vector ::const_iterator _iter688; - for (_iter688 = this->fullTableNames.begin(); _iter688 != this->fullTableNames.end(); ++_iter688) + std::vector ::const_iterator _iter696; + for (_iter696 = this->fullTableNames.begin(); _iter696 != this->fullTableNames.end(); ++_iter696) { - xfer += oprot->writeString((*_iter688)); + xfer += oprot->writeString((*_iter696)); } xfer += oprot->writeListEnd(); } @@ -17024,13 +17290,13 @@ void swap(GetValidWriteIdsRequest &a, GetValidWriteIdsRequest &b) { swap(a.validTxnList, b.validTxnList); } -GetValidWriteIdsRequest::GetValidWriteIdsRequest(const GetValidWriteIdsRequest& other689) { - fullTableNames = other689.fullTableNames; - validTxnList = other689.validTxnList; +GetValidWriteIdsRequest::GetValidWriteIdsRequest(const GetValidWriteIdsRequest& other697) { + fullTableNames = other697.fullTableNames; + validTxnList = other697.validTxnList; } -GetValidWriteIdsRequest& GetValidWriteIdsRequest::operator=(const GetValidWriteIdsRequest& other690) { - fullTableNames = other690.fullTableNames; - validTxnList = other690.validTxnList; +GetValidWriteIdsRequest& GetValidWriteIdsRequest::operator=(const GetValidWriteIdsRequest& other698) { + fullTableNames = other698.fullTableNames; + validTxnList = other698.validTxnList; return *this; } void GetValidWriteIdsRequest::printTo(std::ostream& out) const { @@ -17112,14 +17378,14 @@ uint32_t TableValidWriteIds::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->invalidWriteIds.clear(); - uint32_t _size691; - ::apache::thrift::protocol::TType _etype694; - xfer += iprot->readListBegin(_etype694, _size691); - this->invalidWriteIds.resize(_size691); - uint32_t _i695; - for (_i695 = 0; _i695 < _size691; ++_i695) + uint32_t _size699; + ::apache::thrift::protocol::TType _etype702; + xfer += iprot->readListBegin(_etype702, _size699); + this->invalidWriteIds.resize(_size699); + uint32_t _i703; + for (_i703 = 0; _i703 < _size699; ++_i703) { - xfer += iprot->readI64(this->invalidWriteIds[_i695]); + xfer += iprot->readI64(this->invalidWriteIds[_i703]); } xfer += iprot->readListEnd(); } @@ -17180,10 +17446,10 @@ uint32_t TableValidWriteIds::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("invalidWriteIds", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->invalidWriteIds.size())); - std::vector ::const_iterator _iter696; - for (_iter696 = this->invalidWriteIds.begin(); _iter696 != this->invalidWriteIds.end(); ++_iter696) + std::vector ::const_iterator _iter704; + for (_iter704 = this->invalidWriteIds.begin(); _iter704 != this->invalidWriteIds.end(); ++_iter704) { - xfer += oprot->writeI64((*_iter696)); + xfer += oprot->writeI64((*_iter704)); } xfer += oprot->writeListEnd(); } @@ -17213,21 +17479,21 @@ void swap(TableValidWriteIds &a, TableValidWriteIds &b) { swap(a.__isset, b.__isset); } -TableValidWriteIds::TableValidWriteIds(const TableValidWriteIds& other697) { - fullTableName = other697.fullTableName; - writeIdHighWaterMark = other697.writeIdHighWaterMark; - invalidWriteIds = other697.invalidWriteIds; - minOpenWriteId = other697.minOpenWriteId; - abortedBits = other697.abortedBits; - __isset = other697.__isset; -} -TableValidWriteIds& TableValidWriteIds::operator=(const TableValidWriteIds& other698) { - fullTableName = other698.fullTableName; - writeIdHighWaterMark = other698.writeIdHighWaterMark; - invalidWriteIds = other698.invalidWriteIds; - minOpenWriteId = other698.minOpenWriteId; - abortedBits = other698.abortedBits; - __isset = other698.__isset; +TableValidWriteIds::TableValidWriteIds(const TableValidWriteIds& other705) { + fullTableName = other705.fullTableName; + writeIdHighWaterMark = other705.writeIdHighWaterMark; + invalidWriteIds = other705.invalidWriteIds; + minOpenWriteId = other705.minOpenWriteId; + abortedBits = other705.abortedBits; + __isset = other705.__isset; +} +TableValidWriteIds& TableValidWriteIds::operator=(const TableValidWriteIds& other706) { + fullTableName = other706.fullTableName; + writeIdHighWaterMark = other706.writeIdHighWaterMark; + invalidWriteIds = other706.invalidWriteIds; + minOpenWriteId = other706.minOpenWriteId; + abortedBits = other706.abortedBits; + __isset = other706.__isset; return *this; } void TableValidWriteIds::printTo(std::ostream& out) const { @@ -17276,14 +17542,14 @@ uint32_t GetValidWriteIdsResponse::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tblValidWriteIds.clear(); - uint32_t _size699; - ::apache::thrift::protocol::TType _etype702; - xfer += iprot->readListBegin(_etype702, _size699); - this->tblValidWriteIds.resize(_size699); - uint32_t _i703; - for (_i703 = 0; _i703 < _size699; ++_i703) + uint32_t _size707; + ::apache::thrift::protocol::TType _etype710; + xfer += iprot->readListBegin(_etype710, _size707); + this->tblValidWriteIds.resize(_size707); + uint32_t _i711; + for (_i711 = 0; _i711 < _size707; ++_i711) { - xfer += this->tblValidWriteIds[_i703].read(iprot); + xfer += this->tblValidWriteIds[_i711].read(iprot); } xfer += iprot->readListEnd(); } @@ -17314,10 +17580,10 @@ uint32_t GetValidWriteIdsResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("tblValidWriteIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->tblValidWriteIds.size())); - std::vector ::const_iterator _iter704; - for (_iter704 = this->tblValidWriteIds.begin(); _iter704 != this->tblValidWriteIds.end(); ++_iter704) + std::vector ::const_iterator _iter712; + for (_iter712 = this->tblValidWriteIds.begin(); _iter712 != this->tblValidWriteIds.end(); ++_iter712) { - xfer += (*_iter704).write(oprot); + xfer += (*_iter712).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17333,11 +17599,11 @@ void swap(GetValidWriteIdsResponse &a, GetValidWriteIdsResponse &b) { swap(a.tblValidWriteIds, b.tblValidWriteIds); } -GetValidWriteIdsResponse::GetValidWriteIdsResponse(const GetValidWriteIdsResponse& other705) { - tblValidWriteIds = other705.tblValidWriteIds; +GetValidWriteIdsResponse::GetValidWriteIdsResponse(const GetValidWriteIdsResponse& other713) { + tblValidWriteIds = other713.tblValidWriteIds; } -GetValidWriteIdsResponse& GetValidWriteIdsResponse::operator=(const GetValidWriteIdsResponse& other706) { - tblValidWriteIds = other706.tblValidWriteIds; +GetValidWriteIdsResponse& GetValidWriteIdsResponse::operator=(const GetValidWriteIdsResponse& other714) { + tblValidWriteIds = other714.tblValidWriteIds; return *this; } void GetValidWriteIdsResponse::printTo(std::ostream& out) const { @@ -17418,14 +17684,14 @@ uint32_t AllocateTableWriteIdsRequest::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->txnIds.clear(); - uint32_t _size707; - ::apache::thrift::protocol::TType _etype710; - xfer += iprot->readListBegin(_etype710, _size707); - this->txnIds.resize(_size707); - uint32_t _i711; - for (_i711 = 0; _i711 < _size707; ++_i711) + uint32_t _size715; + ::apache::thrift::protocol::TType _etype718; + xfer += iprot->readListBegin(_etype718, _size715); + this->txnIds.resize(_size715); + uint32_t _i719; + for (_i719 = 0; _i719 < _size715; ++_i719) { - xfer += iprot->readI64(this->txnIds[_i711]); + xfer += iprot->readI64(this->txnIds[_i719]); } xfer += iprot->readListEnd(); } @@ -17446,14 +17712,14 @@ uint32_t AllocateTableWriteIdsRequest::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->srcTxnToWriteIdList.clear(); - uint32_t _size712; - ::apache::thrift::protocol::TType _etype715; - xfer += iprot->readListBegin(_etype715, _size712); - this->srcTxnToWriteIdList.resize(_size712); - uint32_t _i716; - for (_i716 = 0; _i716 < _size712; ++_i716) + uint32_t _size720; + ::apache::thrift::protocol::TType _etype723; + xfer += iprot->readListBegin(_etype723, _size720); + this->srcTxnToWriteIdList.resize(_size720); + uint32_t _i724; + for (_i724 = 0; _i724 < _size720; ++_i724) { - xfer += this->srcTxnToWriteIdList[_i716].read(iprot); + xfer += this->srcTxnToWriteIdList[_i724].read(iprot); } xfer += iprot->readListEnd(); } @@ -17495,10 +17761,10 @@ uint32_t AllocateTableWriteIdsRequest::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("txnIds", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->txnIds.size())); - std::vector ::const_iterator _iter717; - for (_iter717 = this->txnIds.begin(); _iter717 != this->txnIds.end(); ++_iter717) + std::vector ::const_iterator _iter725; + for (_iter725 = this->txnIds.begin(); _iter725 != this->txnIds.end(); ++_iter725) { - xfer += oprot->writeI64((*_iter717)); + xfer += oprot->writeI64((*_iter725)); } xfer += oprot->writeListEnd(); } @@ -17513,10 +17779,10 @@ uint32_t AllocateTableWriteIdsRequest::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("srcTxnToWriteIdList", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->srcTxnToWriteIdList.size())); - std::vector ::const_iterator _iter718; - for (_iter718 = this->srcTxnToWriteIdList.begin(); _iter718 != this->srcTxnToWriteIdList.end(); ++_iter718) + std::vector ::const_iterator _iter726; + for (_iter726 = this->srcTxnToWriteIdList.begin(); _iter726 != this->srcTxnToWriteIdList.end(); ++_iter726) { - xfer += (*_iter718).write(oprot); + xfer += (*_iter726).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17537,21 +17803,21 @@ void swap(AllocateTableWriteIdsRequest &a, AllocateTableWriteIdsRequest &b) { swap(a.__isset, b.__isset); } -AllocateTableWriteIdsRequest::AllocateTableWriteIdsRequest(const AllocateTableWriteIdsRequest& other719) { - dbName = other719.dbName; - tableName = other719.tableName; - txnIds = other719.txnIds; - replPolicy = other719.replPolicy; - srcTxnToWriteIdList = other719.srcTxnToWriteIdList; - __isset = other719.__isset; -} -AllocateTableWriteIdsRequest& AllocateTableWriteIdsRequest::operator=(const AllocateTableWriteIdsRequest& other720) { - dbName = other720.dbName; - tableName = other720.tableName; - txnIds = other720.txnIds; - replPolicy = other720.replPolicy; - srcTxnToWriteIdList = other720.srcTxnToWriteIdList; - __isset = other720.__isset; +AllocateTableWriteIdsRequest::AllocateTableWriteIdsRequest(const AllocateTableWriteIdsRequest& other727) { + dbName = other727.dbName; + tableName = other727.tableName; + txnIds = other727.txnIds; + replPolicy = other727.replPolicy; + srcTxnToWriteIdList = other727.srcTxnToWriteIdList; + __isset = other727.__isset; +} +AllocateTableWriteIdsRequest& AllocateTableWriteIdsRequest::operator=(const AllocateTableWriteIdsRequest& other728) { + dbName = other728.dbName; + tableName = other728.tableName; + txnIds = other728.txnIds; + replPolicy = other728.replPolicy; + srcTxnToWriteIdList = other728.srcTxnToWriteIdList; + __isset = other728.__isset; return *this; } void AllocateTableWriteIdsRequest::printTo(std::ostream& out) const { @@ -17657,13 +17923,13 @@ void swap(TxnToWriteId &a, TxnToWriteId &b) { swap(a.writeId, b.writeId); } -TxnToWriteId::TxnToWriteId(const TxnToWriteId& other721) { - txnId = other721.txnId; - writeId = other721.writeId; +TxnToWriteId::TxnToWriteId(const TxnToWriteId& other729) { + txnId = other729.txnId; + writeId = other729.writeId; } -TxnToWriteId& TxnToWriteId::operator=(const TxnToWriteId& other722) { - txnId = other722.txnId; - writeId = other722.writeId; +TxnToWriteId& TxnToWriteId::operator=(const TxnToWriteId& other730) { + txnId = other730.txnId; + writeId = other730.writeId; return *this; } void TxnToWriteId::printTo(std::ostream& out) const { @@ -17709,14 +17975,14 @@ uint32_t AllocateTableWriteIdsResponse::read(::apache::thrift::protocol::TProtoc if (ftype == ::apache::thrift::protocol::T_LIST) { { this->txnToWriteIds.clear(); - uint32_t _size723; - ::apache::thrift::protocol::TType _etype726; - xfer += iprot->readListBegin(_etype726, _size723); - this->txnToWriteIds.resize(_size723); - uint32_t _i727; - for (_i727 = 0; _i727 < _size723; ++_i727) + uint32_t _size731; + ::apache::thrift::protocol::TType _etype734; + xfer += iprot->readListBegin(_etype734, _size731); + this->txnToWriteIds.resize(_size731); + uint32_t _i735; + for (_i735 = 0; _i735 < _size731; ++_i735) { - xfer += this->txnToWriteIds[_i727].read(iprot); + xfer += this->txnToWriteIds[_i735].read(iprot); } xfer += iprot->readListEnd(); } @@ -17747,10 +18013,10 @@ uint32_t AllocateTableWriteIdsResponse::write(::apache::thrift::protocol::TProto xfer += oprot->writeFieldBegin("txnToWriteIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->txnToWriteIds.size())); - std::vector ::const_iterator _iter728; - for (_iter728 = this->txnToWriteIds.begin(); _iter728 != this->txnToWriteIds.end(); ++_iter728) + std::vector ::const_iterator _iter736; + for (_iter736 = this->txnToWriteIds.begin(); _iter736 != this->txnToWriteIds.end(); ++_iter736) { - xfer += (*_iter728).write(oprot); + xfer += (*_iter736).write(oprot); } xfer += oprot->writeListEnd(); } @@ -17766,11 +18032,11 @@ void swap(AllocateTableWriteIdsResponse &a, AllocateTableWriteIdsResponse &b) { swap(a.txnToWriteIds, b.txnToWriteIds); } -AllocateTableWriteIdsResponse::AllocateTableWriteIdsResponse(const AllocateTableWriteIdsResponse& other729) { - txnToWriteIds = other729.txnToWriteIds; +AllocateTableWriteIdsResponse::AllocateTableWriteIdsResponse(const AllocateTableWriteIdsResponse& other737) { + txnToWriteIds = other737.txnToWriteIds; } -AllocateTableWriteIdsResponse& AllocateTableWriteIdsResponse::operator=(const AllocateTableWriteIdsResponse& other730) { - txnToWriteIds = other730.txnToWriteIds; +AllocateTableWriteIdsResponse& AllocateTableWriteIdsResponse::operator=(const AllocateTableWriteIdsResponse& other738) { + txnToWriteIds = other738.txnToWriteIds; return *this; } void AllocateTableWriteIdsResponse::printTo(std::ostream& out) const { @@ -17848,9 +18114,9 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { { case 1: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast731; - xfer += iprot->readI32(ecast731); - this->type = (LockType::type)ecast731; + int32_t ecast739; + xfer += iprot->readI32(ecast739); + this->type = (LockType::type)ecast739; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -17858,9 +18124,9 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast732; - xfer += iprot->readI32(ecast732); - this->level = (LockLevel::type)ecast732; + int32_t ecast740; + xfer += iprot->readI32(ecast740); + this->level = (LockLevel::type)ecast740; isset_level = true; } else { xfer += iprot->skip(ftype); @@ -17892,9 +18158,9 @@ uint32_t LockComponent::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast733; - xfer += iprot->readI32(ecast733); - this->operationType = (DataOperationType::type)ecast733; + int32_t ecast741; + xfer += iprot->readI32(ecast741); + this->operationType = (DataOperationType::type)ecast741; this->__isset.operationType = true; } else { xfer += iprot->skip(ftype); @@ -17994,27 +18260,27 @@ void swap(LockComponent &a, LockComponent &b) { swap(a.__isset, b.__isset); } -LockComponent::LockComponent(const LockComponent& other734) { - type = other734.type; - level = other734.level; - dbname = other734.dbname; - tablename = other734.tablename; - partitionname = other734.partitionname; - operationType = other734.operationType; - isTransactional = other734.isTransactional; - isDynamicPartitionWrite = other734.isDynamicPartitionWrite; - __isset = other734.__isset; -} -LockComponent& LockComponent::operator=(const LockComponent& other735) { - type = other735.type; - level = other735.level; - dbname = other735.dbname; - tablename = other735.tablename; - partitionname = other735.partitionname; - operationType = other735.operationType; - isTransactional = other735.isTransactional; - isDynamicPartitionWrite = other735.isDynamicPartitionWrite; - __isset = other735.__isset; +LockComponent::LockComponent(const LockComponent& other742) { + type = other742.type; + level = other742.level; + dbname = other742.dbname; + tablename = other742.tablename; + partitionname = other742.partitionname; + operationType = other742.operationType; + isTransactional = other742.isTransactional; + isDynamicPartitionWrite = other742.isDynamicPartitionWrite; + __isset = other742.__isset; +} +LockComponent& LockComponent::operator=(const LockComponent& other743) { + type = other743.type; + level = other743.level; + dbname = other743.dbname; + tablename = other743.tablename; + partitionname = other743.partitionname; + operationType = other743.operationType; + isTransactional = other743.isTransactional; + isDynamicPartitionWrite = other743.isDynamicPartitionWrite; + __isset = other743.__isset; return *this; } void LockComponent::printTo(std::ostream& out) const { @@ -18086,14 +18352,14 @@ uint32_t LockRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->component.clear(); - uint32_t _size736; - ::apache::thrift::protocol::TType _etype739; - xfer += iprot->readListBegin(_etype739, _size736); - this->component.resize(_size736); - uint32_t _i740; - for (_i740 = 0; _i740 < _size736; ++_i740) + uint32_t _size744; + ::apache::thrift::protocol::TType _etype747; + xfer += iprot->readListBegin(_etype747, _size744); + this->component.resize(_size744); + uint32_t _i748; + for (_i748 = 0; _i748 < _size744; ++_i748) { - xfer += this->component[_i740].read(iprot); + xfer += this->component[_i748].read(iprot); } xfer += iprot->readListEnd(); } @@ -18160,10 +18426,10 @@ uint32_t LockRequest::write(::apache::thrift::protocol::TProtocol* oprot) const xfer += oprot->writeFieldBegin("component", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->component.size())); - std::vector ::const_iterator _iter741; - for (_iter741 = this->component.begin(); _iter741 != this->component.end(); ++_iter741) + std::vector ::const_iterator _iter749; + for (_iter749 = this->component.begin(); _iter749 != this->component.end(); ++_iter749) { - xfer += (*_iter741).write(oprot); + xfer += (*_iter749).write(oprot); } xfer += oprot->writeListEnd(); } @@ -18202,21 +18468,21 @@ void swap(LockRequest &a, LockRequest &b) { swap(a.__isset, b.__isset); } -LockRequest::LockRequest(const LockRequest& other742) { - component = other742.component; - txnid = other742.txnid; - user = other742.user; - hostname = other742.hostname; - agentInfo = other742.agentInfo; - __isset = other742.__isset; -} -LockRequest& LockRequest::operator=(const LockRequest& other743) { - component = other743.component; - txnid = other743.txnid; - user = other743.user; - hostname = other743.hostname; - agentInfo = other743.agentInfo; - __isset = other743.__isset; +LockRequest::LockRequest(const LockRequest& other750) { + component = other750.component; + txnid = other750.txnid; + user = other750.user; + hostname = other750.hostname; + agentInfo = other750.agentInfo; + __isset = other750.__isset; +} +LockRequest& LockRequest::operator=(const LockRequest& other751) { + component = other751.component; + txnid = other751.txnid; + user = other751.user; + hostname = other751.hostname; + agentInfo = other751.agentInfo; + __isset = other751.__isset; return *this; } void LockRequest::printTo(std::ostream& out) const { @@ -18276,9 +18542,9 @@ uint32_t LockResponse::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast744; - xfer += iprot->readI32(ecast744); - this->state = (LockState::type)ecast744; + int32_t ecast752; + xfer += iprot->readI32(ecast752); + this->state = (LockState::type)ecast752; isset_state = true; } else { xfer += iprot->skip(ftype); @@ -18324,13 +18590,13 @@ void swap(LockResponse &a, LockResponse &b) { swap(a.state, b.state); } -LockResponse::LockResponse(const LockResponse& other745) { - lockid = other745.lockid; - state = other745.state; +LockResponse::LockResponse(const LockResponse& other753) { + lockid = other753.lockid; + state = other753.state; } -LockResponse& LockResponse::operator=(const LockResponse& other746) { - lockid = other746.lockid; - state = other746.state; +LockResponse& LockResponse::operator=(const LockResponse& other754) { + lockid = other754.lockid; + state = other754.state; return *this; } void LockResponse::printTo(std::ostream& out) const { @@ -18452,17 +18718,17 @@ void swap(CheckLockRequest &a, CheckLockRequest &b) { swap(a.__isset, b.__isset); } -CheckLockRequest::CheckLockRequest(const CheckLockRequest& other747) { - lockid = other747.lockid; - txnid = other747.txnid; - elapsed_ms = other747.elapsed_ms; - __isset = other747.__isset; +CheckLockRequest::CheckLockRequest(const CheckLockRequest& other755) { + lockid = other755.lockid; + txnid = other755.txnid; + elapsed_ms = other755.elapsed_ms; + __isset = other755.__isset; } -CheckLockRequest& CheckLockRequest::operator=(const CheckLockRequest& other748) { - lockid = other748.lockid; - txnid = other748.txnid; - elapsed_ms = other748.elapsed_ms; - __isset = other748.__isset; +CheckLockRequest& CheckLockRequest::operator=(const CheckLockRequest& other756) { + lockid = other756.lockid; + txnid = other756.txnid; + elapsed_ms = other756.elapsed_ms; + __isset = other756.__isset; return *this; } void CheckLockRequest::printTo(std::ostream& out) const { @@ -18546,11 +18812,11 @@ void swap(UnlockRequest &a, UnlockRequest &b) { swap(a.lockid, b.lockid); } -UnlockRequest::UnlockRequest(const UnlockRequest& other749) { - lockid = other749.lockid; +UnlockRequest::UnlockRequest(const UnlockRequest& other757) { + lockid = other757.lockid; } -UnlockRequest& UnlockRequest::operator=(const UnlockRequest& other750) { - lockid = other750.lockid; +UnlockRequest& UnlockRequest::operator=(const UnlockRequest& other758) { + lockid = other758.lockid; return *this; } void UnlockRequest::printTo(std::ostream& out) const { @@ -18689,19 +18955,19 @@ void swap(ShowLocksRequest &a, ShowLocksRequest &b) { swap(a.__isset, b.__isset); } -ShowLocksRequest::ShowLocksRequest(const ShowLocksRequest& other751) { - dbname = other751.dbname; - tablename = other751.tablename; - partname = other751.partname; - isExtended = other751.isExtended; - __isset = other751.__isset; +ShowLocksRequest::ShowLocksRequest(const ShowLocksRequest& other759) { + dbname = other759.dbname; + tablename = other759.tablename; + partname = other759.partname; + isExtended = other759.isExtended; + __isset = other759.__isset; } -ShowLocksRequest& ShowLocksRequest::operator=(const ShowLocksRequest& other752) { - dbname = other752.dbname; - tablename = other752.tablename; - partname = other752.partname; - isExtended = other752.isExtended; - __isset = other752.__isset; +ShowLocksRequest& ShowLocksRequest::operator=(const ShowLocksRequest& other760) { + dbname = other760.dbname; + tablename = other760.tablename; + partname = other760.partname; + isExtended = other760.isExtended; + __isset = other760.__isset; return *this; } void ShowLocksRequest::printTo(std::ostream& out) const { @@ -18854,9 +19120,9 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast753; - xfer += iprot->readI32(ecast753); - this->state = (LockState::type)ecast753; + int32_t ecast761; + xfer += iprot->readI32(ecast761); + this->state = (LockState::type)ecast761; isset_state = true; } else { xfer += iprot->skip(ftype); @@ -18864,9 +19130,9 @@ uint32_t ShowLocksResponseElement::read(::apache::thrift::protocol::TProtocol* i break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast754; - xfer += iprot->readI32(ecast754); - this->type = (LockType::type)ecast754; + int32_t ecast762; + xfer += iprot->readI32(ecast762); + this->type = (LockType::type)ecast762; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -19082,43 +19348,43 @@ void swap(ShowLocksResponseElement &a, ShowLocksResponseElement &b) { swap(a.__isset, b.__isset); } -ShowLocksResponseElement::ShowLocksResponseElement(const ShowLocksResponseElement& other755) { - lockid = other755.lockid; - dbname = other755.dbname; - tablename = other755.tablename; - partname = other755.partname; - state = other755.state; - type = other755.type; - txnid = other755.txnid; - lastheartbeat = other755.lastheartbeat; - acquiredat = other755.acquiredat; - user = other755.user; - hostname = other755.hostname; - heartbeatCount = other755.heartbeatCount; - agentInfo = other755.agentInfo; - blockedByExtId = other755.blockedByExtId; - blockedByIntId = other755.blockedByIntId; - lockIdInternal = other755.lockIdInternal; - __isset = other755.__isset; +ShowLocksResponseElement::ShowLocksResponseElement(const ShowLocksResponseElement& other763) { + lockid = other763.lockid; + dbname = other763.dbname; + tablename = other763.tablename; + partname = other763.partname; + state = other763.state; + type = other763.type; + txnid = other763.txnid; + lastheartbeat = other763.lastheartbeat; + acquiredat = other763.acquiredat; + user = other763.user; + hostname = other763.hostname; + heartbeatCount = other763.heartbeatCount; + agentInfo = other763.agentInfo; + blockedByExtId = other763.blockedByExtId; + blockedByIntId = other763.blockedByIntId; + lockIdInternal = other763.lockIdInternal; + __isset = other763.__isset; } -ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksResponseElement& other756) { - lockid = other756.lockid; - dbname = other756.dbname; - tablename = other756.tablename; - partname = other756.partname; - state = other756.state; - type = other756.type; - txnid = other756.txnid; - lastheartbeat = other756.lastheartbeat; - acquiredat = other756.acquiredat; - user = other756.user; - hostname = other756.hostname; - heartbeatCount = other756.heartbeatCount; - agentInfo = other756.agentInfo; - blockedByExtId = other756.blockedByExtId; - blockedByIntId = other756.blockedByIntId; - lockIdInternal = other756.lockIdInternal; - __isset = other756.__isset; +ShowLocksResponseElement& ShowLocksResponseElement::operator=(const ShowLocksResponseElement& other764) { + lockid = other764.lockid; + dbname = other764.dbname; + tablename = other764.tablename; + partname = other764.partname; + state = other764.state; + type = other764.type; + txnid = other764.txnid; + lastheartbeat = other764.lastheartbeat; + acquiredat = other764.acquiredat; + user = other764.user; + hostname = other764.hostname; + heartbeatCount = other764.heartbeatCount; + agentInfo = other764.agentInfo; + blockedByExtId = other764.blockedByExtId; + blockedByIntId = other764.blockedByIntId; + lockIdInternal = other764.lockIdInternal; + __isset = other764.__isset; return *this; } void ShowLocksResponseElement::printTo(std::ostream& out) const { @@ -19177,14 +19443,14 @@ uint32_t ShowLocksResponse::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->locks.clear(); - uint32_t _size757; - ::apache::thrift::protocol::TType _etype760; - xfer += iprot->readListBegin(_etype760, _size757); - this->locks.resize(_size757); - uint32_t _i761; - for (_i761 = 0; _i761 < _size757; ++_i761) + uint32_t _size765; + ::apache::thrift::protocol::TType _etype768; + xfer += iprot->readListBegin(_etype768, _size765); + this->locks.resize(_size765); + uint32_t _i769; + for (_i769 = 0; _i769 < _size765; ++_i769) { - xfer += this->locks[_i761].read(iprot); + xfer += this->locks[_i769].read(iprot); } xfer += iprot->readListEnd(); } @@ -19213,10 +19479,10 @@ uint32_t ShowLocksResponse::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("locks", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->locks.size())); - std::vector ::const_iterator _iter762; - for (_iter762 = this->locks.begin(); _iter762 != this->locks.end(); ++_iter762) + std::vector ::const_iterator _iter770; + for (_iter770 = this->locks.begin(); _iter770 != this->locks.end(); ++_iter770) { - xfer += (*_iter762).write(oprot); + xfer += (*_iter770).write(oprot); } xfer += oprot->writeListEnd(); } @@ -19233,13 +19499,13 @@ void swap(ShowLocksResponse &a, ShowLocksResponse &b) { swap(a.__isset, b.__isset); } -ShowLocksResponse::ShowLocksResponse(const ShowLocksResponse& other763) { - locks = other763.locks; - __isset = other763.__isset; +ShowLocksResponse::ShowLocksResponse(const ShowLocksResponse& other771) { + locks = other771.locks; + __isset = other771.__isset; } -ShowLocksResponse& ShowLocksResponse::operator=(const ShowLocksResponse& other764) { - locks = other764.locks; - __isset = other764.__isset; +ShowLocksResponse& ShowLocksResponse::operator=(const ShowLocksResponse& other772) { + locks = other772.locks; + __isset = other772.__isset; return *this; } void ShowLocksResponse::printTo(std::ostream& out) const { @@ -19340,15 +19606,15 @@ void swap(HeartbeatRequest &a, HeartbeatRequest &b) { swap(a.__isset, b.__isset); } -HeartbeatRequest::HeartbeatRequest(const HeartbeatRequest& other765) { - lockid = other765.lockid; - txnid = other765.txnid; - __isset = other765.__isset; +HeartbeatRequest::HeartbeatRequest(const HeartbeatRequest& other773) { + lockid = other773.lockid; + txnid = other773.txnid; + __isset = other773.__isset; } -HeartbeatRequest& HeartbeatRequest::operator=(const HeartbeatRequest& other766) { - lockid = other766.lockid; - txnid = other766.txnid; - __isset = other766.__isset; +HeartbeatRequest& HeartbeatRequest::operator=(const HeartbeatRequest& other774) { + lockid = other774.lockid; + txnid = other774.txnid; + __isset = other774.__isset; return *this; } void HeartbeatRequest::printTo(std::ostream& out) const { @@ -19451,13 +19717,13 @@ void swap(HeartbeatTxnRangeRequest &a, HeartbeatTxnRangeRequest &b) { swap(a.max, b.max); } -HeartbeatTxnRangeRequest::HeartbeatTxnRangeRequest(const HeartbeatTxnRangeRequest& other767) { - min = other767.min; - max = other767.max; +HeartbeatTxnRangeRequest::HeartbeatTxnRangeRequest(const HeartbeatTxnRangeRequest& other775) { + min = other775.min; + max = other775.max; } -HeartbeatTxnRangeRequest& HeartbeatTxnRangeRequest::operator=(const HeartbeatTxnRangeRequest& other768) { - min = other768.min; - max = other768.max; +HeartbeatTxnRangeRequest& HeartbeatTxnRangeRequest::operator=(const HeartbeatTxnRangeRequest& other776) { + min = other776.min; + max = other776.max; return *this; } void HeartbeatTxnRangeRequest::printTo(std::ostream& out) const { @@ -19508,15 +19774,15 @@ uint32_t HeartbeatTxnRangeResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_SET) { { this->aborted.clear(); - uint32_t _size769; - ::apache::thrift::protocol::TType _etype772; - xfer += iprot->readSetBegin(_etype772, _size769); - uint32_t _i773; - for (_i773 = 0; _i773 < _size769; ++_i773) + uint32_t _size777; + ::apache::thrift::protocol::TType _etype780; + xfer += iprot->readSetBegin(_etype780, _size777); + uint32_t _i781; + for (_i781 = 0; _i781 < _size777; ++_i781) { - int64_t _elem774; - xfer += iprot->readI64(_elem774); - this->aborted.insert(_elem774); + int64_t _elem782; + xfer += iprot->readI64(_elem782); + this->aborted.insert(_elem782); } xfer += iprot->readSetEnd(); } @@ -19529,15 +19795,15 @@ uint32_t HeartbeatTxnRangeResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_SET) { { this->nosuch.clear(); - uint32_t _size775; - ::apache::thrift::protocol::TType _etype778; - xfer += iprot->readSetBegin(_etype778, _size775); - uint32_t _i779; - for (_i779 = 0; _i779 < _size775; ++_i779) + uint32_t _size783; + ::apache::thrift::protocol::TType _etype786; + xfer += iprot->readSetBegin(_etype786, _size783); + uint32_t _i787; + for (_i787 = 0; _i787 < _size783; ++_i787) { - int64_t _elem780; - xfer += iprot->readI64(_elem780); - this->nosuch.insert(_elem780); + int64_t _elem788; + xfer += iprot->readI64(_elem788); + this->nosuch.insert(_elem788); } xfer += iprot->readSetEnd(); } @@ -19570,10 +19836,10 @@ uint32_t HeartbeatTxnRangeResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("aborted", ::apache::thrift::protocol::T_SET, 1); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_I64, static_cast(this->aborted.size())); - std::set ::const_iterator _iter781; - for (_iter781 = this->aborted.begin(); _iter781 != this->aborted.end(); ++_iter781) + std::set ::const_iterator _iter789; + for (_iter789 = this->aborted.begin(); _iter789 != this->aborted.end(); ++_iter789) { - xfer += oprot->writeI64((*_iter781)); + xfer += oprot->writeI64((*_iter789)); } xfer += oprot->writeSetEnd(); } @@ -19582,10 +19848,10 @@ uint32_t HeartbeatTxnRangeResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("nosuch", ::apache::thrift::protocol::T_SET, 2); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_I64, static_cast(this->nosuch.size())); - std::set ::const_iterator _iter782; - for (_iter782 = this->nosuch.begin(); _iter782 != this->nosuch.end(); ++_iter782) + std::set ::const_iterator _iter790; + for (_iter790 = this->nosuch.begin(); _iter790 != this->nosuch.end(); ++_iter790) { - xfer += oprot->writeI64((*_iter782)); + xfer += oprot->writeI64((*_iter790)); } xfer += oprot->writeSetEnd(); } @@ -19602,13 +19868,13 @@ void swap(HeartbeatTxnRangeResponse &a, HeartbeatTxnRangeResponse &b) { swap(a.nosuch, b.nosuch); } -HeartbeatTxnRangeResponse::HeartbeatTxnRangeResponse(const HeartbeatTxnRangeResponse& other783) { - aborted = other783.aborted; - nosuch = other783.nosuch; +HeartbeatTxnRangeResponse::HeartbeatTxnRangeResponse(const HeartbeatTxnRangeResponse& other791) { + aborted = other791.aborted; + nosuch = other791.nosuch; } -HeartbeatTxnRangeResponse& HeartbeatTxnRangeResponse::operator=(const HeartbeatTxnRangeResponse& other784) { - aborted = other784.aborted; - nosuch = other784.nosuch; +HeartbeatTxnRangeResponse& HeartbeatTxnRangeResponse::operator=(const HeartbeatTxnRangeResponse& other792) { + aborted = other792.aborted; + nosuch = other792.nosuch; return *this; } void HeartbeatTxnRangeResponse::printTo(std::ostream& out) const { @@ -19701,9 +19967,9 @@ uint32_t CompactionRequest::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast785; - xfer += iprot->readI32(ecast785); - this->type = (CompactionType::type)ecast785; + int32_t ecast793; + xfer += iprot->readI32(ecast793); + this->type = (CompactionType::type)ecast793; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -19721,17 +19987,17 @@ uint32_t CompactionRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_MAP) { { this->properties.clear(); - uint32_t _size786; - ::apache::thrift::protocol::TType _ktype787; - ::apache::thrift::protocol::TType _vtype788; - xfer += iprot->readMapBegin(_ktype787, _vtype788, _size786); - uint32_t _i790; - for (_i790 = 0; _i790 < _size786; ++_i790) + uint32_t _size794; + ::apache::thrift::protocol::TType _ktype795; + ::apache::thrift::protocol::TType _vtype796; + xfer += iprot->readMapBegin(_ktype795, _vtype796, _size794); + uint32_t _i798; + for (_i798 = 0; _i798 < _size794; ++_i798) { - std::string _key791; - xfer += iprot->readString(_key791); - std::string& _val792 = this->properties[_key791]; - xfer += iprot->readString(_val792); + std::string _key799; + xfer += iprot->readString(_key799); + std::string& _val800 = this->properties[_key799]; + xfer += iprot->readString(_val800); } xfer += iprot->readMapEnd(); } @@ -19789,11 +20055,11 @@ uint32_t CompactionRequest::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("properties", ::apache::thrift::protocol::T_MAP, 6); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_STRING, ::apache::thrift::protocol::T_STRING, static_cast(this->properties.size())); - std::map ::const_iterator _iter793; - for (_iter793 = this->properties.begin(); _iter793 != this->properties.end(); ++_iter793) + std::map ::const_iterator _iter801; + for (_iter801 = this->properties.begin(); _iter801 != this->properties.end(); ++_iter801) { - xfer += oprot->writeString(_iter793->first); - xfer += oprot->writeString(_iter793->second); + xfer += oprot->writeString(_iter801->first); + xfer += oprot->writeString(_iter801->second); } xfer += oprot->writeMapEnd(); } @@ -19815,23 +20081,23 @@ void swap(CompactionRequest &a, CompactionRequest &b) { swap(a.__isset, b.__isset); } -CompactionRequest::CompactionRequest(const CompactionRequest& other794) { - dbname = other794.dbname; - tablename = other794.tablename; - partitionname = other794.partitionname; - type = other794.type; - runas = other794.runas; - properties = other794.properties; - __isset = other794.__isset; -} -CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other795) { - dbname = other795.dbname; - tablename = other795.tablename; - partitionname = other795.partitionname; - type = other795.type; - runas = other795.runas; - properties = other795.properties; - __isset = other795.__isset; +CompactionRequest::CompactionRequest(const CompactionRequest& other802) { + dbname = other802.dbname; + tablename = other802.tablename; + partitionname = other802.partitionname; + type = other802.type; + runas = other802.runas; + properties = other802.properties; + __isset = other802.__isset; +} +CompactionRequest& CompactionRequest::operator=(const CompactionRequest& other803) { + dbname = other803.dbname; + tablename = other803.tablename; + partitionname = other803.partitionname; + type = other803.type; + runas = other803.runas; + properties = other803.properties; + __isset = other803.__isset; return *this; } void CompactionRequest::printTo(std::ostream& out) const { @@ -19958,15 +20224,15 @@ void swap(CompactionResponse &a, CompactionResponse &b) { swap(a.accepted, b.accepted); } -CompactionResponse::CompactionResponse(const CompactionResponse& other796) { - id = other796.id; - state = other796.state; - accepted = other796.accepted; +CompactionResponse::CompactionResponse(const CompactionResponse& other804) { + id = other804.id; + state = other804.state; + accepted = other804.accepted; } -CompactionResponse& CompactionResponse::operator=(const CompactionResponse& other797) { - id = other797.id; - state = other797.state; - accepted = other797.accepted; +CompactionResponse& CompactionResponse::operator=(const CompactionResponse& other805) { + id = other805.id; + state = other805.state; + accepted = other805.accepted; return *this; } void CompactionResponse::printTo(std::ostream& out) const { @@ -20027,11 +20293,11 @@ void swap(ShowCompactRequest &a, ShowCompactRequest &b) { (void) b; } -ShowCompactRequest::ShowCompactRequest(const ShowCompactRequest& other798) { - (void) other798; +ShowCompactRequest::ShowCompactRequest(const ShowCompactRequest& other806) { + (void) other806; } -ShowCompactRequest& ShowCompactRequest::operator=(const ShowCompactRequest& other799) { - (void) other799; +ShowCompactRequest& ShowCompactRequest::operator=(const ShowCompactRequest& other807) { + (void) other807; return *this; } void ShowCompactRequest::printTo(std::ostream& out) const { @@ -20157,9 +20423,9 @@ uint32_t ShowCompactResponseElement::read(::apache::thrift::protocol::TProtocol* break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast800; - xfer += iprot->readI32(ecast800); - this->type = (CompactionType::type)ecast800; + int32_t ecast808; + xfer += iprot->readI32(ecast808); + this->type = (CompactionType::type)ecast808; isset_type = true; } else { xfer += iprot->skip(ftype); @@ -20346,37 +20612,37 @@ void swap(ShowCompactResponseElement &a, ShowCompactResponseElement &b) { swap(a.__isset, b.__isset); } -ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponseElement& other801) { - dbname = other801.dbname; - tablename = other801.tablename; - partitionname = other801.partitionname; - type = other801.type; - state = other801.state; - workerid = other801.workerid; - start = other801.start; - runAs = other801.runAs; - hightestTxnId = other801.hightestTxnId; - metaInfo = other801.metaInfo; - endTime = other801.endTime; - hadoopJobId = other801.hadoopJobId; - id = other801.id; - __isset = other801.__isset; -} -ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other802) { - dbname = other802.dbname; - tablename = other802.tablename; - partitionname = other802.partitionname; - type = other802.type; - state = other802.state; - workerid = other802.workerid; - start = other802.start; - runAs = other802.runAs; - hightestTxnId = other802.hightestTxnId; - metaInfo = other802.metaInfo; - endTime = other802.endTime; - hadoopJobId = other802.hadoopJobId; - id = other802.id; - __isset = other802.__isset; +ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponseElement& other809) { + dbname = other809.dbname; + tablename = other809.tablename; + partitionname = other809.partitionname; + type = other809.type; + state = other809.state; + workerid = other809.workerid; + start = other809.start; + runAs = other809.runAs; + hightestTxnId = other809.hightestTxnId; + metaInfo = other809.metaInfo; + endTime = other809.endTime; + hadoopJobId = other809.hadoopJobId; + id = other809.id; + __isset = other809.__isset; +} +ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other810) { + dbname = other810.dbname; + tablename = other810.tablename; + partitionname = other810.partitionname; + type = other810.type; + state = other810.state; + workerid = other810.workerid; + start = other810.start; + runAs = other810.runAs; + hightestTxnId = other810.hightestTxnId; + metaInfo = other810.metaInfo; + endTime = other810.endTime; + hadoopJobId = other810.hadoopJobId; + id = other810.id; + __isset = other810.__isset; return *this; } void ShowCompactResponseElement::printTo(std::ostream& out) const { @@ -20433,14 +20699,14 @@ uint32_t ShowCompactResponse::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->compacts.clear(); - uint32_t _size803; - ::apache::thrift::protocol::TType _etype806; - xfer += iprot->readListBegin(_etype806, _size803); - this->compacts.resize(_size803); - uint32_t _i807; - for (_i807 = 0; _i807 < _size803; ++_i807) + uint32_t _size811; + ::apache::thrift::protocol::TType _etype814; + xfer += iprot->readListBegin(_etype814, _size811); + this->compacts.resize(_size811); + uint32_t _i815; + for (_i815 = 0; _i815 < _size811; ++_i815) { - xfer += this->compacts[_i807].read(iprot); + xfer += this->compacts[_i815].read(iprot); } xfer += iprot->readListEnd(); } @@ -20471,10 +20737,10 @@ uint32_t ShowCompactResponse::write(::apache::thrift::protocol::TProtocol* oprot xfer += oprot->writeFieldBegin("compacts", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->compacts.size())); - std::vector ::const_iterator _iter808; - for (_iter808 = this->compacts.begin(); _iter808 != this->compacts.end(); ++_iter808) + std::vector ::const_iterator _iter816; + for (_iter816 = this->compacts.begin(); _iter816 != this->compacts.end(); ++_iter816) { - xfer += (*_iter808).write(oprot); + xfer += (*_iter816).write(oprot); } xfer += oprot->writeListEnd(); } @@ -20490,11 +20756,11 @@ void swap(ShowCompactResponse &a, ShowCompactResponse &b) { swap(a.compacts, b.compacts); } -ShowCompactResponse::ShowCompactResponse(const ShowCompactResponse& other809) { - compacts = other809.compacts; +ShowCompactResponse::ShowCompactResponse(const ShowCompactResponse& other817) { + compacts = other817.compacts; } -ShowCompactResponse& ShowCompactResponse::operator=(const ShowCompactResponse& other810) { - compacts = other810.compacts; +ShowCompactResponse& ShowCompactResponse::operator=(const ShowCompactResponse& other818) { + compacts = other818.compacts; return *this; } void ShowCompactResponse::printTo(std::ostream& out) const { @@ -20596,14 +20862,14 @@ uint32_t AddDynamicPartitions::read(::apache::thrift::protocol::TProtocol* iprot if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionnames.clear(); - uint32_t _size811; - ::apache::thrift::protocol::TType _etype814; - xfer += iprot->readListBegin(_etype814, _size811); - this->partitionnames.resize(_size811); - uint32_t _i815; - for (_i815 = 0; _i815 < _size811; ++_i815) + uint32_t _size819; + ::apache::thrift::protocol::TType _etype822; + xfer += iprot->readListBegin(_etype822, _size819); + this->partitionnames.resize(_size819); + uint32_t _i823; + for (_i823 = 0; _i823 < _size819; ++_i823) { - xfer += iprot->readString(this->partitionnames[_i815]); + xfer += iprot->readString(this->partitionnames[_i823]); } xfer += iprot->readListEnd(); } @@ -20614,9 +20880,9 @@ uint32_t AddDynamicPartitions::read(::apache::thrift::protocol::TProtocol* iprot break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast816; - xfer += iprot->readI32(ecast816); - this->operationType = (DataOperationType::type)ecast816; + int32_t ecast824; + xfer += iprot->readI32(ecast824); + this->operationType = (DataOperationType::type)ecast824; this->__isset.operationType = true; } else { xfer += iprot->skip(ftype); @@ -20668,10 +20934,10 @@ uint32_t AddDynamicPartitions::write(::apache::thrift::protocol::TProtocol* opro xfer += oprot->writeFieldBegin("partitionnames", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionnames.size())); - std::vector ::const_iterator _iter817; - for (_iter817 = this->partitionnames.begin(); _iter817 != this->partitionnames.end(); ++_iter817) + std::vector ::const_iterator _iter825; + for (_iter825 = this->partitionnames.begin(); _iter825 != this->partitionnames.end(); ++_iter825) { - xfer += oprot->writeString((*_iter817)); + xfer += oprot->writeString((*_iter825)); } xfer += oprot->writeListEnd(); } @@ -20698,23 +20964,23 @@ void swap(AddDynamicPartitions &a, AddDynamicPartitions &b) { swap(a.__isset, b.__isset); } -AddDynamicPartitions::AddDynamicPartitions(const AddDynamicPartitions& other818) { - txnid = other818.txnid; - writeid = other818.writeid; - dbname = other818.dbname; - tablename = other818.tablename; - partitionnames = other818.partitionnames; - operationType = other818.operationType; - __isset = other818.__isset; -} -AddDynamicPartitions& AddDynamicPartitions::operator=(const AddDynamicPartitions& other819) { - txnid = other819.txnid; - writeid = other819.writeid; - dbname = other819.dbname; - tablename = other819.tablename; - partitionnames = other819.partitionnames; - operationType = other819.operationType; - __isset = other819.__isset; +AddDynamicPartitions::AddDynamicPartitions(const AddDynamicPartitions& other826) { + txnid = other826.txnid; + writeid = other826.writeid; + dbname = other826.dbname; + tablename = other826.tablename; + partitionnames = other826.partitionnames; + operationType = other826.operationType; + __isset = other826.__isset; +} +AddDynamicPartitions& AddDynamicPartitions::operator=(const AddDynamicPartitions& other827) { + txnid = other827.txnid; + writeid = other827.writeid; + dbname = other827.dbname; + tablename = other827.tablename; + partitionnames = other827.partitionnames; + operationType = other827.operationType; + __isset = other827.__isset; return *this; } void AddDynamicPartitions::printTo(std::ostream& out) const { @@ -20897,23 +21163,23 @@ void swap(BasicTxnInfo &a, BasicTxnInfo &b) { swap(a.__isset, b.__isset); } -BasicTxnInfo::BasicTxnInfo(const BasicTxnInfo& other820) { - isnull = other820.isnull; - time = other820.time; - txnid = other820.txnid; - dbname = other820.dbname; - tablename = other820.tablename; - partitionname = other820.partitionname; - __isset = other820.__isset; -} -BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other821) { - isnull = other821.isnull; - time = other821.time; - txnid = other821.txnid; - dbname = other821.dbname; - tablename = other821.tablename; - partitionname = other821.partitionname; - __isset = other821.__isset; +BasicTxnInfo::BasicTxnInfo(const BasicTxnInfo& other828) { + isnull = other828.isnull; + time = other828.time; + txnid = other828.txnid; + dbname = other828.dbname; + tablename = other828.tablename; + partitionname = other828.partitionname; + __isset = other828.__isset; +} +BasicTxnInfo& BasicTxnInfo::operator=(const BasicTxnInfo& other829) { + isnull = other829.isnull; + time = other829.time; + txnid = other829.txnid; + dbname = other829.dbname; + tablename = other829.tablename; + partitionname = other829.partitionname; + __isset = other829.__isset; return *this; } void BasicTxnInfo::printTo(std::ostream& out) const { @@ -21012,15 +21278,15 @@ uint32_t CreationMetadata::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_SET) { { this->tablesUsed.clear(); - uint32_t _size822; - ::apache::thrift::protocol::TType _etype825; - xfer += iprot->readSetBegin(_etype825, _size822); - uint32_t _i826; - for (_i826 = 0; _i826 < _size822; ++_i826) + uint32_t _size830; + ::apache::thrift::protocol::TType _etype833; + xfer += iprot->readSetBegin(_etype833, _size830); + uint32_t _i834; + for (_i834 = 0; _i834 < _size830; ++_i834) { - std::string _elem827; - xfer += iprot->readString(_elem827); - this->tablesUsed.insert(_elem827); + std::string _elem835; + xfer += iprot->readString(_elem835); + this->tablesUsed.insert(_elem835); } xfer += iprot->readSetEnd(); } @@ -21085,10 +21351,10 @@ uint32_t CreationMetadata::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tablesUsed", ::apache::thrift::protocol::T_SET, 4); { xfer += oprot->writeSetBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tablesUsed.size())); - std::set ::const_iterator _iter828; - for (_iter828 = this->tablesUsed.begin(); _iter828 != this->tablesUsed.end(); ++_iter828) + std::set ::const_iterator _iter836; + for (_iter836 = this->tablesUsed.begin(); _iter836 != this->tablesUsed.end(); ++_iter836) { - xfer += oprot->writeString((*_iter828)); + xfer += oprot->writeString((*_iter836)); } xfer += oprot->writeSetEnd(); } @@ -21120,23 +21386,23 @@ void swap(CreationMetadata &a, CreationMetadata &b) { swap(a.__isset, b.__isset); } -CreationMetadata::CreationMetadata(const CreationMetadata& other829) { - catName = other829.catName; - dbName = other829.dbName; - tblName = other829.tblName; - tablesUsed = other829.tablesUsed; - validTxnList = other829.validTxnList; - materializationTime = other829.materializationTime; - __isset = other829.__isset; -} -CreationMetadata& CreationMetadata::operator=(const CreationMetadata& other830) { - catName = other830.catName; - dbName = other830.dbName; - tblName = other830.tblName; - tablesUsed = other830.tablesUsed; - validTxnList = other830.validTxnList; - materializationTime = other830.materializationTime; - __isset = other830.__isset; +CreationMetadata::CreationMetadata(const CreationMetadata& other837) { + catName = other837.catName; + dbName = other837.dbName; + tblName = other837.tblName; + tablesUsed = other837.tablesUsed; + validTxnList = other837.validTxnList; + materializationTime = other837.materializationTime; + __isset = other837.__isset; +} +CreationMetadata& CreationMetadata::operator=(const CreationMetadata& other838) { + catName = other838.catName; + dbName = other838.dbName; + tblName = other838.tblName; + tablesUsed = other838.tablesUsed; + validTxnList = other838.validTxnList; + materializationTime = other838.materializationTime; + __isset = other838.__isset; return *this; } void CreationMetadata::printTo(std::ostream& out) const { @@ -21243,15 +21509,15 @@ void swap(NotificationEventRequest &a, NotificationEventRequest &b) { swap(a.__isset, b.__isset); } -NotificationEventRequest::NotificationEventRequest(const NotificationEventRequest& other831) { - lastEvent = other831.lastEvent; - maxEvents = other831.maxEvents; - __isset = other831.__isset; +NotificationEventRequest::NotificationEventRequest(const NotificationEventRequest& other839) { + lastEvent = other839.lastEvent; + maxEvents = other839.maxEvents; + __isset = other839.__isset; } -NotificationEventRequest& NotificationEventRequest::operator=(const NotificationEventRequest& other832) { - lastEvent = other832.lastEvent; - maxEvents = other832.maxEvents; - __isset = other832.__isset; +NotificationEventRequest& NotificationEventRequest::operator=(const NotificationEventRequest& other840) { + lastEvent = other840.lastEvent; + maxEvents = other840.maxEvents; + __isset = other840.__isset; return *this; } void NotificationEventRequest::printTo(std::ostream& out) const { @@ -21471,27 +21737,27 @@ void swap(NotificationEvent &a, NotificationEvent &b) { swap(a.__isset, b.__isset); } -NotificationEvent::NotificationEvent(const NotificationEvent& other833) { - eventId = other833.eventId; - eventTime = other833.eventTime; - eventType = other833.eventType; - dbName = other833.dbName; - tableName = other833.tableName; - message = other833.message; - messageFormat = other833.messageFormat; - catName = other833.catName; - __isset = other833.__isset; -} -NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other834) { - eventId = other834.eventId; - eventTime = other834.eventTime; - eventType = other834.eventType; - dbName = other834.dbName; - tableName = other834.tableName; - message = other834.message; - messageFormat = other834.messageFormat; - catName = other834.catName; - __isset = other834.__isset; +NotificationEvent::NotificationEvent(const NotificationEvent& other841) { + eventId = other841.eventId; + eventTime = other841.eventTime; + eventType = other841.eventType; + dbName = other841.dbName; + tableName = other841.tableName; + message = other841.message; + messageFormat = other841.messageFormat; + catName = other841.catName; + __isset = other841.__isset; +} +NotificationEvent& NotificationEvent::operator=(const NotificationEvent& other842) { + eventId = other842.eventId; + eventTime = other842.eventTime; + eventType = other842.eventType; + dbName = other842.dbName; + tableName = other842.tableName; + message = other842.message; + messageFormat = other842.messageFormat; + catName = other842.catName; + __isset = other842.__isset; return *this; } void NotificationEvent::printTo(std::ostream& out) const { @@ -21543,14 +21809,14 @@ uint32_t NotificationEventResponse::read(::apache::thrift::protocol::TProtocol* if (ftype == ::apache::thrift::protocol::T_LIST) { { this->events.clear(); - uint32_t _size835; - ::apache::thrift::protocol::TType _etype838; - xfer += iprot->readListBegin(_etype838, _size835); - this->events.resize(_size835); - uint32_t _i839; - for (_i839 = 0; _i839 < _size835; ++_i839) + uint32_t _size843; + ::apache::thrift::protocol::TType _etype846; + xfer += iprot->readListBegin(_etype846, _size843); + this->events.resize(_size843); + uint32_t _i847; + for (_i847 = 0; _i847 < _size843; ++_i847) { - xfer += this->events[_i839].read(iprot); + xfer += this->events[_i847].read(iprot); } xfer += iprot->readListEnd(); } @@ -21581,10 +21847,10 @@ uint32_t NotificationEventResponse::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("events", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->events.size())); - std::vector ::const_iterator _iter840; - for (_iter840 = this->events.begin(); _iter840 != this->events.end(); ++_iter840) + std::vector ::const_iterator _iter848; + for (_iter848 = this->events.begin(); _iter848 != this->events.end(); ++_iter848) { - xfer += (*_iter840).write(oprot); + xfer += (*_iter848).write(oprot); } xfer += oprot->writeListEnd(); } @@ -21600,11 +21866,11 @@ void swap(NotificationEventResponse &a, NotificationEventResponse &b) { swap(a.events, b.events); } -NotificationEventResponse::NotificationEventResponse(const NotificationEventResponse& other841) { - events = other841.events; +NotificationEventResponse::NotificationEventResponse(const NotificationEventResponse& other849) { + events = other849.events; } -NotificationEventResponse& NotificationEventResponse::operator=(const NotificationEventResponse& other842) { - events = other842.events; +NotificationEventResponse& NotificationEventResponse::operator=(const NotificationEventResponse& other850) { + events = other850.events; return *this; } void NotificationEventResponse::printTo(std::ostream& out) const { @@ -21686,11 +21952,11 @@ void swap(CurrentNotificationEventId &a, CurrentNotificationEventId &b) { swap(a.eventId, b.eventId); } -CurrentNotificationEventId::CurrentNotificationEventId(const CurrentNotificationEventId& other843) { - eventId = other843.eventId; +CurrentNotificationEventId::CurrentNotificationEventId(const CurrentNotificationEventId& other851) { + eventId = other851.eventId; } -CurrentNotificationEventId& CurrentNotificationEventId::operator=(const CurrentNotificationEventId& other844) { - eventId = other844.eventId; +CurrentNotificationEventId& CurrentNotificationEventId::operator=(const CurrentNotificationEventId& other852) { + eventId = other852.eventId; return *this; } void CurrentNotificationEventId::printTo(std::ostream& out) const { @@ -21812,17 +22078,17 @@ void swap(NotificationEventsCountRequest &a, NotificationEventsCountRequest &b) swap(a.__isset, b.__isset); } -NotificationEventsCountRequest::NotificationEventsCountRequest(const NotificationEventsCountRequest& other845) { - fromEventId = other845.fromEventId; - dbName = other845.dbName; - catName = other845.catName; - __isset = other845.__isset; +NotificationEventsCountRequest::NotificationEventsCountRequest(const NotificationEventsCountRequest& other853) { + fromEventId = other853.fromEventId; + dbName = other853.dbName; + catName = other853.catName; + __isset = other853.__isset; } -NotificationEventsCountRequest& NotificationEventsCountRequest::operator=(const NotificationEventsCountRequest& other846) { - fromEventId = other846.fromEventId; - dbName = other846.dbName; - catName = other846.catName; - __isset = other846.__isset; +NotificationEventsCountRequest& NotificationEventsCountRequest::operator=(const NotificationEventsCountRequest& other854) { + fromEventId = other854.fromEventId; + dbName = other854.dbName; + catName = other854.catName; + __isset = other854.__isset; return *this; } void NotificationEventsCountRequest::printTo(std::ostream& out) const { @@ -21906,11 +22172,11 @@ void swap(NotificationEventsCountResponse &a, NotificationEventsCountResponse &b swap(a.eventsCount, b.eventsCount); } -NotificationEventsCountResponse::NotificationEventsCountResponse(const NotificationEventsCountResponse& other847) { - eventsCount = other847.eventsCount; +NotificationEventsCountResponse::NotificationEventsCountResponse(const NotificationEventsCountResponse& other855) { + eventsCount = other855.eventsCount; } -NotificationEventsCountResponse& NotificationEventsCountResponse::operator=(const NotificationEventsCountResponse& other848) { - eventsCount = other848.eventsCount; +NotificationEventsCountResponse& NotificationEventsCountResponse::operator=(const NotificationEventsCountResponse& other856) { + eventsCount = other856.eventsCount; return *this; } void NotificationEventsCountResponse::printTo(std::ostream& out) const { @@ -21939,6 +22205,11 @@ void InsertEventRequestData::__set_filesAddedChecksum(const std::vector & val) { + this->subDirectoryList = val; +__isset.subDirectoryList = true; +} + uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); @@ -21973,14 +22244,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filesAdded.clear(); - uint32_t _size849; - ::apache::thrift::protocol::TType _etype852; - xfer += iprot->readListBegin(_etype852, _size849); - this->filesAdded.resize(_size849); - uint32_t _i853; - for (_i853 = 0; _i853 < _size849; ++_i853) + uint32_t _size857; + ::apache::thrift::protocol::TType _etype860; + xfer += iprot->readListBegin(_etype860, _size857); + this->filesAdded.resize(_size857); + uint32_t _i861; + for (_i861 = 0; _i861 < _size857; ++_i861) { - xfer += iprot->readString(this->filesAdded[_i853]); + xfer += iprot->readString(this->filesAdded[_i861]); } xfer += iprot->readListEnd(); } @@ -21993,14 +22264,14 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->filesAddedChecksum.clear(); - uint32_t _size854; - ::apache::thrift::protocol::TType _etype857; - xfer += iprot->readListBegin(_etype857, _size854); - this->filesAddedChecksum.resize(_size854); - uint32_t _i858; - for (_i858 = 0; _i858 < _size854; ++_i858) + uint32_t _size862; + ::apache::thrift::protocol::TType _etype865; + xfer += iprot->readListBegin(_etype865, _size862); + this->filesAddedChecksum.resize(_size862); + uint32_t _i866; + for (_i866 = 0; _i866 < _size862; ++_i866) { - xfer += iprot->readString(this->filesAddedChecksum[_i858]); + xfer += iprot->readString(this->filesAddedChecksum[_i866]); } xfer += iprot->readListEnd(); } @@ -22009,6 +22280,26 @@ uint32_t InsertEventRequestData::read(::apache::thrift::protocol::TProtocol* ipr xfer += iprot->skip(ftype); } break; + case 4: + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + this->subDirectoryList.clear(); + uint32_t _size867; + ::apache::thrift::protocol::TType _etype870; + xfer += iprot->readListBegin(_etype870, _size867); + this->subDirectoryList.resize(_size867); + uint32_t _i871; + for (_i871 = 0; _i871 < _size867; ++_i871) + { + xfer += iprot->readString(this->subDirectoryList[_i871]); + } + xfer += iprot->readListEnd(); + } + this->__isset.subDirectoryList = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -22036,10 +22327,10 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("filesAdded", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filesAdded.size())); - std::vector ::const_iterator _iter859; - for (_iter859 = this->filesAdded.begin(); _iter859 != this->filesAdded.end(); ++_iter859) + std::vector ::const_iterator _iter872; + for (_iter872 = this->filesAdded.begin(); _iter872 != this->filesAdded.end(); ++_iter872) { - xfer += oprot->writeString((*_iter859)); + xfer += oprot->writeString((*_iter872)); } xfer += oprot->writeListEnd(); } @@ -22049,10 +22340,23 @@ uint32_t InsertEventRequestData::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("filesAddedChecksum", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->filesAddedChecksum.size())); - std::vector ::const_iterator _iter860; - for (_iter860 = this->filesAddedChecksum.begin(); _iter860 != this->filesAddedChecksum.end(); ++_iter860) + std::vector ::const_iterator _iter873; + for (_iter873 = this->filesAddedChecksum.begin(); _iter873 != this->filesAddedChecksum.end(); ++_iter873) + { + xfer += oprot->writeString((*_iter873)); + } + xfer += oprot->writeListEnd(); + } + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.subDirectoryList) { + xfer += oprot->writeFieldBegin("subDirectoryList", ::apache::thrift::protocol::T_LIST, 4); + { + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->subDirectoryList.size())); + std::vector ::const_iterator _iter874; + for (_iter874 = this->subDirectoryList.begin(); _iter874 != this->subDirectoryList.end(); ++_iter874) { - xfer += oprot->writeString((*_iter860)); + xfer += oprot->writeString((*_iter874)); } xfer += oprot->writeListEnd(); } @@ -22068,20 +22372,23 @@ void swap(InsertEventRequestData &a, InsertEventRequestData &b) { swap(a.replace, b.replace); swap(a.filesAdded, b.filesAdded); swap(a.filesAddedChecksum, b.filesAddedChecksum); + swap(a.subDirectoryList, b.subDirectoryList); swap(a.__isset, b.__isset); } -InsertEventRequestData::InsertEventRequestData(const InsertEventRequestData& other861) { - replace = other861.replace; - filesAdded = other861.filesAdded; - filesAddedChecksum = other861.filesAddedChecksum; - __isset = other861.__isset; +InsertEventRequestData::InsertEventRequestData(const InsertEventRequestData& other875) { + replace = other875.replace; + filesAdded = other875.filesAdded; + filesAddedChecksum = other875.filesAddedChecksum; + subDirectoryList = other875.subDirectoryList; + __isset = other875.__isset; } -InsertEventRequestData& InsertEventRequestData::operator=(const InsertEventRequestData& other862) { - replace = other862.replace; - filesAdded = other862.filesAdded; - filesAddedChecksum = other862.filesAddedChecksum; - __isset = other862.__isset; +InsertEventRequestData& InsertEventRequestData::operator=(const InsertEventRequestData& other876) { + replace = other876.replace; + filesAdded = other876.filesAdded; + filesAddedChecksum = other876.filesAddedChecksum; + subDirectoryList = other876.subDirectoryList; + __isset = other876.__isset; return *this; } void InsertEventRequestData::printTo(std::ostream& out) const { @@ -22090,6 +22397,7 @@ void InsertEventRequestData::printTo(std::ostream& out) const { out << "replace="; (__isset.replace ? (out << to_string(replace)) : (out << "")); out << ", " << "filesAdded=" << to_string(filesAdded); out << ", " << "filesAddedChecksum="; (__isset.filesAddedChecksum ? (out << to_string(filesAddedChecksum)) : (out << "")); + out << ", " << "subDirectoryList="; (__isset.subDirectoryList ? (out << to_string(subDirectoryList)) : (out << "")); out << ")"; } @@ -22163,13 +22471,13 @@ void swap(FireEventRequestData &a, FireEventRequestData &b) { swap(a.__isset, b.__isset); } -FireEventRequestData::FireEventRequestData(const FireEventRequestData& other863) { - insertData = other863.insertData; - __isset = other863.__isset; +FireEventRequestData::FireEventRequestData(const FireEventRequestData& other877) { + insertData = other877.insertData; + __isset = other877.__isset; } -FireEventRequestData& FireEventRequestData::operator=(const FireEventRequestData& other864) { - insertData = other864.insertData; - __isset = other864.__isset; +FireEventRequestData& FireEventRequestData::operator=(const FireEventRequestData& other878) { + insertData = other878.insertData; + __isset = other878.__isset; return *this; } void FireEventRequestData::printTo(std::ostream& out) const { @@ -22184,35 +22492,314 @@ FireEventRequest::~FireEventRequest() throw() { } -void FireEventRequest::__set_successful(const bool val) { - this->successful = val; +void FireEventRequest::__set_successful(const bool val) { + this->successful = val; +} + +void FireEventRequest::__set_data(const FireEventRequestData& val) { + this->data = val; +} + +void FireEventRequest::__set_dbName(const std::string& val) { + this->dbName = val; +__isset.dbName = true; +} + +void FireEventRequest::__set_tableName(const std::string& val) { + this->tableName = val; +__isset.tableName = true; +} + +void FireEventRequest::__set_partitionVals(const std::vector & val) { + this->partitionVals = val; +__isset.partitionVals = true; +} + +void FireEventRequest::__set_catName(const std::string& val) { + this->catName = val; +__isset.catName = true; +} + +uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { + + apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + bool isset_successful = false; + bool isset_data = false; + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + switch (fid) + { + case 1: + if (ftype == ::apache::thrift::protocol::T_BOOL) { + xfer += iprot->readBool(this->successful); + isset_successful = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 2: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->data.read(iprot); + isset_data = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 3: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->dbName); + this->__isset.dbName = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 4: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->tableName); + this->__isset.tableName = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 5: + if (ftype == ::apache::thrift::protocol::T_LIST) { + { + this->partitionVals.clear(); + uint32_t _size879; + ::apache::thrift::protocol::TType _etype882; + xfer += iprot->readListBegin(_etype882, _size879); + this->partitionVals.resize(_size879); + uint32_t _i883; + for (_i883 = 0; _i883 < _size879; ++_i883) + { + xfer += iprot->readString(this->partitionVals[_i883]); + } + xfer += iprot->readListEnd(); + } + this->__isset.partitionVals = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 6: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->catName); + this->__isset.catName = true; + } else { + xfer += iprot->skip(ftype); + } + break; + default: + xfer += iprot->skip(ftype); + break; + } + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + if (!isset_successful) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_data) + throw TProtocolException(TProtocolException::INVALID_DATA); + return xfer; +} + +uint32_t FireEventRequest::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("FireEventRequest"); + + xfer += oprot->writeFieldBegin("successful", ::apache::thrift::protocol::T_BOOL, 1); + xfer += oprot->writeBool(this->successful); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("data", ::apache::thrift::protocol::T_STRUCT, 2); + xfer += this->data.write(oprot); + xfer += oprot->writeFieldEnd(); + + if (this->__isset.dbName) { + xfer += oprot->writeFieldBegin("dbName", ::apache::thrift::protocol::T_STRING, 3); + xfer += oprot->writeString(this->dbName); + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.tableName) { + xfer += oprot->writeFieldBegin("tableName", ::apache::thrift::protocol::T_STRING, 4); + xfer += oprot->writeString(this->tableName); + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.partitionVals) { + xfer += oprot->writeFieldBegin("partitionVals", ::apache::thrift::protocol::T_LIST, 5); + { + xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionVals.size())); + std::vector ::const_iterator _iter884; + for (_iter884 = this->partitionVals.begin(); _iter884 != this->partitionVals.end(); ++_iter884) + { + xfer += oprot->writeString((*_iter884)); + } + xfer += oprot->writeListEnd(); + } + xfer += oprot->writeFieldEnd(); + } + if (this->__isset.catName) { + xfer += oprot->writeFieldBegin("catName", ::apache::thrift::protocol::T_STRING, 6); + xfer += oprot->writeString(this->catName); + xfer += oprot->writeFieldEnd(); + } + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + +void swap(FireEventRequest &a, FireEventRequest &b) { + using ::std::swap; + swap(a.successful, b.successful); + swap(a.data, b.data); + swap(a.dbName, b.dbName); + swap(a.tableName, b.tableName); + swap(a.partitionVals, b.partitionVals); + swap(a.catName, b.catName); + swap(a.__isset, b.__isset); +} + +FireEventRequest::FireEventRequest(const FireEventRequest& other885) { + successful = other885.successful; + data = other885.data; + dbName = other885.dbName; + tableName = other885.tableName; + partitionVals = other885.partitionVals; + catName = other885.catName; + __isset = other885.__isset; +} +FireEventRequest& FireEventRequest::operator=(const FireEventRequest& other886) { + successful = other886.successful; + data = other886.data; + dbName = other886.dbName; + tableName = other886.tableName; + partitionVals = other886.partitionVals; + catName = other886.catName; + __isset = other886.__isset; + return *this; +} +void FireEventRequest::printTo(std::ostream& out) const { + using ::apache::thrift::to_string; + out << "FireEventRequest("; + out << "successful=" << to_string(successful); + out << ", " << "data=" << to_string(data); + out << ", " << "dbName="; (__isset.dbName ? (out << to_string(dbName)) : (out << "")); + out << ", " << "tableName="; (__isset.tableName ? (out << to_string(tableName)) : (out << "")); + out << ", " << "partitionVals="; (__isset.partitionVals ? (out << to_string(partitionVals)) : (out << "")); + out << ", " << "catName="; (__isset.catName ? (out << to_string(catName)) : (out << "")); + out << ")"; +} + + +FireEventResponse::~FireEventResponse() throw() { +} + + +uint32_t FireEventResponse::read(::apache::thrift::protocol::TProtocol* iprot) { + + apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); + uint32_t xfer = 0; + std::string fname; + ::apache::thrift::protocol::TType ftype; + int16_t fid; + + xfer += iprot->readStructBegin(fname); + + using ::apache::thrift::protocol::TProtocolException; + + + while (true) + { + xfer += iprot->readFieldBegin(fname, ftype, fid); + if (ftype == ::apache::thrift::protocol::T_STOP) { + break; + } + xfer += iprot->skip(ftype); + xfer += iprot->readFieldEnd(); + } + + xfer += iprot->readStructEnd(); + + return xfer; +} + +uint32_t FireEventResponse::write(::apache::thrift::protocol::TProtocol* oprot) const { + uint32_t xfer = 0; + apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); + xfer += oprot->writeStructBegin("FireEventResponse"); + + xfer += oprot->writeFieldStop(); + xfer += oprot->writeStructEnd(); + return xfer; +} + +void swap(FireEventResponse &a, FireEventResponse &b) { + using ::std::swap; + (void) a; + (void) b; +} + +FireEventResponse::FireEventResponse(const FireEventResponse& other887) { + (void) other887; +} +FireEventResponse& FireEventResponse::operator=(const FireEventResponse& other888) { + (void) other888; + return *this; +} +void FireEventResponse::printTo(std::ostream& out) const { + using ::apache::thrift::to_string; + out << "FireEventResponse("; + out << ")"; +} + + +WriteNotificationLogRequest::~WriteNotificationLogRequest() throw() { +} + + +void WriteNotificationLogRequest::__set_txnId(const int64_t val) { + this->txnId = val; +} + +void WriteNotificationLogRequest::__set_writeId(const int64_t val) { + this->writeId = val; } -void FireEventRequest::__set_data(const FireEventRequestData& val) { - this->data = val; +void WriteNotificationLogRequest::__set_db(const std::string& val) { + this->db = val; } -void FireEventRequest::__set_dbName(const std::string& val) { - this->dbName = val; -__isset.dbName = true; +void WriteNotificationLogRequest::__set_table(const std::string& val) { + this->table = val; } -void FireEventRequest::__set_tableName(const std::string& val) { - this->tableName = val; -__isset.tableName = true; +void WriteNotificationLogRequest::__set_fileInfo(const InsertEventRequestData& val) { + this->fileInfo = val; } -void FireEventRequest::__set_partitionVals(const std::vector & val) { +void WriteNotificationLogRequest::__set_partitionVals(const std::vector & val) { this->partitionVals = val; __isset.partitionVals = true; } -void FireEventRequest::__set_catName(const std::string& val) { - this->catName = val; -__isset.catName = true; -} - -uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t WriteNotificationLogRequest::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -22224,8 +22811,11 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { using ::apache::thrift::protocol::TProtocolException; - bool isset_successful = false; - bool isset_data = false; + bool isset_txnId = false; + bool isset_writeId = false; + bool isset_db = false; + bool isset_table = false; + bool isset_fileInfo = false; while (true) { @@ -22236,49 +22826,57 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { switch (fid) { case 1: - if (ftype == ::apache::thrift::protocol::T_BOOL) { - xfer += iprot->readBool(this->successful); - isset_successful = true; + if (ftype == ::apache::thrift::protocol::T_I64) { + xfer += iprot->readI64(this->txnId); + isset_txnId = true; } else { xfer += iprot->skip(ftype); } break; case 2: - if (ftype == ::apache::thrift::protocol::T_STRUCT) { - xfer += this->data.read(iprot); - isset_data = true; + if (ftype == ::apache::thrift::protocol::T_I64) { + xfer += iprot->readI64(this->writeId); + isset_writeId = true; } else { xfer += iprot->skip(ftype); } break; case 3: if (ftype == ::apache::thrift::protocol::T_STRING) { - xfer += iprot->readString(this->dbName); - this->__isset.dbName = true; + xfer += iprot->readString(this->db); + isset_db = true; } else { xfer += iprot->skip(ftype); } break; case 4: if (ftype == ::apache::thrift::protocol::T_STRING) { - xfer += iprot->readString(this->tableName); - this->__isset.tableName = true; + xfer += iprot->readString(this->table); + isset_table = true; } else { xfer += iprot->skip(ftype); } break; case 5: + if (ftype == ::apache::thrift::protocol::T_STRUCT) { + xfer += this->fileInfo.read(iprot); + isset_fileInfo = true; + } else { + xfer += iprot->skip(ftype); + } + break; + case 6: if (ftype == ::apache::thrift::protocol::T_LIST) { { this->partitionVals.clear(); - uint32_t _size865; - ::apache::thrift::protocol::TType _etype868; - xfer += iprot->readListBegin(_etype868, _size865); - this->partitionVals.resize(_size865); - uint32_t _i869; - for (_i869 = 0; _i869 < _size865; ++_i869) + uint32_t _size889; + ::apache::thrift::protocol::TType _etype892; + xfer += iprot->readListBegin(_etype892, _size889); + this->partitionVals.resize(_size889); + uint32_t _i893; + for (_i893 = 0; _i893 < _size889; ++_i893) { - xfer += iprot->readString(this->partitionVals[_i869]); + xfer += iprot->readString(this->partitionVals[_i893]); } xfer += iprot->readListEnd(); } @@ -22287,14 +22885,6 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { xfer += iprot->skip(ftype); } break; - case 6: - if (ftype == ::apache::thrift::protocol::T_STRING) { - xfer += iprot->readString(this->catName); - this->__isset.catName = true; - } else { - xfer += iprot->skip(ftype); - } - break; default: xfer += iprot->skip(ftype); break; @@ -22304,107 +22894,110 @@ uint32_t FireEventRequest::read(::apache::thrift::protocol::TProtocol* iprot) { xfer += iprot->readStructEnd(); - if (!isset_successful) + if (!isset_txnId) throw TProtocolException(TProtocolException::INVALID_DATA); - if (!isset_data) + if (!isset_writeId) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_db) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_table) + throw TProtocolException(TProtocolException::INVALID_DATA); + if (!isset_fileInfo) throw TProtocolException(TProtocolException::INVALID_DATA); return xfer; } -uint32_t FireEventRequest::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t WriteNotificationLogRequest::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("FireEventRequest"); + xfer += oprot->writeStructBegin("WriteNotificationLogRequest"); - xfer += oprot->writeFieldBegin("successful", ::apache::thrift::protocol::T_BOOL, 1); - xfer += oprot->writeBool(this->successful); + xfer += oprot->writeFieldBegin("txnId", ::apache::thrift::protocol::T_I64, 1); + xfer += oprot->writeI64(this->txnId); xfer += oprot->writeFieldEnd(); - xfer += oprot->writeFieldBegin("data", ::apache::thrift::protocol::T_STRUCT, 2); - xfer += this->data.write(oprot); + xfer += oprot->writeFieldBegin("writeId", ::apache::thrift::protocol::T_I64, 2); + xfer += oprot->writeI64(this->writeId); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("db", ::apache::thrift::protocol::T_STRING, 3); + xfer += oprot->writeString(this->db); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("table", ::apache::thrift::protocol::T_STRING, 4); + xfer += oprot->writeString(this->table); + xfer += oprot->writeFieldEnd(); + + xfer += oprot->writeFieldBegin("fileInfo", ::apache::thrift::protocol::T_STRUCT, 5); + xfer += this->fileInfo.write(oprot); xfer += oprot->writeFieldEnd(); - if (this->__isset.dbName) { - xfer += oprot->writeFieldBegin("dbName", ::apache::thrift::protocol::T_STRING, 3); - xfer += oprot->writeString(this->dbName); - xfer += oprot->writeFieldEnd(); - } - if (this->__isset.tableName) { - xfer += oprot->writeFieldBegin("tableName", ::apache::thrift::protocol::T_STRING, 4); - xfer += oprot->writeString(this->tableName); - xfer += oprot->writeFieldEnd(); - } if (this->__isset.partitionVals) { - xfer += oprot->writeFieldBegin("partitionVals", ::apache::thrift::protocol::T_LIST, 5); + xfer += oprot->writeFieldBegin("partitionVals", ::apache::thrift::protocol::T_LIST, 6); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->partitionVals.size())); - std::vector ::const_iterator _iter870; - for (_iter870 = this->partitionVals.begin(); _iter870 != this->partitionVals.end(); ++_iter870) + std::vector ::const_iterator _iter894; + for (_iter894 = this->partitionVals.begin(); _iter894 != this->partitionVals.end(); ++_iter894) { - xfer += oprot->writeString((*_iter870)); + xfer += oprot->writeString((*_iter894)); } xfer += oprot->writeListEnd(); } xfer += oprot->writeFieldEnd(); } - if (this->__isset.catName) { - xfer += oprot->writeFieldBegin("catName", ::apache::thrift::protocol::T_STRING, 6); - xfer += oprot->writeString(this->catName); - xfer += oprot->writeFieldEnd(); - } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; } -void swap(FireEventRequest &a, FireEventRequest &b) { +void swap(WriteNotificationLogRequest &a, WriteNotificationLogRequest &b) { using ::std::swap; - swap(a.successful, b.successful); - swap(a.data, b.data); - swap(a.dbName, b.dbName); - swap(a.tableName, b.tableName); + swap(a.txnId, b.txnId); + swap(a.writeId, b.writeId); + swap(a.db, b.db); + swap(a.table, b.table); + swap(a.fileInfo, b.fileInfo); swap(a.partitionVals, b.partitionVals); - swap(a.catName, b.catName); swap(a.__isset, b.__isset); } -FireEventRequest::FireEventRequest(const FireEventRequest& other871) { - successful = other871.successful; - data = other871.data; - dbName = other871.dbName; - tableName = other871.tableName; - partitionVals = other871.partitionVals; - catName = other871.catName; - __isset = other871.__isset; -} -FireEventRequest& FireEventRequest::operator=(const FireEventRequest& other872) { - successful = other872.successful; - data = other872.data; - dbName = other872.dbName; - tableName = other872.tableName; - partitionVals = other872.partitionVals; - catName = other872.catName; - __isset = other872.__isset; +WriteNotificationLogRequest::WriteNotificationLogRequest(const WriteNotificationLogRequest& other895) { + txnId = other895.txnId; + writeId = other895.writeId; + db = other895.db; + table = other895.table; + fileInfo = other895.fileInfo; + partitionVals = other895.partitionVals; + __isset = other895.__isset; +} +WriteNotificationLogRequest& WriteNotificationLogRequest::operator=(const WriteNotificationLogRequest& other896) { + txnId = other896.txnId; + writeId = other896.writeId; + db = other896.db; + table = other896.table; + fileInfo = other896.fileInfo; + partitionVals = other896.partitionVals; + __isset = other896.__isset; return *this; } -void FireEventRequest::printTo(std::ostream& out) const { +void WriteNotificationLogRequest::printTo(std::ostream& out) const { using ::apache::thrift::to_string; - out << "FireEventRequest("; - out << "successful=" << to_string(successful); - out << ", " << "data=" << to_string(data); - out << ", " << "dbName="; (__isset.dbName ? (out << to_string(dbName)) : (out << "")); - out << ", " << "tableName="; (__isset.tableName ? (out << to_string(tableName)) : (out << "")); + out << "WriteNotificationLogRequest("; + out << "txnId=" << to_string(txnId); + out << ", " << "writeId=" << to_string(writeId); + out << ", " << "db=" << to_string(db); + out << ", " << "table=" << to_string(table); + out << ", " << "fileInfo=" << to_string(fileInfo); out << ", " << "partitionVals="; (__isset.partitionVals ? (out << to_string(partitionVals)) : (out << "")); - out << ", " << "catName="; (__isset.catName ? (out << to_string(catName)) : (out << "")); out << ")"; } -FireEventResponse::~FireEventResponse() throw() { +WriteNotificationLogResponse::~WriteNotificationLogResponse() throw() { } -uint32_t FireEventResponse::read(::apache::thrift::protocol::TProtocol* iprot) { +uint32_t WriteNotificationLogResponse::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); uint32_t xfer = 0; @@ -22432,32 +23025,32 @@ uint32_t FireEventResponse::read(::apache::thrift::protocol::TProtocol* iprot) { return xfer; } -uint32_t FireEventResponse::write(::apache::thrift::protocol::TProtocol* oprot) const { +uint32_t WriteNotificationLogResponse::write(::apache::thrift::protocol::TProtocol* oprot) const { uint32_t xfer = 0; apache::thrift::protocol::TOutputRecursionTracker tracker(*oprot); - xfer += oprot->writeStructBegin("FireEventResponse"); + xfer += oprot->writeStructBegin("WriteNotificationLogResponse"); xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; } -void swap(FireEventResponse &a, FireEventResponse &b) { +void swap(WriteNotificationLogResponse &a, WriteNotificationLogResponse &b) { using ::std::swap; (void) a; (void) b; } -FireEventResponse::FireEventResponse(const FireEventResponse& other873) { - (void) other873; +WriteNotificationLogResponse::WriteNotificationLogResponse(const WriteNotificationLogResponse& other897) { + (void) other897; } -FireEventResponse& FireEventResponse::operator=(const FireEventResponse& other874) { - (void) other874; +WriteNotificationLogResponse& WriteNotificationLogResponse::operator=(const WriteNotificationLogResponse& other898) { + (void) other898; return *this; } -void FireEventResponse::printTo(std::ostream& out) const { +void WriteNotificationLogResponse::printTo(std::ostream& out) const { using ::apache::thrift::to_string; - out << "FireEventResponse("; + out << "WriteNotificationLogResponse("; out << ")"; } @@ -22552,15 +23145,15 @@ void swap(MetadataPpdResult &a, MetadataPpdResult &b) { swap(a.__isset, b.__isset); } -MetadataPpdResult::MetadataPpdResult(const MetadataPpdResult& other875) { - metadata = other875.metadata; - includeBitset = other875.includeBitset; - __isset = other875.__isset; +MetadataPpdResult::MetadataPpdResult(const MetadataPpdResult& other899) { + metadata = other899.metadata; + includeBitset = other899.includeBitset; + __isset = other899.__isset; } -MetadataPpdResult& MetadataPpdResult::operator=(const MetadataPpdResult& other876) { - metadata = other876.metadata; - includeBitset = other876.includeBitset; - __isset = other876.__isset; +MetadataPpdResult& MetadataPpdResult::operator=(const MetadataPpdResult& other900) { + metadata = other900.metadata; + includeBitset = other900.includeBitset; + __isset = other900.__isset; return *this; } void MetadataPpdResult::printTo(std::ostream& out) const { @@ -22611,17 +23204,17 @@ uint32_t GetFileMetadataByExprResult::read(::apache::thrift::protocol::TProtocol if (ftype == ::apache::thrift::protocol::T_MAP) { { this->metadata.clear(); - uint32_t _size877; - ::apache::thrift::protocol::TType _ktype878; - ::apache::thrift::protocol::TType _vtype879; - xfer += iprot->readMapBegin(_ktype878, _vtype879, _size877); - uint32_t _i881; - for (_i881 = 0; _i881 < _size877; ++_i881) + uint32_t _size901; + ::apache::thrift::protocol::TType _ktype902; + ::apache::thrift::protocol::TType _vtype903; + xfer += iprot->readMapBegin(_ktype902, _vtype903, _size901); + uint32_t _i905; + for (_i905 = 0; _i905 < _size901; ++_i905) { - int64_t _key882; - xfer += iprot->readI64(_key882); - MetadataPpdResult& _val883 = this->metadata[_key882]; - xfer += _val883.read(iprot); + int64_t _key906; + xfer += iprot->readI64(_key906); + MetadataPpdResult& _val907 = this->metadata[_key906]; + xfer += _val907.read(iprot); } xfer += iprot->readMapEnd(); } @@ -22662,11 +23255,11 @@ uint32_t GetFileMetadataByExprResult::write(::apache::thrift::protocol::TProtoco xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRUCT, static_cast(this->metadata.size())); - std::map ::const_iterator _iter884; - for (_iter884 = this->metadata.begin(); _iter884 != this->metadata.end(); ++_iter884) + std::map ::const_iterator _iter908; + for (_iter908 = this->metadata.begin(); _iter908 != this->metadata.end(); ++_iter908) { - xfer += oprot->writeI64(_iter884->first); - xfer += _iter884->second.write(oprot); + xfer += oprot->writeI64(_iter908->first); + xfer += _iter908->second.write(oprot); } xfer += oprot->writeMapEnd(); } @@ -22687,13 +23280,13 @@ void swap(GetFileMetadataByExprResult &a, GetFileMetadataByExprResult &b) { swap(a.isSupported, b.isSupported); } -GetFileMetadataByExprResult::GetFileMetadataByExprResult(const GetFileMetadataByExprResult& other885) { - metadata = other885.metadata; - isSupported = other885.isSupported; +GetFileMetadataByExprResult::GetFileMetadataByExprResult(const GetFileMetadataByExprResult& other909) { + metadata = other909.metadata; + isSupported = other909.isSupported; } -GetFileMetadataByExprResult& GetFileMetadataByExprResult::operator=(const GetFileMetadataByExprResult& other886) { - metadata = other886.metadata; - isSupported = other886.isSupported; +GetFileMetadataByExprResult& GetFileMetadataByExprResult::operator=(const GetFileMetadataByExprResult& other910) { + metadata = other910.metadata; + isSupported = other910.isSupported; return *this; } void GetFileMetadataByExprResult::printTo(std::ostream& out) const { @@ -22754,14 +23347,14 @@ uint32_t GetFileMetadataByExprRequest::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size887; - ::apache::thrift::protocol::TType _etype890; - xfer += iprot->readListBegin(_etype890, _size887); - this->fileIds.resize(_size887); - uint32_t _i891; - for (_i891 = 0; _i891 < _size887; ++_i891) + uint32_t _size911; + ::apache::thrift::protocol::TType _etype914; + xfer += iprot->readListBegin(_etype914, _size911); + this->fileIds.resize(_size911); + uint32_t _i915; + for (_i915 = 0; _i915 < _size911; ++_i915) { - xfer += iprot->readI64(this->fileIds[_i891]); + xfer += iprot->readI64(this->fileIds[_i915]); } xfer += iprot->readListEnd(); } @@ -22788,9 +23381,9 @@ uint32_t GetFileMetadataByExprRequest::read(::apache::thrift::protocol::TProtoco break; case 4: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast892; - xfer += iprot->readI32(ecast892); - this->type = (FileMetadataExprType::type)ecast892; + int32_t ecast916; + xfer += iprot->readI32(ecast916); + this->type = (FileMetadataExprType::type)ecast916; this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -22820,10 +23413,10 @@ uint32_t GetFileMetadataByExprRequest::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter893; - for (_iter893 = this->fileIds.begin(); _iter893 != this->fileIds.end(); ++_iter893) + std::vector ::const_iterator _iter917; + for (_iter917 = this->fileIds.begin(); _iter917 != this->fileIds.end(); ++_iter917) { - xfer += oprot->writeI64((*_iter893)); + xfer += oprot->writeI64((*_iter917)); } xfer += oprot->writeListEnd(); } @@ -22857,19 +23450,19 @@ void swap(GetFileMetadataByExprRequest &a, GetFileMetadataByExprRequest &b) { swap(a.__isset, b.__isset); } -GetFileMetadataByExprRequest::GetFileMetadataByExprRequest(const GetFileMetadataByExprRequest& other894) { - fileIds = other894.fileIds; - expr = other894.expr; - doGetFooters = other894.doGetFooters; - type = other894.type; - __isset = other894.__isset; +GetFileMetadataByExprRequest::GetFileMetadataByExprRequest(const GetFileMetadataByExprRequest& other918) { + fileIds = other918.fileIds; + expr = other918.expr; + doGetFooters = other918.doGetFooters; + type = other918.type; + __isset = other918.__isset; } -GetFileMetadataByExprRequest& GetFileMetadataByExprRequest::operator=(const GetFileMetadataByExprRequest& other895) { - fileIds = other895.fileIds; - expr = other895.expr; - doGetFooters = other895.doGetFooters; - type = other895.type; - __isset = other895.__isset; +GetFileMetadataByExprRequest& GetFileMetadataByExprRequest::operator=(const GetFileMetadataByExprRequest& other919) { + fileIds = other919.fileIds; + expr = other919.expr; + doGetFooters = other919.doGetFooters; + type = other919.type; + __isset = other919.__isset; return *this; } void GetFileMetadataByExprRequest::printTo(std::ostream& out) const { @@ -22922,17 +23515,17 @@ uint32_t GetFileMetadataResult::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_MAP) { { this->metadata.clear(); - uint32_t _size896; - ::apache::thrift::protocol::TType _ktype897; - ::apache::thrift::protocol::TType _vtype898; - xfer += iprot->readMapBegin(_ktype897, _vtype898, _size896); - uint32_t _i900; - for (_i900 = 0; _i900 < _size896; ++_i900) + uint32_t _size920; + ::apache::thrift::protocol::TType _ktype921; + ::apache::thrift::protocol::TType _vtype922; + xfer += iprot->readMapBegin(_ktype921, _vtype922, _size920); + uint32_t _i924; + for (_i924 = 0; _i924 < _size920; ++_i924) { - int64_t _key901; - xfer += iprot->readI64(_key901); - std::string& _val902 = this->metadata[_key901]; - xfer += iprot->readBinary(_val902); + int64_t _key925; + xfer += iprot->readI64(_key925); + std::string& _val926 = this->metadata[_key925]; + xfer += iprot->readBinary(_val926); } xfer += iprot->readMapEnd(); } @@ -22973,11 +23566,11 @@ uint32_t GetFileMetadataResult::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_MAP, 1); { xfer += oprot->writeMapBegin(::apache::thrift::protocol::T_I64, ::apache::thrift::protocol::T_STRING, static_cast(this->metadata.size())); - std::map ::const_iterator _iter903; - for (_iter903 = this->metadata.begin(); _iter903 != this->metadata.end(); ++_iter903) + std::map ::const_iterator _iter927; + for (_iter927 = this->metadata.begin(); _iter927 != this->metadata.end(); ++_iter927) { - xfer += oprot->writeI64(_iter903->first); - xfer += oprot->writeBinary(_iter903->second); + xfer += oprot->writeI64(_iter927->first); + xfer += oprot->writeBinary(_iter927->second); } xfer += oprot->writeMapEnd(); } @@ -22998,13 +23591,13 @@ void swap(GetFileMetadataResult &a, GetFileMetadataResult &b) { swap(a.isSupported, b.isSupported); } -GetFileMetadataResult::GetFileMetadataResult(const GetFileMetadataResult& other904) { - metadata = other904.metadata; - isSupported = other904.isSupported; +GetFileMetadataResult::GetFileMetadataResult(const GetFileMetadataResult& other928) { + metadata = other928.metadata; + isSupported = other928.isSupported; } -GetFileMetadataResult& GetFileMetadataResult::operator=(const GetFileMetadataResult& other905) { - metadata = other905.metadata; - isSupported = other905.isSupported; +GetFileMetadataResult& GetFileMetadataResult::operator=(const GetFileMetadataResult& other929) { + metadata = other929.metadata; + isSupported = other929.isSupported; return *this; } void GetFileMetadataResult::printTo(std::ostream& out) const { @@ -23050,14 +23643,14 @@ uint32_t GetFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size906; - ::apache::thrift::protocol::TType _etype909; - xfer += iprot->readListBegin(_etype909, _size906); - this->fileIds.resize(_size906); - uint32_t _i910; - for (_i910 = 0; _i910 < _size906; ++_i910) + uint32_t _size930; + ::apache::thrift::protocol::TType _etype933; + xfer += iprot->readListBegin(_etype933, _size930); + this->fileIds.resize(_size930); + uint32_t _i934; + for (_i934 = 0; _i934 < _size930; ++_i934) { - xfer += iprot->readI64(this->fileIds[_i910]); + xfer += iprot->readI64(this->fileIds[_i934]); } xfer += iprot->readListEnd(); } @@ -23088,10 +23681,10 @@ uint32_t GetFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter911; - for (_iter911 = this->fileIds.begin(); _iter911 != this->fileIds.end(); ++_iter911) + std::vector ::const_iterator _iter935; + for (_iter935 = this->fileIds.begin(); _iter935 != this->fileIds.end(); ++_iter935) { - xfer += oprot->writeI64((*_iter911)); + xfer += oprot->writeI64((*_iter935)); } xfer += oprot->writeListEnd(); } @@ -23107,11 +23700,11 @@ void swap(GetFileMetadataRequest &a, GetFileMetadataRequest &b) { swap(a.fileIds, b.fileIds); } -GetFileMetadataRequest::GetFileMetadataRequest(const GetFileMetadataRequest& other912) { - fileIds = other912.fileIds; +GetFileMetadataRequest::GetFileMetadataRequest(const GetFileMetadataRequest& other936) { + fileIds = other936.fileIds; } -GetFileMetadataRequest& GetFileMetadataRequest::operator=(const GetFileMetadataRequest& other913) { - fileIds = other913.fileIds; +GetFileMetadataRequest& GetFileMetadataRequest::operator=(const GetFileMetadataRequest& other937) { + fileIds = other937.fileIds; return *this; } void GetFileMetadataRequest::printTo(std::ostream& out) const { @@ -23170,11 +23763,11 @@ void swap(PutFileMetadataResult &a, PutFileMetadataResult &b) { (void) b; } -PutFileMetadataResult::PutFileMetadataResult(const PutFileMetadataResult& other914) { - (void) other914; +PutFileMetadataResult::PutFileMetadataResult(const PutFileMetadataResult& other938) { + (void) other938; } -PutFileMetadataResult& PutFileMetadataResult::operator=(const PutFileMetadataResult& other915) { - (void) other915; +PutFileMetadataResult& PutFileMetadataResult::operator=(const PutFileMetadataResult& other939) { + (void) other939; return *this; } void PutFileMetadataResult::printTo(std::ostream& out) const { @@ -23228,14 +23821,14 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size916; - ::apache::thrift::protocol::TType _etype919; - xfer += iprot->readListBegin(_etype919, _size916); - this->fileIds.resize(_size916); - uint32_t _i920; - for (_i920 = 0; _i920 < _size916; ++_i920) + uint32_t _size940; + ::apache::thrift::protocol::TType _etype943; + xfer += iprot->readListBegin(_etype943, _size940); + this->fileIds.resize(_size940); + uint32_t _i944; + for (_i944 = 0; _i944 < _size940; ++_i944) { - xfer += iprot->readI64(this->fileIds[_i920]); + xfer += iprot->readI64(this->fileIds[_i944]); } xfer += iprot->readListEnd(); } @@ -23248,14 +23841,14 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr if (ftype == ::apache::thrift::protocol::T_LIST) { { this->metadata.clear(); - uint32_t _size921; - ::apache::thrift::protocol::TType _etype924; - xfer += iprot->readListBegin(_etype924, _size921); - this->metadata.resize(_size921); - uint32_t _i925; - for (_i925 = 0; _i925 < _size921; ++_i925) + uint32_t _size945; + ::apache::thrift::protocol::TType _etype948; + xfer += iprot->readListBegin(_etype948, _size945); + this->metadata.resize(_size945); + uint32_t _i949; + for (_i949 = 0; _i949 < _size945; ++_i949) { - xfer += iprot->readBinary(this->metadata[_i925]); + xfer += iprot->readBinary(this->metadata[_i949]); } xfer += iprot->readListEnd(); } @@ -23266,9 +23859,9 @@ uint32_t PutFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* ipr break; case 3: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast926; - xfer += iprot->readI32(ecast926); - this->type = (FileMetadataExprType::type)ecast926; + int32_t ecast950; + xfer += iprot->readI32(ecast950); + this->type = (FileMetadataExprType::type)ecast950; this->__isset.type = true; } else { xfer += iprot->skip(ftype); @@ -23298,10 +23891,10 @@ uint32_t PutFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter927; - for (_iter927 = this->fileIds.begin(); _iter927 != this->fileIds.end(); ++_iter927) + std::vector ::const_iterator _iter951; + for (_iter951 = this->fileIds.begin(); _iter951 != this->fileIds.end(); ++_iter951) { - xfer += oprot->writeI64((*_iter927)); + xfer += oprot->writeI64((*_iter951)); } xfer += oprot->writeListEnd(); } @@ -23310,10 +23903,10 @@ uint32_t PutFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* op xfer += oprot->writeFieldBegin("metadata", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->metadata.size())); - std::vector ::const_iterator _iter928; - for (_iter928 = this->metadata.begin(); _iter928 != this->metadata.end(); ++_iter928) + std::vector ::const_iterator _iter952; + for (_iter952 = this->metadata.begin(); _iter952 != this->metadata.end(); ++_iter952) { - xfer += oprot->writeBinary((*_iter928)); + xfer += oprot->writeBinary((*_iter952)); } xfer += oprot->writeListEnd(); } @@ -23337,17 +23930,17 @@ void swap(PutFileMetadataRequest &a, PutFileMetadataRequest &b) { swap(a.__isset, b.__isset); } -PutFileMetadataRequest::PutFileMetadataRequest(const PutFileMetadataRequest& other929) { - fileIds = other929.fileIds; - metadata = other929.metadata; - type = other929.type; - __isset = other929.__isset; -} -PutFileMetadataRequest& PutFileMetadataRequest::operator=(const PutFileMetadataRequest& other930) { - fileIds = other930.fileIds; - metadata = other930.metadata; - type = other930.type; - __isset = other930.__isset; +PutFileMetadataRequest::PutFileMetadataRequest(const PutFileMetadataRequest& other953) { + fileIds = other953.fileIds; + metadata = other953.metadata; + type = other953.type; + __isset = other953.__isset; +} +PutFileMetadataRequest& PutFileMetadataRequest::operator=(const PutFileMetadataRequest& other954) { + fileIds = other954.fileIds; + metadata = other954.metadata; + type = other954.type; + __isset = other954.__isset; return *this; } void PutFileMetadataRequest::printTo(std::ostream& out) const { @@ -23408,11 +24001,11 @@ void swap(ClearFileMetadataResult &a, ClearFileMetadataResult &b) { (void) b; } -ClearFileMetadataResult::ClearFileMetadataResult(const ClearFileMetadataResult& other931) { - (void) other931; +ClearFileMetadataResult::ClearFileMetadataResult(const ClearFileMetadataResult& other955) { + (void) other955; } -ClearFileMetadataResult& ClearFileMetadataResult::operator=(const ClearFileMetadataResult& other932) { - (void) other932; +ClearFileMetadataResult& ClearFileMetadataResult::operator=(const ClearFileMetadataResult& other956) { + (void) other956; return *this; } void ClearFileMetadataResult::printTo(std::ostream& out) const { @@ -23456,14 +24049,14 @@ uint32_t ClearFileMetadataRequest::read(::apache::thrift::protocol::TProtocol* i if (ftype == ::apache::thrift::protocol::T_LIST) { { this->fileIds.clear(); - uint32_t _size933; - ::apache::thrift::protocol::TType _etype936; - xfer += iprot->readListBegin(_etype936, _size933); - this->fileIds.resize(_size933); - uint32_t _i937; - for (_i937 = 0; _i937 < _size933; ++_i937) + uint32_t _size957; + ::apache::thrift::protocol::TType _etype960; + xfer += iprot->readListBegin(_etype960, _size957); + this->fileIds.resize(_size957); + uint32_t _i961; + for (_i961 = 0; _i961 < _size957; ++_i961) { - xfer += iprot->readI64(this->fileIds[_i937]); + xfer += iprot->readI64(this->fileIds[_i961]); } xfer += iprot->readListEnd(); } @@ -23494,10 +24087,10 @@ uint32_t ClearFileMetadataRequest::write(::apache::thrift::protocol::TProtocol* xfer += oprot->writeFieldBegin("fileIds", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I64, static_cast(this->fileIds.size())); - std::vector ::const_iterator _iter938; - for (_iter938 = this->fileIds.begin(); _iter938 != this->fileIds.end(); ++_iter938) + std::vector ::const_iterator _iter962; + for (_iter962 = this->fileIds.begin(); _iter962 != this->fileIds.end(); ++_iter962) { - xfer += oprot->writeI64((*_iter938)); + xfer += oprot->writeI64((*_iter962)); } xfer += oprot->writeListEnd(); } @@ -23513,11 +24106,11 @@ void swap(ClearFileMetadataRequest &a, ClearFileMetadataRequest &b) { swap(a.fileIds, b.fileIds); } -ClearFileMetadataRequest::ClearFileMetadataRequest(const ClearFileMetadataRequest& other939) { - fileIds = other939.fileIds; +ClearFileMetadataRequest::ClearFileMetadataRequest(const ClearFileMetadataRequest& other963) { + fileIds = other963.fileIds; } -ClearFileMetadataRequest& ClearFileMetadataRequest::operator=(const ClearFileMetadataRequest& other940) { - fileIds = other940.fileIds; +ClearFileMetadataRequest& ClearFileMetadataRequest::operator=(const ClearFileMetadataRequest& other964) { + fileIds = other964.fileIds; return *this; } void ClearFileMetadataRequest::printTo(std::ostream& out) const { @@ -23599,11 +24192,11 @@ void swap(CacheFileMetadataResult &a, CacheFileMetadataResult &b) { swap(a.isSupported, b.isSupported); } -CacheFileMetadataResult::CacheFileMetadataResult(const CacheFileMetadataResult& other941) { - isSupported = other941.isSupported; +CacheFileMetadataResult::CacheFileMetadataResult(const CacheFileMetadataResult& other965) { + isSupported = other965.isSupported; } -CacheFileMetadataResult& CacheFileMetadataResult::operator=(const CacheFileMetadataResult& other942) { - isSupported = other942.isSupported; +CacheFileMetadataResult& CacheFileMetadataResult::operator=(const CacheFileMetadataResult& other966) { + isSupported = other966.isSupported; return *this; } void CacheFileMetadataResult::printTo(std::ostream& out) const { @@ -23744,19 +24337,19 @@ void swap(CacheFileMetadataRequest &a, CacheFileMetadataRequest &b) { swap(a.__isset, b.__isset); } -CacheFileMetadataRequest::CacheFileMetadataRequest(const CacheFileMetadataRequest& other943) { - dbName = other943.dbName; - tblName = other943.tblName; - partName = other943.partName; - isAllParts = other943.isAllParts; - __isset = other943.__isset; +CacheFileMetadataRequest::CacheFileMetadataRequest(const CacheFileMetadataRequest& other967) { + dbName = other967.dbName; + tblName = other967.tblName; + partName = other967.partName; + isAllParts = other967.isAllParts; + __isset = other967.__isset; } -CacheFileMetadataRequest& CacheFileMetadataRequest::operator=(const CacheFileMetadataRequest& other944) { - dbName = other944.dbName; - tblName = other944.tblName; - partName = other944.partName; - isAllParts = other944.isAllParts; - __isset = other944.__isset; +CacheFileMetadataRequest& CacheFileMetadataRequest::operator=(const CacheFileMetadataRequest& other968) { + dbName = other968.dbName; + tblName = other968.tblName; + partName = other968.partName; + isAllParts = other968.isAllParts; + __isset = other968.__isset; return *this; } void CacheFileMetadataRequest::printTo(std::ostream& out) const { @@ -23804,14 +24397,14 @@ uint32_t GetAllFunctionsResponse::read(::apache::thrift::protocol::TProtocol* ip if (ftype == ::apache::thrift::protocol::T_LIST) { { this->functions.clear(); - uint32_t _size945; - ::apache::thrift::protocol::TType _etype948; - xfer += iprot->readListBegin(_etype948, _size945); - this->functions.resize(_size945); - uint32_t _i949; - for (_i949 = 0; _i949 < _size945; ++_i949) + uint32_t _size969; + ::apache::thrift::protocol::TType _etype972; + xfer += iprot->readListBegin(_etype972, _size969); + this->functions.resize(_size969); + uint32_t _i973; + for (_i973 = 0; _i973 < _size969; ++_i973) { - xfer += this->functions[_i949].read(iprot); + xfer += this->functions[_i973].read(iprot); } xfer += iprot->readListEnd(); } @@ -23841,10 +24434,10 @@ uint32_t GetAllFunctionsResponse::write(::apache::thrift::protocol::TProtocol* o xfer += oprot->writeFieldBegin("functions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->functions.size())); - std::vector ::const_iterator _iter950; - for (_iter950 = this->functions.begin(); _iter950 != this->functions.end(); ++_iter950) + std::vector ::const_iterator _iter974; + for (_iter974 = this->functions.begin(); _iter974 != this->functions.end(); ++_iter974) { - xfer += (*_iter950).write(oprot); + xfer += (*_iter974).write(oprot); } xfer += oprot->writeListEnd(); } @@ -23861,13 +24454,13 @@ void swap(GetAllFunctionsResponse &a, GetAllFunctionsResponse &b) { swap(a.__isset, b.__isset); } -GetAllFunctionsResponse::GetAllFunctionsResponse(const GetAllFunctionsResponse& other951) { - functions = other951.functions; - __isset = other951.__isset; +GetAllFunctionsResponse::GetAllFunctionsResponse(const GetAllFunctionsResponse& other975) { + functions = other975.functions; + __isset = other975.__isset; } -GetAllFunctionsResponse& GetAllFunctionsResponse::operator=(const GetAllFunctionsResponse& other952) { - functions = other952.functions; - __isset = other952.__isset; +GetAllFunctionsResponse& GetAllFunctionsResponse::operator=(const GetAllFunctionsResponse& other976) { + functions = other976.functions; + __isset = other976.__isset; return *this; } void GetAllFunctionsResponse::printTo(std::ostream& out) const { @@ -23912,16 +24505,16 @@ uint32_t ClientCapabilities::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->values.clear(); - uint32_t _size953; - ::apache::thrift::protocol::TType _etype956; - xfer += iprot->readListBegin(_etype956, _size953); - this->values.resize(_size953); - uint32_t _i957; - for (_i957 = 0; _i957 < _size953; ++_i957) + uint32_t _size977; + ::apache::thrift::protocol::TType _etype980; + xfer += iprot->readListBegin(_etype980, _size977); + this->values.resize(_size977); + uint32_t _i981; + for (_i981 = 0; _i981 < _size977; ++_i981) { - int32_t ecast958; - xfer += iprot->readI32(ecast958); - this->values[_i957] = (ClientCapability::type)ecast958; + int32_t ecast982; + xfer += iprot->readI32(ecast982); + this->values[_i981] = (ClientCapability::type)ecast982; } xfer += iprot->readListEnd(); } @@ -23952,10 +24545,10 @@ uint32_t ClientCapabilities::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("values", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_I32, static_cast(this->values.size())); - std::vector ::const_iterator _iter959; - for (_iter959 = this->values.begin(); _iter959 != this->values.end(); ++_iter959) + std::vector ::const_iterator _iter983; + for (_iter983 = this->values.begin(); _iter983 != this->values.end(); ++_iter983) { - xfer += oprot->writeI32((int32_t)(*_iter959)); + xfer += oprot->writeI32((int32_t)(*_iter983)); } xfer += oprot->writeListEnd(); } @@ -23971,11 +24564,11 @@ void swap(ClientCapabilities &a, ClientCapabilities &b) { swap(a.values, b.values); } -ClientCapabilities::ClientCapabilities(const ClientCapabilities& other960) { - values = other960.values; +ClientCapabilities::ClientCapabilities(const ClientCapabilities& other984) { + values = other984.values; } -ClientCapabilities& ClientCapabilities::operator=(const ClientCapabilities& other961) { - values = other961.values; +ClientCapabilities& ClientCapabilities::operator=(const ClientCapabilities& other985) { + values = other985.values; return *this; } void ClientCapabilities::printTo(std::ostream& out) const { @@ -24116,19 +24709,19 @@ void swap(GetTableRequest &a, GetTableRequest &b) { swap(a.__isset, b.__isset); } -GetTableRequest::GetTableRequest(const GetTableRequest& other962) { - dbName = other962.dbName; - tblName = other962.tblName; - capabilities = other962.capabilities; - catName = other962.catName; - __isset = other962.__isset; +GetTableRequest::GetTableRequest(const GetTableRequest& other986) { + dbName = other986.dbName; + tblName = other986.tblName; + capabilities = other986.capabilities; + catName = other986.catName; + __isset = other986.__isset; } -GetTableRequest& GetTableRequest::operator=(const GetTableRequest& other963) { - dbName = other963.dbName; - tblName = other963.tblName; - capabilities = other963.capabilities; - catName = other963.catName; - __isset = other963.__isset; +GetTableRequest& GetTableRequest::operator=(const GetTableRequest& other987) { + dbName = other987.dbName; + tblName = other987.tblName; + capabilities = other987.capabilities; + catName = other987.catName; + __isset = other987.__isset; return *this; } void GetTableRequest::printTo(std::ostream& out) const { @@ -24213,11 +24806,11 @@ void swap(GetTableResult &a, GetTableResult &b) { swap(a.table, b.table); } -GetTableResult::GetTableResult(const GetTableResult& other964) { - table = other964.table; +GetTableResult::GetTableResult(const GetTableResult& other988) { + table = other988.table; } -GetTableResult& GetTableResult::operator=(const GetTableResult& other965) { - table = other965.table; +GetTableResult& GetTableResult::operator=(const GetTableResult& other989) { + table = other989.table; return *this; } void GetTableResult::printTo(std::ostream& out) const { @@ -24285,14 +24878,14 @@ uint32_t GetTablesRequest::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tblNames.clear(); - uint32_t _size966; - ::apache::thrift::protocol::TType _etype969; - xfer += iprot->readListBegin(_etype969, _size966); - this->tblNames.resize(_size966); - uint32_t _i970; - for (_i970 = 0; _i970 < _size966; ++_i970) + uint32_t _size990; + ::apache::thrift::protocol::TType _etype993; + xfer += iprot->readListBegin(_etype993, _size990); + this->tblNames.resize(_size990); + uint32_t _i994; + for (_i994 = 0; _i994 < _size990; ++_i994) { - xfer += iprot->readString(this->tblNames[_i970]); + xfer += iprot->readString(this->tblNames[_i994]); } xfer += iprot->readListEnd(); } @@ -24344,10 +24937,10 @@ uint32_t GetTablesRequest::write(::apache::thrift::protocol::TProtocol* oprot) c xfer += oprot->writeFieldBegin("tblNames", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->tblNames.size())); - std::vector ::const_iterator _iter971; - for (_iter971 = this->tblNames.begin(); _iter971 != this->tblNames.end(); ++_iter971) + std::vector ::const_iterator _iter995; + for (_iter995 = this->tblNames.begin(); _iter995 != this->tblNames.end(); ++_iter995) { - xfer += oprot->writeString((*_iter971)); + xfer += oprot->writeString((*_iter995)); } xfer += oprot->writeListEnd(); } @@ -24377,19 +24970,19 @@ void swap(GetTablesRequest &a, GetTablesRequest &b) { swap(a.__isset, b.__isset); } -GetTablesRequest::GetTablesRequest(const GetTablesRequest& other972) { - dbName = other972.dbName; - tblNames = other972.tblNames; - capabilities = other972.capabilities; - catName = other972.catName; - __isset = other972.__isset; +GetTablesRequest::GetTablesRequest(const GetTablesRequest& other996) { + dbName = other996.dbName; + tblNames = other996.tblNames; + capabilities = other996.capabilities; + catName = other996.catName; + __isset = other996.__isset; } -GetTablesRequest& GetTablesRequest::operator=(const GetTablesRequest& other973) { - dbName = other973.dbName; - tblNames = other973.tblNames; - capabilities = other973.capabilities; - catName = other973.catName; - __isset = other973.__isset; +GetTablesRequest& GetTablesRequest::operator=(const GetTablesRequest& other997) { + dbName = other997.dbName; + tblNames = other997.tblNames; + capabilities = other997.capabilities; + catName = other997.catName; + __isset = other997.__isset; return *this; } void GetTablesRequest::printTo(std::ostream& out) const { @@ -24437,14 +25030,14 @@ uint32_t GetTablesResult::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->tables.clear(); - uint32_t _size974; - ::apache::thrift::protocol::TType _etype977; - xfer += iprot->readListBegin(_etype977, _size974); - this->tables.resize(_size974); - uint32_t _i978; - for (_i978 = 0; _i978 < _size974; ++_i978) + uint32_t _size998; + ::apache::thrift::protocol::TType _etype1001; + xfer += iprot->readListBegin(_etype1001, _size998); + this->tables.resize(_size998); + uint32_t _i1002; + for (_i1002 = 0; _i1002 < _size998; ++_i1002) { - xfer += this->tables[_i978].read(iprot); + xfer += this->tables[_i1002].read(iprot); } xfer += iprot->readListEnd(); } @@ -24475,10 +25068,10 @@ uint32_t GetTablesResult::write(::apache::thrift::protocol::TProtocol* oprot) co xfer += oprot->writeFieldBegin("tables", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->tables.size())); - std::vector
::const_iterator _iter979; - for (_iter979 = this->tables.begin(); _iter979 != this->tables.end(); ++_iter979) + std::vector
::const_iterator _iter1003; + for (_iter1003 = this->tables.begin(); _iter1003 != this->tables.end(); ++_iter1003) { - xfer += (*_iter979).write(oprot); + xfer += (*_iter1003).write(oprot); } xfer += oprot->writeListEnd(); } @@ -24494,11 +25087,11 @@ void swap(GetTablesResult &a, GetTablesResult &b) { swap(a.tables, b.tables); } -GetTablesResult::GetTablesResult(const GetTablesResult& other980) { - tables = other980.tables; +GetTablesResult::GetTablesResult(const GetTablesResult& other1004) { + tables = other1004.tables; } -GetTablesResult& GetTablesResult::operator=(const GetTablesResult& other981) { - tables = other981.tables; +GetTablesResult& GetTablesResult::operator=(const GetTablesResult& other1005) { + tables = other1005.tables; return *this; } void GetTablesResult::printTo(std::ostream& out) const { @@ -24600,13 +25193,13 @@ void swap(CmRecycleRequest &a, CmRecycleRequest &b) { swap(a.purge, b.purge); } -CmRecycleRequest::CmRecycleRequest(const CmRecycleRequest& other982) { - dataPath = other982.dataPath; - purge = other982.purge; +CmRecycleRequest::CmRecycleRequest(const CmRecycleRequest& other1006) { + dataPath = other1006.dataPath; + purge = other1006.purge; } -CmRecycleRequest& CmRecycleRequest::operator=(const CmRecycleRequest& other983) { - dataPath = other983.dataPath; - purge = other983.purge; +CmRecycleRequest& CmRecycleRequest::operator=(const CmRecycleRequest& other1007) { + dataPath = other1007.dataPath; + purge = other1007.purge; return *this; } void CmRecycleRequest::printTo(std::ostream& out) const { @@ -24666,11 +25259,11 @@ void swap(CmRecycleResponse &a, CmRecycleResponse &b) { (void) b; } -CmRecycleResponse::CmRecycleResponse(const CmRecycleResponse& other984) { - (void) other984; +CmRecycleResponse::CmRecycleResponse(const CmRecycleResponse& other1008) { + (void) other1008; } -CmRecycleResponse& CmRecycleResponse::operator=(const CmRecycleResponse& other985) { - (void) other985; +CmRecycleResponse& CmRecycleResponse::operator=(const CmRecycleResponse& other1009) { + (void) other1009; return *this; } void CmRecycleResponse::printTo(std::ostream& out) const { @@ -24830,21 +25423,21 @@ void swap(TableMeta &a, TableMeta &b) { swap(a.__isset, b.__isset); } -TableMeta::TableMeta(const TableMeta& other986) { - dbName = other986.dbName; - tableName = other986.tableName; - tableType = other986.tableType; - comments = other986.comments; - catName = other986.catName; - __isset = other986.__isset; -} -TableMeta& TableMeta::operator=(const TableMeta& other987) { - dbName = other987.dbName; - tableName = other987.tableName; - tableType = other987.tableType; - comments = other987.comments; - catName = other987.catName; - __isset = other987.__isset; +TableMeta::TableMeta(const TableMeta& other1010) { + dbName = other1010.dbName; + tableName = other1010.tableName; + tableType = other1010.tableType; + comments = other1010.comments; + catName = other1010.catName; + __isset = other1010.__isset; +} +TableMeta& TableMeta::operator=(const TableMeta& other1011) { + dbName = other1011.dbName; + tableName = other1011.tableName; + tableType = other1011.tableType; + comments = other1011.comments; + catName = other1011.catName; + __isset = other1011.__isset; return *this; } void TableMeta::printTo(std::ostream& out) const { @@ -24930,11 +25523,11 @@ void swap(Materialization &a, Materialization &b) { swap(a.sourceTablesUpdateDeleteModified, b.sourceTablesUpdateDeleteModified); } -Materialization::Materialization(const Materialization& other988) { - sourceTablesUpdateDeleteModified = other988.sourceTablesUpdateDeleteModified; +Materialization::Materialization(const Materialization& other1012) { + sourceTablesUpdateDeleteModified = other1012.sourceTablesUpdateDeleteModified; } -Materialization& Materialization::operator=(const Materialization& other989) { - sourceTablesUpdateDeleteModified = other989.sourceTablesUpdateDeleteModified; +Materialization& Materialization::operator=(const Materialization& other1013) { + sourceTablesUpdateDeleteModified = other1013.sourceTablesUpdateDeleteModified; return *this; } void Materialization::printTo(std::ostream& out) const { @@ -25000,9 +25593,9 @@ uint32_t WMResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast990; - xfer += iprot->readI32(ecast990); - this->status = (WMResourcePlanStatus::type)ecast990; + int32_t ecast1014; + xfer += iprot->readI32(ecast1014); + this->status = (WMResourcePlanStatus::type)ecast1014; this->__isset.status = true; } else { xfer += iprot->skip(ftype); @@ -25076,19 +25669,19 @@ void swap(WMResourcePlan &a, WMResourcePlan &b) { swap(a.__isset, b.__isset); } -WMResourcePlan::WMResourcePlan(const WMResourcePlan& other991) { - name = other991.name; - status = other991.status; - queryParallelism = other991.queryParallelism; - defaultPoolPath = other991.defaultPoolPath; - __isset = other991.__isset; +WMResourcePlan::WMResourcePlan(const WMResourcePlan& other1015) { + name = other1015.name; + status = other1015.status; + queryParallelism = other1015.queryParallelism; + defaultPoolPath = other1015.defaultPoolPath; + __isset = other1015.__isset; } -WMResourcePlan& WMResourcePlan::operator=(const WMResourcePlan& other992) { - name = other992.name; - status = other992.status; - queryParallelism = other992.queryParallelism; - defaultPoolPath = other992.defaultPoolPath; - __isset = other992.__isset; +WMResourcePlan& WMResourcePlan::operator=(const WMResourcePlan& other1016) { + name = other1016.name; + status = other1016.status; + queryParallelism = other1016.queryParallelism; + defaultPoolPath = other1016.defaultPoolPath; + __isset = other1016.__isset; return *this; } void WMResourcePlan::printTo(std::ostream& out) const { @@ -25167,9 +25760,9 @@ uint32_t WMNullableResourcePlan::read(::apache::thrift::protocol::TProtocol* ipr break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast993; - xfer += iprot->readI32(ecast993); - this->status = (WMResourcePlanStatus::type)ecast993; + int32_t ecast1017; + xfer += iprot->readI32(ecast1017); + this->status = (WMResourcePlanStatus::type)ecast1017; this->__isset.status = true; } else { xfer += iprot->skip(ftype); @@ -25270,23 +25863,23 @@ void swap(WMNullableResourcePlan &a, WMNullableResourcePlan &b) { swap(a.__isset, b.__isset); } -WMNullableResourcePlan::WMNullableResourcePlan(const WMNullableResourcePlan& other994) { - name = other994.name; - status = other994.status; - queryParallelism = other994.queryParallelism; - isSetQueryParallelism = other994.isSetQueryParallelism; - defaultPoolPath = other994.defaultPoolPath; - isSetDefaultPoolPath = other994.isSetDefaultPoolPath; - __isset = other994.__isset; -} -WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResourcePlan& other995) { - name = other995.name; - status = other995.status; - queryParallelism = other995.queryParallelism; - isSetQueryParallelism = other995.isSetQueryParallelism; - defaultPoolPath = other995.defaultPoolPath; - isSetDefaultPoolPath = other995.isSetDefaultPoolPath; - __isset = other995.__isset; +WMNullableResourcePlan::WMNullableResourcePlan(const WMNullableResourcePlan& other1018) { + name = other1018.name; + status = other1018.status; + queryParallelism = other1018.queryParallelism; + isSetQueryParallelism = other1018.isSetQueryParallelism; + defaultPoolPath = other1018.defaultPoolPath; + isSetDefaultPoolPath = other1018.isSetDefaultPoolPath; + __isset = other1018.__isset; +} +WMNullableResourcePlan& WMNullableResourcePlan::operator=(const WMNullableResourcePlan& other1019) { + name = other1019.name; + status = other1019.status; + queryParallelism = other1019.queryParallelism; + isSetQueryParallelism = other1019.isSetQueryParallelism; + defaultPoolPath = other1019.defaultPoolPath; + isSetDefaultPoolPath = other1019.isSetDefaultPoolPath; + __isset = other1019.__isset; return *this; } void WMNullableResourcePlan::printTo(std::ostream& out) const { @@ -25451,21 +26044,21 @@ void swap(WMPool &a, WMPool &b) { swap(a.__isset, b.__isset); } -WMPool::WMPool(const WMPool& other996) { - resourcePlanName = other996.resourcePlanName; - poolPath = other996.poolPath; - allocFraction = other996.allocFraction; - queryParallelism = other996.queryParallelism; - schedulingPolicy = other996.schedulingPolicy; - __isset = other996.__isset; -} -WMPool& WMPool::operator=(const WMPool& other997) { - resourcePlanName = other997.resourcePlanName; - poolPath = other997.poolPath; - allocFraction = other997.allocFraction; - queryParallelism = other997.queryParallelism; - schedulingPolicy = other997.schedulingPolicy; - __isset = other997.__isset; +WMPool::WMPool(const WMPool& other1020) { + resourcePlanName = other1020.resourcePlanName; + poolPath = other1020.poolPath; + allocFraction = other1020.allocFraction; + queryParallelism = other1020.queryParallelism; + schedulingPolicy = other1020.schedulingPolicy; + __isset = other1020.__isset; +} +WMPool& WMPool::operator=(const WMPool& other1021) { + resourcePlanName = other1021.resourcePlanName; + poolPath = other1021.poolPath; + allocFraction = other1021.allocFraction; + queryParallelism = other1021.queryParallelism; + schedulingPolicy = other1021.schedulingPolicy; + __isset = other1021.__isset; return *this; } void WMPool::printTo(std::ostream& out) const { @@ -25648,23 +26241,23 @@ void swap(WMNullablePool &a, WMNullablePool &b) { swap(a.__isset, b.__isset); } -WMNullablePool::WMNullablePool(const WMNullablePool& other998) { - resourcePlanName = other998.resourcePlanName; - poolPath = other998.poolPath; - allocFraction = other998.allocFraction; - queryParallelism = other998.queryParallelism; - schedulingPolicy = other998.schedulingPolicy; - isSetSchedulingPolicy = other998.isSetSchedulingPolicy; - __isset = other998.__isset; -} -WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other999) { - resourcePlanName = other999.resourcePlanName; - poolPath = other999.poolPath; - allocFraction = other999.allocFraction; - queryParallelism = other999.queryParallelism; - schedulingPolicy = other999.schedulingPolicy; - isSetSchedulingPolicy = other999.isSetSchedulingPolicy; - __isset = other999.__isset; +WMNullablePool::WMNullablePool(const WMNullablePool& other1022) { + resourcePlanName = other1022.resourcePlanName; + poolPath = other1022.poolPath; + allocFraction = other1022.allocFraction; + queryParallelism = other1022.queryParallelism; + schedulingPolicy = other1022.schedulingPolicy; + isSetSchedulingPolicy = other1022.isSetSchedulingPolicy; + __isset = other1022.__isset; +} +WMNullablePool& WMNullablePool::operator=(const WMNullablePool& other1023) { + resourcePlanName = other1023.resourcePlanName; + poolPath = other1023.poolPath; + allocFraction = other1023.allocFraction; + queryParallelism = other1023.queryParallelism; + schedulingPolicy = other1023.schedulingPolicy; + isSetSchedulingPolicy = other1023.isSetSchedulingPolicy; + __isset = other1023.__isset; return *this; } void WMNullablePool::printTo(std::ostream& out) const { @@ -25829,21 +26422,21 @@ void swap(WMTrigger &a, WMTrigger &b) { swap(a.__isset, b.__isset); } -WMTrigger::WMTrigger(const WMTrigger& other1000) { - resourcePlanName = other1000.resourcePlanName; - triggerName = other1000.triggerName; - triggerExpression = other1000.triggerExpression; - actionExpression = other1000.actionExpression; - isInUnmanaged = other1000.isInUnmanaged; - __isset = other1000.__isset; -} -WMTrigger& WMTrigger::operator=(const WMTrigger& other1001) { - resourcePlanName = other1001.resourcePlanName; - triggerName = other1001.triggerName; - triggerExpression = other1001.triggerExpression; - actionExpression = other1001.actionExpression; - isInUnmanaged = other1001.isInUnmanaged; - __isset = other1001.__isset; +WMTrigger::WMTrigger(const WMTrigger& other1024) { + resourcePlanName = other1024.resourcePlanName; + triggerName = other1024.triggerName; + triggerExpression = other1024.triggerExpression; + actionExpression = other1024.actionExpression; + isInUnmanaged = other1024.isInUnmanaged; + __isset = other1024.__isset; +} +WMTrigger& WMTrigger::operator=(const WMTrigger& other1025) { + resourcePlanName = other1025.resourcePlanName; + triggerName = other1025.triggerName; + triggerExpression = other1025.triggerExpression; + actionExpression = other1025.actionExpression; + isInUnmanaged = other1025.isInUnmanaged; + __isset = other1025.__isset; return *this; } void WMTrigger::printTo(std::ostream& out) const { @@ -26008,21 +26601,21 @@ void swap(WMMapping &a, WMMapping &b) { swap(a.__isset, b.__isset); } -WMMapping::WMMapping(const WMMapping& other1002) { - resourcePlanName = other1002.resourcePlanName; - entityType = other1002.entityType; - entityName = other1002.entityName; - poolPath = other1002.poolPath; - ordering = other1002.ordering; - __isset = other1002.__isset; -} -WMMapping& WMMapping::operator=(const WMMapping& other1003) { - resourcePlanName = other1003.resourcePlanName; - entityType = other1003.entityType; - entityName = other1003.entityName; - poolPath = other1003.poolPath; - ordering = other1003.ordering; - __isset = other1003.__isset; +WMMapping::WMMapping(const WMMapping& other1026) { + resourcePlanName = other1026.resourcePlanName; + entityType = other1026.entityType; + entityName = other1026.entityName; + poolPath = other1026.poolPath; + ordering = other1026.ordering; + __isset = other1026.__isset; +} +WMMapping& WMMapping::operator=(const WMMapping& other1027) { + resourcePlanName = other1027.resourcePlanName; + entityType = other1027.entityType; + entityName = other1027.entityName; + poolPath = other1027.poolPath; + ordering = other1027.ordering; + __isset = other1027.__isset; return *this; } void WMMapping::printTo(std::ostream& out) const { @@ -26128,13 +26721,13 @@ void swap(WMPoolTrigger &a, WMPoolTrigger &b) { swap(a.trigger, b.trigger); } -WMPoolTrigger::WMPoolTrigger(const WMPoolTrigger& other1004) { - pool = other1004.pool; - trigger = other1004.trigger; +WMPoolTrigger::WMPoolTrigger(const WMPoolTrigger& other1028) { + pool = other1028.pool; + trigger = other1028.trigger; } -WMPoolTrigger& WMPoolTrigger::operator=(const WMPoolTrigger& other1005) { - pool = other1005.pool; - trigger = other1005.trigger; +WMPoolTrigger& WMPoolTrigger::operator=(const WMPoolTrigger& other1029) { + pool = other1029.pool; + trigger = other1029.trigger; return *this; } void WMPoolTrigger::printTo(std::ostream& out) const { @@ -26208,14 +26801,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->pools.clear(); - uint32_t _size1006; - ::apache::thrift::protocol::TType _etype1009; - xfer += iprot->readListBegin(_etype1009, _size1006); - this->pools.resize(_size1006); - uint32_t _i1010; - for (_i1010 = 0; _i1010 < _size1006; ++_i1010) + uint32_t _size1030; + ::apache::thrift::protocol::TType _etype1033; + xfer += iprot->readListBegin(_etype1033, _size1030); + this->pools.resize(_size1030); + uint32_t _i1034; + for (_i1034 = 0; _i1034 < _size1030; ++_i1034) { - xfer += this->pools[_i1010].read(iprot); + xfer += this->pools[_i1034].read(iprot); } xfer += iprot->readListEnd(); } @@ -26228,14 +26821,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->mappings.clear(); - uint32_t _size1011; - ::apache::thrift::protocol::TType _etype1014; - xfer += iprot->readListBegin(_etype1014, _size1011); - this->mappings.resize(_size1011); - uint32_t _i1015; - for (_i1015 = 0; _i1015 < _size1011; ++_i1015) + uint32_t _size1035; + ::apache::thrift::protocol::TType _etype1038; + xfer += iprot->readListBegin(_etype1038, _size1035); + this->mappings.resize(_size1035); + uint32_t _i1039; + for (_i1039 = 0; _i1039 < _size1035; ++_i1039) { - xfer += this->mappings[_i1015].read(iprot); + xfer += this->mappings[_i1039].read(iprot); } xfer += iprot->readListEnd(); } @@ -26248,14 +26841,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->triggers.clear(); - uint32_t _size1016; - ::apache::thrift::protocol::TType _etype1019; - xfer += iprot->readListBegin(_etype1019, _size1016); - this->triggers.resize(_size1016); - uint32_t _i1020; - for (_i1020 = 0; _i1020 < _size1016; ++_i1020) + uint32_t _size1040; + ::apache::thrift::protocol::TType _etype1043; + xfer += iprot->readListBegin(_etype1043, _size1040); + this->triggers.resize(_size1040); + uint32_t _i1044; + for (_i1044 = 0; _i1044 < _size1040; ++_i1044) { - xfer += this->triggers[_i1020].read(iprot); + xfer += this->triggers[_i1044].read(iprot); } xfer += iprot->readListEnd(); } @@ -26268,14 +26861,14 @@ uint32_t WMFullResourcePlan::read(::apache::thrift::protocol::TProtocol* iprot) if (ftype == ::apache::thrift::protocol::T_LIST) { { this->poolTriggers.clear(); - uint32_t _size1021; - ::apache::thrift::protocol::TType _etype1024; - xfer += iprot->readListBegin(_etype1024, _size1021); - this->poolTriggers.resize(_size1021); - uint32_t _i1025; - for (_i1025 = 0; _i1025 < _size1021; ++_i1025) + uint32_t _size1045; + ::apache::thrift::protocol::TType _etype1048; + xfer += iprot->readListBegin(_etype1048, _size1045); + this->poolTriggers.resize(_size1045); + uint32_t _i1049; + for (_i1049 = 0; _i1049 < _size1045; ++_i1049) { - xfer += this->poolTriggers[_i1025].read(iprot); + xfer += this->poolTriggers[_i1049].read(iprot); } xfer += iprot->readListEnd(); } @@ -26312,10 +26905,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("pools", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->pools.size())); - std::vector ::const_iterator _iter1026; - for (_iter1026 = this->pools.begin(); _iter1026 != this->pools.end(); ++_iter1026) + std::vector ::const_iterator _iter1050; + for (_iter1050 = this->pools.begin(); _iter1050 != this->pools.end(); ++_iter1050) { - xfer += (*_iter1026).write(oprot); + xfer += (*_iter1050).write(oprot); } xfer += oprot->writeListEnd(); } @@ -26325,10 +26918,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("mappings", ::apache::thrift::protocol::T_LIST, 3); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->mappings.size())); - std::vector ::const_iterator _iter1027; - for (_iter1027 = this->mappings.begin(); _iter1027 != this->mappings.end(); ++_iter1027) + std::vector ::const_iterator _iter1051; + for (_iter1051 = this->mappings.begin(); _iter1051 != this->mappings.end(); ++_iter1051) { - xfer += (*_iter1027).write(oprot); + xfer += (*_iter1051).write(oprot); } xfer += oprot->writeListEnd(); } @@ -26338,10 +26931,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("triggers", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->triggers.size())); - std::vector ::const_iterator _iter1028; - for (_iter1028 = this->triggers.begin(); _iter1028 != this->triggers.end(); ++_iter1028) + std::vector ::const_iterator _iter1052; + for (_iter1052 = this->triggers.begin(); _iter1052 != this->triggers.end(); ++_iter1052) { - xfer += (*_iter1028).write(oprot); + xfer += (*_iter1052).write(oprot); } xfer += oprot->writeListEnd(); } @@ -26351,10 +26944,10 @@ uint32_t WMFullResourcePlan::write(::apache::thrift::protocol::TProtocol* oprot) xfer += oprot->writeFieldBegin("poolTriggers", ::apache::thrift::protocol::T_LIST, 5); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->poolTriggers.size())); - std::vector ::const_iterator _iter1029; - for (_iter1029 = this->poolTriggers.begin(); _iter1029 != this->poolTriggers.end(); ++_iter1029) + std::vector ::const_iterator _iter1053; + for (_iter1053 = this->poolTriggers.begin(); _iter1053 != this->poolTriggers.end(); ++_iter1053) { - xfer += (*_iter1029).write(oprot); + xfer += (*_iter1053).write(oprot); } xfer += oprot->writeListEnd(); } @@ -26375,21 +26968,21 @@ void swap(WMFullResourcePlan &a, WMFullResourcePlan &b) { swap(a.__isset, b.__isset); } -WMFullResourcePlan::WMFullResourcePlan(const WMFullResourcePlan& other1030) { - plan = other1030.plan; - pools = other1030.pools; - mappings = other1030.mappings; - triggers = other1030.triggers; - poolTriggers = other1030.poolTriggers; - __isset = other1030.__isset; -} -WMFullResourcePlan& WMFullResourcePlan::operator=(const WMFullResourcePlan& other1031) { - plan = other1031.plan; - pools = other1031.pools; - mappings = other1031.mappings; - triggers = other1031.triggers; - poolTriggers = other1031.poolTriggers; - __isset = other1031.__isset; +WMFullResourcePlan::WMFullResourcePlan(const WMFullResourcePlan& other1054) { + plan = other1054.plan; + pools = other1054.pools; + mappings = other1054.mappings; + triggers = other1054.triggers; + poolTriggers = other1054.poolTriggers; + __isset = other1054.__isset; +} +WMFullResourcePlan& WMFullResourcePlan::operator=(const WMFullResourcePlan& other1055) { + plan = other1055.plan; + pools = other1055.pools; + mappings = other1055.mappings; + triggers = other1055.triggers; + poolTriggers = other1055.poolTriggers; + __isset = other1055.__isset; return *this; } void WMFullResourcePlan::printTo(std::ostream& out) const { @@ -26494,15 +27087,15 @@ void swap(WMCreateResourcePlanRequest &a, WMCreateResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMCreateResourcePlanRequest::WMCreateResourcePlanRequest(const WMCreateResourcePlanRequest& other1032) { - resourcePlan = other1032.resourcePlan; - copyFrom = other1032.copyFrom; - __isset = other1032.__isset; +WMCreateResourcePlanRequest::WMCreateResourcePlanRequest(const WMCreateResourcePlanRequest& other1056) { + resourcePlan = other1056.resourcePlan; + copyFrom = other1056.copyFrom; + __isset = other1056.__isset; } -WMCreateResourcePlanRequest& WMCreateResourcePlanRequest::operator=(const WMCreateResourcePlanRequest& other1033) { - resourcePlan = other1033.resourcePlan; - copyFrom = other1033.copyFrom; - __isset = other1033.__isset; +WMCreateResourcePlanRequest& WMCreateResourcePlanRequest::operator=(const WMCreateResourcePlanRequest& other1057) { + resourcePlan = other1057.resourcePlan; + copyFrom = other1057.copyFrom; + __isset = other1057.__isset; return *this; } void WMCreateResourcePlanRequest::printTo(std::ostream& out) const { @@ -26562,11 +27155,11 @@ void swap(WMCreateResourcePlanResponse &a, WMCreateResourcePlanResponse &b) { (void) b; } -WMCreateResourcePlanResponse::WMCreateResourcePlanResponse(const WMCreateResourcePlanResponse& other1034) { - (void) other1034; +WMCreateResourcePlanResponse::WMCreateResourcePlanResponse(const WMCreateResourcePlanResponse& other1058) { + (void) other1058; } -WMCreateResourcePlanResponse& WMCreateResourcePlanResponse::operator=(const WMCreateResourcePlanResponse& other1035) { - (void) other1035; +WMCreateResourcePlanResponse& WMCreateResourcePlanResponse::operator=(const WMCreateResourcePlanResponse& other1059) { + (void) other1059; return *this; } void WMCreateResourcePlanResponse::printTo(std::ostream& out) const { @@ -26624,11 +27217,11 @@ void swap(WMGetActiveResourcePlanRequest &a, WMGetActiveResourcePlanRequest &b) (void) b; } -WMGetActiveResourcePlanRequest::WMGetActiveResourcePlanRequest(const WMGetActiveResourcePlanRequest& other1036) { - (void) other1036; +WMGetActiveResourcePlanRequest::WMGetActiveResourcePlanRequest(const WMGetActiveResourcePlanRequest& other1060) { + (void) other1060; } -WMGetActiveResourcePlanRequest& WMGetActiveResourcePlanRequest::operator=(const WMGetActiveResourcePlanRequest& other1037) { - (void) other1037; +WMGetActiveResourcePlanRequest& WMGetActiveResourcePlanRequest::operator=(const WMGetActiveResourcePlanRequest& other1061) { + (void) other1061; return *this; } void WMGetActiveResourcePlanRequest::printTo(std::ostream& out) const { @@ -26709,13 +27302,13 @@ void swap(WMGetActiveResourcePlanResponse &a, WMGetActiveResourcePlanResponse &b swap(a.__isset, b.__isset); } -WMGetActiveResourcePlanResponse::WMGetActiveResourcePlanResponse(const WMGetActiveResourcePlanResponse& other1038) { - resourcePlan = other1038.resourcePlan; - __isset = other1038.__isset; +WMGetActiveResourcePlanResponse::WMGetActiveResourcePlanResponse(const WMGetActiveResourcePlanResponse& other1062) { + resourcePlan = other1062.resourcePlan; + __isset = other1062.__isset; } -WMGetActiveResourcePlanResponse& WMGetActiveResourcePlanResponse::operator=(const WMGetActiveResourcePlanResponse& other1039) { - resourcePlan = other1039.resourcePlan; - __isset = other1039.__isset; +WMGetActiveResourcePlanResponse& WMGetActiveResourcePlanResponse::operator=(const WMGetActiveResourcePlanResponse& other1063) { + resourcePlan = other1063.resourcePlan; + __isset = other1063.__isset; return *this; } void WMGetActiveResourcePlanResponse::printTo(std::ostream& out) const { @@ -26797,13 +27390,13 @@ void swap(WMGetResourcePlanRequest &a, WMGetResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMGetResourcePlanRequest::WMGetResourcePlanRequest(const WMGetResourcePlanRequest& other1040) { - resourcePlanName = other1040.resourcePlanName; - __isset = other1040.__isset; +WMGetResourcePlanRequest::WMGetResourcePlanRequest(const WMGetResourcePlanRequest& other1064) { + resourcePlanName = other1064.resourcePlanName; + __isset = other1064.__isset; } -WMGetResourcePlanRequest& WMGetResourcePlanRequest::operator=(const WMGetResourcePlanRequest& other1041) { - resourcePlanName = other1041.resourcePlanName; - __isset = other1041.__isset; +WMGetResourcePlanRequest& WMGetResourcePlanRequest::operator=(const WMGetResourcePlanRequest& other1065) { + resourcePlanName = other1065.resourcePlanName; + __isset = other1065.__isset; return *this; } void WMGetResourcePlanRequest::printTo(std::ostream& out) const { @@ -26885,13 +27478,13 @@ void swap(WMGetResourcePlanResponse &a, WMGetResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMGetResourcePlanResponse::WMGetResourcePlanResponse(const WMGetResourcePlanResponse& other1042) { - resourcePlan = other1042.resourcePlan; - __isset = other1042.__isset; +WMGetResourcePlanResponse::WMGetResourcePlanResponse(const WMGetResourcePlanResponse& other1066) { + resourcePlan = other1066.resourcePlan; + __isset = other1066.__isset; } -WMGetResourcePlanResponse& WMGetResourcePlanResponse::operator=(const WMGetResourcePlanResponse& other1043) { - resourcePlan = other1043.resourcePlan; - __isset = other1043.__isset; +WMGetResourcePlanResponse& WMGetResourcePlanResponse::operator=(const WMGetResourcePlanResponse& other1067) { + resourcePlan = other1067.resourcePlan; + __isset = other1067.__isset; return *this; } void WMGetResourcePlanResponse::printTo(std::ostream& out) const { @@ -26950,11 +27543,11 @@ void swap(WMGetAllResourcePlanRequest &a, WMGetAllResourcePlanRequest &b) { (void) b; } -WMGetAllResourcePlanRequest::WMGetAllResourcePlanRequest(const WMGetAllResourcePlanRequest& other1044) { - (void) other1044; +WMGetAllResourcePlanRequest::WMGetAllResourcePlanRequest(const WMGetAllResourcePlanRequest& other1068) { + (void) other1068; } -WMGetAllResourcePlanRequest& WMGetAllResourcePlanRequest::operator=(const WMGetAllResourcePlanRequest& other1045) { - (void) other1045; +WMGetAllResourcePlanRequest& WMGetAllResourcePlanRequest::operator=(const WMGetAllResourcePlanRequest& other1069) { + (void) other1069; return *this; } void WMGetAllResourcePlanRequest::printTo(std::ostream& out) const { @@ -26998,14 +27591,14 @@ uint32_t WMGetAllResourcePlanResponse::read(::apache::thrift::protocol::TProtoco if (ftype == ::apache::thrift::protocol::T_LIST) { { this->resourcePlans.clear(); - uint32_t _size1046; - ::apache::thrift::protocol::TType _etype1049; - xfer += iprot->readListBegin(_etype1049, _size1046); - this->resourcePlans.resize(_size1046); - uint32_t _i1050; - for (_i1050 = 0; _i1050 < _size1046; ++_i1050) + uint32_t _size1070; + ::apache::thrift::protocol::TType _etype1073; + xfer += iprot->readListBegin(_etype1073, _size1070); + this->resourcePlans.resize(_size1070); + uint32_t _i1074; + for (_i1074 = 0; _i1074 < _size1070; ++_i1074) { - xfer += this->resourcePlans[_i1050].read(iprot); + xfer += this->resourcePlans[_i1074].read(iprot); } xfer += iprot->readListEnd(); } @@ -27035,10 +27628,10 @@ uint32_t WMGetAllResourcePlanResponse::write(::apache::thrift::protocol::TProtoc xfer += oprot->writeFieldBegin("resourcePlans", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->resourcePlans.size())); - std::vector ::const_iterator _iter1051; - for (_iter1051 = this->resourcePlans.begin(); _iter1051 != this->resourcePlans.end(); ++_iter1051) + std::vector ::const_iterator _iter1075; + for (_iter1075 = this->resourcePlans.begin(); _iter1075 != this->resourcePlans.end(); ++_iter1075) { - xfer += (*_iter1051).write(oprot); + xfer += (*_iter1075).write(oprot); } xfer += oprot->writeListEnd(); } @@ -27055,13 +27648,13 @@ void swap(WMGetAllResourcePlanResponse &a, WMGetAllResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMGetAllResourcePlanResponse::WMGetAllResourcePlanResponse(const WMGetAllResourcePlanResponse& other1052) { - resourcePlans = other1052.resourcePlans; - __isset = other1052.__isset; +WMGetAllResourcePlanResponse::WMGetAllResourcePlanResponse(const WMGetAllResourcePlanResponse& other1076) { + resourcePlans = other1076.resourcePlans; + __isset = other1076.__isset; } -WMGetAllResourcePlanResponse& WMGetAllResourcePlanResponse::operator=(const WMGetAllResourcePlanResponse& other1053) { - resourcePlans = other1053.resourcePlans; - __isset = other1053.__isset; +WMGetAllResourcePlanResponse& WMGetAllResourcePlanResponse::operator=(const WMGetAllResourcePlanResponse& other1077) { + resourcePlans = other1077.resourcePlans; + __isset = other1077.__isset; return *this; } void WMGetAllResourcePlanResponse::printTo(std::ostream& out) const { @@ -27219,21 +27812,21 @@ void swap(WMAlterResourcePlanRequest &a, WMAlterResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMAlterResourcePlanRequest::WMAlterResourcePlanRequest(const WMAlterResourcePlanRequest& other1054) { - resourcePlanName = other1054.resourcePlanName; - resourcePlan = other1054.resourcePlan; - isEnableAndActivate = other1054.isEnableAndActivate; - isForceDeactivate = other1054.isForceDeactivate; - isReplace = other1054.isReplace; - __isset = other1054.__isset; +WMAlterResourcePlanRequest::WMAlterResourcePlanRequest(const WMAlterResourcePlanRequest& other1078) { + resourcePlanName = other1078.resourcePlanName; + resourcePlan = other1078.resourcePlan; + isEnableAndActivate = other1078.isEnableAndActivate; + isForceDeactivate = other1078.isForceDeactivate; + isReplace = other1078.isReplace; + __isset = other1078.__isset; } -WMAlterResourcePlanRequest& WMAlterResourcePlanRequest::operator=(const WMAlterResourcePlanRequest& other1055) { - resourcePlanName = other1055.resourcePlanName; - resourcePlan = other1055.resourcePlan; - isEnableAndActivate = other1055.isEnableAndActivate; - isForceDeactivate = other1055.isForceDeactivate; - isReplace = other1055.isReplace; - __isset = other1055.__isset; +WMAlterResourcePlanRequest& WMAlterResourcePlanRequest::operator=(const WMAlterResourcePlanRequest& other1079) { + resourcePlanName = other1079.resourcePlanName; + resourcePlan = other1079.resourcePlan; + isEnableAndActivate = other1079.isEnableAndActivate; + isForceDeactivate = other1079.isForceDeactivate; + isReplace = other1079.isReplace; + __isset = other1079.__isset; return *this; } void WMAlterResourcePlanRequest::printTo(std::ostream& out) const { @@ -27319,13 +27912,13 @@ void swap(WMAlterResourcePlanResponse &a, WMAlterResourcePlanResponse &b) { swap(a.__isset, b.__isset); } -WMAlterResourcePlanResponse::WMAlterResourcePlanResponse(const WMAlterResourcePlanResponse& other1056) { - fullResourcePlan = other1056.fullResourcePlan; - __isset = other1056.__isset; +WMAlterResourcePlanResponse::WMAlterResourcePlanResponse(const WMAlterResourcePlanResponse& other1080) { + fullResourcePlan = other1080.fullResourcePlan; + __isset = other1080.__isset; } -WMAlterResourcePlanResponse& WMAlterResourcePlanResponse::operator=(const WMAlterResourcePlanResponse& other1057) { - fullResourcePlan = other1057.fullResourcePlan; - __isset = other1057.__isset; +WMAlterResourcePlanResponse& WMAlterResourcePlanResponse::operator=(const WMAlterResourcePlanResponse& other1081) { + fullResourcePlan = other1081.fullResourcePlan; + __isset = other1081.__isset; return *this; } void WMAlterResourcePlanResponse::printTo(std::ostream& out) const { @@ -27407,13 +28000,13 @@ void swap(WMValidateResourcePlanRequest &a, WMValidateResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMValidateResourcePlanRequest::WMValidateResourcePlanRequest(const WMValidateResourcePlanRequest& other1058) { - resourcePlanName = other1058.resourcePlanName; - __isset = other1058.__isset; +WMValidateResourcePlanRequest::WMValidateResourcePlanRequest(const WMValidateResourcePlanRequest& other1082) { + resourcePlanName = other1082.resourcePlanName; + __isset = other1082.__isset; } -WMValidateResourcePlanRequest& WMValidateResourcePlanRequest::operator=(const WMValidateResourcePlanRequest& other1059) { - resourcePlanName = other1059.resourcePlanName; - __isset = other1059.__isset; +WMValidateResourcePlanRequest& WMValidateResourcePlanRequest::operator=(const WMValidateResourcePlanRequest& other1083) { + resourcePlanName = other1083.resourcePlanName; + __isset = other1083.__isset; return *this; } void WMValidateResourcePlanRequest::printTo(std::ostream& out) const { @@ -27463,14 +28056,14 @@ uint32_t WMValidateResourcePlanResponse::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->errors.clear(); - uint32_t _size1060; - ::apache::thrift::protocol::TType _etype1063; - xfer += iprot->readListBegin(_etype1063, _size1060); - this->errors.resize(_size1060); - uint32_t _i1064; - for (_i1064 = 0; _i1064 < _size1060; ++_i1064) + uint32_t _size1084; + ::apache::thrift::protocol::TType _etype1087; + xfer += iprot->readListBegin(_etype1087, _size1084); + this->errors.resize(_size1084); + uint32_t _i1088; + for (_i1088 = 0; _i1088 < _size1084; ++_i1088) { - xfer += iprot->readString(this->errors[_i1064]); + xfer += iprot->readString(this->errors[_i1088]); } xfer += iprot->readListEnd(); } @@ -27483,14 +28076,14 @@ uint32_t WMValidateResourcePlanResponse::read(::apache::thrift::protocol::TProto if (ftype == ::apache::thrift::protocol::T_LIST) { { this->warnings.clear(); - uint32_t _size1065; - ::apache::thrift::protocol::TType _etype1068; - xfer += iprot->readListBegin(_etype1068, _size1065); - this->warnings.resize(_size1065); - uint32_t _i1069; - for (_i1069 = 0; _i1069 < _size1065; ++_i1069) + uint32_t _size1089; + ::apache::thrift::protocol::TType _etype1092; + xfer += iprot->readListBegin(_etype1092, _size1089); + this->warnings.resize(_size1089); + uint32_t _i1093; + for (_i1093 = 0; _i1093 < _size1089; ++_i1093) { - xfer += iprot->readString(this->warnings[_i1069]); + xfer += iprot->readString(this->warnings[_i1093]); } xfer += iprot->readListEnd(); } @@ -27520,10 +28113,10 @@ uint32_t WMValidateResourcePlanResponse::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("errors", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->errors.size())); - std::vector ::const_iterator _iter1070; - for (_iter1070 = this->errors.begin(); _iter1070 != this->errors.end(); ++_iter1070) + std::vector ::const_iterator _iter1094; + for (_iter1094 = this->errors.begin(); _iter1094 != this->errors.end(); ++_iter1094) { - xfer += oprot->writeString((*_iter1070)); + xfer += oprot->writeString((*_iter1094)); } xfer += oprot->writeListEnd(); } @@ -27533,10 +28126,10 @@ uint32_t WMValidateResourcePlanResponse::write(::apache::thrift::protocol::TProt xfer += oprot->writeFieldBegin("warnings", ::apache::thrift::protocol::T_LIST, 2); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRING, static_cast(this->warnings.size())); - std::vector ::const_iterator _iter1071; - for (_iter1071 = this->warnings.begin(); _iter1071 != this->warnings.end(); ++_iter1071) + std::vector ::const_iterator _iter1095; + for (_iter1095 = this->warnings.begin(); _iter1095 != this->warnings.end(); ++_iter1095) { - xfer += oprot->writeString((*_iter1071)); + xfer += oprot->writeString((*_iter1095)); } xfer += oprot->writeListEnd(); } @@ -27554,15 +28147,15 @@ void swap(WMValidateResourcePlanResponse &a, WMValidateResourcePlanResponse &b) swap(a.__isset, b.__isset); } -WMValidateResourcePlanResponse::WMValidateResourcePlanResponse(const WMValidateResourcePlanResponse& other1072) { - errors = other1072.errors; - warnings = other1072.warnings; - __isset = other1072.__isset; +WMValidateResourcePlanResponse::WMValidateResourcePlanResponse(const WMValidateResourcePlanResponse& other1096) { + errors = other1096.errors; + warnings = other1096.warnings; + __isset = other1096.__isset; } -WMValidateResourcePlanResponse& WMValidateResourcePlanResponse::operator=(const WMValidateResourcePlanResponse& other1073) { - errors = other1073.errors; - warnings = other1073.warnings; - __isset = other1073.__isset; +WMValidateResourcePlanResponse& WMValidateResourcePlanResponse::operator=(const WMValidateResourcePlanResponse& other1097) { + errors = other1097.errors; + warnings = other1097.warnings; + __isset = other1097.__isset; return *this; } void WMValidateResourcePlanResponse::printTo(std::ostream& out) const { @@ -27645,13 +28238,13 @@ void swap(WMDropResourcePlanRequest &a, WMDropResourcePlanRequest &b) { swap(a.__isset, b.__isset); } -WMDropResourcePlanRequest::WMDropResourcePlanRequest(const WMDropResourcePlanRequest& other1074) { - resourcePlanName = other1074.resourcePlanName; - __isset = other1074.__isset; +WMDropResourcePlanRequest::WMDropResourcePlanRequest(const WMDropResourcePlanRequest& other1098) { + resourcePlanName = other1098.resourcePlanName; + __isset = other1098.__isset; } -WMDropResourcePlanRequest& WMDropResourcePlanRequest::operator=(const WMDropResourcePlanRequest& other1075) { - resourcePlanName = other1075.resourcePlanName; - __isset = other1075.__isset; +WMDropResourcePlanRequest& WMDropResourcePlanRequest::operator=(const WMDropResourcePlanRequest& other1099) { + resourcePlanName = other1099.resourcePlanName; + __isset = other1099.__isset; return *this; } void WMDropResourcePlanRequest::printTo(std::ostream& out) const { @@ -27710,11 +28303,11 @@ void swap(WMDropResourcePlanResponse &a, WMDropResourcePlanResponse &b) { (void) b; } -WMDropResourcePlanResponse::WMDropResourcePlanResponse(const WMDropResourcePlanResponse& other1076) { - (void) other1076; +WMDropResourcePlanResponse::WMDropResourcePlanResponse(const WMDropResourcePlanResponse& other1100) { + (void) other1100; } -WMDropResourcePlanResponse& WMDropResourcePlanResponse::operator=(const WMDropResourcePlanResponse& other1077) { - (void) other1077; +WMDropResourcePlanResponse& WMDropResourcePlanResponse::operator=(const WMDropResourcePlanResponse& other1101) { + (void) other1101; return *this; } void WMDropResourcePlanResponse::printTo(std::ostream& out) const { @@ -27795,13 +28388,13 @@ void swap(WMCreateTriggerRequest &a, WMCreateTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMCreateTriggerRequest::WMCreateTriggerRequest(const WMCreateTriggerRequest& other1078) { - trigger = other1078.trigger; - __isset = other1078.__isset; +WMCreateTriggerRequest::WMCreateTriggerRequest(const WMCreateTriggerRequest& other1102) { + trigger = other1102.trigger; + __isset = other1102.__isset; } -WMCreateTriggerRequest& WMCreateTriggerRequest::operator=(const WMCreateTriggerRequest& other1079) { - trigger = other1079.trigger; - __isset = other1079.__isset; +WMCreateTriggerRequest& WMCreateTriggerRequest::operator=(const WMCreateTriggerRequest& other1103) { + trigger = other1103.trigger; + __isset = other1103.__isset; return *this; } void WMCreateTriggerRequest::printTo(std::ostream& out) const { @@ -27860,11 +28453,11 @@ void swap(WMCreateTriggerResponse &a, WMCreateTriggerResponse &b) { (void) b; } -WMCreateTriggerResponse::WMCreateTriggerResponse(const WMCreateTriggerResponse& other1080) { - (void) other1080; +WMCreateTriggerResponse::WMCreateTriggerResponse(const WMCreateTriggerResponse& other1104) { + (void) other1104; } -WMCreateTriggerResponse& WMCreateTriggerResponse::operator=(const WMCreateTriggerResponse& other1081) { - (void) other1081; +WMCreateTriggerResponse& WMCreateTriggerResponse::operator=(const WMCreateTriggerResponse& other1105) { + (void) other1105; return *this; } void WMCreateTriggerResponse::printTo(std::ostream& out) const { @@ -27945,13 +28538,13 @@ void swap(WMAlterTriggerRequest &a, WMAlterTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMAlterTriggerRequest::WMAlterTriggerRequest(const WMAlterTriggerRequest& other1082) { - trigger = other1082.trigger; - __isset = other1082.__isset; +WMAlterTriggerRequest::WMAlterTriggerRequest(const WMAlterTriggerRequest& other1106) { + trigger = other1106.trigger; + __isset = other1106.__isset; } -WMAlterTriggerRequest& WMAlterTriggerRequest::operator=(const WMAlterTriggerRequest& other1083) { - trigger = other1083.trigger; - __isset = other1083.__isset; +WMAlterTriggerRequest& WMAlterTriggerRequest::operator=(const WMAlterTriggerRequest& other1107) { + trigger = other1107.trigger; + __isset = other1107.__isset; return *this; } void WMAlterTriggerRequest::printTo(std::ostream& out) const { @@ -28010,11 +28603,11 @@ void swap(WMAlterTriggerResponse &a, WMAlterTriggerResponse &b) { (void) b; } -WMAlterTriggerResponse::WMAlterTriggerResponse(const WMAlterTriggerResponse& other1084) { - (void) other1084; +WMAlterTriggerResponse::WMAlterTriggerResponse(const WMAlterTriggerResponse& other1108) { + (void) other1108; } -WMAlterTriggerResponse& WMAlterTriggerResponse::operator=(const WMAlterTriggerResponse& other1085) { - (void) other1085; +WMAlterTriggerResponse& WMAlterTriggerResponse::operator=(const WMAlterTriggerResponse& other1109) { + (void) other1109; return *this; } void WMAlterTriggerResponse::printTo(std::ostream& out) const { @@ -28114,15 +28707,15 @@ void swap(WMDropTriggerRequest &a, WMDropTriggerRequest &b) { swap(a.__isset, b.__isset); } -WMDropTriggerRequest::WMDropTriggerRequest(const WMDropTriggerRequest& other1086) { - resourcePlanName = other1086.resourcePlanName; - triggerName = other1086.triggerName; - __isset = other1086.__isset; +WMDropTriggerRequest::WMDropTriggerRequest(const WMDropTriggerRequest& other1110) { + resourcePlanName = other1110.resourcePlanName; + triggerName = other1110.triggerName; + __isset = other1110.__isset; } -WMDropTriggerRequest& WMDropTriggerRequest::operator=(const WMDropTriggerRequest& other1087) { - resourcePlanName = other1087.resourcePlanName; - triggerName = other1087.triggerName; - __isset = other1087.__isset; +WMDropTriggerRequest& WMDropTriggerRequest::operator=(const WMDropTriggerRequest& other1111) { + resourcePlanName = other1111.resourcePlanName; + triggerName = other1111.triggerName; + __isset = other1111.__isset; return *this; } void WMDropTriggerRequest::printTo(std::ostream& out) const { @@ -28182,11 +28775,11 @@ void swap(WMDropTriggerResponse &a, WMDropTriggerResponse &b) { (void) b; } -WMDropTriggerResponse::WMDropTriggerResponse(const WMDropTriggerResponse& other1088) { - (void) other1088; +WMDropTriggerResponse::WMDropTriggerResponse(const WMDropTriggerResponse& other1112) { + (void) other1112; } -WMDropTriggerResponse& WMDropTriggerResponse::operator=(const WMDropTriggerResponse& other1089) { - (void) other1089; +WMDropTriggerResponse& WMDropTriggerResponse::operator=(const WMDropTriggerResponse& other1113) { + (void) other1113; return *this; } void WMDropTriggerResponse::printTo(std::ostream& out) const { @@ -28267,13 +28860,13 @@ void swap(WMGetTriggersForResourePlanRequest &a, WMGetTriggersForResourePlanRequ swap(a.__isset, b.__isset); } -WMGetTriggersForResourePlanRequest::WMGetTriggersForResourePlanRequest(const WMGetTriggersForResourePlanRequest& other1090) { - resourcePlanName = other1090.resourcePlanName; - __isset = other1090.__isset; +WMGetTriggersForResourePlanRequest::WMGetTriggersForResourePlanRequest(const WMGetTriggersForResourePlanRequest& other1114) { + resourcePlanName = other1114.resourcePlanName; + __isset = other1114.__isset; } -WMGetTriggersForResourePlanRequest& WMGetTriggersForResourePlanRequest::operator=(const WMGetTriggersForResourePlanRequest& other1091) { - resourcePlanName = other1091.resourcePlanName; - __isset = other1091.__isset; +WMGetTriggersForResourePlanRequest& WMGetTriggersForResourePlanRequest::operator=(const WMGetTriggersForResourePlanRequest& other1115) { + resourcePlanName = other1115.resourcePlanName; + __isset = other1115.__isset; return *this; } void WMGetTriggersForResourePlanRequest::printTo(std::ostream& out) const { @@ -28318,14 +28911,14 @@ uint32_t WMGetTriggersForResourePlanResponse::read(::apache::thrift::protocol::T if (ftype == ::apache::thrift::protocol::T_LIST) { { this->triggers.clear(); - uint32_t _size1092; - ::apache::thrift::protocol::TType _etype1095; - xfer += iprot->readListBegin(_etype1095, _size1092); - this->triggers.resize(_size1092); - uint32_t _i1096; - for (_i1096 = 0; _i1096 < _size1092; ++_i1096) + uint32_t _size1116; + ::apache::thrift::protocol::TType _etype1119; + xfer += iprot->readListBegin(_etype1119, _size1116); + this->triggers.resize(_size1116); + uint32_t _i1120; + for (_i1120 = 0; _i1120 < _size1116; ++_i1120) { - xfer += this->triggers[_i1096].read(iprot); + xfer += this->triggers[_i1120].read(iprot); } xfer += iprot->readListEnd(); } @@ -28355,10 +28948,10 @@ uint32_t WMGetTriggersForResourePlanResponse::write(::apache::thrift::protocol:: xfer += oprot->writeFieldBegin("triggers", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->triggers.size())); - std::vector ::const_iterator _iter1097; - for (_iter1097 = this->triggers.begin(); _iter1097 != this->triggers.end(); ++_iter1097) + std::vector ::const_iterator _iter1121; + for (_iter1121 = this->triggers.begin(); _iter1121 != this->triggers.end(); ++_iter1121) { - xfer += (*_iter1097).write(oprot); + xfer += (*_iter1121).write(oprot); } xfer += oprot->writeListEnd(); } @@ -28375,13 +28968,13 @@ void swap(WMGetTriggersForResourePlanResponse &a, WMGetTriggersForResourePlanRes swap(a.__isset, b.__isset); } -WMGetTriggersForResourePlanResponse::WMGetTriggersForResourePlanResponse(const WMGetTriggersForResourePlanResponse& other1098) { - triggers = other1098.triggers; - __isset = other1098.__isset; +WMGetTriggersForResourePlanResponse::WMGetTriggersForResourePlanResponse(const WMGetTriggersForResourePlanResponse& other1122) { + triggers = other1122.triggers; + __isset = other1122.__isset; } -WMGetTriggersForResourePlanResponse& WMGetTriggersForResourePlanResponse::operator=(const WMGetTriggersForResourePlanResponse& other1099) { - triggers = other1099.triggers; - __isset = other1099.__isset; +WMGetTriggersForResourePlanResponse& WMGetTriggersForResourePlanResponse::operator=(const WMGetTriggersForResourePlanResponse& other1123) { + triggers = other1123.triggers; + __isset = other1123.__isset; return *this; } void WMGetTriggersForResourePlanResponse::printTo(std::ostream& out) const { @@ -28463,13 +29056,13 @@ void swap(WMCreatePoolRequest &a, WMCreatePoolRequest &b) { swap(a.__isset, b.__isset); } -WMCreatePoolRequest::WMCreatePoolRequest(const WMCreatePoolRequest& other1100) { - pool = other1100.pool; - __isset = other1100.__isset; +WMCreatePoolRequest::WMCreatePoolRequest(const WMCreatePoolRequest& other1124) { + pool = other1124.pool; + __isset = other1124.__isset; } -WMCreatePoolRequest& WMCreatePoolRequest::operator=(const WMCreatePoolRequest& other1101) { - pool = other1101.pool; - __isset = other1101.__isset; +WMCreatePoolRequest& WMCreatePoolRequest::operator=(const WMCreatePoolRequest& other1125) { + pool = other1125.pool; + __isset = other1125.__isset; return *this; } void WMCreatePoolRequest::printTo(std::ostream& out) const { @@ -28528,11 +29121,11 @@ void swap(WMCreatePoolResponse &a, WMCreatePoolResponse &b) { (void) b; } -WMCreatePoolResponse::WMCreatePoolResponse(const WMCreatePoolResponse& other1102) { - (void) other1102; +WMCreatePoolResponse::WMCreatePoolResponse(const WMCreatePoolResponse& other1126) { + (void) other1126; } -WMCreatePoolResponse& WMCreatePoolResponse::operator=(const WMCreatePoolResponse& other1103) { - (void) other1103; +WMCreatePoolResponse& WMCreatePoolResponse::operator=(const WMCreatePoolResponse& other1127) { + (void) other1127; return *this; } void WMCreatePoolResponse::printTo(std::ostream& out) const { @@ -28632,15 +29225,15 @@ void swap(WMAlterPoolRequest &a, WMAlterPoolRequest &b) { swap(a.__isset, b.__isset); } -WMAlterPoolRequest::WMAlterPoolRequest(const WMAlterPoolRequest& other1104) { - pool = other1104.pool; - poolPath = other1104.poolPath; - __isset = other1104.__isset; +WMAlterPoolRequest::WMAlterPoolRequest(const WMAlterPoolRequest& other1128) { + pool = other1128.pool; + poolPath = other1128.poolPath; + __isset = other1128.__isset; } -WMAlterPoolRequest& WMAlterPoolRequest::operator=(const WMAlterPoolRequest& other1105) { - pool = other1105.pool; - poolPath = other1105.poolPath; - __isset = other1105.__isset; +WMAlterPoolRequest& WMAlterPoolRequest::operator=(const WMAlterPoolRequest& other1129) { + pool = other1129.pool; + poolPath = other1129.poolPath; + __isset = other1129.__isset; return *this; } void WMAlterPoolRequest::printTo(std::ostream& out) const { @@ -28700,11 +29293,11 @@ void swap(WMAlterPoolResponse &a, WMAlterPoolResponse &b) { (void) b; } -WMAlterPoolResponse::WMAlterPoolResponse(const WMAlterPoolResponse& other1106) { - (void) other1106; +WMAlterPoolResponse::WMAlterPoolResponse(const WMAlterPoolResponse& other1130) { + (void) other1130; } -WMAlterPoolResponse& WMAlterPoolResponse::operator=(const WMAlterPoolResponse& other1107) { - (void) other1107; +WMAlterPoolResponse& WMAlterPoolResponse::operator=(const WMAlterPoolResponse& other1131) { + (void) other1131; return *this; } void WMAlterPoolResponse::printTo(std::ostream& out) const { @@ -28804,15 +29397,15 @@ void swap(WMDropPoolRequest &a, WMDropPoolRequest &b) { swap(a.__isset, b.__isset); } -WMDropPoolRequest::WMDropPoolRequest(const WMDropPoolRequest& other1108) { - resourcePlanName = other1108.resourcePlanName; - poolPath = other1108.poolPath; - __isset = other1108.__isset; +WMDropPoolRequest::WMDropPoolRequest(const WMDropPoolRequest& other1132) { + resourcePlanName = other1132.resourcePlanName; + poolPath = other1132.poolPath; + __isset = other1132.__isset; } -WMDropPoolRequest& WMDropPoolRequest::operator=(const WMDropPoolRequest& other1109) { - resourcePlanName = other1109.resourcePlanName; - poolPath = other1109.poolPath; - __isset = other1109.__isset; +WMDropPoolRequest& WMDropPoolRequest::operator=(const WMDropPoolRequest& other1133) { + resourcePlanName = other1133.resourcePlanName; + poolPath = other1133.poolPath; + __isset = other1133.__isset; return *this; } void WMDropPoolRequest::printTo(std::ostream& out) const { @@ -28872,11 +29465,11 @@ void swap(WMDropPoolResponse &a, WMDropPoolResponse &b) { (void) b; } -WMDropPoolResponse::WMDropPoolResponse(const WMDropPoolResponse& other1110) { - (void) other1110; +WMDropPoolResponse::WMDropPoolResponse(const WMDropPoolResponse& other1134) { + (void) other1134; } -WMDropPoolResponse& WMDropPoolResponse::operator=(const WMDropPoolResponse& other1111) { - (void) other1111; +WMDropPoolResponse& WMDropPoolResponse::operator=(const WMDropPoolResponse& other1135) { + (void) other1135; return *this; } void WMDropPoolResponse::printTo(std::ostream& out) const { @@ -28976,15 +29569,15 @@ void swap(WMCreateOrUpdateMappingRequest &a, WMCreateOrUpdateMappingRequest &b) swap(a.__isset, b.__isset); } -WMCreateOrUpdateMappingRequest::WMCreateOrUpdateMappingRequest(const WMCreateOrUpdateMappingRequest& other1112) { - mapping = other1112.mapping; - update = other1112.update; - __isset = other1112.__isset; +WMCreateOrUpdateMappingRequest::WMCreateOrUpdateMappingRequest(const WMCreateOrUpdateMappingRequest& other1136) { + mapping = other1136.mapping; + update = other1136.update; + __isset = other1136.__isset; } -WMCreateOrUpdateMappingRequest& WMCreateOrUpdateMappingRequest::operator=(const WMCreateOrUpdateMappingRequest& other1113) { - mapping = other1113.mapping; - update = other1113.update; - __isset = other1113.__isset; +WMCreateOrUpdateMappingRequest& WMCreateOrUpdateMappingRequest::operator=(const WMCreateOrUpdateMappingRequest& other1137) { + mapping = other1137.mapping; + update = other1137.update; + __isset = other1137.__isset; return *this; } void WMCreateOrUpdateMappingRequest::printTo(std::ostream& out) const { @@ -29044,11 +29637,11 @@ void swap(WMCreateOrUpdateMappingResponse &a, WMCreateOrUpdateMappingResponse &b (void) b; } -WMCreateOrUpdateMappingResponse::WMCreateOrUpdateMappingResponse(const WMCreateOrUpdateMappingResponse& other1114) { - (void) other1114; +WMCreateOrUpdateMappingResponse::WMCreateOrUpdateMappingResponse(const WMCreateOrUpdateMappingResponse& other1138) { + (void) other1138; } -WMCreateOrUpdateMappingResponse& WMCreateOrUpdateMappingResponse::operator=(const WMCreateOrUpdateMappingResponse& other1115) { - (void) other1115; +WMCreateOrUpdateMappingResponse& WMCreateOrUpdateMappingResponse::operator=(const WMCreateOrUpdateMappingResponse& other1139) { + (void) other1139; return *this; } void WMCreateOrUpdateMappingResponse::printTo(std::ostream& out) const { @@ -29129,13 +29722,13 @@ void swap(WMDropMappingRequest &a, WMDropMappingRequest &b) { swap(a.__isset, b.__isset); } -WMDropMappingRequest::WMDropMappingRequest(const WMDropMappingRequest& other1116) { - mapping = other1116.mapping; - __isset = other1116.__isset; +WMDropMappingRequest::WMDropMappingRequest(const WMDropMappingRequest& other1140) { + mapping = other1140.mapping; + __isset = other1140.__isset; } -WMDropMappingRequest& WMDropMappingRequest::operator=(const WMDropMappingRequest& other1117) { - mapping = other1117.mapping; - __isset = other1117.__isset; +WMDropMappingRequest& WMDropMappingRequest::operator=(const WMDropMappingRequest& other1141) { + mapping = other1141.mapping; + __isset = other1141.__isset; return *this; } void WMDropMappingRequest::printTo(std::ostream& out) const { @@ -29194,11 +29787,11 @@ void swap(WMDropMappingResponse &a, WMDropMappingResponse &b) { (void) b; } -WMDropMappingResponse::WMDropMappingResponse(const WMDropMappingResponse& other1118) { - (void) other1118; +WMDropMappingResponse::WMDropMappingResponse(const WMDropMappingResponse& other1142) { + (void) other1142; } -WMDropMappingResponse& WMDropMappingResponse::operator=(const WMDropMappingResponse& other1119) { - (void) other1119; +WMDropMappingResponse& WMDropMappingResponse::operator=(const WMDropMappingResponse& other1143) { + (void) other1143; return *this; } void WMDropMappingResponse::printTo(std::ostream& out) const { @@ -29336,19 +29929,19 @@ void swap(WMCreateOrDropTriggerToPoolMappingRequest &a, WMCreateOrDropTriggerToP swap(a.__isset, b.__isset); } -WMCreateOrDropTriggerToPoolMappingRequest::WMCreateOrDropTriggerToPoolMappingRequest(const WMCreateOrDropTriggerToPoolMappingRequest& other1120) { - resourcePlanName = other1120.resourcePlanName; - triggerName = other1120.triggerName; - poolPath = other1120.poolPath; - drop = other1120.drop; - __isset = other1120.__isset; +WMCreateOrDropTriggerToPoolMappingRequest::WMCreateOrDropTriggerToPoolMappingRequest(const WMCreateOrDropTriggerToPoolMappingRequest& other1144) { + resourcePlanName = other1144.resourcePlanName; + triggerName = other1144.triggerName; + poolPath = other1144.poolPath; + drop = other1144.drop; + __isset = other1144.__isset; } -WMCreateOrDropTriggerToPoolMappingRequest& WMCreateOrDropTriggerToPoolMappingRequest::operator=(const WMCreateOrDropTriggerToPoolMappingRequest& other1121) { - resourcePlanName = other1121.resourcePlanName; - triggerName = other1121.triggerName; - poolPath = other1121.poolPath; - drop = other1121.drop; - __isset = other1121.__isset; +WMCreateOrDropTriggerToPoolMappingRequest& WMCreateOrDropTriggerToPoolMappingRequest::operator=(const WMCreateOrDropTriggerToPoolMappingRequest& other1145) { + resourcePlanName = other1145.resourcePlanName; + triggerName = other1145.triggerName; + poolPath = other1145.poolPath; + drop = other1145.drop; + __isset = other1145.__isset; return *this; } void WMCreateOrDropTriggerToPoolMappingRequest::printTo(std::ostream& out) const { @@ -29410,11 +30003,11 @@ void swap(WMCreateOrDropTriggerToPoolMappingResponse &a, WMCreateOrDropTriggerTo (void) b; } -WMCreateOrDropTriggerToPoolMappingResponse::WMCreateOrDropTriggerToPoolMappingResponse(const WMCreateOrDropTriggerToPoolMappingResponse& other1122) { - (void) other1122; +WMCreateOrDropTriggerToPoolMappingResponse::WMCreateOrDropTriggerToPoolMappingResponse(const WMCreateOrDropTriggerToPoolMappingResponse& other1146) { + (void) other1146; } -WMCreateOrDropTriggerToPoolMappingResponse& WMCreateOrDropTriggerToPoolMappingResponse::operator=(const WMCreateOrDropTriggerToPoolMappingResponse& other1123) { - (void) other1123; +WMCreateOrDropTriggerToPoolMappingResponse& WMCreateOrDropTriggerToPoolMappingResponse::operator=(const WMCreateOrDropTriggerToPoolMappingResponse& other1147) { + (void) other1147; return *this; } void WMCreateOrDropTriggerToPoolMappingResponse::printTo(std::ostream& out) const { @@ -29489,9 +30082,9 @@ uint32_t ISchema::read(::apache::thrift::protocol::TProtocol* iprot) { { case 1: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1124; - xfer += iprot->readI32(ecast1124); - this->schemaType = (SchemaType::type)ecast1124; + int32_t ecast1148; + xfer += iprot->readI32(ecast1148); + this->schemaType = (SchemaType::type)ecast1148; this->__isset.schemaType = true; } else { xfer += iprot->skip(ftype); @@ -29523,9 +30116,9 @@ uint32_t ISchema::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1125; - xfer += iprot->readI32(ecast1125); - this->compatibility = (SchemaCompatibility::type)ecast1125; + int32_t ecast1149; + xfer += iprot->readI32(ecast1149); + this->compatibility = (SchemaCompatibility::type)ecast1149; this->__isset.compatibility = true; } else { xfer += iprot->skip(ftype); @@ -29533,9 +30126,9 @@ uint32_t ISchema::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 6: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1126; - xfer += iprot->readI32(ecast1126); - this->validationLevel = (SchemaValidation::type)ecast1126; + int32_t ecast1150; + xfer += iprot->readI32(ecast1150); + this->validationLevel = (SchemaValidation::type)ecast1150; this->__isset.validationLevel = true; } else { xfer += iprot->skip(ftype); @@ -29639,29 +30232,29 @@ void swap(ISchema &a, ISchema &b) { swap(a.__isset, b.__isset); } -ISchema::ISchema(const ISchema& other1127) { - schemaType = other1127.schemaType; - name = other1127.name; - catName = other1127.catName; - dbName = other1127.dbName; - compatibility = other1127.compatibility; - validationLevel = other1127.validationLevel; - canEvolve = other1127.canEvolve; - schemaGroup = other1127.schemaGroup; - description = other1127.description; - __isset = other1127.__isset; -} -ISchema& ISchema::operator=(const ISchema& other1128) { - schemaType = other1128.schemaType; - name = other1128.name; - catName = other1128.catName; - dbName = other1128.dbName; - compatibility = other1128.compatibility; - validationLevel = other1128.validationLevel; - canEvolve = other1128.canEvolve; - schemaGroup = other1128.schemaGroup; - description = other1128.description; - __isset = other1128.__isset; +ISchema::ISchema(const ISchema& other1151) { + schemaType = other1151.schemaType; + name = other1151.name; + catName = other1151.catName; + dbName = other1151.dbName; + compatibility = other1151.compatibility; + validationLevel = other1151.validationLevel; + canEvolve = other1151.canEvolve; + schemaGroup = other1151.schemaGroup; + description = other1151.description; + __isset = other1151.__isset; +} +ISchema& ISchema::operator=(const ISchema& other1152) { + schemaType = other1152.schemaType; + name = other1152.name; + catName = other1152.catName; + dbName = other1152.dbName; + compatibility = other1152.compatibility; + validationLevel = other1152.validationLevel; + canEvolve = other1152.canEvolve; + schemaGroup = other1152.schemaGroup; + description = other1152.description; + __isset = other1152.__isset; return *this; } void ISchema::printTo(std::ostream& out) const { @@ -29783,17 +30376,17 @@ void swap(ISchemaName &a, ISchemaName &b) { swap(a.__isset, b.__isset); } -ISchemaName::ISchemaName(const ISchemaName& other1129) { - catName = other1129.catName; - dbName = other1129.dbName; - schemaName = other1129.schemaName; - __isset = other1129.__isset; +ISchemaName::ISchemaName(const ISchemaName& other1153) { + catName = other1153.catName; + dbName = other1153.dbName; + schemaName = other1153.schemaName; + __isset = other1153.__isset; } -ISchemaName& ISchemaName::operator=(const ISchemaName& other1130) { - catName = other1130.catName; - dbName = other1130.dbName; - schemaName = other1130.schemaName; - __isset = other1130.__isset; +ISchemaName& ISchemaName::operator=(const ISchemaName& other1154) { + catName = other1154.catName; + dbName = other1154.dbName; + schemaName = other1154.schemaName; + __isset = other1154.__isset; return *this; } void ISchemaName::printTo(std::ostream& out) const { @@ -29892,15 +30485,15 @@ void swap(AlterISchemaRequest &a, AlterISchemaRequest &b) { swap(a.__isset, b.__isset); } -AlterISchemaRequest::AlterISchemaRequest(const AlterISchemaRequest& other1131) { - name = other1131.name; - newSchema = other1131.newSchema; - __isset = other1131.__isset; +AlterISchemaRequest::AlterISchemaRequest(const AlterISchemaRequest& other1155) { + name = other1155.name; + newSchema = other1155.newSchema; + __isset = other1155.__isset; } -AlterISchemaRequest& AlterISchemaRequest::operator=(const AlterISchemaRequest& other1132) { - name = other1132.name; - newSchema = other1132.newSchema; - __isset = other1132.__isset; +AlterISchemaRequest& AlterISchemaRequest::operator=(const AlterISchemaRequest& other1156) { + name = other1156.name; + newSchema = other1156.newSchema; + __isset = other1156.__isset; return *this; } void AlterISchemaRequest::printTo(std::ostream& out) const { @@ -30011,14 +30604,14 @@ uint32_t SchemaVersion::read(::apache::thrift::protocol::TProtocol* iprot) { if (ftype == ::apache::thrift::protocol::T_LIST) { { this->cols.clear(); - uint32_t _size1133; - ::apache::thrift::protocol::TType _etype1136; - xfer += iprot->readListBegin(_etype1136, _size1133); - this->cols.resize(_size1133); - uint32_t _i1137; - for (_i1137 = 0; _i1137 < _size1133; ++_i1137) + uint32_t _size1157; + ::apache::thrift::protocol::TType _etype1160; + xfer += iprot->readListBegin(_etype1160, _size1157); + this->cols.resize(_size1157); + uint32_t _i1161; + for (_i1161 = 0; _i1161 < _size1157; ++_i1161) { - xfer += this->cols[_i1137].read(iprot); + xfer += this->cols[_i1161].read(iprot); } xfer += iprot->readListEnd(); } @@ -30029,9 +30622,9 @@ uint32_t SchemaVersion::read(::apache::thrift::protocol::TProtocol* iprot) { break; case 5: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1138; - xfer += iprot->readI32(ecast1138); - this->state = (SchemaVersionState::type)ecast1138; + int32_t ecast1162; + xfer += iprot->readI32(ecast1162); + this->state = (SchemaVersionState::type)ecast1162; this->__isset.state = true; } else { xfer += iprot->skip(ftype); @@ -30109,10 +30702,10 @@ uint32_t SchemaVersion::write(::apache::thrift::protocol::TProtocol* oprot) cons xfer += oprot->writeFieldBegin("cols", ::apache::thrift::protocol::T_LIST, 4); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->cols.size())); - std::vector ::const_iterator _iter1139; - for (_iter1139 = this->cols.begin(); _iter1139 != this->cols.end(); ++_iter1139) + std::vector ::const_iterator _iter1163; + for (_iter1163 = this->cols.begin(); _iter1163 != this->cols.end(); ++_iter1163) { - xfer += (*_iter1139).write(oprot); + xfer += (*_iter1163).write(oprot); } xfer += oprot->writeListEnd(); } @@ -30168,31 +30761,31 @@ void swap(SchemaVersion &a, SchemaVersion &b) { swap(a.__isset, b.__isset); } -SchemaVersion::SchemaVersion(const SchemaVersion& other1140) { - schema = other1140.schema; - version = other1140.version; - createdAt = other1140.createdAt; - cols = other1140.cols; - state = other1140.state; - description = other1140.description; - schemaText = other1140.schemaText; - fingerprint = other1140.fingerprint; - name = other1140.name; - serDe = other1140.serDe; - __isset = other1140.__isset; -} -SchemaVersion& SchemaVersion::operator=(const SchemaVersion& other1141) { - schema = other1141.schema; - version = other1141.version; - createdAt = other1141.createdAt; - cols = other1141.cols; - state = other1141.state; - description = other1141.description; - schemaText = other1141.schemaText; - fingerprint = other1141.fingerprint; - name = other1141.name; - serDe = other1141.serDe; - __isset = other1141.__isset; +SchemaVersion::SchemaVersion(const SchemaVersion& other1164) { + schema = other1164.schema; + version = other1164.version; + createdAt = other1164.createdAt; + cols = other1164.cols; + state = other1164.state; + description = other1164.description; + schemaText = other1164.schemaText; + fingerprint = other1164.fingerprint; + name = other1164.name; + serDe = other1164.serDe; + __isset = other1164.__isset; +} +SchemaVersion& SchemaVersion::operator=(const SchemaVersion& other1165) { + schema = other1165.schema; + version = other1165.version; + createdAt = other1165.createdAt; + cols = other1165.cols; + state = other1165.state; + description = other1165.description; + schemaText = other1165.schemaText; + fingerprint = other1165.fingerprint; + name = other1165.name; + serDe = other1165.serDe; + __isset = other1165.__isset; return *this; } void SchemaVersion::printTo(std::ostream& out) const { @@ -30298,15 +30891,15 @@ void swap(SchemaVersionDescriptor &a, SchemaVersionDescriptor &b) { swap(a.__isset, b.__isset); } -SchemaVersionDescriptor::SchemaVersionDescriptor(const SchemaVersionDescriptor& other1142) { - schema = other1142.schema; - version = other1142.version; - __isset = other1142.__isset; +SchemaVersionDescriptor::SchemaVersionDescriptor(const SchemaVersionDescriptor& other1166) { + schema = other1166.schema; + version = other1166.version; + __isset = other1166.__isset; } -SchemaVersionDescriptor& SchemaVersionDescriptor::operator=(const SchemaVersionDescriptor& other1143) { - schema = other1143.schema; - version = other1143.version; - __isset = other1143.__isset; +SchemaVersionDescriptor& SchemaVersionDescriptor::operator=(const SchemaVersionDescriptor& other1167) { + schema = other1167.schema; + version = other1167.version; + __isset = other1167.__isset; return *this; } void SchemaVersionDescriptor::printTo(std::ostream& out) const { @@ -30427,17 +31020,17 @@ void swap(FindSchemasByColsRqst &a, FindSchemasByColsRqst &b) { swap(a.__isset, b.__isset); } -FindSchemasByColsRqst::FindSchemasByColsRqst(const FindSchemasByColsRqst& other1144) { - colName = other1144.colName; - colNamespace = other1144.colNamespace; - type = other1144.type; - __isset = other1144.__isset; +FindSchemasByColsRqst::FindSchemasByColsRqst(const FindSchemasByColsRqst& other1168) { + colName = other1168.colName; + colNamespace = other1168.colNamespace; + type = other1168.type; + __isset = other1168.__isset; } -FindSchemasByColsRqst& FindSchemasByColsRqst::operator=(const FindSchemasByColsRqst& other1145) { - colName = other1145.colName; - colNamespace = other1145.colNamespace; - type = other1145.type; - __isset = other1145.__isset; +FindSchemasByColsRqst& FindSchemasByColsRqst::operator=(const FindSchemasByColsRqst& other1169) { + colName = other1169.colName; + colNamespace = other1169.colNamespace; + type = other1169.type; + __isset = other1169.__isset; return *this; } void FindSchemasByColsRqst::printTo(std::ostream& out) const { @@ -30483,14 +31076,14 @@ uint32_t FindSchemasByColsResp::read(::apache::thrift::protocol::TProtocol* ipro if (ftype == ::apache::thrift::protocol::T_LIST) { { this->schemaVersions.clear(); - uint32_t _size1146; - ::apache::thrift::protocol::TType _etype1149; - xfer += iprot->readListBegin(_etype1149, _size1146); - this->schemaVersions.resize(_size1146); - uint32_t _i1150; - for (_i1150 = 0; _i1150 < _size1146; ++_i1150) + uint32_t _size1170; + ::apache::thrift::protocol::TType _etype1173; + xfer += iprot->readListBegin(_etype1173, _size1170); + this->schemaVersions.resize(_size1170); + uint32_t _i1174; + for (_i1174 = 0; _i1174 < _size1170; ++_i1174) { - xfer += this->schemaVersions[_i1150].read(iprot); + xfer += this->schemaVersions[_i1174].read(iprot); } xfer += iprot->readListEnd(); } @@ -30519,10 +31112,10 @@ uint32_t FindSchemasByColsResp::write(::apache::thrift::protocol::TProtocol* opr xfer += oprot->writeFieldBegin("schemaVersions", ::apache::thrift::protocol::T_LIST, 1); { xfer += oprot->writeListBegin(::apache::thrift::protocol::T_STRUCT, static_cast(this->schemaVersions.size())); - std::vector ::const_iterator _iter1151; - for (_iter1151 = this->schemaVersions.begin(); _iter1151 != this->schemaVersions.end(); ++_iter1151) + std::vector ::const_iterator _iter1175; + for (_iter1175 = this->schemaVersions.begin(); _iter1175 != this->schemaVersions.end(); ++_iter1175) { - xfer += (*_iter1151).write(oprot); + xfer += (*_iter1175).write(oprot); } xfer += oprot->writeListEnd(); } @@ -30539,13 +31132,13 @@ void swap(FindSchemasByColsResp &a, FindSchemasByColsResp &b) { swap(a.__isset, b.__isset); } -FindSchemasByColsResp::FindSchemasByColsResp(const FindSchemasByColsResp& other1152) { - schemaVersions = other1152.schemaVersions; - __isset = other1152.__isset; +FindSchemasByColsResp::FindSchemasByColsResp(const FindSchemasByColsResp& other1176) { + schemaVersions = other1176.schemaVersions; + __isset = other1176.__isset; } -FindSchemasByColsResp& FindSchemasByColsResp::operator=(const FindSchemasByColsResp& other1153) { - schemaVersions = other1153.schemaVersions; - __isset = other1153.__isset; +FindSchemasByColsResp& FindSchemasByColsResp::operator=(const FindSchemasByColsResp& other1177) { + schemaVersions = other1177.schemaVersions; + __isset = other1177.__isset; return *this; } void FindSchemasByColsResp::printTo(std::ostream& out) const { @@ -30642,15 +31235,15 @@ void swap(MapSchemaVersionToSerdeRequest &a, MapSchemaVersionToSerdeRequest &b) swap(a.__isset, b.__isset); } -MapSchemaVersionToSerdeRequest::MapSchemaVersionToSerdeRequest(const MapSchemaVersionToSerdeRequest& other1154) { - schemaVersion = other1154.schemaVersion; - serdeName = other1154.serdeName; - __isset = other1154.__isset; +MapSchemaVersionToSerdeRequest::MapSchemaVersionToSerdeRequest(const MapSchemaVersionToSerdeRequest& other1178) { + schemaVersion = other1178.schemaVersion; + serdeName = other1178.serdeName; + __isset = other1178.__isset; } -MapSchemaVersionToSerdeRequest& MapSchemaVersionToSerdeRequest::operator=(const MapSchemaVersionToSerdeRequest& other1155) { - schemaVersion = other1155.schemaVersion; - serdeName = other1155.serdeName; - __isset = other1155.__isset; +MapSchemaVersionToSerdeRequest& MapSchemaVersionToSerdeRequest::operator=(const MapSchemaVersionToSerdeRequest& other1179) { + schemaVersion = other1179.schemaVersion; + serdeName = other1179.serdeName; + __isset = other1179.__isset; return *this; } void MapSchemaVersionToSerdeRequest::printTo(std::ostream& out) const { @@ -30705,9 +31298,9 @@ uint32_t SetSchemaVersionStateRequest::read(::apache::thrift::protocol::TProtoco break; case 2: if (ftype == ::apache::thrift::protocol::T_I32) { - int32_t ecast1156; - xfer += iprot->readI32(ecast1156); - this->state = (SchemaVersionState::type)ecast1156; + int32_t ecast1180; + xfer += iprot->readI32(ecast1180); + this->state = (SchemaVersionState::type)ecast1180; this->__isset.state = true; } else { xfer += iprot->skip(ftype); @@ -30750,15 +31343,15 @@ void swap(SetSchemaVersionStateRequest &a, SetSchemaVersionStateRequest &b) { swap(a.__isset, b.__isset); } -SetSchemaVersionStateRequest::SetSchemaVersionStateRequest(const SetSchemaVersionStateRequest& other1157) { - schemaVersion = other1157.schemaVersion; - state = other1157.state; - __isset = other1157.__isset; +SetSchemaVersionStateRequest::SetSchemaVersionStateRequest(const SetSchemaVersionStateRequest& other1181) { + schemaVersion = other1181.schemaVersion; + state = other1181.state; + __isset = other1181.__isset; } -SetSchemaVersionStateRequest& SetSchemaVersionStateRequest::operator=(const SetSchemaVersionStateRequest& other1158) { - schemaVersion = other1158.schemaVersion; - state = other1158.state; - __isset = other1158.__isset; +SetSchemaVersionStateRequest& SetSchemaVersionStateRequest::operator=(const SetSchemaVersionStateRequest& other1182) { + schemaVersion = other1182.schemaVersion; + state = other1182.state; + __isset = other1182.__isset; return *this; } void SetSchemaVersionStateRequest::printTo(std::ostream& out) const { @@ -30839,13 +31432,13 @@ void swap(GetSerdeRequest &a, GetSerdeRequest &b) { swap(a.__isset, b.__isset); } -GetSerdeRequest::GetSerdeRequest(const GetSerdeRequest& other1159) { - serdeName = other1159.serdeName; - __isset = other1159.__isset; +GetSerdeRequest::GetSerdeRequest(const GetSerdeRequest& other1183) { + serdeName = other1183.serdeName; + __isset = other1183.__isset; } -GetSerdeRequest& GetSerdeRequest::operator=(const GetSerdeRequest& other1160) { - serdeName = other1160.serdeName; - __isset = other1160.__isset; +GetSerdeRequest& GetSerdeRequest::operator=(const GetSerdeRequest& other1184) { + serdeName = other1184.serdeName; + __isset = other1184.__isset; return *this; } void GetSerdeRequest::printTo(std::ostream& out) const { @@ -30967,17 +31560,17 @@ void swap(RuntimeStat &a, RuntimeStat &b) { swap(a.__isset, b.__isset); } -RuntimeStat::RuntimeStat(const RuntimeStat& other1161) { - createTime = other1161.createTime; - weight = other1161.weight; - payload = other1161.payload; - __isset = other1161.__isset; +RuntimeStat::RuntimeStat(const RuntimeStat& other1185) { + createTime = other1185.createTime; + weight = other1185.weight; + payload = other1185.payload; + __isset = other1185.__isset; } -RuntimeStat& RuntimeStat::operator=(const RuntimeStat& other1162) { - createTime = other1162.createTime; - weight = other1162.weight; - payload = other1162.payload; - __isset = other1162.__isset; +RuntimeStat& RuntimeStat::operator=(const RuntimeStat& other1186) { + createTime = other1186.createTime; + weight = other1186.weight; + payload = other1186.payload; + __isset = other1186.__isset; return *this; } void RuntimeStat::printTo(std::ostream& out) const { @@ -31081,13 +31674,13 @@ void swap(GetRuntimeStatsRequest &a, GetRuntimeStatsRequest &b) { swap(a.maxCreateTime, b.maxCreateTime); } -GetRuntimeStatsRequest::GetRuntimeStatsRequest(const GetRuntimeStatsRequest& other1163) { - maxWeight = other1163.maxWeight; - maxCreateTime = other1163.maxCreateTime; +GetRuntimeStatsRequest::GetRuntimeStatsRequest(const GetRuntimeStatsRequest& other1187) { + maxWeight = other1187.maxWeight; + maxCreateTime = other1187.maxCreateTime; } -GetRuntimeStatsRequest& GetRuntimeStatsRequest::operator=(const GetRuntimeStatsRequest& other1164) { - maxWeight = other1164.maxWeight; - maxCreateTime = other1164.maxCreateTime; +GetRuntimeStatsRequest& GetRuntimeStatsRequest::operator=(const GetRuntimeStatsRequest& other1188) { + maxWeight = other1188.maxWeight; + maxCreateTime = other1188.maxCreateTime; return *this; } void GetRuntimeStatsRequest::printTo(std::ostream& out) const { @@ -31168,13 +31761,13 @@ void swap(MetaException &a, MetaException &b) { swap(a.__isset, b.__isset); } -MetaException::MetaException(const MetaException& other1165) : TException() { - message = other1165.message; - __isset = other1165.__isset; +MetaException::MetaException(const MetaException& other1189) : TException() { + message = other1189.message; + __isset = other1189.__isset; } -MetaException& MetaException::operator=(const MetaException& other1166) { - message = other1166.message; - __isset = other1166.__isset; +MetaException& MetaException::operator=(const MetaException& other1190) { + message = other1190.message; + __isset = other1190.__isset; return *this; } void MetaException::printTo(std::ostream& out) const { @@ -31265,13 +31858,13 @@ void swap(UnknownTableException &a, UnknownTableException &b) { swap(a.__isset, b.__isset); } -UnknownTableException::UnknownTableException(const UnknownTableException& other1167) : TException() { - message = other1167.message; - __isset = other1167.__isset; +UnknownTableException::UnknownTableException(const UnknownTableException& other1191) : TException() { + message = other1191.message; + __isset = other1191.__isset; } -UnknownTableException& UnknownTableException::operator=(const UnknownTableException& other1168) { - message = other1168.message; - __isset = other1168.__isset; +UnknownTableException& UnknownTableException::operator=(const UnknownTableException& other1192) { + message = other1192.message; + __isset = other1192.__isset; return *this; } void UnknownTableException::printTo(std::ostream& out) const { @@ -31362,13 +31955,13 @@ void swap(UnknownDBException &a, UnknownDBException &b) { swap(a.__isset, b.__isset); } -UnknownDBException::UnknownDBException(const UnknownDBException& other1169) : TException() { - message = other1169.message; - __isset = other1169.__isset; +UnknownDBException::UnknownDBException(const UnknownDBException& other1193) : TException() { + message = other1193.message; + __isset = other1193.__isset; } -UnknownDBException& UnknownDBException::operator=(const UnknownDBException& other1170) { - message = other1170.message; - __isset = other1170.__isset; +UnknownDBException& UnknownDBException::operator=(const UnknownDBException& other1194) { + message = other1194.message; + __isset = other1194.__isset; return *this; } void UnknownDBException::printTo(std::ostream& out) const { @@ -31459,13 +32052,13 @@ void swap(AlreadyExistsException &a, AlreadyExistsException &b) { swap(a.__isset, b.__isset); } -AlreadyExistsException::AlreadyExistsException(const AlreadyExistsException& other1171) : TException() { - message = other1171.message; - __isset = other1171.__isset; +AlreadyExistsException::AlreadyExistsException(const AlreadyExistsException& other1195) : TException() { + message = other1195.message; + __isset = other1195.__isset; } -AlreadyExistsException& AlreadyExistsException::operator=(const AlreadyExistsException& other1172) { - message = other1172.message; - __isset = other1172.__isset; +AlreadyExistsException& AlreadyExistsException::operator=(const AlreadyExistsException& other1196) { + message = other1196.message; + __isset = other1196.__isset; return *this; } void AlreadyExistsException::printTo(std::ostream& out) const { @@ -31556,13 +32149,13 @@ void swap(InvalidPartitionException &a, InvalidPartitionException &b) { swap(a.__isset, b.__isset); } -InvalidPartitionException::InvalidPartitionException(const InvalidPartitionException& other1173) : TException() { - message = other1173.message; - __isset = other1173.__isset; +InvalidPartitionException::InvalidPartitionException(const InvalidPartitionException& other1197) : TException() { + message = other1197.message; + __isset = other1197.__isset; } -InvalidPartitionException& InvalidPartitionException::operator=(const InvalidPartitionException& other1174) { - message = other1174.message; - __isset = other1174.__isset; +InvalidPartitionException& InvalidPartitionException::operator=(const InvalidPartitionException& other1198) { + message = other1198.message; + __isset = other1198.__isset; return *this; } void InvalidPartitionException::printTo(std::ostream& out) const { @@ -31653,13 +32246,13 @@ void swap(UnknownPartitionException &a, UnknownPartitionException &b) { swap(a.__isset, b.__isset); } -UnknownPartitionException::UnknownPartitionException(const UnknownPartitionException& other1175) : TException() { - message = other1175.message; - __isset = other1175.__isset; +UnknownPartitionException::UnknownPartitionException(const UnknownPartitionException& other1199) : TException() { + message = other1199.message; + __isset = other1199.__isset; } -UnknownPartitionException& UnknownPartitionException::operator=(const UnknownPartitionException& other1176) { - message = other1176.message; - __isset = other1176.__isset; +UnknownPartitionException& UnknownPartitionException::operator=(const UnknownPartitionException& other1200) { + message = other1200.message; + __isset = other1200.__isset; return *this; } void UnknownPartitionException::printTo(std::ostream& out) const { @@ -31750,13 +32343,13 @@ void swap(InvalidObjectException &a, InvalidObjectException &b) { swap(a.__isset, b.__isset); } -InvalidObjectException::InvalidObjectException(const InvalidObjectException& other1177) : TException() { - message = other1177.message; - __isset = other1177.__isset; +InvalidObjectException::InvalidObjectException(const InvalidObjectException& other1201) : TException() { + message = other1201.message; + __isset = other1201.__isset; } -InvalidObjectException& InvalidObjectException::operator=(const InvalidObjectException& other1178) { - message = other1178.message; - __isset = other1178.__isset; +InvalidObjectException& InvalidObjectException::operator=(const InvalidObjectException& other1202) { + message = other1202.message; + __isset = other1202.__isset; return *this; } void InvalidObjectException::printTo(std::ostream& out) const { @@ -31847,13 +32440,13 @@ void swap(NoSuchObjectException &a, NoSuchObjectException &b) { swap(a.__isset, b.__isset); } -NoSuchObjectException::NoSuchObjectException(const NoSuchObjectException& other1179) : TException() { - message = other1179.message; - __isset = other1179.__isset; +NoSuchObjectException::NoSuchObjectException(const NoSuchObjectException& other1203) : TException() { + message = other1203.message; + __isset = other1203.__isset; } -NoSuchObjectException& NoSuchObjectException::operator=(const NoSuchObjectException& other1180) { - message = other1180.message; - __isset = other1180.__isset; +NoSuchObjectException& NoSuchObjectException::operator=(const NoSuchObjectException& other1204) { + message = other1204.message; + __isset = other1204.__isset; return *this; } void NoSuchObjectException::printTo(std::ostream& out) const { @@ -31944,13 +32537,13 @@ void swap(InvalidOperationException &a, InvalidOperationException &b) { swap(a.__isset, b.__isset); } -InvalidOperationException::InvalidOperationException(const InvalidOperationException& other1181) : TException() { - message = other1181.message; - __isset = other1181.__isset; +InvalidOperationException::InvalidOperationException(const InvalidOperationException& other1205) : TException() { + message = other1205.message; + __isset = other1205.__isset; } -InvalidOperationException& InvalidOperationException::operator=(const InvalidOperationException& other1182) { - message = other1182.message; - __isset = other1182.__isset; +InvalidOperationException& InvalidOperationException::operator=(const InvalidOperationException& other1206) { + message = other1206.message; + __isset = other1206.__isset; return *this; } void InvalidOperationException::printTo(std::ostream& out) const { @@ -32041,13 +32634,13 @@ void swap(ConfigValSecurityException &a, ConfigValSecurityException &b) { swap(a.__isset, b.__isset); } -ConfigValSecurityException::ConfigValSecurityException(const ConfigValSecurityException& other1183) : TException() { - message = other1183.message; - __isset = other1183.__isset; +ConfigValSecurityException::ConfigValSecurityException(const ConfigValSecurityException& other1207) : TException() { + message = other1207.message; + __isset = other1207.__isset; } -ConfigValSecurityException& ConfigValSecurityException::operator=(const ConfigValSecurityException& other1184) { - message = other1184.message; - __isset = other1184.__isset; +ConfigValSecurityException& ConfigValSecurityException::operator=(const ConfigValSecurityException& other1208) { + message = other1208.message; + __isset = other1208.__isset; return *this; } void ConfigValSecurityException::printTo(std::ostream& out) const { @@ -32138,13 +32731,13 @@ void swap(InvalidInputException &a, InvalidInputException &b) { swap(a.__isset, b.__isset); } -InvalidInputException::InvalidInputException(const InvalidInputException& other1185) : TException() { - message = other1185.message; - __isset = other1185.__isset; +InvalidInputException::InvalidInputException(const InvalidInputException& other1209) : TException() { + message = other1209.message; + __isset = other1209.__isset; } -InvalidInputException& InvalidInputException::operator=(const InvalidInputException& other1186) { - message = other1186.message; - __isset = other1186.__isset; +InvalidInputException& InvalidInputException::operator=(const InvalidInputException& other1210) { + message = other1210.message; + __isset = other1210.__isset; return *this; } void InvalidInputException::printTo(std::ostream& out) const { @@ -32235,13 +32828,13 @@ void swap(NoSuchTxnException &a, NoSuchTxnException &b) { swap(a.__isset, b.__isset); } -NoSuchTxnException::NoSuchTxnException(const NoSuchTxnException& other1187) : TException() { - message = other1187.message; - __isset = other1187.__isset; +NoSuchTxnException::NoSuchTxnException(const NoSuchTxnException& other1211) : TException() { + message = other1211.message; + __isset = other1211.__isset; } -NoSuchTxnException& NoSuchTxnException::operator=(const NoSuchTxnException& other1188) { - message = other1188.message; - __isset = other1188.__isset; +NoSuchTxnException& NoSuchTxnException::operator=(const NoSuchTxnException& other1212) { + message = other1212.message; + __isset = other1212.__isset; return *this; } void NoSuchTxnException::printTo(std::ostream& out) const { @@ -32332,13 +32925,13 @@ void swap(TxnAbortedException &a, TxnAbortedException &b) { swap(a.__isset, b.__isset); } -TxnAbortedException::TxnAbortedException(const TxnAbortedException& other1189) : TException() { - message = other1189.message; - __isset = other1189.__isset; +TxnAbortedException::TxnAbortedException(const TxnAbortedException& other1213) : TException() { + message = other1213.message; + __isset = other1213.__isset; } -TxnAbortedException& TxnAbortedException::operator=(const TxnAbortedException& other1190) { - message = other1190.message; - __isset = other1190.__isset; +TxnAbortedException& TxnAbortedException::operator=(const TxnAbortedException& other1214) { + message = other1214.message; + __isset = other1214.__isset; return *this; } void TxnAbortedException::printTo(std::ostream& out) const { @@ -32429,13 +33022,13 @@ void swap(TxnOpenException &a, TxnOpenException &b) { swap(a.__isset, b.__isset); } -TxnOpenException::TxnOpenException(const TxnOpenException& other1191) : TException() { - message = other1191.message; - __isset = other1191.__isset; +TxnOpenException::TxnOpenException(const TxnOpenException& other1215) : TException() { + message = other1215.message; + __isset = other1215.__isset; } -TxnOpenException& TxnOpenException::operator=(const TxnOpenException& other1192) { - message = other1192.message; - __isset = other1192.__isset; +TxnOpenException& TxnOpenException::operator=(const TxnOpenException& other1216) { + message = other1216.message; + __isset = other1216.__isset; return *this; } void TxnOpenException::printTo(std::ostream& out) const { @@ -32526,13 +33119,13 @@ void swap(NoSuchLockException &a, NoSuchLockException &b) { swap(a.__isset, b.__isset); } -NoSuchLockException::NoSuchLockException(const NoSuchLockException& other1193) : TException() { - message = other1193.message; - __isset = other1193.__isset; +NoSuchLockException::NoSuchLockException(const NoSuchLockException& other1217) : TException() { + message = other1217.message; + __isset = other1217.__isset; } -NoSuchLockException& NoSuchLockException::operator=(const NoSuchLockException& other1194) { - message = other1194.message; - __isset = other1194.__isset; +NoSuchLockException& NoSuchLockException::operator=(const NoSuchLockException& other1218) { + message = other1218.message; + __isset = other1218.__isset; return *this; } void NoSuchLockException::printTo(std::ostream& out) const { diff --git a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h index 756f954e68..d0c299b8b0 100644 --- a/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h +++ b/standalone-metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h @@ -445,6 +445,8 @@ class AbortTxnsRequest; class CommitTxnRequest; +class WriteEventInfo; + class ReplTblWriteIdStateRequest; class GetValidWriteIdsRequest; @@ -517,6 +519,10 @@ class FireEventRequest; class FireEventResponse; +class WriteNotificationLogRequest; + +class WriteNotificationLogResponse; + class MetadataPpdResult; class GetFileMetadataByExprResult; @@ -6985,8 +6991,9 @@ inline std::ostream& operator<<(std::ostream& out, const AbortTxnsRequest& obj) } typedef struct _CommitTxnRequest__isset { - _CommitTxnRequest__isset() : replPolicy(false) {} + _CommitTxnRequest__isset() : replPolicy(false), writeEventInfos(false) {} bool replPolicy :1; + bool writeEventInfos :1; } _CommitTxnRequest__isset; class CommitTxnRequest { @@ -7000,6 +7007,7 @@ class CommitTxnRequest { virtual ~CommitTxnRequest() throw(); int64_t txnid; std::string replPolicy; + std::vector writeEventInfos; _CommitTxnRequest__isset __isset; @@ -7007,6 +7015,8 @@ class CommitTxnRequest { void __set_replPolicy(const std::string& val); + void __set_writeEventInfos(const std::vector & val); + bool operator == (const CommitTxnRequest & rhs) const { if (!(txnid == rhs.txnid)) @@ -7015,6 +7025,10 @@ class CommitTxnRequest { return false; else if (__isset.replPolicy && !(replPolicy == rhs.replPolicy)) return false; + if (__isset.writeEventInfos != rhs.__isset.writeEventInfos) + return false; + else if (__isset.writeEventInfos && !(writeEventInfos == rhs.writeEventInfos)) + return false; return true; } bool operator != (const CommitTxnRequest &rhs) const { @@ -7037,6 +7051,90 @@ inline std::ostream& operator<<(std::ostream& out, const CommitTxnRequest& obj) return out; } +typedef struct _WriteEventInfo__isset { + _WriteEventInfo__isset() : partition(false), tableObj(false), partitionObj(false) {} + bool partition :1; + bool tableObj :1; + bool partitionObj :1; +} _WriteEventInfo__isset; + +class WriteEventInfo { + public: + + WriteEventInfo(const WriteEventInfo&); + WriteEventInfo& operator=(const WriteEventInfo&); + WriteEventInfo() : writeId(0), database(), table(), files(), partition(), tableObj(), partitionObj() { + } + + virtual ~WriteEventInfo() throw(); + int64_t writeId; + std::string database; + std::string table; + std::string files; + std::string partition; + std::string tableObj; + std::string partitionObj; + + _WriteEventInfo__isset __isset; + + void __set_writeId(const int64_t val); + + void __set_database(const std::string& val); + + void __set_table(const std::string& val); + + void __set_files(const std::string& val); + + void __set_partition(const std::string& val); + + void __set_tableObj(const std::string& val); + + void __set_partitionObj(const std::string& val); + + bool operator == (const WriteEventInfo & rhs) const + { + if (!(writeId == rhs.writeId)) + return false; + if (!(database == rhs.database)) + return false; + if (!(table == rhs.table)) + return false; + if (!(files == rhs.files)) + return false; + if (__isset.partition != rhs.__isset.partition) + return false; + else if (__isset.partition && !(partition == rhs.partition)) + return false; + if (__isset.tableObj != rhs.__isset.tableObj) + return false; + else if (__isset.tableObj && !(tableObj == rhs.tableObj)) + return false; + if (__isset.partitionObj != rhs.__isset.partitionObj) + return false; + else if (__isset.partitionObj && !(partitionObj == rhs.partitionObj)) + return false; + return true; + } + bool operator != (const WriteEventInfo &rhs) const { + return !(*this == rhs); + } + + bool operator < (const WriteEventInfo & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + + virtual void printTo(std::ostream& out) const; +}; + +void swap(WriteEventInfo &a, WriteEventInfo &b); + +inline std::ostream& operator<<(std::ostream& out, const WriteEventInfo& obj) +{ + obj.printTo(out); + return out; +} + typedef struct _ReplTblWriteIdStateRequest__isset { _ReplTblWriteIdStateRequest__isset() : partNames(false) {} bool partNames :1; @@ -9043,9 +9141,10 @@ inline std::ostream& operator<<(std::ostream& out, const NotificationEventsCount } typedef struct _InsertEventRequestData__isset { - _InsertEventRequestData__isset() : replace(false), filesAddedChecksum(false) {} + _InsertEventRequestData__isset() : replace(false), filesAddedChecksum(false), subDirectoryList(false) {} bool replace :1; bool filesAddedChecksum :1; + bool subDirectoryList :1; } _InsertEventRequestData__isset; class InsertEventRequestData { @@ -9060,6 +9159,7 @@ class InsertEventRequestData { bool replace; std::vector filesAdded; std::vector filesAddedChecksum; + std::vector subDirectoryList; _InsertEventRequestData__isset __isset; @@ -9069,6 +9169,8 @@ class InsertEventRequestData { void __set_filesAddedChecksum(const std::vector & val); + void __set_subDirectoryList(const std::vector & val); + bool operator == (const InsertEventRequestData & rhs) const { if (__isset.replace != rhs.__isset.replace) @@ -9081,6 +9183,10 @@ class InsertEventRequestData { return false; else if (__isset.filesAddedChecksum && !(filesAddedChecksum == rhs.filesAddedChecksum)) return false; + if (__isset.subDirectoryList != rhs.__isset.subDirectoryList) + return false; + else if (__isset.subDirectoryList && !(subDirectoryList == rhs.subDirectoryList)) + return false; return true; } bool operator != (const InsertEventRequestData &rhs) const { @@ -9266,6 +9372,114 @@ inline std::ostream& operator<<(std::ostream& out, const FireEventResponse& obj) return out; } +typedef struct _WriteNotificationLogRequest__isset { + _WriteNotificationLogRequest__isset() : partitionVals(false) {} + bool partitionVals :1; +} _WriteNotificationLogRequest__isset; + +class WriteNotificationLogRequest { + public: + + WriteNotificationLogRequest(const WriteNotificationLogRequest&); + WriteNotificationLogRequest& operator=(const WriteNotificationLogRequest&); + WriteNotificationLogRequest() : txnId(0), writeId(0), db(), table() { + } + + virtual ~WriteNotificationLogRequest() throw(); + int64_t txnId; + int64_t writeId; + std::string db; + std::string table; + InsertEventRequestData fileInfo; + std::vector partitionVals; + + _WriteNotificationLogRequest__isset __isset; + + void __set_txnId(const int64_t val); + + void __set_writeId(const int64_t val); + + void __set_db(const std::string& val); + + void __set_table(const std::string& val); + + void __set_fileInfo(const InsertEventRequestData& val); + + void __set_partitionVals(const std::vector & val); + + bool operator == (const WriteNotificationLogRequest & rhs) const + { + if (!(txnId == rhs.txnId)) + return false; + if (!(writeId == rhs.writeId)) + return false; + if (!(db == rhs.db)) + return false; + if (!(table == rhs.table)) + return false; + if (!(fileInfo == rhs.fileInfo)) + return false; + if (__isset.partitionVals != rhs.__isset.partitionVals) + return false; + else if (__isset.partitionVals && !(partitionVals == rhs.partitionVals)) + return false; + return true; + } + bool operator != (const WriteNotificationLogRequest &rhs) const { + return !(*this == rhs); + } + + bool operator < (const WriteNotificationLogRequest & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + + virtual void printTo(std::ostream& out) const; +}; + +void swap(WriteNotificationLogRequest &a, WriteNotificationLogRequest &b); + +inline std::ostream& operator<<(std::ostream& out, const WriteNotificationLogRequest& obj) +{ + obj.printTo(out); + return out; +} + + +class WriteNotificationLogResponse { + public: + + WriteNotificationLogResponse(const WriteNotificationLogResponse&); + WriteNotificationLogResponse& operator=(const WriteNotificationLogResponse&); + WriteNotificationLogResponse() { + } + + virtual ~WriteNotificationLogResponse() throw(); + + bool operator == (const WriteNotificationLogResponse & /* rhs */) const + { + return true; + } + bool operator != (const WriteNotificationLogResponse &rhs) const { + return !(*this == rhs); + } + + bool operator < (const WriteNotificationLogResponse & ) const; + + uint32_t read(::apache::thrift::protocol::TProtocol* iprot); + uint32_t write(::apache::thrift::protocol::TProtocol* oprot) const; + + virtual void printTo(std::ostream& out) const; +}; + +void swap(WriteNotificationLogResponse &a, WriteNotificationLogResponse &b); + +inline std::ostream& operator<<(std::ostream& out, const WriteNotificationLogResponse& obj) +{ + obj.printTo(out); + return out; +} + typedef struct _MetadataPpdResult__isset { _MetadataPpdResult__isset() : metadata(false), includeBitset(false) {} bool metadata :1; diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java index 1dcc8707b5..3ce72e9ce6 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddDynamicPartitions.java @@ -816,13 +816,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, AddDynamicPartition case 5: // PARTITIONNAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list700 = iprot.readListBegin(); - struct.partitionnames = new ArrayList(_list700.size); - String _elem701; - for (int _i702 = 0; _i702 < _list700.size; ++_i702) + org.apache.thrift.protocol.TList _list708 = iprot.readListBegin(); + struct.partitionnames = new ArrayList(_list708.size); + String _elem709; + for (int _i710 = 0; _i710 < _list708.size; ++_i710) { - _elem701 = iprot.readString(); - struct.partitionnames.add(_elem701); + _elem709 = iprot.readString(); + struct.partitionnames.add(_elem709); } iprot.readListEnd(); } @@ -872,9 +872,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, AddDynamicPartitio oprot.writeFieldBegin(PARTITIONNAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.partitionnames.size())); - for (String _iter703 : struct.partitionnames) + for (String _iter711 : struct.partitionnames) { - oprot.writeString(_iter703); + oprot.writeString(_iter711); } oprot.writeListEnd(); } @@ -910,9 +910,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, AddDynamicPartition oprot.writeString(struct.tablename); { oprot.writeI32(struct.partitionnames.size()); - for (String _iter704 : struct.partitionnames) + for (String _iter712 : struct.partitionnames) { - oprot.writeString(_iter704); + oprot.writeString(_iter712); } } BitSet optionals = new BitSet(); @@ -937,13 +937,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, AddDynamicPartitions struct.tablename = iprot.readString(); struct.setTablenameIsSet(true); { - org.apache.thrift.protocol.TList _list705 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.partitionnames = new ArrayList(_list705.size); - String _elem706; - for (int _i707 = 0; _i707 < _list705.size; ++_i707) + org.apache.thrift.protocol.TList _list713 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partitionnames = new ArrayList(_list713.size); + String _elem714; + for (int _i715 = 0; _i715 < _list713.size; ++_i715) { - _elem706 = iprot.readString(); - struct.partitionnames.add(_elem706); + _elem714 = iprot.readString(); + struct.partitionnames.add(_elem714); } } struct.setPartitionnamesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java index fa33963799..a0b47a9c67 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsRequest.java @@ -716,13 +716,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, AllocateTableWriteI case 3: // TXN_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list626 = iprot.readListBegin(); - struct.txnIds = new ArrayList(_list626.size); - long _elem627; - for (int _i628 = 0; _i628 < _list626.size; ++_i628) + org.apache.thrift.protocol.TList _list634 = iprot.readListBegin(); + struct.txnIds = new ArrayList(_list634.size); + long _elem635; + for (int _i636 = 0; _i636 < _list634.size; ++_i636) { - _elem627 = iprot.readI64(); - struct.txnIds.add(_elem627); + _elem635 = iprot.readI64(); + struct.txnIds.add(_elem635); } iprot.readListEnd(); } @@ -742,14 +742,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, AllocateTableWriteI case 5: // SRC_TXN_TO_WRITE_ID_LIST if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list629 = iprot.readListBegin(); - struct.srcTxnToWriteIdList = new ArrayList(_list629.size); - TxnToWriteId _elem630; - for (int _i631 = 0; _i631 < _list629.size; ++_i631) + org.apache.thrift.protocol.TList _list637 = iprot.readListBegin(); + struct.srcTxnToWriteIdList = new ArrayList(_list637.size); + TxnToWriteId _elem638; + for (int _i639 = 0; _i639 < _list637.size; ++_i639) { - _elem630 = new TxnToWriteId(); - _elem630.read(iprot); - struct.srcTxnToWriteIdList.add(_elem630); + _elem638 = new TxnToWriteId(); + _elem638.read(iprot); + struct.srcTxnToWriteIdList.add(_elem638); } iprot.readListEnd(); } @@ -786,9 +786,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, AllocateTableWrite oprot.writeFieldBegin(TXN_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.txnIds.size())); - for (long _iter632 : struct.txnIds) + for (long _iter640 : struct.txnIds) { - oprot.writeI64(_iter632); + oprot.writeI64(_iter640); } oprot.writeListEnd(); } @@ -807,9 +807,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, AllocateTableWrite oprot.writeFieldBegin(SRC_TXN_TO_WRITE_ID_LIST_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.srcTxnToWriteIdList.size())); - for (TxnToWriteId _iter633 : struct.srcTxnToWriteIdList) + for (TxnToWriteId _iter641 : struct.srcTxnToWriteIdList) { - _iter633.write(oprot); + _iter641.write(oprot); } oprot.writeListEnd(); } @@ -849,9 +849,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteI if (struct.isSetTxnIds()) { { oprot.writeI32(struct.txnIds.size()); - for (long _iter634 : struct.txnIds) + for (long _iter642 : struct.txnIds) { - oprot.writeI64(_iter634); + oprot.writeI64(_iter642); } } } @@ -861,9 +861,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteI if (struct.isSetSrcTxnToWriteIdList()) { { oprot.writeI32(struct.srcTxnToWriteIdList.size()); - for (TxnToWriteId _iter635 : struct.srcTxnToWriteIdList) + for (TxnToWriteId _iter643 : struct.srcTxnToWriteIdList) { - _iter635.write(oprot); + _iter643.write(oprot); } } } @@ -879,13 +879,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteId BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list636 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.txnIds = new ArrayList(_list636.size); - long _elem637; - for (int _i638 = 0; _i638 < _list636.size; ++_i638) + org.apache.thrift.protocol.TList _list644 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.txnIds = new ArrayList(_list644.size); + long _elem645; + for (int _i646 = 0; _i646 < _list644.size; ++_i646) { - _elem637 = iprot.readI64(); - struct.txnIds.add(_elem637); + _elem645 = iprot.readI64(); + struct.txnIds.add(_elem645); } } struct.setTxnIdsIsSet(true); @@ -896,14 +896,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteId } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list639 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.srcTxnToWriteIdList = new ArrayList(_list639.size); - TxnToWriteId _elem640; - for (int _i641 = 0; _i641 < _list639.size; ++_i641) + org.apache.thrift.protocol.TList _list647 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.srcTxnToWriteIdList = new ArrayList(_list647.size); + TxnToWriteId _elem648; + for (int _i649 = 0; _i649 < _list647.size; ++_i649) { - _elem640 = new TxnToWriteId(); - _elem640.read(iprot); - struct.srcTxnToWriteIdList.add(_elem640); + _elem648 = new TxnToWriteId(); + _elem648.read(iprot); + struct.srcTxnToWriteIdList.add(_elem648); } } struct.setSrcTxnToWriteIdListIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java index 20dc757dc7..13df26d5fd 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AllocateTableWriteIdsResponse.java @@ -354,14 +354,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, AllocateTableWriteI case 1: // TXN_TO_WRITE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list642 = iprot.readListBegin(); - struct.txnToWriteIds = new ArrayList(_list642.size); - TxnToWriteId _elem643; - for (int _i644 = 0; _i644 < _list642.size; ++_i644) + org.apache.thrift.protocol.TList _list650 = iprot.readListBegin(); + struct.txnToWriteIds = new ArrayList(_list650.size); + TxnToWriteId _elem651; + for (int _i652 = 0; _i652 < _list650.size; ++_i652) { - _elem643 = new TxnToWriteId(); - _elem643.read(iprot); - struct.txnToWriteIds.add(_elem643); + _elem651 = new TxnToWriteId(); + _elem651.read(iprot); + struct.txnToWriteIds.add(_elem651); } iprot.readListEnd(); } @@ -387,9 +387,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, AllocateTableWrite oprot.writeFieldBegin(TXN_TO_WRITE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.txnToWriteIds.size())); - for (TxnToWriteId _iter645 : struct.txnToWriteIds) + for (TxnToWriteId _iter653 : struct.txnToWriteIds) { - _iter645.write(oprot); + _iter653.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteI TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.txnToWriteIds.size()); - for (TxnToWriteId _iter646 : struct.txnToWriteIds) + for (TxnToWriteId _iter654 : struct.txnToWriteIds) { - _iter646.write(oprot); + _iter654.write(oprot); } } } @@ -425,14 +425,14 @@ public void write(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteI public void read(org.apache.thrift.protocol.TProtocol prot, AllocateTableWriteIdsResponse struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list647 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.txnToWriteIds = new ArrayList(_list647.size); - TxnToWriteId _elem648; - for (int _i649 = 0; _i649 < _list647.size; ++_i649) + org.apache.thrift.protocol.TList _list655 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.txnToWriteIds = new ArrayList(_list655.size); + TxnToWriteId _elem656; + for (int _i657 = 0; _i657 < _list655.size; ++_i657) { - _elem648 = new TxnToWriteId(); - _elem648.read(iprot); - struct.txnToWriteIds.add(_elem648); + _elem656 = new TxnToWriteId(); + _elem656.read(iprot); + struct.txnToWriteIds.add(_elem656); } } struct.setTxnToWriteIdsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClearFileMetadataRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClearFileMetadataRequest.java index 470a0706a2..1af1628a41 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClearFileMetadataRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClearFileMetadataRequest.java @@ -351,13 +351,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ClearFileMetadataRe case 1: // FILE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list800 = iprot.readListBegin(); - struct.fileIds = new ArrayList(_list800.size); - long _elem801; - for (int _i802 = 0; _i802 < _list800.size; ++_i802) + org.apache.thrift.protocol.TList _list824 = iprot.readListBegin(); + struct.fileIds = new ArrayList(_list824.size); + long _elem825; + for (int _i826 = 0; _i826 < _list824.size; ++_i826) { - _elem801 = iprot.readI64(); - struct.fileIds.add(_elem801); + _elem825 = iprot.readI64(); + struct.fileIds.add(_elem825); } iprot.readListEnd(); } @@ -383,9 +383,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ClearFileMetadataR oprot.writeFieldBegin(FILE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.fileIds.size())); - for (long _iter803 : struct.fileIds) + for (long _iter827 : struct.fileIds) { - oprot.writeI64(_iter803); + oprot.writeI64(_iter827); } oprot.writeListEnd(); } @@ -410,9 +410,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ClearFileMetadataRe TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.fileIds.size()); - for (long _iter804 : struct.fileIds) + for (long _iter828 : struct.fileIds) { - oprot.writeI64(_iter804); + oprot.writeI64(_iter828); } } } @@ -421,13 +421,13 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ClearFileMetadataRe public void read(org.apache.thrift.protocol.TProtocol prot, ClearFileMetadataRequest struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list805 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.fileIds = new ArrayList(_list805.size); - long _elem806; - for (int _i807 = 0; _i807 < _list805.size; ++_i807) + org.apache.thrift.protocol.TList _list829 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.fileIds = new ArrayList(_list829.size); + long _elem830; + for (int _i831 = 0; _i831 < _list829.size; ++_i831) { - _elem806 = iprot.readI64(); - struct.fileIds.add(_elem806); + _elem830 = iprot.readI64(); + struct.fileIds.add(_elem830); } } struct.setFileIdsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClientCapabilities.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClientCapabilities.java index af48583c18..4cd04f1d92 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClientCapabilities.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ClientCapabilities.java @@ -354,13 +354,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ClientCapabilities case 1: // VALUES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list816 = iprot.readListBegin(); - struct.values = new ArrayList(_list816.size); - ClientCapability _elem817; - for (int _i818 = 0; _i818 < _list816.size; ++_i818) + org.apache.thrift.protocol.TList _list840 = iprot.readListBegin(); + struct.values = new ArrayList(_list840.size); + ClientCapability _elem841; + for (int _i842 = 0; _i842 < _list840.size; ++_i842) { - _elem817 = org.apache.hadoop.hive.metastore.api.ClientCapability.findByValue(iprot.readI32()); - struct.values.add(_elem817); + _elem841 = org.apache.hadoop.hive.metastore.api.ClientCapability.findByValue(iprot.readI32()); + struct.values.add(_elem841); } iprot.readListEnd(); } @@ -386,9 +386,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ClientCapabilities oprot.writeFieldBegin(VALUES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I32, struct.values.size())); - for (ClientCapability _iter819 : struct.values) + for (ClientCapability _iter843 : struct.values) { - oprot.writeI32(_iter819.getValue()); + oprot.writeI32(_iter843.getValue()); } oprot.writeListEnd(); } @@ -413,9 +413,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ClientCapabilities TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.values.size()); - for (ClientCapability _iter820 : struct.values) + for (ClientCapability _iter844 : struct.values) { - oprot.writeI32(_iter820.getValue()); + oprot.writeI32(_iter844.getValue()); } } } @@ -424,13 +424,13 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ClientCapabilities public void read(org.apache.thrift.protocol.TProtocol prot, ClientCapabilities struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list821 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I32, iprot.readI32()); - struct.values = new ArrayList(_list821.size); - ClientCapability _elem822; - for (int _i823 = 0; _i823 < _list821.size; ++_i823) + org.apache.thrift.protocol.TList _list845 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I32, iprot.readI32()); + struct.values = new ArrayList(_list845.size); + ClientCapability _elem846; + for (int _i847 = 0; _i847 < _list845.size; ++_i847) { - _elem822 = org.apache.hadoop.hive.metastore.api.ClientCapability.findByValue(iprot.readI32()); - struct.values.add(_elem822); + _elem846 = org.apache.hadoop.hive.metastore.api.ClientCapability.findByValue(iprot.readI32()); + struct.values.add(_elem846); } } struct.setValuesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CommitTxnRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CommitTxnRequest.java index 3c15f84942..f29595886b 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CommitTxnRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CommitTxnRequest.java @@ -40,6 +40,7 @@ private static final org.apache.thrift.protocol.TField TXNID_FIELD_DESC = new org.apache.thrift.protocol.TField("txnid", org.apache.thrift.protocol.TType.I64, (short)1); private static final org.apache.thrift.protocol.TField REPL_POLICY_FIELD_DESC = new org.apache.thrift.protocol.TField("replPolicy", org.apache.thrift.protocol.TType.STRING, (short)2); + private static final org.apache.thrift.protocol.TField WRITE_EVENT_INFOS_FIELD_DESC = new org.apache.thrift.protocol.TField("writeEventInfos", org.apache.thrift.protocol.TType.LIST, (short)3); private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); static { @@ -49,11 +50,13 @@ private long txnid; // required private String replPolicy; // optional + private List writeEventInfos; // optional /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { TXNID((short)1, "txnid"), - REPL_POLICY((short)2, "replPolicy"); + REPL_POLICY((short)2, "replPolicy"), + WRITE_EVENT_INFOS((short)3, "writeEventInfos"); private static final Map byName = new HashMap(); @@ -72,6 +75,8 @@ public static _Fields findByThriftId(int fieldId) { return TXNID; case 2: // REPL_POLICY return REPL_POLICY; + case 3: // WRITE_EVENT_INFOS + return WRITE_EVENT_INFOS; default: return null; } @@ -114,7 +119,7 @@ public String getFieldName() { // isset id assignments private static final int __TXNID_ISSET_ID = 0; private byte __isset_bitfield = 0; - private static final _Fields optionals[] = {_Fields.REPL_POLICY}; + private static final _Fields optionals[] = {_Fields.REPL_POLICY,_Fields.WRITE_EVENT_INFOS}; public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); @@ -122,6 +127,9 @@ public String getFieldName() { new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); tmpMap.put(_Fields.REPL_POLICY, new org.apache.thrift.meta_data.FieldMetaData("replPolicy", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.WRITE_EVENT_INFOS, new org.apache.thrift.meta_data.FieldMetaData("writeEventInfos", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.ListMetaData(org.apache.thrift.protocol.TType.LIST, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRUCT , "WriteEventInfo")))); metaDataMap = Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(CommitTxnRequest.class, metaDataMap); } @@ -146,6 +154,13 @@ public CommitTxnRequest(CommitTxnRequest other) { if (other.isSetReplPolicy()) { this.replPolicy = other.replPolicy; } + if (other.isSetWriteEventInfos()) { + List __this__writeEventInfos = new ArrayList(other.writeEventInfos.size()); + for (WriteEventInfo other_element : other.writeEventInfos) { + __this__writeEventInfos.add(other_element); + } + this.writeEventInfos = __this__writeEventInfos; + } } public CommitTxnRequest deepCopy() { @@ -157,6 +172,7 @@ public void clear() { setTxnidIsSet(false); this.txnid = 0; this.replPolicy = null; + this.writeEventInfos = null; } public long getTxnid() { @@ -204,6 +220,44 @@ public void setReplPolicyIsSet(boolean value) { } } + public int getWriteEventInfosSize() { + return (this.writeEventInfos == null) ? 0 : this.writeEventInfos.size(); + } + + public java.util.Iterator getWriteEventInfosIterator() { + return (this.writeEventInfos == null) ? null : this.writeEventInfos.iterator(); + } + + public void addToWriteEventInfos(WriteEventInfo elem) { + if (this.writeEventInfos == null) { + this.writeEventInfos = new ArrayList(); + } + this.writeEventInfos.add(elem); + } + + public List getWriteEventInfos() { + return this.writeEventInfos; + } + + public void setWriteEventInfos(List writeEventInfos) { + this.writeEventInfos = writeEventInfos; + } + + public void unsetWriteEventInfos() { + this.writeEventInfos = null; + } + + /** Returns true if field writeEventInfos is set (has been assigned a value) and false otherwise */ + public boolean isSetWriteEventInfos() { + return this.writeEventInfos != null; + } + + public void setWriteEventInfosIsSet(boolean value) { + if (!value) { + this.writeEventInfos = null; + } + } + public void setFieldValue(_Fields field, Object value) { switch (field) { case TXNID: @@ -222,6 +276,14 @@ public void setFieldValue(_Fields field, Object value) { } break; + case WRITE_EVENT_INFOS: + if (value == null) { + unsetWriteEventInfos(); + } else { + setWriteEventInfos((List)value); + } + break; + } } @@ -233,6 +295,9 @@ public Object getFieldValue(_Fields field) { case REPL_POLICY: return getReplPolicy(); + case WRITE_EVENT_INFOS: + return getWriteEventInfos(); + } throw new IllegalStateException(); } @@ -248,6 +313,8 @@ public boolean isSet(_Fields field) { return isSetTxnid(); case REPL_POLICY: return isSetReplPolicy(); + case WRITE_EVENT_INFOS: + return isSetWriteEventInfos(); } throw new IllegalStateException(); } @@ -283,6 +350,15 @@ public boolean equals(CommitTxnRequest that) { return false; } + boolean this_present_writeEventInfos = true && this.isSetWriteEventInfos(); + boolean that_present_writeEventInfos = true && that.isSetWriteEventInfos(); + if (this_present_writeEventInfos || that_present_writeEventInfos) { + if (!(this_present_writeEventInfos && that_present_writeEventInfos)) + return false; + if (!this.writeEventInfos.equals(that.writeEventInfos)) + return false; + } + return true; } @@ -300,6 +376,11 @@ public int hashCode() { if (present_replPolicy) list.add(replPolicy); + boolean present_writeEventInfos = true && (isSetWriteEventInfos()); + list.add(present_writeEventInfos); + if (present_writeEventInfos) + list.add(writeEventInfos); + return list.hashCode(); } @@ -331,6 +412,16 @@ public int compareTo(CommitTxnRequest other) { return lastComparison; } } + lastComparison = Boolean.valueOf(isSetWriteEventInfos()).compareTo(other.isSetWriteEventInfos()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetWriteEventInfos()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.writeEventInfos, other.writeEventInfos); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -364,6 +455,16 @@ public String toString() { } first = false; } + if (isSetWriteEventInfos()) { + if (!first) sb.append(", "); + sb.append("writeEventInfos:"); + if (this.writeEventInfos == null) { + sb.append("null"); + } else { + sb.append(this.writeEventInfos); + } + first = false; + } sb.append(")"); return sb.toString(); } @@ -429,6 +530,25 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, CommitTxnRequest st org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 3: // WRITE_EVENT_INFOS + if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { + { + org.apache.thrift.protocol.TList _list594 = iprot.readListBegin(); + struct.writeEventInfos = new ArrayList(_list594.size); + WriteEventInfo _elem595; + for (int _i596 = 0; _i596 < _list594.size; ++_i596) + { + _elem595 = new WriteEventInfo(); + _elem595.read(iprot); + struct.writeEventInfos.add(_elem595); + } + iprot.readListEnd(); + } + struct.setWriteEventInfosIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -452,6 +572,20 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, CommitTxnRequest s oprot.writeFieldEnd(); } } + if (struct.writeEventInfos != null) { + if (struct.isSetWriteEventInfos()) { + oprot.writeFieldBegin(WRITE_EVENT_INFOS_FIELD_DESC); + { + oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.writeEventInfos.size())); + for (WriteEventInfo _iter597 : struct.writeEventInfos) + { + _iter597.write(oprot); + } + oprot.writeListEnd(); + } + oprot.writeFieldEnd(); + } + } oprot.writeFieldStop(); oprot.writeStructEnd(); } @@ -474,10 +608,22 @@ public void write(org.apache.thrift.protocol.TProtocol prot, CommitTxnRequest st if (struct.isSetReplPolicy()) { optionals.set(0); } - oprot.writeBitSet(optionals, 1); + if (struct.isSetWriteEventInfos()) { + optionals.set(1); + } + oprot.writeBitSet(optionals, 2); if (struct.isSetReplPolicy()) { oprot.writeString(struct.replPolicy); } + if (struct.isSetWriteEventInfos()) { + { + oprot.writeI32(struct.writeEventInfos.size()); + for (WriteEventInfo _iter598 : struct.writeEventInfos) + { + _iter598.write(oprot); + } + } + } } @Override @@ -485,11 +631,25 @@ public void read(org.apache.thrift.protocol.TProtocol prot, CommitTxnRequest str TTupleProtocol iprot = (TTupleProtocol) prot; struct.txnid = iprot.readI64(); struct.setTxnidIsSet(true); - BitSet incoming = iprot.readBitSet(1); + BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { struct.replPolicy = iprot.readString(); struct.setReplPolicyIsSet(true); } + if (incoming.get(1)) { + { + org.apache.thrift.protocol.TList _list599 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.writeEventInfos = new ArrayList(_list599.size); + WriteEventInfo _elem600; + for (int _i601 = 0; _i601 < _list599.size; ++_i601) + { + _elem600 = new WriteEventInfo(); + _elem600.read(iprot); + struct.writeEventInfos.add(_elem600); + } + } + struct.setWriteEventInfosIsSet(true); + } } } diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CompactionRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CompactionRequest.java index 31f2e144a9..57eb5efffb 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CompactionRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CompactionRequest.java @@ -814,15 +814,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, CompactionRequest s case 6: // PROPERTIES if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map682 = iprot.readMapBegin(); - struct.properties = new HashMap(2*_map682.size); - String _key683; - String _val684; - for (int _i685 = 0; _i685 < _map682.size; ++_i685) + org.apache.thrift.protocol.TMap _map690 = iprot.readMapBegin(); + struct.properties = new HashMap(2*_map690.size); + String _key691; + String _val692; + for (int _i693 = 0; _i693 < _map690.size; ++_i693) { - _key683 = iprot.readString(); - _val684 = iprot.readString(); - struct.properties.put(_key683, _val684); + _key691 = iprot.readString(); + _val692 = iprot.readString(); + struct.properties.put(_key691, _val692); } iprot.readMapEnd(); } @@ -878,10 +878,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, CompactionRequest oprot.writeFieldBegin(PROPERTIES_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, struct.properties.size())); - for (Map.Entry _iter686 : struct.properties.entrySet()) + for (Map.Entry _iter694 : struct.properties.entrySet()) { - oprot.writeString(_iter686.getKey()); - oprot.writeString(_iter686.getValue()); + oprot.writeString(_iter694.getKey()); + oprot.writeString(_iter694.getValue()); } oprot.writeMapEnd(); } @@ -928,10 +928,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, CompactionRequest s if (struct.isSetProperties()) { { oprot.writeI32(struct.properties.size()); - for (Map.Entry _iter687 : struct.properties.entrySet()) + for (Map.Entry _iter695 : struct.properties.entrySet()) { - oprot.writeString(_iter687.getKey()); - oprot.writeString(_iter687.getValue()); + oprot.writeString(_iter695.getKey()); + oprot.writeString(_iter695.getValue()); } } } @@ -957,15 +957,15 @@ public void read(org.apache.thrift.protocol.TProtocol prot, CompactionRequest st } if (incoming.get(2)) { { - org.apache.thrift.protocol.TMap _map688 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.properties = new HashMap(2*_map688.size); - String _key689; - String _val690; - for (int _i691 = 0; _i691 < _map688.size; ++_i691) + org.apache.thrift.protocol.TMap _map696 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.properties = new HashMap(2*_map696.size); + String _key697; + String _val698; + for (int _i699 = 0; _i699 < _map696.size; ++_i699) { - _key689 = iprot.readString(); - _val690 = iprot.readString(); - struct.properties.put(_key689, _val690); + _key697 = iprot.readString(); + _val698 = iprot.readString(); + struct.properties.put(_key697, _val698); } } struct.setPropertiesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CreationMetadata.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CreationMetadata.java index 684a0bcba6..281dada8a6 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CreationMetadata.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/CreationMetadata.java @@ -792,13 +792,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, CreationMetadata st case 4: // TABLES_USED if (schemeField.type == org.apache.thrift.protocol.TType.SET) { { - org.apache.thrift.protocol.TSet _set708 = iprot.readSetBegin(); - struct.tablesUsed = new HashSet(2*_set708.size); - String _elem709; - for (int _i710 = 0; _i710 < _set708.size; ++_i710) + org.apache.thrift.protocol.TSet _set716 = iprot.readSetBegin(); + struct.tablesUsed = new HashSet(2*_set716.size); + String _elem717; + for (int _i718 = 0; _i718 < _set716.size; ++_i718) { - _elem709 = iprot.readString(); - struct.tablesUsed.add(_elem709); + _elem717 = iprot.readString(); + struct.tablesUsed.add(_elem717); } iprot.readSetEnd(); } @@ -855,9 +855,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, CreationMetadata s oprot.writeFieldBegin(TABLES_USED_FIELD_DESC); { oprot.writeSetBegin(new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.STRING, struct.tablesUsed.size())); - for (String _iter711 : struct.tablesUsed) + for (String _iter719 : struct.tablesUsed) { - oprot.writeString(_iter711); + oprot.writeString(_iter719); } oprot.writeSetEnd(); } @@ -897,9 +897,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, CreationMetadata st oprot.writeString(struct.tblName); { oprot.writeI32(struct.tablesUsed.size()); - for (String _iter712 : struct.tablesUsed) + for (String _iter720 : struct.tablesUsed) { - oprot.writeString(_iter712); + oprot.writeString(_iter720); } } BitSet optionals = new BitSet(); @@ -928,13 +928,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, CreationMetadata str struct.tblName = iprot.readString(); struct.setTblNameIsSet(true); { - org.apache.thrift.protocol.TSet _set713 = new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.tablesUsed = new HashSet(2*_set713.size); - String _elem714; - for (int _i715 = 0; _i715 < _set713.size; ++_i715) + org.apache.thrift.protocol.TSet _set721 = new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.tablesUsed = new HashSet(2*_set721.size); + String _elem722; + for (int _i723 = 0; _i723 < _set721.size; ++_i723) { - _elem714 = iprot.readString(); - struct.tablesUsed.add(_elem714); + _elem722 = iprot.readString(); + struct.tablesUsed.add(_elem722); } } struct.setTablesUsedIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FindSchemasByColsResp.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FindSchemasByColsResp.java index bb640865b3..79d9fc6409 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FindSchemasByColsResp.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FindSchemasByColsResp.java @@ -350,14 +350,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, FindSchemasByColsRe case 1: // SCHEMA_VERSIONS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list912 = iprot.readListBegin(); - struct.schemaVersions = new ArrayList(_list912.size); - SchemaVersionDescriptor _elem913; - for (int _i914 = 0; _i914 < _list912.size; ++_i914) + org.apache.thrift.protocol.TList _list936 = iprot.readListBegin(); + struct.schemaVersions = new ArrayList(_list936.size); + SchemaVersionDescriptor _elem937; + for (int _i938 = 0; _i938 < _list936.size; ++_i938) { - _elem913 = new SchemaVersionDescriptor(); - _elem913.read(iprot); - struct.schemaVersions.add(_elem913); + _elem937 = new SchemaVersionDescriptor(); + _elem937.read(iprot); + struct.schemaVersions.add(_elem937); } iprot.readListEnd(); } @@ -383,9 +383,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, FindSchemasByColsR oprot.writeFieldBegin(SCHEMA_VERSIONS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.schemaVersions.size())); - for (SchemaVersionDescriptor _iter915 : struct.schemaVersions) + for (SchemaVersionDescriptor _iter939 : struct.schemaVersions) { - _iter915.write(oprot); + _iter939.write(oprot); } oprot.writeListEnd(); } @@ -416,9 +416,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, FindSchemasByColsRe if (struct.isSetSchemaVersions()) { { oprot.writeI32(struct.schemaVersions.size()); - for (SchemaVersionDescriptor _iter916 : struct.schemaVersions) + for (SchemaVersionDescriptor _iter940 : struct.schemaVersions) { - _iter916.write(oprot); + _iter940.write(oprot); } } } @@ -430,14 +430,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, FindSchemasByColsRes BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list917 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.schemaVersions = new ArrayList(_list917.size); - SchemaVersionDescriptor _elem918; - for (int _i919 = 0; _i919 < _list917.size; ++_i919) + org.apache.thrift.protocol.TList _list941 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.schemaVersions = new ArrayList(_list941.size); + SchemaVersionDescriptor _elem942; + for (int _i943 = 0; _i943 < _list941.size; ++_i943) { - _elem918 = new SchemaVersionDescriptor(); - _elem918.read(iprot); - struct.schemaVersions.add(_elem918); + _elem942 = new SchemaVersionDescriptor(); + _elem942.read(iprot); + struct.schemaVersions.add(_elem942); } } struct.setSchemaVersionsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FireEventRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FireEventRequest.java index 7b0ec6c20a..2560922cfb 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FireEventRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/FireEventRequest.java @@ -794,13 +794,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, FireEventRequest st case 5: // PARTITION_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list740 = iprot.readListBegin(); - struct.partitionVals = new ArrayList(_list740.size); - String _elem741; - for (int _i742 = 0; _i742 < _list740.size; ++_i742) + org.apache.thrift.protocol.TList _list756 = iprot.readListBegin(); + struct.partitionVals = new ArrayList(_list756.size); + String _elem757; + for (int _i758 = 0; _i758 < _list756.size; ++_i758) { - _elem741 = iprot.readString(); - struct.partitionVals.add(_elem741); + _elem757 = iprot.readString(); + struct.partitionVals.add(_elem757); } iprot.readListEnd(); } @@ -857,9 +857,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, FireEventRequest s oprot.writeFieldBegin(PARTITION_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.partitionVals.size())); - for (String _iter743 : struct.partitionVals) + for (String _iter759 : struct.partitionVals) { - oprot.writeString(_iter743); + oprot.writeString(_iter759); } oprot.writeListEnd(); } @@ -915,9 +915,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, FireEventRequest st if (struct.isSetPartitionVals()) { { oprot.writeI32(struct.partitionVals.size()); - for (String _iter744 : struct.partitionVals) + for (String _iter760 : struct.partitionVals) { - oprot.writeString(_iter744); + oprot.writeString(_iter760); } } } @@ -945,13 +945,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, FireEventRequest str } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list745 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.partitionVals = new ArrayList(_list745.size); - String _elem746; - for (int _i747 = 0; _i747 < _list745.size; ++_i747) + org.apache.thrift.protocol.TList _list761 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partitionVals = new ArrayList(_list761.size); + String _elem762; + for (int _i763 = 0; _i763 < _list761.size; ++_i763) { - _elem746 = iprot.readString(); - struct.partitionVals.add(_elem746); + _elem762 = iprot.readString(); + struct.partitionVals.add(_elem762); } } struct.setPartitionValsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetAllFunctionsResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetAllFunctionsResponse.java index 544ba19c24..f68afe8494 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetAllFunctionsResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetAllFunctionsResponse.java @@ -346,14 +346,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetAllFunctionsResp case 1: // FUNCTIONS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list808 = iprot.readListBegin(); - struct.functions = new ArrayList(_list808.size); - Function _elem809; - for (int _i810 = 0; _i810 < _list808.size; ++_i810) + org.apache.thrift.protocol.TList _list832 = iprot.readListBegin(); + struct.functions = new ArrayList(_list832.size); + Function _elem833; + for (int _i834 = 0; _i834 < _list832.size; ++_i834) { - _elem809 = new Function(); - _elem809.read(iprot); - struct.functions.add(_elem809); + _elem833 = new Function(); + _elem833.read(iprot); + struct.functions.add(_elem833); } iprot.readListEnd(); } @@ -380,9 +380,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetAllFunctionsRes oprot.writeFieldBegin(FUNCTIONS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.functions.size())); - for (Function _iter811 : struct.functions) + for (Function _iter835 : struct.functions) { - _iter811.write(oprot); + _iter835.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetAllFunctionsResp if (struct.isSetFunctions()) { { oprot.writeI32(struct.functions.size()); - for (Function _iter812 : struct.functions) + for (Function _iter836 : struct.functions) { - _iter812.write(oprot); + _iter836.write(oprot); } } } @@ -428,14 +428,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, GetAllFunctionsRespo BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list813 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.functions = new ArrayList(_list813.size); - Function _elem814; - for (int _i815 = 0; _i815 < _list813.size; ++_i815) + org.apache.thrift.protocol.TList _list837 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.functions = new ArrayList(_list837.size); + Function _elem838; + for (int _i839 = 0; _i839 < _list837.size; ++_i839) { - _elem814 = new Function(); - _elem814.read(iprot); - struct.functions.add(_elem814); + _elem838 = new Function(); + _elem838.read(iprot); + struct.functions.add(_elem838); } } struct.setFunctionsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprRequest.java index 0a94f2f899..836f35fe95 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprRequest.java @@ -619,13 +619,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetFileMetadataByEx case 1: // FILE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list758 = iprot.readListBegin(); - struct.fileIds = new ArrayList(_list758.size); - long _elem759; - for (int _i760 = 0; _i760 < _list758.size; ++_i760) + org.apache.thrift.protocol.TList _list782 = iprot.readListBegin(); + struct.fileIds = new ArrayList(_list782.size); + long _elem783; + for (int _i784 = 0; _i784 < _list782.size; ++_i784) { - _elem759 = iprot.readI64(); - struct.fileIds.add(_elem759); + _elem783 = iprot.readI64(); + struct.fileIds.add(_elem783); } iprot.readListEnd(); } @@ -675,9 +675,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetFileMetadataByE oprot.writeFieldBegin(FILE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.fileIds.size())); - for (long _iter761 : struct.fileIds) + for (long _iter785 : struct.fileIds) { - oprot.writeI64(_iter761); + oprot.writeI64(_iter785); } oprot.writeListEnd(); } @@ -719,9 +719,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataByEx TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.fileIds.size()); - for (long _iter762 : struct.fileIds) + for (long _iter786 : struct.fileIds) { - oprot.writeI64(_iter762); + oprot.writeI64(_iter786); } } oprot.writeBinary(struct.expr); @@ -745,13 +745,13 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataByEx public void read(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataByExprRequest struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list763 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.fileIds = new ArrayList(_list763.size); - long _elem764; - for (int _i765 = 0; _i765 < _list763.size; ++_i765) + org.apache.thrift.protocol.TList _list787 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.fileIds = new ArrayList(_list787.size); + long _elem788; + for (int _i789 = 0; _i789 < _list787.size; ++_i789) { - _elem764 = iprot.readI64(); - struct.fileIds.add(_elem764); + _elem788 = iprot.readI64(); + struct.fileIds.add(_elem788); } } struct.setFileIdsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprResult.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprResult.java index e07d2e5698..17f0ee523d 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprResult.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataByExprResult.java @@ -444,16 +444,16 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetFileMetadataByEx case 1: // METADATA if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map748 = iprot.readMapBegin(); - struct.metadata = new HashMap(2*_map748.size); - long _key749; - MetadataPpdResult _val750; - for (int _i751 = 0; _i751 < _map748.size; ++_i751) + org.apache.thrift.protocol.TMap _map772 = iprot.readMapBegin(); + struct.metadata = new HashMap(2*_map772.size); + long _key773; + MetadataPpdResult _val774; + for (int _i775 = 0; _i775 < _map772.size; ++_i775) { - _key749 = iprot.readI64(); - _val750 = new MetadataPpdResult(); - _val750.read(iprot); - struct.metadata.put(_key749, _val750); + _key773 = iprot.readI64(); + _val774 = new MetadataPpdResult(); + _val774.read(iprot); + struct.metadata.put(_key773, _val774); } iprot.readMapEnd(); } @@ -487,10 +487,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetFileMetadataByE oprot.writeFieldBegin(METADATA_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.I64, org.apache.thrift.protocol.TType.STRUCT, struct.metadata.size())); - for (Map.Entry _iter752 : struct.metadata.entrySet()) + for (Map.Entry _iter776 : struct.metadata.entrySet()) { - oprot.writeI64(_iter752.getKey()); - _iter752.getValue().write(oprot); + oprot.writeI64(_iter776.getKey()); + _iter776.getValue().write(oprot); } oprot.writeMapEnd(); } @@ -518,10 +518,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataByEx TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.metadata.size()); - for (Map.Entry _iter753 : struct.metadata.entrySet()) + for (Map.Entry _iter777 : struct.metadata.entrySet()) { - oprot.writeI64(_iter753.getKey()); - _iter753.getValue().write(oprot); + oprot.writeI64(_iter777.getKey()); + _iter777.getValue().write(oprot); } } oprot.writeBool(struct.isSupported); @@ -531,16 +531,16 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataByEx public void read(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataByExprResult struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TMap _map754 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.I64, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.metadata = new HashMap(2*_map754.size); - long _key755; - MetadataPpdResult _val756; - for (int _i757 = 0; _i757 < _map754.size; ++_i757) + org.apache.thrift.protocol.TMap _map778 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.I64, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.metadata = new HashMap(2*_map778.size); + long _key779; + MetadataPpdResult _val780; + for (int _i781 = 0; _i781 < _map778.size; ++_i781) { - _key755 = iprot.readI64(); - _val756 = new MetadataPpdResult(); - _val756.read(iprot); - struct.metadata.put(_key755, _val756); + _key779 = iprot.readI64(); + _val780 = new MetadataPpdResult(); + _val780.read(iprot); + struct.metadata.put(_key779, _val780); } } struct.setMetadataIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataRequest.java index ebb663938d..12b439283a 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataRequest.java @@ -351,13 +351,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetFileMetadataRequ case 1: // FILE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list776 = iprot.readListBegin(); - struct.fileIds = new ArrayList(_list776.size); - long _elem777; - for (int _i778 = 0; _i778 < _list776.size; ++_i778) + org.apache.thrift.protocol.TList _list800 = iprot.readListBegin(); + struct.fileIds = new ArrayList(_list800.size); + long _elem801; + for (int _i802 = 0; _i802 < _list800.size; ++_i802) { - _elem777 = iprot.readI64(); - struct.fileIds.add(_elem777); + _elem801 = iprot.readI64(); + struct.fileIds.add(_elem801); } iprot.readListEnd(); } @@ -383,9 +383,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetFileMetadataReq oprot.writeFieldBegin(FILE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.fileIds.size())); - for (long _iter779 : struct.fileIds) + for (long _iter803 : struct.fileIds) { - oprot.writeI64(_iter779); + oprot.writeI64(_iter803); } oprot.writeListEnd(); } @@ -410,9 +410,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataRequ TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.fileIds.size()); - for (long _iter780 : struct.fileIds) + for (long _iter804 : struct.fileIds) { - oprot.writeI64(_iter780); + oprot.writeI64(_iter804); } } } @@ -421,13 +421,13 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataRequ public void read(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataRequest struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list781 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.fileIds = new ArrayList(_list781.size); - long _elem782; - for (int _i783 = 0; _i783 < _list781.size; ++_i783) + org.apache.thrift.protocol.TList _list805 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.fileIds = new ArrayList(_list805.size); + long _elem806; + for (int _i807 = 0; _i807 < _list805.size; ++_i807) { - _elem782 = iprot.readI64(); - struct.fileIds.add(_elem782); + _elem806 = iprot.readI64(); + struct.fileIds.add(_elem806); } } struct.setFileIdsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataResult.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataResult.java index 67981cd70a..65708d759d 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataResult.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetFileMetadataResult.java @@ -433,15 +433,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetFileMetadataResu case 1: // METADATA if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map766 = iprot.readMapBegin(); - struct.metadata = new HashMap(2*_map766.size); - long _key767; - ByteBuffer _val768; - for (int _i769 = 0; _i769 < _map766.size; ++_i769) + org.apache.thrift.protocol.TMap _map790 = iprot.readMapBegin(); + struct.metadata = new HashMap(2*_map790.size); + long _key791; + ByteBuffer _val792; + for (int _i793 = 0; _i793 < _map790.size; ++_i793) { - _key767 = iprot.readI64(); - _val768 = iprot.readBinary(); - struct.metadata.put(_key767, _val768); + _key791 = iprot.readI64(); + _val792 = iprot.readBinary(); + struct.metadata.put(_key791, _val792); } iprot.readMapEnd(); } @@ -475,10 +475,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetFileMetadataRes oprot.writeFieldBegin(METADATA_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.I64, org.apache.thrift.protocol.TType.STRING, struct.metadata.size())); - for (Map.Entry _iter770 : struct.metadata.entrySet()) + for (Map.Entry _iter794 : struct.metadata.entrySet()) { - oprot.writeI64(_iter770.getKey()); - oprot.writeBinary(_iter770.getValue()); + oprot.writeI64(_iter794.getKey()); + oprot.writeBinary(_iter794.getValue()); } oprot.writeMapEnd(); } @@ -506,10 +506,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataResu TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.metadata.size()); - for (Map.Entry _iter771 : struct.metadata.entrySet()) + for (Map.Entry _iter795 : struct.metadata.entrySet()) { - oprot.writeI64(_iter771.getKey()); - oprot.writeBinary(_iter771.getValue()); + oprot.writeI64(_iter795.getKey()); + oprot.writeBinary(_iter795.getValue()); } } oprot.writeBool(struct.isSupported); @@ -519,15 +519,15 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataResu public void read(org.apache.thrift.protocol.TProtocol prot, GetFileMetadataResult struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TMap _map772 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.I64, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.metadata = new HashMap(2*_map772.size); - long _key773; - ByteBuffer _val774; - for (int _i775 = 0; _i775 < _map772.size; ++_i775) + org.apache.thrift.protocol.TMap _map796 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.I64, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.metadata = new HashMap(2*_map796.size); + long _key797; + ByteBuffer _val798; + for (int _i799 = 0; _i799 < _map796.size; ++_i799) { - _key773 = iprot.readI64(); - _val774 = iprot.readBinary(); - struct.metadata.put(_key773, _val774); + _key797 = iprot.readI64(); + _val798 = iprot.readBinary(); + struct.metadata.put(_key797, _val798); } } struct.setMetadataIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesRequest.java index 6a78b77761..09ca865ba8 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesRequest.java @@ -606,13 +606,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetTablesRequest st case 2: // TBL_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list824 = iprot.readListBegin(); - struct.tblNames = new ArrayList(_list824.size); - String _elem825; - for (int _i826 = 0; _i826 < _list824.size; ++_i826) + org.apache.thrift.protocol.TList _list848 = iprot.readListBegin(); + struct.tblNames = new ArrayList(_list848.size); + String _elem849; + for (int _i850 = 0; _i850 < _list848.size; ++_i850) { - _elem825 = iprot.readString(); - struct.tblNames.add(_elem825); + _elem849 = iprot.readString(); + struct.tblNames.add(_elem849); } iprot.readListEnd(); } @@ -661,9 +661,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetTablesRequest s oprot.writeFieldBegin(TBL_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.tblNames.size())); - for (String _iter827 : struct.tblNames) + for (String _iter851 : struct.tblNames) { - oprot.writeString(_iter827); + oprot.writeString(_iter851); } oprot.writeListEnd(); } @@ -716,9 +716,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetTablesRequest st if (struct.isSetTblNames()) { { oprot.writeI32(struct.tblNames.size()); - for (String _iter828 : struct.tblNames) + for (String _iter852 : struct.tblNames) { - oprot.writeString(_iter828); + oprot.writeString(_iter852); } } } @@ -738,13 +738,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, GetTablesRequest str BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list829 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.tblNames = new ArrayList(_list829.size); - String _elem830; - for (int _i831 = 0; _i831 < _list829.size; ++_i831) + org.apache.thrift.protocol.TList _list853 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.tblNames = new ArrayList(_list853.size); + String _elem854; + for (int _i855 = 0; _i855 < _list853.size; ++_i855) { - _elem830 = iprot.readString(); - struct.tblNames.add(_elem830); + _elem854 = iprot.readString(); + struct.tblNames.add(_elem854); } } struct.setTblNamesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesResult.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesResult.java index 13be2edb2d..72256e64f5 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesResult.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetTablesResult.java @@ -354,14 +354,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetTablesResult str case 1: // TABLES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list832 = iprot.readListBegin(); - struct.tables = new ArrayList
(_list832.size); - Table _elem833; - for (int _i834 = 0; _i834 < _list832.size; ++_i834) + org.apache.thrift.protocol.TList _list856 = iprot.readListBegin(); + struct.tables = new ArrayList
(_list856.size); + Table _elem857; + for (int _i858 = 0; _i858 < _list856.size; ++_i858) { - _elem833 = new Table(); - _elem833.read(iprot); - struct.tables.add(_elem833); + _elem857 = new Table(); + _elem857.read(iprot); + struct.tables.add(_elem857); } iprot.readListEnd(); } @@ -387,9 +387,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetTablesResult st oprot.writeFieldBegin(TABLES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.tables.size())); - for (Table _iter835 : struct.tables) + for (Table _iter859 : struct.tables) { - _iter835.write(oprot); + _iter859.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetTablesResult str TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.tables.size()); - for (Table _iter836 : struct.tables) + for (Table _iter860 : struct.tables) { - _iter836.write(oprot); + _iter860.write(oprot); } } } @@ -425,14 +425,14 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetTablesResult str public void read(org.apache.thrift.protocol.TProtocol prot, GetTablesResult struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list837 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.tables = new ArrayList
(_list837.size); - Table _elem838; - for (int _i839 = 0; _i839 < _list837.size; ++_i839) + org.apache.thrift.protocol.TList _list861 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.tables = new ArrayList
(_list861.size); + Table _elem862; + for (int _i863 = 0; _i863 < _list861.size; ++_i863) { - _elem838 = new Table(); - _elem838.read(iprot); - struct.tables.add(_elem838); + _elem862 = new Table(); + _elem862.read(iprot); + struct.tables.add(_elem862); } } struct.setTablesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsRequest.java index 27b6cf8669..af62ca14ed 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsRequest.java @@ -436,13 +436,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetValidWriteIdsReq case 1: // FULL_TABLE_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list602 = iprot.readListBegin(); - struct.fullTableNames = new ArrayList(_list602.size); - String _elem603; - for (int _i604 = 0; _i604 < _list602.size; ++_i604) + org.apache.thrift.protocol.TList _list610 = iprot.readListBegin(); + struct.fullTableNames = new ArrayList(_list610.size); + String _elem611; + for (int _i612 = 0; _i612 < _list610.size; ++_i612) { - _elem603 = iprot.readString(); - struct.fullTableNames.add(_elem603); + _elem611 = iprot.readString(); + struct.fullTableNames.add(_elem611); } iprot.readListEnd(); } @@ -476,9 +476,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetValidWriteIdsRe oprot.writeFieldBegin(FULL_TABLE_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.fullTableNames.size())); - for (String _iter605 : struct.fullTableNames) + for (String _iter613 : struct.fullTableNames) { - oprot.writeString(_iter605); + oprot.writeString(_iter613); } oprot.writeListEnd(); } @@ -508,9 +508,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetValidWriteIdsReq TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.fullTableNames.size()); - for (String _iter606 : struct.fullTableNames) + for (String _iter614 : struct.fullTableNames) { - oprot.writeString(_iter606); + oprot.writeString(_iter614); } } oprot.writeString(struct.validTxnList); @@ -520,13 +520,13 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetValidWriteIdsReq public void read(org.apache.thrift.protocol.TProtocol prot, GetValidWriteIdsRequest struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list607 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.fullTableNames = new ArrayList(_list607.size); - String _elem608; - for (int _i609 = 0; _i609 < _list607.size; ++_i609) + org.apache.thrift.protocol.TList _list615 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.fullTableNames = new ArrayList(_list615.size); + String _elem616; + for (int _i617 = 0; _i617 < _list615.size; ++_i617) { - _elem608 = iprot.readString(); - struct.fullTableNames.add(_elem608); + _elem616 = iprot.readString(); + struct.fullTableNames.add(_elem616); } } struct.setFullTableNamesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsResponse.java index 7a1bbc7cc1..615a422256 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/GetValidWriteIdsResponse.java @@ -354,14 +354,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, GetValidWriteIdsRes case 1: // TBL_VALID_WRITE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list618 = iprot.readListBegin(); - struct.tblValidWriteIds = new ArrayList(_list618.size); - TableValidWriteIds _elem619; - for (int _i620 = 0; _i620 < _list618.size; ++_i620) + org.apache.thrift.protocol.TList _list626 = iprot.readListBegin(); + struct.tblValidWriteIds = new ArrayList(_list626.size); + TableValidWriteIds _elem627; + for (int _i628 = 0; _i628 < _list626.size; ++_i628) { - _elem619 = new TableValidWriteIds(); - _elem619.read(iprot); - struct.tblValidWriteIds.add(_elem619); + _elem627 = new TableValidWriteIds(); + _elem627.read(iprot); + struct.tblValidWriteIds.add(_elem627); } iprot.readListEnd(); } @@ -387,9 +387,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, GetValidWriteIdsRe oprot.writeFieldBegin(TBL_VALID_WRITE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.tblValidWriteIds.size())); - for (TableValidWriteIds _iter621 : struct.tblValidWriteIds) + for (TableValidWriteIds _iter629 : struct.tblValidWriteIds) { - _iter621.write(oprot); + _iter629.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetValidWriteIdsRes TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.tblValidWriteIds.size()); - for (TableValidWriteIds _iter622 : struct.tblValidWriteIds) + for (TableValidWriteIds _iter630 : struct.tblValidWriteIds) { - _iter622.write(oprot); + _iter630.write(oprot); } } } @@ -425,14 +425,14 @@ public void write(org.apache.thrift.protocol.TProtocol prot, GetValidWriteIdsRes public void read(org.apache.thrift.protocol.TProtocol prot, GetValidWriteIdsResponse struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list623 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.tblValidWriteIds = new ArrayList(_list623.size); - TableValidWriteIds _elem624; - for (int _i625 = 0; _i625 < _list623.size; ++_i625) + org.apache.thrift.protocol.TList _list631 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.tblValidWriteIds = new ArrayList(_list631.size); + TableValidWriteIds _elem632; + for (int _i633 = 0; _i633 < _list631.size; ++_i633) { - _elem624 = new TableValidWriteIds(); - _elem624.read(iprot); - struct.tblValidWriteIds.add(_elem624); + _elem632 = new TableValidWriteIds(); + _elem632.read(iprot); + struct.tblValidWriteIds.add(_elem632); } } struct.setTblValidWriteIdsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/HeartbeatTxnRangeResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/HeartbeatTxnRangeResponse.java index 4999215989..a3dceab951 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/HeartbeatTxnRangeResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/HeartbeatTxnRangeResponse.java @@ -453,13 +453,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, HeartbeatTxnRangeRe case 1: // ABORTED if (schemeField.type == org.apache.thrift.protocol.TType.SET) { { - org.apache.thrift.protocol.TSet _set666 = iprot.readSetBegin(); - struct.aborted = new HashSet(2*_set666.size); - long _elem667; - for (int _i668 = 0; _i668 < _set666.size; ++_i668) + org.apache.thrift.protocol.TSet _set674 = iprot.readSetBegin(); + struct.aborted = new HashSet(2*_set674.size); + long _elem675; + for (int _i676 = 0; _i676 < _set674.size; ++_i676) { - _elem667 = iprot.readI64(); - struct.aborted.add(_elem667); + _elem675 = iprot.readI64(); + struct.aborted.add(_elem675); } iprot.readSetEnd(); } @@ -471,13 +471,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, HeartbeatTxnRangeRe case 2: // NOSUCH if (schemeField.type == org.apache.thrift.protocol.TType.SET) { { - org.apache.thrift.protocol.TSet _set669 = iprot.readSetBegin(); - struct.nosuch = new HashSet(2*_set669.size); - long _elem670; - for (int _i671 = 0; _i671 < _set669.size; ++_i671) + org.apache.thrift.protocol.TSet _set677 = iprot.readSetBegin(); + struct.nosuch = new HashSet(2*_set677.size); + long _elem678; + for (int _i679 = 0; _i679 < _set677.size; ++_i679) { - _elem670 = iprot.readI64(); - struct.nosuch.add(_elem670); + _elem678 = iprot.readI64(); + struct.nosuch.add(_elem678); } iprot.readSetEnd(); } @@ -503,9 +503,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, HeartbeatTxnRangeR oprot.writeFieldBegin(ABORTED_FIELD_DESC); { oprot.writeSetBegin(new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.I64, struct.aborted.size())); - for (long _iter672 : struct.aborted) + for (long _iter680 : struct.aborted) { - oprot.writeI64(_iter672); + oprot.writeI64(_iter680); } oprot.writeSetEnd(); } @@ -515,9 +515,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, HeartbeatTxnRangeR oprot.writeFieldBegin(NOSUCH_FIELD_DESC); { oprot.writeSetBegin(new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.I64, struct.nosuch.size())); - for (long _iter673 : struct.nosuch) + for (long _iter681 : struct.nosuch) { - oprot.writeI64(_iter673); + oprot.writeI64(_iter681); } oprot.writeSetEnd(); } @@ -542,16 +542,16 @@ public void write(org.apache.thrift.protocol.TProtocol prot, HeartbeatTxnRangeRe TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.aborted.size()); - for (long _iter674 : struct.aborted) + for (long _iter682 : struct.aborted) { - oprot.writeI64(_iter674); + oprot.writeI64(_iter682); } } { oprot.writeI32(struct.nosuch.size()); - for (long _iter675 : struct.nosuch) + for (long _iter683 : struct.nosuch) { - oprot.writeI64(_iter675); + oprot.writeI64(_iter683); } } } @@ -560,24 +560,24 @@ public void write(org.apache.thrift.protocol.TProtocol prot, HeartbeatTxnRangeRe public void read(org.apache.thrift.protocol.TProtocol prot, HeartbeatTxnRangeResponse struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TSet _set676 = new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.aborted = new HashSet(2*_set676.size); - long _elem677; - for (int _i678 = 0; _i678 < _set676.size; ++_i678) + org.apache.thrift.protocol.TSet _set684 = new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.aborted = new HashSet(2*_set684.size); + long _elem685; + for (int _i686 = 0; _i686 < _set684.size; ++_i686) { - _elem677 = iprot.readI64(); - struct.aborted.add(_elem677); + _elem685 = iprot.readI64(); + struct.aborted.add(_elem685); } } struct.setAbortedIsSet(true); { - org.apache.thrift.protocol.TSet _set679 = new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.nosuch = new HashSet(2*_set679.size); - long _elem680; - for (int _i681 = 0; _i681 < _set679.size; ++_i681) + org.apache.thrift.protocol.TSet _set687 = new org.apache.thrift.protocol.TSet(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.nosuch = new HashSet(2*_set687.size); + long _elem688; + for (int _i689 = 0; _i689 < _set687.size; ++_i689) { - _elem680 = iprot.readI64(); - struct.nosuch.add(_elem680); + _elem688 = iprot.readI64(); + struct.nosuch.add(_elem688); } } struct.setNosuchIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/InsertEventRequestData.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/InsertEventRequestData.java index 0a240e0286..4a9824b908 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/InsertEventRequestData.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/InsertEventRequestData.java @@ -41,6 +41,7 @@ private static final org.apache.thrift.protocol.TField REPLACE_FIELD_DESC = new org.apache.thrift.protocol.TField("replace", org.apache.thrift.protocol.TType.BOOL, (short)1); private static final org.apache.thrift.protocol.TField FILES_ADDED_FIELD_DESC = new org.apache.thrift.protocol.TField("filesAdded", org.apache.thrift.protocol.TType.LIST, (short)2); private static final org.apache.thrift.protocol.TField FILES_ADDED_CHECKSUM_FIELD_DESC = new org.apache.thrift.protocol.TField("filesAddedChecksum", org.apache.thrift.protocol.TType.LIST, (short)3); + private static final org.apache.thrift.protocol.TField SUB_DIRECTORY_LIST_FIELD_DESC = new org.apache.thrift.protocol.TField("subDirectoryList", org.apache.thrift.protocol.TType.LIST, (short)4); private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); static { @@ -51,12 +52,14 @@ private boolean replace; // optional private List filesAdded; // required private List filesAddedChecksum; // optional + private List subDirectoryList; // optional /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { REPLACE((short)1, "replace"), FILES_ADDED((short)2, "filesAdded"), - FILES_ADDED_CHECKSUM((short)3, "filesAddedChecksum"); + FILES_ADDED_CHECKSUM((short)3, "filesAddedChecksum"), + SUB_DIRECTORY_LIST((short)4, "subDirectoryList"); private static final Map byName = new HashMap(); @@ -77,6 +80,8 @@ public static _Fields findByThriftId(int fieldId) { return FILES_ADDED; case 3: // FILES_ADDED_CHECKSUM return FILES_ADDED_CHECKSUM; + case 4: // SUB_DIRECTORY_LIST + return SUB_DIRECTORY_LIST; default: return null; } @@ -119,7 +124,7 @@ public String getFieldName() { // isset id assignments private static final int __REPLACE_ISSET_ID = 0; private byte __isset_bitfield = 0; - private static final _Fields optionals[] = {_Fields.REPLACE,_Fields.FILES_ADDED_CHECKSUM}; + private static final _Fields optionals[] = {_Fields.REPLACE,_Fields.FILES_ADDED_CHECKSUM,_Fields.SUB_DIRECTORY_LIST}; public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); @@ -131,6 +136,9 @@ public String getFieldName() { tmpMap.put(_Fields.FILES_ADDED_CHECKSUM, new org.apache.thrift.meta_data.FieldMetaData("filesAddedChecksum", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.ListMetaData(org.apache.thrift.protocol.TType.LIST, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)))); + tmpMap.put(_Fields.SUB_DIRECTORY_LIST, new org.apache.thrift.meta_data.FieldMetaData("subDirectoryList", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.ListMetaData(org.apache.thrift.protocol.TType.LIST, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)))); metaDataMap = Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(InsertEventRequestData.class, metaDataMap); } @@ -159,6 +167,10 @@ public InsertEventRequestData(InsertEventRequestData other) { List __this__filesAddedChecksum = new ArrayList(other.filesAddedChecksum); this.filesAddedChecksum = __this__filesAddedChecksum; } + if (other.isSetSubDirectoryList()) { + List __this__subDirectoryList = new ArrayList(other.subDirectoryList); + this.subDirectoryList = __this__subDirectoryList; + } } public InsertEventRequestData deepCopy() { @@ -171,6 +183,7 @@ public void clear() { this.replace = false; this.filesAdded = null; this.filesAddedChecksum = null; + this.subDirectoryList = null; } public boolean isReplace() { @@ -271,6 +284,44 @@ public void setFilesAddedChecksumIsSet(boolean value) { } } + public int getSubDirectoryListSize() { + return (this.subDirectoryList == null) ? 0 : this.subDirectoryList.size(); + } + + public java.util.Iterator getSubDirectoryListIterator() { + return (this.subDirectoryList == null) ? null : this.subDirectoryList.iterator(); + } + + public void addToSubDirectoryList(String elem) { + if (this.subDirectoryList == null) { + this.subDirectoryList = new ArrayList(); + } + this.subDirectoryList.add(elem); + } + + public List getSubDirectoryList() { + return this.subDirectoryList; + } + + public void setSubDirectoryList(List subDirectoryList) { + this.subDirectoryList = subDirectoryList; + } + + public void unsetSubDirectoryList() { + this.subDirectoryList = null; + } + + /** Returns true if field subDirectoryList is set (has been assigned a value) and false otherwise */ + public boolean isSetSubDirectoryList() { + return this.subDirectoryList != null; + } + + public void setSubDirectoryListIsSet(boolean value) { + if (!value) { + this.subDirectoryList = null; + } + } + public void setFieldValue(_Fields field, Object value) { switch (field) { case REPLACE: @@ -297,6 +348,14 @@ public void setFieldValue(_Fields field, Object value) { } break; + case SUB_DIRECTORY_LIST: + if (value == null) { + unsetSubDirectoryList(); + } else { + setSubDirectoryList((List)value); + } + break; + } } @@ -311,6 +370,9 @@ public Object getFieldValue(_Fields field) { case FILES_ADDED_CHECKSUM: return getFilesAddedChecksum(); + case SUB_DIRECTORY_LIST: + return getSubDirectoryList(); + } throw new IllegalStateException(); } @@ -328,6 +390,8 @@ public boolean isSet(_Fields field) { return isSetFilesAdded(); case FILES_ADDED_CHECKSUM: return isSetFilesAddedChecksum(); + case SUB_DIRECTORY_LIST: + return isSetSubDirectoryList(); } throw new IllegalStateException(); } @@ -372,6 +436,15 @@ public boolean equals(InsertEventRequestData that) { return false; } + boolean this_present_subDirectoryList = true && this.isSetSubDirectoryList(); + boolean that_present_subDirectoryList = true && that.isSetSubDirectoryList(); + if (this_present_subDirectoryList || that_present_subDirectoryList) { + if (!(this_present_subDirectoryList && that_present_subDirectoryList)) + return false; + if (!this.subDirectoryList.equals(that.subDirectoryList)) + return false; + } + return true; } @@ -394,6 +467,11 @@ public int hashCode() { if (present_filesAddedChecksum) list.add(filesAddedChecksum); + boolean present_subDirectoryList = true && (isSetSubDirectoryList()); + list.add(present_subDirectoryList); + if (present_subDirectoryList) + list.add(subDirectoryList); + return list.hashCode(); } @@ -435,6 +513,16 @@ public int compareTo(InsertEventRequestData other) { return lastComparison; } } + lastComparison = Boolean.valueOf(isSetSubDirectoryList()).compareTo(other.isSetSubDirectoryList()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetSubDirectoryList()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.subDirectoryList, other.subDirectoryList); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -478,6 +566,16 @@ public String toString() { } first = false; } + if (isSetSubDirectoryList()) { + if (!first) sb.append(", "); + sb.append("subDirectoryList:"); + if (this.subDirectoryList == null) { + sb.append("null"); + } else { + sb.append(this.subDirectoryList); + } + first = false; + } sb.append(")"); return sb.toString(); } @@ -538,13 +636,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, InsertEventRequestD case 2: // FILES_ADDED if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list724 = iprot.readListBegin(); - struct.filesAdded = new ArrayList(_list724.size); - String _elem725; - for (int _i726 = 0; _i726 < _list724.size; ++_i726) + org.apache.thrift.protocol.TList _list732 = iprot.readListBegin(); + struct.filesAdded = new ArrayList(_list732.size); + String _elem733; + for (int _i734 = 0; _i734 < _list732.size; ++_i734) { - _elem725 = iprot.readString(); - struct.filesAdded.add(_elem725); + _elem733 = iprot.readString(); + struct.filesAdded.add(_elem733); } iprot.readListEnd(); } @@ -556,13 +654,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, InsertEventRequestD case 3: // FILES_ADDED_CHECKSUM if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list727 = iprot.readListBegin(); - struct.filesAddedChecksum = new ArrayList(_list727.size); - String _elem728; - for (int _i729 = 0; _i729 < _list727.size; ++_i729) + org.apache.thrift.protocol.TList _list735 = iprot.readListBegin(); + struct.filesAddedChecksum = new ArrayList(_list735.size); + String _elem736; + for (int _i737 = 0; _i737 < _list735.size; ++_i737) { - _elem728 = iprot.readString(); - struct.filesAddedChecksum.add(_elem728); + _elem736 = iprot.readString(); + struct.filesAddedChecksum.add(_elem736); } iprot.readListEnd(); } @@ -571,6 +669,24 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, InsertEventRequestD org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 4: // SUB_DIRECTORY_LIST + if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { + { + org.apache.thrift.protocol.TList _list738 = iprot.readListBegin(); + struct.subDirectoryList = new ArrayList(_list738.size); + String _elem739; + for (int _i740 = 0; _i740 < _list738.size; ++_i740) + { + _elem739 = iprot.readString(); + struct.subDirectoryList.add(_elem739); + } + iprot.readListEnd(); + } + struct.setSubDirectoryListIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -593,9 +709,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, InsertEventRequest oprot.writeFieldBegin(FILES_ADDED_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.filesAdded.size())); - for (String _iter730 : struct.filesAdded) + for (String _iter741 : struct.filesAdded) { - oprot.writeString(_iter730); + oprot.writeString(_iter741); } oprot.writeListEnd(); } @@ -606,9 +722,23 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, InsertEventRequest oprot.writeFieldBegin(FILES_ADDED_CHECKSUM_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.filesAddedChecksum.size())); - for (String _iter731 : struct.filesAddedChecksum) + for (String _iter742 : struct.filesAddedChecksum) { - oprot.writeString(_iter731); + oprot.writeString(_iter742); + } + oprot.writeListEnd(); + } + oprot.writeFieldEnd(); + } + } + if (struct.subDirectoryList != null) { + if (struct.isSetSubDirectoryList()) { + oprot.writeFieldBegin(SUB_DIRECTORY_LIST_FIELD_DESC); + { + oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.subDirectoryList.size())); + for (String _iter743 : struct.subDirectoryList) + { + oprot.writeString(_iter743); } oprot.writeListEnd(); } @@ -634,9 +764,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, InsertEventRequestD TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.filesAdded.size()); - for (String _iter732 : struct.filesAdded) + for (String _iter744 : struct.filesAdded) { - oprot.writeString(_iter732); + oprot.writeString(_iter744); } } BitSet optionals = new BitSet(); @@ -646,16 +776,28 @@ public void write(org.apache.thrift.protocol.TProtocol prot, InsertEventRequestD if (struct.isSetFilesAddedChecksum()) { optionals.set(1); } - oprot.writeBitSet(optionals, 2); + if (struct.isSetSubDirectoryList()) { + optionals.set(2); + } + oprot.writeBitSet(optionals, 3); if (struct.isSetReplace()) { oprot.writeBool(struct.replace); } if (struct.isSetFilesAddedChecksum()) { { oprot.writeI32(struct.filesAddedChecksum.size()); - for (String _iter733 : struct.filesAddedChecksum) + for (String _iter745 : struct.filesAddedChecksum) + { + oprot.writeString(_iter745); + } + } + } + if (struct.isSetSubDirectoryList()) { + { + oprot.writeI32(struct.subDirectoryList.size()); + for (String _iter746 : struct.subDirectoryList) { - oprot.writeString(_iter733); + oprot.writeString(_iter746); } } } @@ -665,34 +807,47 @@ public void write(org.apache.thrift.protocol.TProtocol prot, InsertEventRequestD public void read(org.apache.thrift.protocol.TProtocol prot, InsertEventRequestData struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list734 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.filesAdded = new ArrayList(_list734.size); - String _elem735; - for (int _i736 = 0; _i736 < _list734.size; ++_i736) + org.apache.thrift.protocol.TList _list747 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.filesAdded = new ArrayList(_list747.size); + String _elem748; + for (int _i749 = 0; _i749 < _list747.size; ++_i749) { - _elem735 = iprot.readString(); - struct.filesAdded.add(_elem735); + _elem748 = iprot.readString(); + struct.filesAdded.add(_elem748); } } struct.setFilesAddedIsSet(true); - BitSet incoming = iprot.readBitSet(2); + BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { struct.replace = iprot.readBool(); struct.setReplaceIsSet(true); } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list737 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.filesAddedChecksum = new ArrayList(_list737.size); - String _elem738; - for (int _i739 = 0; _i739 < _list737.size; ++_i739) + org.apache.thrift.protocol.TList _list750 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.filesAddedChecksum = new ArrayList(_list750.size); + String _elem751; + for (int _i752 = 0; _i752 < _list750.size; ++_i752) { - _elem738 = iprot.readString(); - struct.filesAddedChecksum.add(_elem738); + _elem751 = iprot.readString(); + struct.filesAddedChecksum.add(_elem751); } } struct.setFilesAddedChecksumIsSet(true); } + if (incoming.get(2)) { + { + org.apache.thrift.protocol.TList _list753 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.subDirectoryList = new ArrayList(_list753.size); + String _elem754; + for (int _i755 = 0; _i755 < _list753.size; ++_i755) + { + _elem754 = iprot.readString(); + struct.subDirectoryList.add(_elem754); + } + } + struct.setSubDirectoryListIsSet(true); + } } } diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockRequest.java index d0dc21c319..d4eed3254c 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/LockRequest.java @@ -689,14 +689,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, LockRequest struct) case 1: // COMPONENT if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list650 = iprot.readListBegin(); - struct.component = new ArrayList(_list650.size); - LockComponent _elem651; - for (int _i652 = 0; _i652 < _list650.size; ++_i652) + org.apache.thrift.protocol.TList _list658 = iprot.readListBegin(); + struct.component = new ArrayList(_list658.size); + LockComponent _elem659; + for (int _i660 = 0; _i660 < _list658.size; ++_i660) { - _elem651 = new LockComponent(); - _elem651.read(iprot); - struct.component.add(_elem651); + _elem659 = new LockComponent(); + _elem659.read(iprot); + struct.component.add(_elem659); } iprot.readListEnd(); } @@ -754,9 +754,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, LockRequest struct oprot.writeFieldBegin(COMPONENT_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.component.size())); - for (LockComponent _iter653 : struct.component) + for (LockComponent _iter661 : struct.component) { - _iter653.write(oprot); + _iter661.write(oprot); } oprot.writeListEnd(); } @@ -803,9 +803,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, LockRequest struct) TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.component.size()); - for (LockComponent _iter654 : struct.component) + for (LockComponent _iter662 : struct.component) { - _iter654.write(oprot); + _iter662.write(oprot); } } oprot.writeString(struct.user); @@ -830,14 +830,14 @@ public void write(org.apache.thrift.protocol.TProtocol prot, LockRequest struct) public void read(org.apache.thrift.protocol.TProtocol prot, LockRequest struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list655 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.component = new ArrayList(_list655.size); - LockComponent _elem656; - for (int _i657 = 0; _i657 < _list655.size; ++_i657) + org.apache.thrift.protocol.TList _list663 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.component = new ArrayList(_list663.size); + LockComponent _elem664; + for (int _i665 = 0; _i665 < _list663.size; ++_i665) { - _elem656 = new LockComponent(); - _elem656.read(iprot); - struct.component.add(_elem656); + _elem664 = new LockComponent(); + _elem664.read(iprot); + struct.component.add(_elem664); } } struct.setComponentIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/NotificationEventResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/NotificationEventResponse.java index 0c850faea7..9228c39d74 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/NotificationEventResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/NotificationEventResponse.java @@ -354,14 +354,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, NotificationEventRe case 1: // EVENTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list716 = iprot.readListBegin(); - struct.events = new ArrayList(_list716.size); - NotificationEvent _elem717; - for (int _i718 = 0; _i718 < _list716.size; ++_i718) + org.apache.thrift.protocol.TList _list724 = iprot.readListBegin(); + struct.events = new ArrayList(_list724.size); + NotificationEvent _elem725; + for (int _i726 = 0; _i726 < _list724.size; ++_i726) { - _elem717 = new NotificationEvent(); - _elem717.read(iprot); - struct.events.add(_elem717); + _elem725 = new NotificationEvent(); + _elem725.read(iprot); + struct.events.add(_elem725); } iprot.readListEnd(); } @@ -387,9 +387,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, NotificationEventR oprot.writeFieldBegin(EVENTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.events.size())); - for (NotificationEvent _iter719 : struct.events) + for (NotificationEvent _iter727 : struct.events) { - _iter719.write(oprot); + _iter727.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, NotificationEventRe TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.events.size()); - for (NotificationEvent _iter720 : struct.events) + for (NotificationEvent _iter728 : struct.events) { - _iter720.write(oprot); + _iter728.write(oprot); } } } @@ -425,14 +425,14 @@ public void write(org.apache.thrift.protocol.TProtocol prot, NotificationEventRe public void read(org.apache.thrift.protocol.TProtocol prot, NotificationEventResponse struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list721 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.events = new ArrayList(_list721.size); - NotificationEvent _elem722; - for (int _i723 = 0; _i723 < _list721.size; ++_i723) + org.apache.thrift.protocol.TList _list729 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.events = new ArrayList(_list729.size); + NotificationEvent _elem730; + for (int _i731 = 0; _i731 < _list729.size; ++_i731) { - _elem722 = new NotificationEvent(); - _elem722.read(iprot); - struct.events.add(_elem722); + _elem730 = new NotificationEvent(); + _elem730.read(iprot); + struct.events.add(_elem730); } } struct.setEventsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/PutFileMetadataRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/PutFileMetadataRequest.java index 77c260d919..7d9ebbae76 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/PutFileMetadataRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/PutFileMetadataRequest.java @@ -547,13 +547,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, PutFileMetadataRequ case 1: // FILE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list784 = iprot.readListBegin(); - struct.fileIds = new ArrayList(_list784.size); - long _elem785; - for (int _i786 = 0; _i786 < _list784.size; ++_i786) + org.apache.thrift.protocol.TList _list808 = iprot.readListBegin(); + struct.fileIds = new ArrayList(_list808.size); + long _elem809; + for (int _i810 = 0; _i810 < _list808.size; ++_i810) { - _elem785 = iprot.readI64(); - struct.fileIds.add(_elem785); + _elem809 = iprot.readI64(); + struct.fileIds.add(_elem809); } iprot.readListEnd(); } @@ -565,13 +565,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, PutFileMetadataRequ case 2: // METADATA if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list787 = iprot.readListBegin(); - struct.metadata = new ArrayList(_list787.size); - ByteBuffer _elem788; - for (int _i789 = 0; _i789 < _list787.size; ++_i789) + org.apache.thrift.protocol.TList _list811 = iprot.readListBegin(); + struct.metadata = new ArrayList(_list811.size); + ByteBuffer _elem812; + for (int _i813 = 0; _i813 < _list811.size; ++_i813) { - _elem788 = iprot.readBinary(); - struct.metadata.add(_elem788); + _elem812 = iprot.readBinary(); + struct.metadata.add(_elem812); } iprot.readListEnd(); } @@ -605,9 +605,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, PutFileMetadataReq oprot.writeFieldBegin(FILE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.fileIds.size())); - for (long _iter790 : struct.fileIds) + for (long _iter814 : struct.fileIds) { - oprot.writeI64(_iter790); + oprot.writeI64(_iter814); } oprot.writeListEnd(); } @@ -617,9 +617,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, PutFileMetadataReq oprot.writeFieldBegin(METADATA_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.metadata.size())); - for (ByteBuffer _iter791 : struct.metadata) + for (ByteBuffer _iter815 : struct.metadata) { - oprot.writeBinary(_iter791); + oprot.writeBinary(_iter815); } oprot.writeListEnd(); } @@ -651,16 +651,16 @@ public void write(org.apache.thrift.protocol.TProtocol prot, PutFileMetadataRequ TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.fileIds.size()); - for (long _iter792 : struct.fileIds) + for (long _iter816 : struct.fileIds) { - oprot.writeI64(_iter792); + oprot.writeI64(_iter816); } } { oprot.writeI32(struct.metadata.size()); - for (ByteBuffer _iter793 : struct.metadata) + for (ByteBuffer _iter817 : struct.metadata) { - oprot.writeBinary(_iter793); + oprot.writeBinary(_iter817); } } BitSet optionals = new BitSet(); @@ -677,24 +677,24 @@ public void write(org.apache.thrift.protocol.TProtocol prot, PutFileMetadataRequ public void read(org.apache.thrift.protocol.TProtocol prot, PutFileMetadataRequest struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list794 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.fileIds = new ArrayList(_list794.size); - long _elem795; - for (int _i796 = 0; _i796 < _list794.size; ++_i796) + org.apache.thrift.protocol.TList _list818 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.fileIds = new ArrayList(_list818.size); + long _elem819; + for (int _i820 = 0; _i820 < _list818.size; ++_i820) { - _elem795 = iprot.readI64(); - struct.fileIds.add(_elem795); + _elem819 = iprot.readI64(); + struct.fileIds.add(_elem819); } } struct.setFileIdsIsSet(true); { - org.apache.thrift.protocol.TList _list797 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.metadata = new ArrayList(_list797.size); - ByteBuffer _elem798; - for (int _i799 = 0; _i799 < _list797.size; ++_i799) + org.apache.thrift.protocol.TList _list821 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.metadata = new ArrayList(_list821.size); + ByteBuffer _elem822; + for (int _i823 = 0; _i823 < _list821.size; ++_i823) { - _elem798 = iprot.readBinary(); - struct.metadata.add(_elem798); + _elem822 = iprot.readBinary(); + struct.metadata.add(_elem822); } } struct.setMetadataIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ReplTblWriteIdStateRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ReplTblWriteIdStateRequest.java index 97bb8a47ac..0aeca145db 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ReplTblWriteIdStateRequest.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ReplTblWriteIdStateRequest.java @@ -813,13 +813,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ReplTblWriteIdState case 6: // PART_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list594 = iprot.readListBegin(); - struct.partNames = new ArrayList(_list594.size); - String _elem595; - for (int _i596 = 0; _i596 < _list594.size; ++_i596) + org.apache.thrift.protocol.TList _list602 = iprot.readListBegin(); + struct.partNames = new ArrayList(_list602.size); + String _elem603; + for (int _i604 = 0; _i604 < _list602.size; ++_i604) { - _elem595 = iprot.readString(); - struct.partNames.add(_elem595); + _elem603 = iprot.readString(); + struct.partNames.add(_elem603); } iprot.readListEnd(); } @@ -871,9 +871,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ReplTblWriteIdStat oprot.writeFieldBegin(PART_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.partNames.size())); - for (String _iter597 : struct.partNames) + for (String _iter605 : struct.partNames) { - oprot.writeString(_iter597); + oprot.writeString(_iter605); } oprot.writeListEnd(); } @@ -910,9 +910,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ReplTblWriteIdState if (struct.isSetPartNames()) { { oprot.writeI32(struct.partNames.size()); - for (String _iter598 : struct.partNames) + for (String _iter606 : struct.partNames) { - oprot.writeString(_iter598); + oprot.writeString(_iter606); } } } @@ -934,13 +934,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, ReplTblWriteIdStateR BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list599 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.partNames = new ArrayList(_list599.size); - String _elem600; - for (int _i601 = 0; _i601 < _list599.size; ++_i601) + org.apache.thrift.protocol.TList _list607 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partNames = new ArrayList(_list607.size); + String _elem608; + for (int _i609 = 0; _i609 < _list607.size; ++_i609) { - _elem600 = iprot.readString(); - struct.partNames.add(_elem600); + _elem608 = iprot.readString(); + struct.partNames.add(_elem608); } } struct.setPartNamesIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/SchemaVersion.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/SchemaVersion.java index 62bc3b4bb4..935af04b35 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/SchemaVersion.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/SchemaVersion.java @@ -1119,14 +1119,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, SchemaVersion struc case 4: // COLS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list904 = iprot.readListBegin(); - struct.cols = new ArrayList(_list904.size); - FieldSchema _elem905; - for (int _i906 = 0; _i906 < _list904.size; ++_i906) + org.apache.thrift.protocol.TList _list928 = iprot.readListBegin(); + struct.cols = new ArrayList(_list928.size); + FieldSchema _elem929; + for (int _i930 = 0; _i930 < _list928.size; ++_i930) { - _elem905 = new FieldSchema(); - _elem905.read(iprot); - struct.cols.add(_elem905); + _elem929 = new FieldSchema(); + _elem929.read(iprot); + struct.cols.add(_elem929); } iprot.readListEnd(); } @@ -1212,9 +1212,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, SchemaVersion stru oprot.writeFieldBegin(COLS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.cols.size())); - for (FieldSchema _iter907 : struct.cols) + for (FieldSchema _iter931 : struct.cols) { - _iter907.write(oprot); + _iter931.write(oprot); } oprot.writeListEnd(); } @@ -1323,9 +1323,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, SchemaVersion struc if (struct.isSetCols()) { { oprot.writeI32(struct.cols.size()); - for (FieldSchema _iter908 : struct.cols) + for (FieldSchema _iter932 : struct.cols) { - _iter908.write(oprot); + _iter932.write(oprot); } } } @@ -1368,14 +1368,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, SchemaVersion struct } if (incoming.get(3)) { { - org.apache.thrift.protocol.TList _list909 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.cols = new ArrayList(_list909.size); - FieldSchema _elem910; - for (int _i911 = 0; _i911 < _list909.size; ++_i911) + org.apache.thrift.protocol.TList _list933 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.cols = new ArrayList(_list933.size); + FieldSchema _elem934; + for (int _i935 = 0; _i935 < _list933.size; ++_i935) { - _elem910 = new FieldSchema(); - _elem910.read(iprot); - struct.cols.add(_elem910); + _elem934 = new FieldSchema(); + _elem934.read(iprot); + struct.cols.add(_elem934); } } struct.setColsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponse.java index d7e5132ea9..9fb037f6b8 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponse.java @@ -354,14 +354,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ShowCompactResponse case 1: // COMPACTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list692 = iprot.readListBegin(); - struct.compacts = new ArrayList(_list692.size); - ShowCompactResponseElement _elem693; - for (int _i694 = 0; _i694 < _list692.size; ++_i694) + org.apache.thrift.protocol.TList _list700 = iprot.readListBegin(); + struct.compacts = new ArrayList(_list700.size); + ShowCompactResponseElement _elem701; + for (int _i702 = 0; _i702 < _list700.size; ++_i702) { - _elem693 = new ShowCompactResponseElement(); - _elem693.read(iprot); - struct.compacts.add(_elem693); + _elem701 = new ShowCompactResponseElement(); + _elem701.read(iprot); + struct.compacts.add(_elem701); } iprot.readListEnd(); } @@ -387,9 +387,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ShowCompactRespons oprot.writeFieldBegin(COMPACTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.compacts.size())); - for (ShowCompactResponseElement _iter695 : struct.compacts) + for (ShowCompactResponseElement _iter703 : struct.compacts) { - _iter695.write(oprot); + _iter703.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponse TTupleProtocol oprot = (TTupleProtocol) prot; { oprot.writeI32(struct.compacts.size()); - for (ShowCompactResponseElement _iter696 : struct.compacts) + for (ShowCompactResponseElement _iter704 : struct.compacts) { - _iter696.write(oprot); + _iter704.write(oprot); } } } @@ -425,14 +425,14 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponse public void read(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponse struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; { - org.apache.thrift.protocol.TList _list697 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.compacts = new ArrayList(_list697.size); - ShowCompactResponseElement _elem698; - for (int _i699 = 0; _i699 < _list697.size; ++_i699) + org.apache.thrift.protocol.TList _list705 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.compacts = new ArrayList(_list705.size); + ShowCompactResponseElement _elem706; + for (int _i707 = 0; _i707 < _list705.size; ++_i707) { - _elem698 = new ShowCompactResponseElement(); - _elem698.read(iprot); - struct.compacts.add(_elem698); + _elem706 = new ShowCompactResponseElement(); + _elem706.read(iprot); + struct.compacts.add(_elem706); } } struct.setCompactsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponse.java index 0e1009ce1a..e0db2f789c 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowLocksResponse.java @@ -350,14 +350,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ShowLocksResponse s case 1: // LOCKS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list658 = iprot.readListBegin(); - struct.locks = new ArrayList(_list658.size); - ShowLocksResponseElement _elem659; - for (int _i660 = 0; _i660 < _list658.size; ++_i660) + org.apache.thrift.protocol.TList _list666 = iprot.readListBegin(); + struct.locks = new ArrayList(_list666.size); + ShowLocksResponseElement _elem667; + for (int _i668 = 0; _i668 < _list666.size; ++_i668) { - _elem659 = new ShowLocksResponseElement(); - _elem659.read(iprot); - struct.locks.add(_elem659); + _elem667 = new ShowLocksResponseElement(); + _elem667.read(iprot); + struct.locks.add(_elem667); } iprot.readListEnd(); } @@ -383,9 +383,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ShowLocksResponse oprot.writeFieldBegin(LOCKS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.locks.size())); - for (ShowLocksResponseElement _iter661 : struct.locks) + for (ShowLocksResponseElement _iter669 : struct.locks) { - _iter661.write(oprot); + _iter669.write(oprot); } oprot.writeListEnd(); } @@ -416,9 +416,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowLocksResponse s if (struct.isSetLocks()) { { oprot.writeI32(struct.locks.size()); - for (ShowLocksResponseElement _iter662 : struct.locks) + for (ShowLocksResponseElement _iter670 : struct.locks) { - _iter662.write(oprot); + _iter670.write(oprot); } } } @@ -430,14 +430,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, ShowLocksResponse st BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list663 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.locks = new ArrayList(_list663.size); - ShowLocksResponseElement _elem664; - for (int _i665 = 0; _i665 < _list663.size; ++_i665) + org.apache.thrift.protocol.TList _list671 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.locks = new ArrayList(_list671.size); + ShowLocksResponseElement _elem672; + for (int _i673 = 0; _i673 < _list671.size; ++_i673) { - _elem664 = new ShowLocksResponseElement(); - _elem664.read(iprot); - struct.locks.add(_elem664); + _elem672 = new ShowLocksResponseElement(); + _elem672.read(iprot); + struct.locks.add(_elem672); } } struct.setLocksIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/TableValidWriteIds.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/TableValidWriteIds.java index 20f225d007..de15fc6be2 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/TableValidWriteIds.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/TableValidWriteIds.java @@ -708,13 +708,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, TableValidWriteIds case 3: // INVALID_WRITE_IDS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list610 = iprot.readListBegin(); - struct.invalidWriteIds = new ArrayList(_list610.size); - long _elem611; - for (int _i612 = 0; _i612 < _list610.size; ++_i612) + org.apache.thrift.protocol.TList _list618 = iprot.readListBegin(); + struct.invalidWriteIds = new ArrayList(_list618.size); + long _elem619; + for (int _i620 = 0; _i620 < _list618.size; ++_i620) { - _elem611 = iprot.readI64(); - struct.invalidWriteIds.add(_elem611); + _elem619 = iprot.readI64(); + struct.invalidWriteIds.add(_elem619); } iprot.readListEnd(); } @@ -764,9 +764,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, TableValidWriteIds oprot.writeFieldBegin(INVALID_WRITE_IDS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, struct.invalidWriteIds.size())); - for (long _iter613 : struct.invalidWriteIds) + for (long _iter621 : struct.invalidWriteIds) { - oprot.writeI64(_iter613); + oprot.writeI64(_iter621); } oprot.writeListEnd(); } @@ -803,9 +803,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, TableValidWriteIds oprot.writeI64(struct.writeIdHighWaterMark); { oprot.writeI32(struct.invalidWriteIds.size()); - for (long _iter614 : struct.invalidWriteIds) + for (long _iter622 : struct.invalidWriteIds) { - oprot.writeI64(_iter614); + oprot.writeI64(_iter622); } } oprot.writeBinary(struct.abortedBits); @@ -827,13 +827,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, TableValidWriteIds s struct.writeIdHighWaterMark = iprot.readI64(); struct.setWriteIdHighWaterMarkIsSet(true); { - org.apache.thrift.protocol.TList _list615 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); - struct.invalidWriteIds = new ArrayList(_list615.size); - long _elem616; - for (int _i617 = 0; _i617 < _list615.size; ++_i617) + org.apache.thrift.protocol.TList _list623 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.I64, iprot.readI32()); + struct.invalidWriteIds = new ArrayList(_list623.size); + long _elem624; + for (int _i625 = 0; _i625 < _list623.size; ++_i625) { - _elem616 = iprot.readI64(); - struct.invalidWriteIds.add(_elem616); + _elem624 = iprot.readI64(); + struct.invalidWriteIds.add(_elem624); } } struct.setInvalidWriteIdsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java index b7a9d587af..24ffadb00b 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java @@ -372,6 +372,8 @@ public void flushCache() throws org.apache.thrift.TException; + public WriteNotificationLogResponse add_write_notification_log(WriteNotificationLogRequest rqst) throws org.apache.thrift.TException; + public CmRecycleResponse cm_recycle(CmRecycleRequest request) throws MetaException, org.apache.thrift.TException; public GetFileMetadataByExprResult get_file_metadata_by_expr(GetFileMetadataByExprRequest req) throws org.apache.thrift.TException; @@ -790,6 +792,8 @@ public void flushCache(org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; + public void add_write_notification_log(WriteNotificationLogRequest rqst, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; + public void cm_recycle(CmRecycleRequest request, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; public void get_file_metadata_by_expr(GetFileMetadataByExprRequest req, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException; @@ -5644,6 +5648,29 @@ public void recv_flushCache() throws org.apache.thrift.TException return; } + public WriteNotificationLogResponse add_write_notification_log(WriteNotificationLogRequest rqst) throws org.apache.thrift.TException + { + send_add_write_notification_log(rqst); + return recv_add_write_notification_log(); + } + + public void send_add_write_notification_log(WriteNotificationLogRequest rqst) throws org.apache.thrift.TException + { + add_write_notification_log_args args = new add_write_notification_log_args(); + args.setRqst(rqst); + sendBase("add_write_notification_log", args); + } + + public WriteNotificationLogResponse recv_add_write_notification_log() throws org.apache.thrift.TException + { + add_write_notification_log_result result = new add_write_notification_log_result(); + receiveBase(result, "add_write_notification_log"); + if (result.isSetSuccess()) { + return result.success; + } + throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "add_write_notification_log failed: unknown result"); + } + public CmRecycleResponse cm_recycle(CmRecycleRequest request) throws MetaException, org.apache.thrift.TException { send_cm_recycle(request); @@ -12619,6 +12646,38 @@ public void getResult() throws org.apache.thrift.TException { } } + public void add_write_notification_log(WriteNotificationLogRequest rqst, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { + checkReady(); + add_write_notification_log_call method_call = new add_write_notification_log_call(rqst, resultHandler, this, ___protocolFactory, ___transport); + this.___currentMethod = method_call; + ___manager.call(method_call); + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_write_notification_log_call extends org.apache.thrift.async.TAsyncMethodCall { + private WriteNotificationLogRequest rqst; + public add_write_notification_log_call(WriteNotificationLogRequest rqst, org.apache.thrift.async.AsyncMethodCallback resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException { + super(client, protocolFactory, transport, resultHandler, false); + this.rqst = rqst; + } + + public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException { + prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("add_write_notification_log", org.apache.thrift.protocol.TMessageType.CALL, 0)); + add_write_notification_log_args args = new add_write_notification_log_args(); + args.setRqst(rqst); + args.write(prot); + prot.writeMessageEnd(); + } + + public WriteNotificationLogResponse getResult() throws org.apache.thrift.TException { + if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) { + throw new IllegalStateException("Method call not finished!"); + } + org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array()); + org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport); + return (new Client(prot)).recv_add_write_notification_log(); + } + } + public void cm_recycle(CmRecycleRequest request, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws org.apache.thrift.TException { checkReady(); cm_recycle_call method_call = new cm_recycle_call(request, resultHandler, this, ___protocolFactory, ___transport); @@ -14150,6 +14209,7 @@ protected Processor(I iface, Map extends org.apache.thrift.ProcessFunction { + public add_write_notification_log() { + super("add_write_notification_log"); + } + + public add_write_notification_log_args getEmptyArgsInstance() { + return new add_write_notification_log_args(); + } + + protected boolean isOneway() { + return false; + } + + public add_write_notification_log_result getResult(I iface, add_write_notification_log_args args) throws org.apache.thrift.TException { + add_write_notification_log_result result = new add_write_notification_log_result(); + result.success = iface.add_write_notification_log(args.rqst); + return result; + } + } + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class cm_recycle extends org.apache.thrift.ProcessFunction { public cm_recycle() { super("cm_recycle"); @@ -19683,6 +19763,7 @@ protected AsyncProcessor(I iface, Map extends org.apache.thrift.AsyncProcessFunction { + public add_write_notification_log() { + super("add_write_notification_log"); + } + + public add_write_notification_log_args getEmptyArgsInstance() { + return new add_write_notification_log_args(); + } + + public AsyncMethodCallback getResultHandler(final AsyncFrameBuffer fb, final int seqid) { + final org.apache.thrift.AsyncProcessFunction fcall = this; + return new AsyncMethodCallback() { + public void onComplete(WriteNotificationLogResponse o) { + add_write_notification_log_result result = new add_write_notification_log_result(); + result.success = o; + try { + fcall.sendResponse(fb,result, org.apache.thrift.protocol.TMessageType.REPLY,seqid); + return; + } catch (Exception e) { + LOGGER.error("Exception writing to internal frame buffer", e); + } + fb.close(); + } + public void onError(Exception e) { + byte msgType = org.apache.thrift.protocol.TMessageType.REPLY; + org.apache.thrift.TBase msg; + add_write_notification_log_result result = new add_write_notification_log_result(); + { + msgType = org.apache.thrift.protocol.TMessageType.EXCEPTION; + msg = (org.apache.thrift.TBase)new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.INTERNAL_ERROR, e.getMessage()); + } + try { + fcall.sendResponse(fb,msg,msgType,seqid); + return; + } catch (Exception ex) { + LOGGER.error("Exception writing to internal frame buffer", ex); + } + fb.close(); + } + }; + } + + protected boolean isOneway() { + return false; + } + + public void start(I iface, add_write_notification_log_args args, org.apache.thrift.async.AsyncMethodCallback resultHandler) throws TException { + iface.add_write_notification_log(args.rqst,resultHandler); + } + } + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class cm_recycle extends org.apache.thrift.AsyncProcessFunction { public cm_recycle() { super("cm_recycle"); @@ -42252,13 +42384,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_databases_resul case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list920 = iprot.readListBegin(); - struct.success = new ArrayList(_list920.size); - String _elem921; - for (int _i922 = 0; _i922 < _list920.size; ++_i922) + org.apache.thrift.protocol.TList _list944 = iprot.readListBegin(); + struct.success = new ArrayList(_list944.size); + String _elem945; + for (int _i946 = 0; _i946 < _list944.size; ++_i946) { - _elem921 = iprot.readString(); - struct.success.add(_elem921); + _elem945 = iprot.readString(); + struct.success.add(_elem945); } iprot.readListEnd(); } @@ -42293,9 +42425,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_databases_resu oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter923 : struct.success) + for (String _iter947 : struct.success) { - oprot.writeString(_iter923); + oprot.writeString(_iter947); } oprot.writeListEnd(); } @@ -42334,9 +42466,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_databases_resul if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter924 : struct.success) + for (String _iter948 : struct.success) { - oprot.writeString(_iter924); + oprot.writeString(_iter948); } } } @@ -42351,13 +42483,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_databases_result BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list925 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list925.size); - String _elem926; - for (int _i927 = 0; _i927 < _list925.size; ++_i927) + org.apache.thrift.protocol.TList _list949 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list949.size); + String _elem950; + for (int _i951 = 0; _i951 < _list949.size; ++_i951) { - _elem926 = iprot.readString(); - struct.success.add(_elem926); + _elem950 = iprot.readString(); + struct.success.add(_elem950); } } struct.setSuccessIsSet(true); @@ -43011,13 +43143,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_all_databases_r case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list928 = iprot.readListBegin(); - struct.success = new ArrayList(_list928.size); - String _elem929; - for (int _i930 = 0; _i930 < _list928.size; ++_i930) + org.apache.thrift.protocol.TList _list952 = iprot.readListBegin(); + struct.success = new ArrayList(_list952.size); + String _elem953; + for (int _i954 = 0; _i954 < _list952.size; ++_i954) { - _elem929 = iprot.readString(); - struct.success.add(_elem929); + _elem953 = iprot.readString(); + struct.success.add(_elem953); } iprot.readListEnd(); } @@ -43052,9 +43184,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_all_databases_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter931 : struct.success) + for (String _iter955 : struct.success) { - oprot.writeString(_iter931); + oprot.writeString(_iter955); } oprot.writeListEnd(); } @@ -43093,9 +43225,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_all_databases_r if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter932 : struct.success) + for (String _iter956 : struct.success) { - oprot.writeString(_iter932); + oprot.writeString(_iter956); } } } @@ -43110,13 +43242,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_all_databases_re BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list933 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list933.size); - String _elem934; - for (int _i935 = 0; _i935 < _list933.size; ++_i935) + org.apache.thrift.protocol.TList _list957 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list957.size); + String _elem958; + for (int _i959 = 0; _i959 < _list957.size; ++_i959) { - _elem934 = iprot.readString(); - struct.success.add(_elem934); + _elem958 = iprot.readString(); + struct.success.add(_elem958); } } struct.setSuccessIsSet(true); @@ -47723,16 +47855,16 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_type_all_result case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map936 = iprot.readMapBegin(); - struct.success = new HashMap(2*_map936.size); - String _key937; - Type _val938; - for (int _i939 = 0; _i939 < _map936.size; ++_i939) + org.apache.thrift.protocol.TMap _map960 = iprot.readMapBegin(); + struct.success = new HashMap(2*_map960.size); + String _key961; + Type _val962; + for (int _i963 = 0; _i963 < _map960.size; ++_i963) { - _key937 = iprot.readString(); - _val938 = new Type(); - _val938.read(iprot); - struct.success.put(_key937, _val938); + _key961 = iprot.readString(); + _val962 = new Type(); + _val962.read(iprot); + struct.success.put(_key961, _val962); } iprot.readMapEnd(); } @@ -47767,10 +47899,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_type_all_resul oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Map.Entry _iter940 : struct.success.entrySet()) + for (Map.Entry _iter964 : struct.success.entrySet()) { - oprot.writeString(_iter940.getKey()); - _iter940.getValue().write(oprot); + oprot.writeString(_iter964.getKey()); + _iter964.getValue().write(oprot); } oprot.writeMapEnd(); } @@ -47809,10 +47941,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_type_all_result if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Map.Entry _iter941 : struct.success.entrySet()) + for (Map.Entry _iter965 : struct.success.entrySet()) { - oprot.writeString(_iter941.getKey()); - _iter941.getValue().write(oprot); + oprot.writeString(_iter965.getKey()); + _iter965.getValue().write(oprot); } } } @@ -47827,16 +47959,16 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_type_all_result BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TMap _map942 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new HashMap(2*_map942.size); - String _key943; - Type _val944; - for (int _i945 = 0; _i945 < _map942.size; ++_i945) + org.apache.thrift.protocol.TMap _map966 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new HashMap(2*_map966.size); + String _key967; + Type _val968; + for (int _i969 = 0; _i969 < _map966.size; ++_i969) { - _key943 = iprot.readString(); - _val944 = new Type(); - _val944.read(iprot); - struct.success.put(_key943, _val944); + _key967 = iprot.readString(); + _val968 = new Type(); + _val968.read(iprot); + struct.success.put(_key967, _val968); } } struct.setSuccessIsSet(true); @@ -48871,14 +49003,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_fields_result s case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list946 = iprot.readListBegin(); - struct.success = new ArrayList(_list946.size); - FieldSchema _elem947; - for (int _i948 = 0; _i948 < _list946.size; ++_i948) + org.apache.thrift.protocol.TList _list970 = iprot.readListBegin(); + struct.success = new ArrayList(_list970.size); + FieldSchema _elem971; + for (int _i972 = 0; _i972 < _list970.size; ++_i972) { - _elem947 = new FieldSchema(); - _elem947.read(iprot); - struct.success.add(_elem947); + _elem971 = new FieldSchema(); + _elem971.read(iprot); + struct.success.add(_elem971); } iprot.readListEnd(); } @@ -48931,9 +49063,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_fields_result oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (FieldSchema _iter949 : struct.success) + for (FieldSchema _iter973 : struct.success) { - _iter949.write(oprot); + _iter973.write(oprot); } oprot.writeListEnd(); } @@ -48988,9 +49120,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_fields_result s if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (FieldSchema _iter950 : struct.success) + for (FieldSchema _iter974 : struct.success) { - _iter950.write(oprot); + _iter974.write(oprot); } } } @@ -49011,14 +49143,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_fields_result st BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list951 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list951.size); - FieldSchema _elem952; - for (int _i953 = 0; _i953 < _list951.size; ++_i953) + org.apache.thrift.protocol.TList _list975 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list975.size); + FieldSchema _elem976; + for (int _i977 = 0; _i977 < _list975.size; ++_i977) { - _elem952 = new FieldSchema(); - _elem952.read(iprot); - struct.success.add(_elem952); + _elem976 = new FieldSchema(); + _elem976.read(iprot); + struct.success.add(_elem976); } } struct.setSuccessIsSet(true); @@ -50172,14 +50304,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_fields_with_env case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list954 = iprot.readListBegin(); - struct.success = new ArrayList(_list954.size); - FieldSchema _elem955; - for (int _i956 = 0; _i956 < _list954.size; ++_i956) + org.apache.thrift.protocol.TList _list978 = iprot.readListBegin(); + struct.success = new ArrayList(_list978.size); + FieldSchema _elem979; + for (int _i980 = 0; _i980 < _list978.size; ++_i980) { - _elem955 = new FieldSchema(); - _elem955.read(iprot); - struct.success.add(_elem955); + _elem979 = new FieldSchema(); + _elem979.read(iprot); + struct.success.add(_elem979); } iprot.readListEnd(); } @@ -50232,9 +50364,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_fields_with_en oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (FieldSchema _iter957 : struct.success) + for (FieldSchema _iter981 : struct.success) { - _iter957.write(oprot); + _iter981.write(oprot); } oprot.writeListEnd(); } @@ -50289,9 +50421,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_fields_with_env if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (FieldSchema _iter958 : struct.success) + for (FieldSchema _iter982 : struct.success) { - _iter958.write(oprot); + _iter982.write(oprot); } } } @@ -50312,14 +50444,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_fields_with_envi BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list959 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list959.size); - FieldSchema _elem960; - for (int _i961 = 0; _i961 < _list959.size; ++_i961) + org.apache.thrift.protocol.TList _list983 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list983.size); + FieldSchema _elem984; + for (int _i985 = 0; _i985 < _list983.size; ++_i985) { - _elem960 = new FieldSchema(); - _elem960.read(iprot); - struct.success.add(_elem960); + _elem984 = new FieldSchema(); + _elem984.read(iprot); + struct.success.add(_elem984); } } struct.setSuccessIsSet(true); @@ -51364,14 +51496,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_schema_result s case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list962 = iprot.readListBegin(); - struct.success = new ArrayList(_list962.size); - FieldSchema _elem963; - for (int _i964 = 0; _i964 < _list962.size; ++_i964) + org.apache.thrift.protocol.TList _list986 = iprot.readListBegin(); + struct.success = new ArrayList(_list986.size); + FieldSchema _elem987; + for (int _i988 = 0; _i988 < _list986.size; ++_i988) { - _elem963 = new FieldSchema(); - _elem963.read(iprot); - struct.success.add(_elem963); + _elem987 = new FieldSchema(); + _elem987.read(iprot); + struct.success.add(_elem987); } iprot.readListEnd(); } @@ -51424,9 +51556,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_schema_result oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (FieldSchema _iter965 : struct.success) + for (FieldSchema _iter989 : struct.success) { - _iter965.write(oprot); + _iter989.write(oprot); } oprot.writeListEnd(); } @@ -51481,9 +51613,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_schema_result s if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (FieldSchema _iter966 : struct.success) + for (FieldSchema _iter990 : struct.success) { - _iter966.write(oprot); + _iter990.write(oprot); } } } @@ -51504,14 +51636,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_schema_result st BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list967 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list967.size); - FieldSchema _elem968; - for (int _i969 = 0; _i969 < _list967.size; ++_i969) + org.apache.thrift.protocol.TList _list991 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list991.size); + FieldSchema _elem992; + for (int _i993 = 0; _i993 < _list991.size; ++_i993) { - _elem968 = new FieldSchema(); - _elem968.read(iprot); - struct.success.add(_elem968); + _elem992 = new FieldSchema(); + _elem992.read(iprot); + struct.success.add(_elem992); } } struct.setSuccessIsSet(true); @@ -52665,14 +52797,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_schema_with_env case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list970 = iprot.readListBegin(); - struct.success = new ArrayList(_list970.size); - FieldSchema _elem971; - for (int _i972 = 0; _i972 < _list970.size; ++_i972) + org.apache.thrift.protocol.TList _list994 = iprot.readListBegin(); + struct.success = new ArrayList(_list994.size); + FieldSchema _elem995; + for (int _i996 = 0; _i996 < _list994.size; ++_i996) { - _elem971 = new FieldSchema(); - _elem971.read(iprot); - struct.success.add(_elem971); + _elem995 = new FieldSchema(); + _elem995.read(iprot); + struct.success.add(_elem995); } iprot.readListEnd(); } @@ -52725,9 +52857,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_schema_with_en oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (FieldSchema _iter973 : struct.success) + for (FieldSchema _iter997 : struct.success) { - _iter973.write(oprot); + _iter997.write(oprot); } oprot.writeListEnd(); } @@ -52782,9 +52914,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_schema_with_env if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (FieldSchema _iter974 : struct.success) + for (FieldSchema _iter998 : struct.success) { - _iter974.write(oprot); + _iter998.write(oprot); } } } @@ -52805,14 +52937,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_schema_with_envi BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list975 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list975.size); - FieldSchema _elem976; - for (int _i977 = 0; _i977 < _list975.size; ++_i977) + org.apache.thrift.protocol.TList _list999 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list999.size); + FieldSchema _elem1000; + for (int _i1001 = 0; _i1001 < _list999.size; ++_i1001) { - _elem976 = new FieldSchema(); - _elem976.read(iprot); - struct.success.add(_elem976); + _elem1000 = new FieldSchema(); + _elem1000.read(iprot); + struct.success.add(_elem1000); } } struct.setSuccessIsSet(true); @@ -55941,14 +56073,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, create_table_with_c case 2: // PRIMARY_KEYS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list978 = iprot.readListBegin(); - struct.primaryKeys = new ArrayList(_list978.size); - SQLPrimaryKey _elem979; - for (int _i980 = 0; _i980 < _list978.size; ++_i980) + org.apache.thrift.protocol.TList _list1002 = iprot.readListBegin(); + struct.primaryKeys = new ArrayList(_list1002.size); + SQLPrimaryKey _elem1003; + for (int _i1004 = 0; _i1004 < _list1002.size; ++_i1004) { - _elem979 = new SQLPrimaryKey(); - _elem979.read(iprot); - struct.primaryKeys.add(_elem979); + _elem1003 = new SQLPrimaryKey(); + _elem1003.read(iprot); + struct.primaryKeys.add(_elem1003); } iprot.readListEnd(); } @@ -55960,14 +56092,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, create_table_with_c case 3: // FOREIGN_KEYS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list981 = iprot.readListBegin(); - struct.foreignKeys = new ArrayList(_list981.size); - SQLForeignKey _elem982; - for (int _i983 = 0; _i983 < _list981.size; ++_i983) + org.apache.thrift.protocol.TList _list1005 = iprot.readListBegin(); + struct.foreignKeys = new ArrayList(_list1005.size); + SQLForeignKey _elem1006; + for (int _i1007 = 0; _i1007 < _list1005.size; ++_i1007) { - _elem982 = new SQLForeignKey(); - _elem982.read(iprot); - struct.foreignKeys.add(_elem982); + _elem1006 = new SQLForeignKey(); + _elem1006.read(iprot); + struct.foreignKeys.add(_elem1006); } iprot.readListEnd(); } @@ -55979,14 +56111,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, create_table_with_c case 4: // UNIQUE_CONSTRAINTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list984 = iprot.readListBegin(); - struct.uniqueConstraints = new ArrayList(_list984.size); - SQLUniqueConstraint _elem985; - for (int _i986 = 0; _i986 < _list984.size; ++_i986) + org.apache.thrift.protocol.TList _list1008 = iprot.readListBegin(); + struct.uniqueConstraints = new ArrayList(_list1008.size); + SQLUniqueConstraint _elem1009; + for (int _i1010 = 0; _i1010 < _list1008.size; ++_i1010) { - _elem985 = new SQLUniqueConstraint(); - _elem985.read(iprot); - struct.uniqueConstraints.add(_elem985); + _elem1009 = new SQLUniqueConstraint(); + _elem1009.read(iprot); + struct.uniqueConstraints.add(_elem1009); } iprot.readListEnd(); } @@ -55998,14 +56130,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, create_table_with_c case 5: // NOT_NULL_CONSTRAINTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list987 = iprot.readListBegin(); - struct.notNullConstraints = new ArrayList(_list987.size); - SQLNotNullConstraint _elem988; - for (int _i989 = 0; _i989 < _list987.size; ++_i989) + org.apache.thrift.protocol.TList _list1011 = iprot.readListBegin(); + struct.notNullConstraints = new ArrayList(_list1011.size); + SQLNotNullConstraint _elem1012; + for (int _i1013 = 0; _i1013 < _list1011.size; ++_i1013) { - _elem988 = new SQLNotNullConstraint(); - _elem988.read(iprot); - struct.notNullConstraints.add(_elem988); + _elem1012 = new SQLNotNullConstraint(); + _elem1012.read(iprot); + struct.notNullConstraints.add(_elem1012); } iprot.readListEnd(); } @@ -56017,14 +56149,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, create_table_with_c case 6: // DEFAULT_CONSTRAINTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list990 = iprot.readListBegin(); - struct.defaultConstraints = new ArrayList(_list990.size); - SQLDefaultConstraint _elem991; - for (int _i992 = 0; _i992 < _list990.size; ++_i992) + org.apache.thrift.protocol.TList _list1014 = iprot.readListBegin(); + struct.defaultConstraints = new ArrayList(_list1014.size); + SQLDefaultConstraint _elem1015; + for (int _i1016 = 0; _i1016 < _list1014.size; ++_i1016) { - _elem991 = new SQLDefaultConstraint(); - _elem991.read(iprot); - struct.defaultConstraints.add(_elem991); + _elem1015 = new SQLDefaultConstraint(); + _elem1015.read(iprot); + struct.defaultConstraints.add(_elem1015); } iprot.readListEnd(); } @@ -56036,14 +56168,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, create_table_with_c case 7: // CHECK_CONSTRAINTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list993 = iprot.readListBegin(); - struct.checkConstraints = new ArrayList(_list993.size); - SQLCheckConstraint _elem994; - for (int _i995 = 0; _i995 < _list993.size; ++_i995) + org.apache.thrift.protocol.TList _list1017 = iprot.readListBegin(); + struct.checkConstraints = new ArrayList(_list1017.size); + SQLCheckConstraint _elem1018; + for (int _i1019 = 0; _i1019 < _list1017.size; ++_i1019) { - _elem994 = new SQLCheckConstraint(); - _elem994.read(iprot); - struct.checkConstraints.add(_elem994); + _elem1018 = new SQLCheckConstraint(); + _elem1018.read(iprot); + struct.checkConstraints.add(_elem1018); } iprot.readListEnd(); } @@ -56074,9 +56206,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, create_table_with_ oprot.writeFieldBegin(PRIMARY_KEYS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.primaryKeys.size())); - for (SQLPrimaryKey _iter996 : struct.primaryKeys) + for (SQLPrimaryKey _iter1020 : struct.primaryKeys) { - _iter996.write(oprot); + _iter1020.write(oprot); } oprot.writeListEnd(); } @@ -56086,9 +56218,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, create_table_with_ oprot.writeFieldBegin(FOREIGN_KEYS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.foreignKeys.size())); - for (SQLForeignKey _iter997 : struct.foreignKeys) + for (SQLForeignKey _iter1021 : struct.foreignKeys) { - _iter997.write(oprot); + _iter1021.write(oprot); } oprot.writeListEnd(); } @@ -56098,9 +56230,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, create_table_with_ oprot.writeFieldBegin(UNIQUE_CONSTRAINTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.uniqueConstraints.size())); - for (SQLUniqueConstraint _iter998 : struct.uniqueConstraints) + for (SQLUniqueConstraint _iter1022 : struct.uniqueConstraints) { - _iter998.write(oprot); + _iter1022.write(oprot); } oprot.writeListEnd(); } @@ -56110,9 +56242,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, create_table_with_ oprot.writeFieldBegin(NOT_NULL_CONSTRAINTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.notNullConstraints.size())); - for (SQLNotNullConstraint _iter999 : struct.notNullConstraints) + for (SQLNotNullConstraint _iter1023 : struct.notNullConstraints) { - _iter999.write(oprot); + _iter1023.write(oprot); } oprot.writeListEnd(); } @@ -56122,9 +56254,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, create_table_with_ oprot.writeFieldBegin(DEFAULT_CONSTRAINTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.defaultConstraints.size())); - for (SQLDefaultConstraint _iter1000 : struct.defaultConstraints) + for (SQLDefaultConstraint _iter1024 : struct.defaultConstraints) { - _iter1000.write(oprot); + _iter1024.write(oprot); } oprot.writeListEnd(); } @@ -56134,9 +56266,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, create_table_with_ oprot.writeFieldBegin(CHECK_CONSTRAINTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.checkConstraints.size())); - for (SQLCheckConstraint _iter1001 : struct.checkConstraints) + for (SQLCheckConstraint _iter1025 : struct.checkConstraints) { - _iter1001.write(oprot); + _iter1025.write(oprot); } oprot.writeListEnd(); } @@ -56188,54 +56320,54 @@ public void write(org.apache.thrift.protocol.TProtocol prot, create_table_with_c if (struct.isSetPrimaryKeys()) { { oprot.writeI32(struct.primaryKeys.size()); - for (SQLPrimaryKey _iter1002 : struct.primaryKeys) + for (SQLPrimaryKey _iter1026 : struct.primaryKeys) { - _iter1002.write(oprot); + _iter1026.write(oprot); } } } if (struct.isSetForeignKeys()) { { oprot.writeI32(struct.foreignKeys.size()); - for (SQLForeignKey _iter1003 : struct.foreignKeys) + for (SQLForeignKey _iter1027 : struct.foreignKeys) { - _iter1003.write(oprot); + _iter1027.write(oprot); } } } if (struct.isSetUniqueConstraints()) { { oprot.writeI32(struct.uniqueConstraints.size()); - for (SQLUniqueConstraint _iter1004 : struct.uniqueConstraints) + for (SQLUniqueConstraint _iter1028 : struct.uniqueConstraints) { - _iter1004.write(oprot); + _iter1028.write(oprot); } } } if (struct.isSetNotNullConstraints()) { { oprot.writeI32(struct.notNullConstraints.size()); - for (SQLNotNullConstraint _iter1005 : struct.notNullConstraints) + for (SQLNotNullConstraint _iter1029 : struct.notNullConstraints) { - _iter1005.write(oprot); + _iter1029.write(oprot); } } } if (struct.isSetDefaultConstraints()) { { oprot.writeI32(struct.defaultConstraints.size()); - for (SQLDefaultConstraint _iter1006 : struct.defaultConstraints) + for (SQLDefaultConstraint _iter1030 : struct.defaultConstraints) { - _iter1006.write(oprot); + _iter1030.write(oprot); } } } if (struct.isSetCheckConstraints()) { { oprot.writeI32(struct.checkConstraints.size()); - for (SQLCheckConstraint _iter1007 : struct.checkConstraints) + for (SQLCheckConstraint _iter1031 : struct.checkConstraints) { - _iter1007.write(oprot); + _iter1031.write(oprot); } } } @@ -56252,84 +56384,84 @@ public void read(org.apache.thrift.protocol.TProtocol prot, create_table_with_co } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list1008 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.primaryKeys = new ArrayList(_list1008.size); - SQLPrimaryKey _elem1009; - for (int _i1010 = 0; _i1010 < _list1008.size; ++_i1010) + org.apache.thrift.protocol.TList _list1032 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.primaryKeys = new ArrayList(_list1032.size); + SQLPrimaryKey _elem1033; + for (int _i1034 = 0; _i1034 < _list1032.size; ++_i1034) { - _elem1009 = new SQLPrimaryKey(); - _elem1009.read(iprot); - struct.primaryKeys.add(_elem1009); + _elem1033 = new SQLPrimaryKey(); + _elem1033.read(iprot); + struct.primaryKeys.add(_elem1033); } } struct.setPrimaryKeysIsSet(true); } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1011 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.foreignKeys = new ArrayList(_list1011.size); - SQLForeignKey _elem1012; - for (int _i1013 = 0; _i1013 < _list1011.size; ++_i1013) + org.apache.thrift.protocol.TList _list1035 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.foreignKeys = new ArrayList(_list1035.size); + SQLForeignKey _elem1036; + for (int _i1037 = 0; _i1037 < _list1035.size; ++_i1037) { - _elem1012 = new SQLForeignKey(); - _elem1012.read(iprot); - struct.foreignKeys.add(_elem1012); + _elem1036 = new SQLForeignKey(); + _elem1036.read(iprot); + struct.foreignKeys.add(_elem1036); } } struct.setForeignKeysIsSet(true); } if (incoming.get(3)) { { - org.apache.thrift.protocol.TList _list1014 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.uniqueConstraints = new ArrayList(_list1014.size); - SQLUniqueConstraint _elem1015; - for (int _i1016 = 0; _i1016 < _list1014.size; ++_i1016) + org.apache.thrift.protocol.TList _list1038 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.uniqueConstraints = new ArrayList(_list1038.size); + SQLUniqueConstraint _elem1039; + for (int _i1040 = 0; _i1040 < _list1038.size; ++_i1040) { - _elem1015 = new SQLUniqueConstraint(); - _elem1015.read(iprot); - struct.uniqueConstraints.add(_elem1015); + _elem1039 = new SQLUniqueConstraint(); + _elem1039.read(iprot); + struct.uniqueConstraints.add(_elem1039); } } struct.setUniqueConstraintsIsSet(true); } if (incoming.get(4)) { { - org.apache.thrift.protocol.TList _list1017 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.notNullConstraints = new ArrayList(_list1017.size); - SQLNotNullConstraint _elem1018; - for (int _i1019 = 0; _i1019 < _list1017.size; ++_i1019) + org.apache.thrift.protocol.TList _list1041 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.notNullConstraints = new ArrayList(_list1041.size); + SQLNotNullConstraint _elem1042; + for (int _i1043 = 0; _i1043 < _list1041.size; ++_i1043) { - _elem1018 = new SQLNotNullConstraint(); - _elem1018.read(iprot); - struct.notNullConstraints.add(_elem1018); + _elem1042 = new SQLNotNullConstraint(); + _elem1042.read(iprot); + struct.notNullConstraints.add(_elem1042); } } struct.setNotNullConstraintsIsSet(true); } if (incoming.get(5)) { { - org.apache.thrift.protocol.TList _list1020 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.defaultConstraints = new ArrayList(_list1020.size); - SQLDefaultConstraint _elem1021; - for (int _i1022 = 0; _i1022 < _list1020.size; ++_i1022) + org.apache.thrift.protocol.TList _list1044 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.defaultConstraints = new ArrayList(_list1044.size); + SQLDefaultConstraint _elem1045; + for (int _i1046 = 0; _i1046 < _list1044.size; ++_i1046) { - _elem1021 = new SQLDefaultConstraint(); - _elem1021.read(iprot); - struct.defaultConstraints.add(_elem1021); + _elem1045 = new SQLDefaultConstraint(); + _elem1045.read(iprot); + struct.defaultConstraints.add(_elem1045); } } struct.setDefaultConstraintsIsSet(true); } if (incoming.get(6)) { { - org.apache.thrift.protocol.TList _list1023 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.checkConstraints = new ArrayList(_list1023.size); - SQLCheckConstraint _elem1024; - for (int _i1025 = 0; _i1025 < _list1023.size; ++_i1025) + org.apache.thrift.protocol.TList _list1047 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.checkConstraints = new ArrayList(_list1047.size); + SQLCheckConstraint _elem1048; + for (int _i1049 = 0; _i1049 < _list1047.size; ++_i1049) { - _elem1024 = new SQLCheckConstraint(); - _elem1024.read(iprot); - struct.checkConstraints.add(_elem1024); + _elem1048 = new SQLCheckConstraint(); + _elem1048.read(iprot); + struct.checkConstraints.add(_elem1048); } } struct.setCheckConstraintsIsSet(true); @@ -65479,13 +65611,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, truncate_table_args case 3: // PART_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1026 = iprot.readListBegin(); - struct.partNames = new ArrayList(_list1026.size); - String _elem1027; - for (int _i1028 = 0; _i1028 < _list1026.size; ++_i1028) + org.apache.thrift.protocol.TList _list1050 = iprot.readListBegin(); + struct.partNames = new ArrayList(_list1050.size); + String _elem1051; + for (int _i1052 = 0; _i1052 < _list1050.size; ++_i1052) { - _elem1027 = iprot.readString(); - struct.partNames.add(_elem1027); + _elem1051 = iprot.readString(); + struct.partNames.add(_elem1051); } iprot.readListEnd(); } @@ -65521,9 +65653,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, truncate_table_arg oprot.writeFieldBegin(PART_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.partNames.size())); - for (String _iter1029 : struct.partNames) + for (String _iter1053 : struct.partNames) { - oprot.writeString(_iter1029); + oprot.writeString(_iter1053); } oprot.writeListEnd(); } @@ -65566,9 +65698,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, truncate_table_args if (struct.isSetPartNames()) { { oprot.writeI32(struct.partNames.size()); - for (String _iter1030 : struct.partNames) + for (String _iter1054 : struct.partNames) { - oprot.writeString(_iter1030); + oprot.writeString(_iter1054); } } } @@ -65588,13 +65720,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, truncate_table_args } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1031 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.partNames = new ArrayList(_list1031.size); - String _elem1032; - for (int _i1033 = 0; _i1033 < _list1031.size; ++_i1033) + org.apache.thrift.protocol.TList _list1055 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partNames = new ArrayList(_list1055.size); + String _elem1056; + for (int _i1057 = 0; _i1057 < _list1055.size; ++_i1057) { - _elem1032 = iprot.readString(); - struct.partNames.add(_elem1032); + _elem1056 = iprot.readString(); + struct.partNames.add(_elem1056); } } struct.setPartNamesIsSet(true); @@ -66819,13 +66951,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_tables_result s case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1034 = iprot.readListBegin(); - struct.success = new ArrayList(_list1034.size); - String _elem1035; - for (int _i1036 = 0; _i1036 < _list1034.size; ++_i1036) + org.apache.thrift.protocol.TList _list1058 = iprot.readListBegin(); + struct.success = new ArrayList(_list1058.size); + String _elem1059; + for (int _i1060 = 0; _i1060 < _list1058.size; ++_i1060) { - _elem1035 = iprot.readString(); - struct.success.add(_elem1035); + _elem1059 = iprot.readString(); + struct.success.add(_elem1059); } iprot.readListEnd(); } @@ -66860,9 +66992,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_tables_result oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1037 : struct.success) + for (String _iter1061 : struct.success) { - oprot.writeString(_iter1037); + oprot.writeString(_iter1061); } oprot.writeListEnd(); } @@ -66901,9 +67033,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_tables_result s if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1038 : struct.success) + for (String _iter1062 : struct.success) { - oprot.writeString(_iter1038); + oprot.writeString(_iter1062); } } } @@ -66918,13 +67050,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_tables_result st BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1039 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1039.size); - String _elem1040; - for (int _i1041 = 0; _i1041 < _list1039.size; ++_i1041) + org.apache.thrift.protocol.TList _list1063 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1063.size); + String _elem1064; + for (int _i1065 = 0; _i1065 < _list1063.size; ++_i1065) { - _elem1040 = iprot.readString(); - struct.success.add(_elem1040); + _elem1064 = iprot.readString(); + struct.success.add(_elem1064); } } struct.setSuccessIsSet(true); @@ -67898,13 +68030,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_tables_by_type_ case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1042 = iprot.readListBegin(); - struct.success = new ArrayList(_list1042.size); - String _elem1043; - for (int _i1044 = 0; _i1044 < _list1042.size; ++_i1044) + org.apache.thrift.protocol.TList _list1066 = iprot.readListBegin(); + struct.success = new ArrayList(_list1066.size); + String _elem1067; + for (int _i1068 = 0; _i1068 < _list1066.size; ++_i1068) { - _elem1043 = iprot.readString(); - struct.success.add(_elem1043); + _elem1067 = iprot.readString(); + struct.success.add(_elem1067); } iprot.readListEnd(); } @@ -67939,9 +68071,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_tables_by_type oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1045 : struct.success) + for (String _iter1069 : struct.success) { - oprot.writeString(_iter1045); + oprot.writeString(_iter1069); } oprot.writeListEnd(); } @@ -67980,9 +68112,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_tables_by_type_ if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1046 : struct.success) + for (String _iter1070 : struct.success) { - oprot.writeString(_iter1046); + oprot.writeString(_iter1070); } } } @@ -67997,13 +68129,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_tables_by_type_r BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1047 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1047.size); - String _elem1048; - for (int _i1049 = 0; _i1049 < _list1047.size; ++_i1049) + org.apache.thrift.protocol.TList _list1071 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1071.size); + String _elem1072; + for (int _i1073 = 0; _i1073 < _list1071.size; ++_i1073) { - _elem1048 = iprot.readString(); - struct.success.add(_elem1048); + _elem1072 = iprot.readString(); + struct.success.add(_elem1072); } } struct.setSuccessIsSet(true); @@ -68769,13 +68901,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_materialized_vi case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1050 = iprot.readListBegin(); - struct.success = new ArrayList(_list1050.size); - String _elem1051; - for (int _i1052 = 0; _i1052 < _list1050.size; ++_i1052) + org.apache.thrift.protocol.TList _list1074 = iprot.readListBegin(); + struct.success = new ArrayList(_list1074.size); + String _elem1075; + for (int _i1076 = 0; _i1076 < _list1074.size; ++_i1076) { - _elem1051 = iprot.readString(); - struct.success.add(_elem1051); + _elem1075 = iprot.readString(); + struct.success.add(_elem1075); } iprot.readListEnd(); } @@ -68810,9 +68942,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_materialized_v oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1053 : struct.success) + for (String _iter1077 : struct.success) { - oprot.writeString(_iter1053); + oprot.writeString(_iter1077); } oprot.writeListEnd(); } @@ -68851,9 +68983,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_materialized_vi if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1054 : struct.success) + for (String _iter1078 : struct.success) { - oprot.writeString(_iter1054); + oprot.writeString(_iter1078); } } } @@ -68868,13 +69000,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_materialized_vie BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1055 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1055.size); - String _elem1056; - for (int _i1057 = 0; _i1057 < _list1055.size; ++_i1057) + org.apache.thrift.protocol.TList _list1079 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1079.size); + String _elem1080; + for (int _i1081 = 0; _i1081 < _list1079.size; ++_i1081) { - _elem1056 = iprot.readString(); - struct.success.add(_elem1056); + _elem1080 = iprot.readString(); + struct.success.add(_elem1080); } } struct.setSuccessIsSet(true); @@ -69379,13 +69511,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_table_meta_args case 3: // TBL_TYPES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1058 = iprot.readListBegin(); - struct.tbl_types = new ArrayList(_list1058.size); - String _elem1059; - for (int _i1060 = 0; _i1060 < _list1058.size; ++_i1060) + org.apache.thrift.protocol.TList _list1082 = iprot.readListBegin(); + struct.tbl_types = new ArrayList(_list1082.size); + String _elem1083; + for (int _i1084 = 0; _i1084 < _list1082.size; ++_i1084) { - _elem1059 = iprot.readString(); - struct.tbl_types.add(_elem1059); + _elem1083 = iprot.readString(); + struct.tbl_types.add(_elem1083); } iprot.readListEnd(); } @@ -69421,9 +69553,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_table_meta_arg oprot.writeFieldBegin(TBL_TYPES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.tbl_types.size())); - for (String _iter1061 : struct.tbl_types) + for (String _iter1085 : struct.tbl_types) { - oprot.writeString(_iter1061); + oprot.writeString(_iter1085); } oprot.writeListEnd(); } @@ -69466,9 +69598,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_table_meta_args if (struct.isSetTbl_types()) { { oprot.writeI32(struct.tbl_types.size()); - for (String _iter1062 : struct.tbl_types) + for (String _iter1086 : struct.tbl_types) { - oprot.writeString(_iter1062); + oprot.writeString(_iter1086); } } } @@ -69488,13 +69620,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_table_meta_args } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1063 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.tbl_types = new ArrayList(_list1063.size); - String _elem1064; - for (int _i1065 = 0; _i1065 < _list1063.size; ++_i1065) + org.apache.thrift.protocol.TList _list1087 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.tbl_types = new ArrayList(_list1087.size); + String _elem1088; + for (int _i1089 = 0; _i1089 < _list1087.size; ++_i1089) { - _elem1064 = iprot.readString(); - struct.tbl_types.add(_elem1064); + _elem1088 = iprot.readString(); + struct.tbl_types.add(_elem1088); } } struct.setTbl_typesIsSet(true); @@ -69900,14 +70032,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_table_meta_resu case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1066 = iprot.readListBegin(); - struct.success = new ArrayList(_list1066.size); - TableMeta _elem1067; - for (int _i1068 = 0; _i1068 < _list1066.size; ++_i1068) + org.apache.thrift.protocol.TList _list1090 = iprot.readListBegin(); + struct.success = new ArrayList(_list1090.size); + TableMeta _elem1091; + for (int _i1092 = 0; _i1092 < _list1090.size; ++_i1092) { - _elem1067 = new TableMeta(); - _elem1067.read(iprot); - struct.success.add(_elem1067); + _elem1091 = new TableMeta(); + _elem1091.read(iprot); + struct.success.add(_elem1091); } iprot.readListEnd(); } @@ -69942,9 +70074,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_table_meta_res oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (TableMeta _iter1069 : struct.success) + for (TableMeta _iter1093 : struct.success) { - _iter1069.write(oprot); + _iter1093.write(oprot); } oprot.writeListEnd(); } @@ -69983,9 +70115,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_table_meta_resu if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (TableMeta _iter1070 : struct.success) + for (TableMeta _iter1094 : struct.success) { - _iter1070.write(oprot); + _iter1094.write(oprot); } } } @@ -70000,14 +70132,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_table_meta_resul BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1071 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1071.size); - TableMeta _elem1072; - for (int _i1073 = 0; _i1073 < _list1071.size; ++_i1073) + org.apache.thrift.protocol.TList _list1095 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1095.size); + TableMeta _elem1096; + for (int _i1097 = 0; _i1097 < _list1095.size; ++_i1097) { - _elem1072 = new TableMeta(); - _elem1072.read(iprot); - struct.success.add(_elem1072); + _elem1096 = new TableMeta(); + _elem1096.read(iprot); + struct.success.add(_elem1096); } } struct.setSuccessIsSet(true); @@ -70773,13 +70905,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_all_tables_resu case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1074 = iprot.readListBegin(); - struct.success = new ArrayList(_list1074.size); - String _elem1075; - for (int _i1076 = 0; _i1076 < _list1074.size; ++_i1076) + org.apache.thrift.protocol.TList _list1098 = iprot.readListBegin(); + struct.success = new ArrayList(_list1098.size); + String _elem1099; + for (int _i1100 = 0; _i1100 < _list1098.size; ++_i1100) { - _elem1075 = iprot.readString(); - struct.success.add(_elem1075); + _elem1099 = iprot.readString(); + struct.success.add(_elem1099); } iprot.readListEnd(); } @@ -70814,9 +70946,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_all_tables_res oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1077 : struct.success) + for (String _iter1101 : struct.success) { - oprot.writeString(_iter1077); + oprot.writeString(_iter1101); } oprot.writeListEnd(); } @@ -70855,9 +70987,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_all_tables_resu if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1078 : struct.success) + for (String _iter1102 : struct.success) { - oprot.writeString(_iter1078); + oprot.writeString(_iter1102); } } } @@ -70872,13 +71004,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_all_tables_resul BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1079 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1079.size); - String _elem1080; - for (int _i1081 = 0; _i1081 < _list1079.size; ++_i1081) + org.apache.thrift.protocol.TList _list1103 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1103.size); + String _elem1104; + for (int _i1105 = 0; _i1105 < _list1103.size; ++_i1105) { - _elem1080 = iprot.readString(); - struct.success.add(_elem1080); + _elem1104 = iprot.readString(); + struct.success.add(_elem1104); } } struct.setSuccessIsSet(true); @@ -72331,13 +72463,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_table_objects_b case 2: // TBL_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1082 = iprot.readListBegin(); - struct.tbl_names = new ArrayList(_list1082.size); - String _elem1083; - for (int _i1084 = 0; _i1084 < _list1082.size; ++_i1084) + org.apache.thrift.protocol.TList _list1106 = iprot.readListBegin(); + struct.tbl_names = new ArrayList(_list1106.size); + String _elem1107; + for (int _i1108 = 0; _i1108 < _list1106.size; ++_i1108) { - _elem1083 = iprot.readString(); - struct.tbl_names.add(_elem1083); + _elem1107 = iprot.readString(); + struct.tbl_names.add(_elem1107); } iprot.readListEnd(); } @@ -72368,9 +72500,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_table_objects_ oprot.writeFieldBegin(TBL_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.tbl_names.size())); - for (String _iter1085 : struct.tbl_names) + for (String _iter1109 : struct.tbl_names) { - oprot.writeString(_iter1085); + oprot.writeString(_iter1109); } oprot.writeListEnd(); } @@ -72407,9 +72539,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_table_objects_b if (struct.isSetTbl_names()) { { oprot.writeI32(struct.tbl_names.size()); - for (String _iter1086 : struct.tbl_names) + for (String _iter1110 : struct.tbl_names) { - oprot.writeString(_iter1086); + oprot.writeString(_iter1110); } } } @@ -72425,13 +72557,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_table_objects_by } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list1087 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.tbl_names = new ArrayList(_list1087.size); - String _elem1088; - for (int _i1089 = 0; _i1089 < _list1087.size; ++_i1089) + org.apache.thrift.protocol.TList _list1111 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.tbl_names = new ArrayList(_list1111.size); + String _elem1112; + for (int _i1113 = 0; _i1113 < _list1111.size; ++_i1113) { - _elem1088 = iprot.readString(); - struct.tbl_names.add(_elem1088); + _elem1112 = iprot.readString(); + struct.tbl_names.add(_elem1112); } } struct.setTbl_namesIsSet(true); @@ -72756,14 +72888,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_table_objects_b case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1090 = iprot.readListBegin(); - struct.success = new ArrayList
(_list1090.size); - Table _elem1091; - for (int _i1092 = 0; _i1092 < _list1090.size; ++_i1092) + org.apache.thrift.protocol.TList _list1114 = iprot.readListBegin(); + struct.success = new ArrayList
(_list1114.size); + Table _elem1115; + for (int _i1116 = 0; _i1116 < _list1114.size; ++_i1116) { - _elem1091 = new Table(); - _elem1091.read(iprot); - struct.success.add(_elem1091); + _elem1115 = new Table(); + _elem1115.read(iprot); + struct.success.add(_elem1115); } iprot.readListEnd(); } @@ -72789,9 +72921,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_table_objects_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Table _iter1093 : struct.success) + for (Table _iter1117 : struct.success) { - _iter1093.write(oprot); + _iter1117.write(oprot); } oprot.writeListEnd(); } @@ -72822,9 +72954,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_table_objects_b if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Table _iter1094 : struct.success) + for (Table _iter1118 : struct.success) { - _iter1094.write(oprot); + _iter1118.write(oprot); } } } @@ -72836,14 +72968,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_table_objects_by BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1095 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList
(_list1095.size); - Table _elem1096; - for (int _i1097 = 0; _i1097 < _list1095.size; ++_i1097) + org.apache.thrift.protocol.TList _list1119 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList
(_list1119.size); + Table _elem1120; + for (int _i1121 = 0; _i1121 < _list1119.size; ++_i1121) { - _elem1096 = new Table(); - _elem1096.read(iprot); - struct.success.add(_elem1096); + _elem1120 = new Table(); + _elem1120.read(iprot); + struct.success.add(_elem1120); } } struct.setSuccessIsSet(true); @@ -78351,13 +78483,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_table_names_by_ case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1098 = iprot.readListBegin(); - struct.success = new ArrayList(_list1098.size); - String _elem1099; - for (int _i1100 = 0; _i1100 < _list1098.size; ++_i1100) + org.apache.thrift.protocol.TList _list1122 = iprot.readListBegin(); + struct.success = new ArrayList(_list1122.size); + String _elem1123; + for (int _i1124 = 0; _i1124 < _list1122.size; ++_i1124) { - _elem1099 = iprot.readString(); - struct.success.add(_elem1099); + _elem1123 = iprot.readString(); + struct.success.add(_elem1123); } iprot.readListEnd(); } @@ -78410,9 +78542,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_table_names_by oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1101 : struct.success) + for (String _iter1125 : struct.success) { - oprot.writeString(_iter1101); + oprot.writeString(_iter1125); } oprot.writeListEnd(); } @@ -78467,9 +78599,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_table_names_by_ if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1102 : struct.success) + for (String _iter1126 : struct.success) { - oprot.writeString(_iter1102); + oprot.writeString(_iter1126); } } } @@ -78490,13 +78622,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_table_names_by_f BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1103 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1103.size); - String _elem1104; - for (int _i1105 = 0; _i1105 < _list1103.size; ++_i1105) + org.apache.thrift.protocol.TList _list1127 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1127.size); + String _elem1128; + for (int _i1129 = 0; _i1129 < _list1127.size; ++_i1129) { - _elem1104 = iprot.readString(); - struct.success.add(_elem1104); + _elem1128 = iprot.readString(); + struct.success.add(_elem1128); } } struct.setSuccessIsSet(true); @@ -84355,14 +84487,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, add_partitions_args case 1: // NEW_PARTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1106 = iprot.readListBegin(); - struct.new_parts = new ArrayList(_list1106.size); - Partition _elem1107; - for (int _i1108 = 0; _i1108 < _list1106.size; ++_i1108) + org.apache.thrift.protocol.TList _list1130 = iprot.readListBegin(); + struct.new_parts = new ArrayList(_list1130.size); + Partition _elem1131; + for (int _i1132 = 0; _i1132 < _list1130.size; ++_i1132) { - _elem1107 = new Partition(); - _elem1107.read(iprot); - struct.new_parts.add(_elem1107); + _elem1131 = new Partition(); + _elem1131.read(iprot); + struct.new_parts.add(_elem1131); } iprot.readListEnd(); } @@ -84388,9 +84520,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, add_partitions_arg oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size())); - for (Partition _iter1109 : struct.new_parts) + for (Partition _iter1133 : struct.new_parts) { - _iter1109.write(oprot); + _iter1133.write(oprot); } oprot.writeListEnd(); } @@ -84421,9 +84553,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, add_partitions_args if (struct.isSetNew_parts()) { { oprot.writeI32(struct.new_parts.size()); - for (Partition _iter1110 : struct.new_parts) + for (Partition _iter1134 : struct.new_parts) { - _iter1110.write(oprot); + _iter1134.write(oprot); } } } @@ -84435,14 +84567,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, add_partitions_args BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1111 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.new_parts = new ArrayList(_list1111.size); - Partition _elem1112; - for (int _i1113 = 0; _i1113 < _list1111.size; ++_i1113) + org.apache.thrift.protocol.TList _list1135 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.new_parts = new ArrayList(_list1135.size); + Partition _elem1136; + for (int _i1137 = 0; _i1137 < _list1135.size; ++_i1137) { - _elem1112 = new Partition(); - _elem1112.read(iprot); - struct.new_parts.add(_elem1112); + _elem1136 = new Partition(); + _elem1136.read(iprot); + struct.new_parts.add(_elem1136); } } struct.setNew_partsIsSet(true); @@ -85443,14 +85575,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, add_partitions_pspe case 1: // NEW_PARTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1114 = iprot.readListBegin(); - struct.new_parts = new ArrayList(_list1114.size); - PartitionSpec _elem1115; - for (int _i1116 = 0; _i1116 < _list1114.size; ++_i1116) + org.apache.thrift.protocol.TList _list1138 = iprot.readListBegin(); + struct.new_parts = new ArrayList(_list1138.size); + PartitionSpec _elem1139; + for (int _i1140 = 0; _i1140 < _list1138.size; ++_i1140) { - _elem1115 = new PartitionSpec(); - _elem1115.read(iprot); - struct.new_parts.add(_elem1115); + _elem1139 = new PartitionSpec(); + _elem1139.read(iprot); + struct.new_parts.add(_elem1139); } iprot.readListEnd(); } @@ -85476,9 +85608,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, add_partitions_psp oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size())); - for (PartitionSpec _iter1117 : struct.new_parts) + for (PartitionSpec _iter1141 : struct.new_parts) { - _iter1117.write(oprot); + _iter1141.write(oprot); } oprot.writeListEnd(); } @@ -85509,9 +85641,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, add_partitions_pspe if (struct.isSetNew_parts()) { { oprot.writeI32(struct.new_parts.size()); - for (PartitionSpec _iter1118 : struct.new_parts) + for (PartitionSpec _iter1142 : struct.new_parts) { - _iter1118.write(oprot); + _iter1142.write(oprot); } } } @@ -85523,14 +85655,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, add_partitions_pspec BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1119 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.new_parts = new ArrayList(_list1119.size); - PartitionSpec _elem1120; - for (int _i1121 = 0; _i1121 < _list1119.size; ++_i1121) + org.apache.thrift.protocol.TList _list1143 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.new_parts = new ArrayList(_list1143.size); + PartitionSpec _elem1144; + for (int _i1145 = 0; _i1145 < _list1143.size; ++_i1145) { - _elem1120 = new PartitionSpec(); - _elem1120.read(iprot); - struct.new_parts.add(_elem1120); + _elem1144 = new PartitionSpec(); + _elem1144.read(iprot); + struct.new_parts.add(_elem1144); } } struct.setNew_partsIsSet(true); @@ -86706,13 +86838,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, append_partition_ar case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1122 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1122.size); - String _elem1123; - for (int _i1124 = 0; _i1124 < _list1122.size; ++_i1124) + org.apache.thrift.protocol.TList _list1146 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1146.size); + String _elem1147; + for (int _i1148 = 0; _i1148 < _list1146.size; ++_i1148) { - _elem1123 = iprot.readString(); - struct.part_vals.add(_elem1123); + _elem1147 = iprot.readString(); + struct.part_vals.add(_elem1147); } iprot.readListEnd(); } @@ -86748,9 +86880,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, append_partition_a oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1125 : struct.part_vals) + for (String _iter1149 : struct.part_vals) { - oprot.writeString(_iter1125); + oprot.writeString(_iter1149); } oprot.writeListEnd(); } @@ -86793,9 +86925,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_ar if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1126 : struct.part_vals) + for (String _iter1150 : struct.part_vals) { - oprot.writeString(_iter1126); + oprot.writeString(_iter1150); } } } @@ -86815,13 +86947,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_arg } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1127 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1127.size); - String _elem1128; - for (int _i1129 = 0; _i1129 < _list1127.size; ++_i1129) + org.apache.thrift.protocol.TList _list1151 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1151.size); + String _elem1152; + for (int _i1153 = 0; _i1153 < _list1151.size; ++_i1153) { - _elem1128 = iprot.readString(); - struct.part_vals.add(_elem1128); + _elem1152 = iprot.readString(); + struct.part_vals.add(_elem1152); } } struct.setPart_valsIsSet(true); @@ -89130,13 +89262,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, append_partition_wi case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1130 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1130.size); - String _elem1131; - for (int _i1132 = 0; _i1132 < _list1130.size; ++_i1132) + org.apache.thrift.protocol.TList _list1154 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1154.size); + String _elem1155; + for (int _i1156 = 0; _i1156 < _list1154.size; ++_i1156) { - _elem1131 = iprot.readString(); - struct.part_vals.add(_elem1131); + _elem1155 = iprot.readString(); + struct.part_vals.add(_elem1155); } iprot.readListEnd(); } @@ -89181,9 +89313,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, append_partition_w oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1133 : struct.part_vals) + for (String _iter1157 : struct.part_vals) { - oprot.writeString(_iter1133); + oprot.writeString(_iter1157); } oprot.writeListEnd(); } @@ -89234,9 +89366,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_wi if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1134 : struct.part_vals) + for (String _iter1158 : struct.part_vals) { - oprot.writeString(_iter1134); + oprot.writeString(_iter1158); } } } @@ -89259,13 +89391,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_wit } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1135 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1135.size); - String _elem1136; - for (int _i1137 = 0; _i1137 < _list1135.size; ++_i1137) + org.apache.thrift.protocol.TList _list1159 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1159.size); + String _elem1160; + for (int _i1161 = 0; _i1161 < _list1159.size; ++_i1161) { - _elem1136 = iprot.readString(); - struct.part_vals.add(_elem1136); + _elem1160 = iprot.readString(); + struct.part_vals.add(_elem1160); } } struct.setPart_valsIsSet(true); @@ -93135,13 +93267,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, drop_partition_args case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1138 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1138.size); - String _elem1139; - for (int _i1140 = 0; _i1140 < _list1138.size; ++_i1140) + org.apache.thrift.protocol.TList _list1162 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1162.size); + String _elem1163; + for (int _i1164 = 0; _i1164 < _list1162.size; ++_i1164) { - _elem1139 = iprot.readString(); - struct.part_vals.add(_elem1139); + _elem1163 = iprot.readString(); + struct.part_vals.add(_elem1163); } iprot.readListEnd(); } @@ -93185,9 +93317,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, drop_partition_arg oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1141 : struct.part_vals) + for (String _iter1165 : struct.part_vals) { - oprot.writeString(_iter1141); + oprot.writeString(_iter1165); } oprot.writeListEnd(); } @@ -93236,9 +93368,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, drop_partition_args if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1142 : struct.part_vals) + for (String _iter1166 : struct.part_vals) { - oprot.writeString(_iter1142); + oprot.writeString(_iter1166); } } } @@ -93261,13 +93393,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, drop_partition_args } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1143 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1143.size); - String _elem1144; - for (int _i1145 = 0; _i1145 < _list1143.size; ++_i1145) + org.apache.thrift.protocol.TList _list1167 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1167.size); + String _elem1168; + for (int _i1169 = 0; _i1169 < _list1167.size; ++_i1169) { - _elem1144 = iprot.readString(); - struct.part_vals.add(_elem1144); + _elem1168 = iprot.readString(); + struct.part_vals.add(_elem1168); } } struct.setPart_valsIsSet(true); @@ -94506,13 +94638,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, drop_partition_with case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1146 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1146.size); - String _elem1147; - for (int _i1148 = 0; _i1148 < _list1146.size; ++_i1148) + org.apache.thrift.protocol.TList _list1170 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1170.size); + String _elem1171; + for (int _i1172 = 0; _i1172 < _list1170.size; ++_i1172) { - _elem1147 = iprot.readString(); - struct.part_vals.add(_elem1147); + _elem1171 = iprot.readString(); + struct.part_vals.add(_elem1171); } iprot.readListEnd(); } @@ -94565,9 +94697,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, drop_partition_wit oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1149 : struct.part_vals) + for (String _iter1173 : struct.part_vals) { - oprot.writeString(_iter1149); + oprot.writeString(_iter1173); } oprot.writeListEnd(); } @@ -94624,9 +94756,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, drop_partition_with if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1150 : struct.part_vals) + for (String _iter1174 : struct.part_vals) { - oprot.writeString(_iter1150); + oprot.writeString(_iter1174); } } } @@ -94652,13 +94784,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, drop_partition_with_ } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1151 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1151.size); - String _elem1152; - for (int _i1153 = 0; _i1153 < _list1151.size; ++_i1153) + org.apache.thrift.protocol.TList _list1175 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1175.size); + String _elem1176; + for (int _i1177 = 0; _i1177 < _list1175.size; ++_i1177) { - _elem1152 = iprot.readString(); - struct.part_vals.add(_elem1152); + _elem1176 = iprot.readString(); + struct.part_vals.add(_elem1176); } } struct.setPart_valsIsSet(true); @@ -99260,13 +99392,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partition_args case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1154 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1154.size); - String _elem1155; - for (int _i1156 = 0; _i1156 < _list1154.size; ++_i1156) + org.apache.thrift.protocol.TList _list1178 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1178.size); + String _elem1179; + for (int _i1180 = 0; _i1180 < _list1178.size; ++_i1180) { - _elem1155 = iprot.readString(); - struct.part_vals.add(_elem1155); + _elem1179 = iprot.readString(); + struct.part_vals.add(_elem1179); } iprot.readListEnd(); } @@ -99302,9 +99434,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partition_args oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1157 : struct.part_vals) + for (String _iter1181 : struct.part_vals) { - oprot.writeString(_iter1157); + oprot.writeString(_iter1181); } oprot.writeListEnd(); } @@ -99347,9 +99479,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partition_args if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1158 : struct.part_vals) + for (String _iter1182 : struct.part_vals) { - oprot.writeString(_iter1158); + oprot.writeString(_iter1182); } } } @@ -99369,13 +99501,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partition_args s } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1159 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1159.size); - String _elem1160; - for (int _i1161 = 0; _i1161 < _list1159.size; ++_i1161) + org.apache.thrift.protocol.TList _list1183 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1183.size); + String _elem1184; + for (int _i1185 = 0; _i1185 < _list1183.size; ++_i1185) { - _elem1160 = iprot.readString(); - struct.part_vals.add(_elem1160); + _elem1184 = iprot.readString(); + struct.part_vals.add(_elem1184); } } struct.setPart_valsIsSet(true); @@ -100593,15 +100725,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, exchange_partition_ case 1: // PARTITION_SPECS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map1162 = iprot.readMapBegin(); - struct.partitionSpecs = new HashMap(2*_map1162.size); - String _key1163; - String _val1164; - for (int _i1165 = 0; _i1165 < _map1162.size; ++_i1165) + org.apache.thrift.protocol.TMap _map1186 = iprot.readMapBegin(); + struct.partitionSpecs = new HashMap(2*_map1186.size); + String _key1187; + String _val1188; + for (int _i1189 = 0; _i1189 < _map1186.size; ++_i1189) { - _key1163 = iprot.readString(); - _val1164 = iprot.readString(); - struct.partitionSpecs.put(_key1163, _val1164); + _key1187 = iprot.readString(); + _val1188 = iprot.readString(); + struct.partitionSpecs.put(_key1187, _val1188); } iprot.readMapEnd(); } @@ -100659,10 +100791,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, exchange_partition oprot.writeFieldBegin(PARTITION_SPECS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, struct.partitionSpecs.size())); - for (Map.Entry _iter1166 : struct.partitionSpecs.entrySet()) + for (Map.Entry _iter1190 : struct.partitionSpecs.entrySet()) { - oprot.writeString(_iter1166.getKey()); - oprot.writeString(_iter1166.getValue()); + oprot.writeString(_iter1190.getKey()); + oprot.writeString(_iter1190.getValue()); } oprot.writeMapEnd(); } @@ -100725,10 +100857,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, exchange_partition_ if (struct.isSetPartitionSpecs()) { { oprot.writeI32(struct.partitionSpecs.size()); - for (Map.Entry _iter1167 : struct.partitionSpecs.entrySet()) + for (Map.Entry _iter1191 : struct.partitionSpecs.entrySet()) { - oprot.writeString(_iter1167.getKey()); - oprot.writeString(_iter1167.getValue()); + oprot.writeString(_iter1191.getKey()); + oprot.writeString(_iter1191.getValue()); } } } @@ -100752,15 +100884,15 @@ public void read(org.apache.thrift.protocol.TProtocol prot, exchange_partition_a BitSet incoming = iprot.readBitSet(5); if (incoming.get(0)) { { - org.apache.thrift.protocol.TMap _map1168 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.partitionSpecs = new HashMap(2*_map1168.size); - String _key1169; - String _val1170; - for (int _i1171 = 0; _i1171 < _map1168.size; ++_i1171) + org.apache.thrift.protocol.TMap _map1192 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partitionSpecs = new HashMap(2*_map1192.size); + String _key1193; + String _val1194; + for (int _i1195 = 0; _i1195 < _map1192.size; ++_i1195) { - _key1169 = iprot.readString(); - _val1170 = iprot.readString(); - struct.partitionSpecs.put(_key1169, _val1170); + _key1193 = iprot.readString(); + _val1194 = iprot.readString(); + struct.partitionSpecs.put(_key1193, _val1194); } } struct.setPartitionSpecsIsSet(true); @@ -102206,15 +102338,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, exchange_partitions case 1: // PARTITION_SPECS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map1172 = iprot.readMapBegin(); - struct.partitionSpecs = new HashMap(2*_map1172.size); - String _key1173; - String _val1174; - for (int _i1175 = 0; _i1175 < _map1172.size; ++_i1175) + org.apache.thrift.protocol.TMap _map1196 = iprot.readMapBegin(); + struct.partitionSpecs = new HashMap(2*_map1196.size); + String _key1197; + String _val1198; + for (int _i1199 = 0; _i1199 < _map1196.size; ++_i1199) { - _key1173 = iprot.readString(); - _val1174 = iprot.readString(); - struct.partitionSpecs.put(_key1173, _val1174); + _key1197 = iprot.readString(); + _val1198 = iprot.readString(); + struct.partitionSpecs.put(_key1197, _val1198); } iprot.readMapEnd(); } @@ -102272,10 +102404,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, exchange_partition oprot.writeFieldBegin(PARTITION_SPECS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, struct.partitionSpecs.size())); - for (Map.Entry _iter1176 : struct.partitionSpecs.entrySet()) + for (Map.Entry _iter1200 : struct.partitionSpecs.entrySet()) { - oprot.writeString(_iter1176.getKey()); - oprot.writeString(_iter1176.getValue()); + oprot.writeString(_iter1200.getKey()); + oprot.writeString(_iter1200.getValue()); } oprot.writeMapEnd(); } @@ -102338,10 +102470,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, exchange_partitions if (struct.isSetPartitionSpecs()) { { oprot.writeI32(struct.partitionSpecs.size()); - for (Map.Entry _iter1177 : struct.partitionSpecs.entrySet()) + for (Map.Entry _iter1201 : struct.partitionSpecs.entrySet()) { - oprot.writeString(_iter1177.getKey()); - oprot.writeString(_iter1177.getValue()); + oprot.writeString(_iter1201.getKey()); + oprot.writeString(_iter1201.getValue()); } } } @@ -102365,15 +102497,15 @@ public void read(org.apache.thrift.protocol.TProtocol prot, exchange_partitions_ BitSet incoming = iprot.readBitSet(5); if (incoming.get(0)) { { - org.apache.thrift.protocol.TMap _map1178 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.partitionSpecs = new HashMap(2*_map1178.size); - String _key1179; - String _val1180; - for (int _i1181 = 0; _i1181 < _map1178.size; ++_i1181) + org.apache.thrift.protocol.TMap _map1202 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partitionSpecs = new HashMap(2*_map1202.size); + String _key1203; + String _val1204; + for (int _i1205 = 0; _i1205 < _map1202.size; ++_i1205) { - _key1179 = iprot.readString(); - _val1180 = iprot.readString(); - struct.partitionSpecs.put(_key1179, _val1180); + _key1203 = iprot.readString(); + _val1204 = iprot.readString(); + struct.partitionSpecs.put(_key1203, _val1204); } } struct.setPartitionSpecsIsSet(true); @@ -103038,14 +103170,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, exchange_partitions case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1182 = iprot.readListBegin(); - struct.success = new ArrayList(_list1182.size); - Partition _elem1183; - for (int _i1184 = 0; _i1184 < _list1182.size; ++_i1184) + org.apache.thrift.protocol.TList _list1206 = iprot.readListBegin(); + struct.success = new ArrayList(_list1206.size); + Partition _elem1207; + for (int _i1208 = 0; _i1208 < _list1206.size; ++_i1208) { - _elem1183 = new Partition(); - _elem1183.read(iprot); - struct.success.add(_elem1183); + _elem1207 = new Partition(); + _elem1207.read(iprot); + struct.success.add(_elem1207); } iprot.readListEnd(); } @@ -103107,9 +103239,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, exchange_partition oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1185 : struct.success) + for (Partition _iter1209 : struct.success) { - _iter1185.write(oprot); + _iter1209.write(oprot); } oprot.writeListEnd(); } @@ -103172,9 +103304,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, exchange_partitions if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1186 : struct.success) + for (Partition _iter1210 : struct.success) { - _iter1186.write(oprot); + _iter1210.write(oprot); } } } @@ -103198,14 +103330,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, exchange_partitions_ BitSet incoming = iprot.readBitSet(5); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1187 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1187.size); - Partition _elem1188; - for (int _i1189 = 0; _i1189 < _list1187.size; ++_i1189) + org.apache.thrift.protocol.TList _list1211 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1211.size); + Partition _elem1212; + for (int _i1213 = 0; _i1213 < _list1211.size; ++_i1213) { - _elem1188 = new Partition(); - _elem1188.read(iprot); - struct.success.add(_elem1188); + _elem1212 = new Partition(); + _elem1212.read(iprot); + struct.success.add(_elem1212); } } struct.setSuccessIsSet(true); @@ -103904,13 +104036,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partition_with_ case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1190 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1190.size); - String _elem1191; - for (int _i1192 = 0; _i1192 < _list1190.size; ++_i1192) + org.apache.thrift.protocol.TList _list1214 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1214.size); + String _elem1215; + for (int _i1216 = 0; _i1216 < _list1214.size; ++_i1216) { - _elem1191 = iprot.readString(); - struct.part_vals.add(_elem1191); + _elem1215 = iprot.readString(); + struct.part_vals.add(_elem1215); } iprot.readListEnd(); } @@ -103930,13 +104062,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partition_with_ case 5: // GROUP_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1193 = iprot.readListBegin(); - struct.group_names = new ArrayList(_list1193.size); - String _elem1194; - for (int _i1195 = 0; _i1195 < _list1193.size; ++_i1195) + org.apache.thrift.protocol.TList _list1217 = iprot.readListBegin(); + struct.group_names = new ArrayList(_list1217.size); + String _elem1218; + for (int _i1219 = 0; _i1219 < _list1217.size; ++_i1219) { - _elem1194 = iprot.readString(); - struct.group_names.add(_elem1194); + _elem1218 = iprot.readString(); + struct.group_names.add(_elem1218); } iprot.readListEnd(); } @@ -103972,9 +104104,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partition_with oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1196 : struct.part_vals) + for (String _iter1220 : struct.part_vals) { - oprot.writeString(_iter1196); + oprot.writeString(_iter1220); } oprot.writeListEnd(); } @@ -103989,9 +104121,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partition_with oprot.writeFieldBegin(GROUP_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.group_names.size())); - for (String _iter1197 : struct.group_names) + for (String _iter1221 : struct.group_names) { - oprot.writeString(_iter1197); + oprot.writeString(_iter1221); } oprot.writeListEnd(); } @@ -104040,9 +104172,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partition_with_ if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1198 : struct.part_vals) + for (String _iter1222 : struct.part_vals) { - oprot.writeString(_iter1198); + oprot.writeString(_iter1222); } } } @@ -104052,9 +104184,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partition_with_ if (struct.isSetGroup_names()) { { oprot.writeI32(struct.group_names.size()); - for (String _iter1199 : struct.group_names) + for (String _iter1223 : struct.group_names) { - oprot.writeString(_iter1199); + oprot.writeString(_iter1223); } } } @@ -104074,13 +104206,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partition_with_a } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1200 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1200.size); - String _elem1201; - for (int _i1202 = 0; _i1202 < _list1200.size; ++_i1202) + org.apache.thrift.protocol.TList _list1224 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1224.size); + String _elem1225; + for (int _i1226 = 0; _i1226 < _list1224.size; ++_i1226) { - _elem1201 = iprot.readString(); - struct.part_vals.add(_elem1201); + _elem1225 = iprot.readString(); + struct.part_vals.add(_elem1225); } } struct.setPart_valsIsSet(true); @@ -104091,13 +104223,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partition_with_a } if (incoming.get(4)) { { - org.apache.thrift.protocol.TList _list1203 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.group_names = new ArrayList(_list1203.size); - String _elem1204; - for (int _i1205 = 0; _i1205 < _list1203.size; ++_i1205) + org.apache.thrift.protocol.TList _list1227 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.group_names = new ArrayList(_list1227.size); + String _elem1228; + for (int _i1229 = 0; _i1229 < _list1227.size; ++_i1229) { - _elem1204 = iprot.readString(); - struct.group_names.add(_elem1204); + _elem1228 = iprot.readString(); + struct.group_names.add(_elem1228); } } struct.setGroup_namesIsSet(true); @@ -106866,14 +106998,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_resu case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1206 = iprot.readListBegin(); - struct.success = new ArrayList(_list1206.size); - Partition _elem1207; - for (int _i1208 = 0; _i1208 < _list1206.size; ++_i1208) + org.apache.thrift.protocol.TList _list1230 = iprot.readListBegin(); + struct.success = new ArrayList(_list1230.size); + Partition _elem1231; + for (int _i1232 = 0; _i1232 < _list1230.size; ++_i1232) { - _elem1207 = new Partition(); - _elem1207.read(iprot); - struct.success.add(_elem1207); + _elem1231 = new Partition(); + _elem1231.read(iprot); + struct.success.add(_elem1231); } iprot.readListEnd(); } @@ -106917,9 +107049,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_res oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1209 : struct.success) + for (Partition _iter1233 : struct.success) { - _iter1209.write(oprot); + _iter1233.write(oprot); } oprot.writeListEnd(); } @@ -106966,9 +107098,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_resu if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1210 : struct.success) + for (Partition _iter1234 : struct.success) { - _iter1210.write(oprot); + _iter1234.write(oprot); } } } @@ -106986,14 +107118,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_resul BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1211 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1211.size); - Partition _elem1212; - for (int _i1213 = 0; _i1213 < _list1211.size; ++_i1213) + org.apache.thrift.protocol.TList _list1235 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1235.size); + Partition _elem1236; + for (int _i1237 = 0; _i1237 < _list1235.size; ++_i1237) { - _elem1212 = new Partition(); - _elem1212.read(iprot); - struct.success.add(_elem1212); + _elem1236 = new Partition(); + _elem1236.read(iprot); + struct.success.add(_elem1236); } } struct.setSuccessIsSet(true); @@ -107683,13 +107815,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_with case 5: // GROUP_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1214 = iprot.readListBegin(); - struct.group_names = new ArrayList(_list1214.size); - String _elem1215; - for (int _i1216 = 0; _i1216 < _list1214.size; ++_i1216) + org.apache.thrift.protocol.TList _list1238 = iprot.readListBegin(); + struct.group_names = new ArrayList(_list1238.size); + String _elem1239; + for (int _i1240 = 0; _i1240 < _list1238.size; ++_i1240) { - _elem1215 = iprot.readString(); - struct.group_names.add(_elem1215); + _elem1239 = iprot.readString(); + struct.group_names.add(_elem1239); } iprot.readListEnd(); } @@ -107733,9 +107865,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_wit oprot.writeFieldBegin(GROUP_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.group_names.size())); - for (String _iter1217 : struct.group_names) + for (String _iter1241 : struct.group_names) { - oprot.writeString(_iter1217); + oprot.writeString(_iter1241); } oprot.writeListEnd(); } @@ -107790,9 +107922,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_with if (struct.isSetGroup_names()) { { oprot.writeI32(struct.group_names.size()); - for (String _iter1218 : struct.group_names) + for (String _iter1242 : struct.group_names) { - oprot.writeString(_iter1218); + oprot.writeString(_iter1242); } } } @@ -107820,13 +107952,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_with_ } if (incoming.get(4)) { { - org.apache.thrift.protocol.TList _list1219 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.group_names = new ArrayList(_list1219.size); - String _elem1220; - for (int _i1221 = 0; _i1221 < _list1219.size; ++_i1221) + org.apache.thrift.protocol.TList _list1243 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.group_names = new ArrayList(_list1243.size); + String _elem1244; + for (int _i1245 = 0; _i1245 < _list1243.size; ++_i1245) { - _elem1220 = iprot.readString(); - struct.group_names.add(_elem1220); + _elem1244 = iprot.readString(); + struct.group_names.add(_elem1244); } } struct.setGroup_namesIsSet(true); @@ -108313,14 +108445,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_with case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1222 = iprot.readListBegin(); - struct.success = new ArrayList(_list1222.size); - Partition _elem1223; - for (int _i1224 = 0; _i1224 < _list1222.size; ++_i1224) + org.apache.thrift.protocol.TList _list1246 = iprot.readListBegin(); + struct.success = new ArrayList(_list1246.size); + Partition _elem1247; + for (int _i1248 = 0; _i1248 < _list1246.size; ++_i1248) { - _elem1223 = new Partition(); - _elem1223.read(iprot); - struct.success.add(_elem1223); + _elem1247 = new Partition(); + _elem1247.read(iprot); + struct.success.add(_elem1247); } iprot.readListEnd(); } @@ -108364,9 +108496,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_wit oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1225 : struct.success) + for (Partition _iter1249 : struct.success) { - _iter1225.write(oprot); + _iter1249.write(oprot); } oprot.writeListEnd(); } @@ -108413,9 +108545,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_with if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1226 : struct.success) + for (Partition _iter1250 : struct.success) { - _iter1226.write(oprot); + _iter1250.write(oprot); } } } @@ -108433,14 +108565,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_with_ BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1227 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1227.size); - Partition _elem1228; - for (int _i1229 = 0; _i1229 < _list1227.size; ++_i1229) + org.apache.thrift.protocol.TList _list1251 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1251.size); + Partition _elem1252; + for (int _i1253 = 0; _i1253 < _list1251.size; ++_i1253) { - _elem1228 = new Partition(); - _elem1228.read(iprot); - struct.success.add(_elem1228); + _elem1252 = new Partition(); + _elem1252.read(iprot); + struct.success.add(_elem1252); } } struct.setSuccessIsSet(true); @@ -109503,14 +109635,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_pspe case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1230 = iprot.readListBegin(); - struct.success = new ArrayList(_list1230.size); - PartitionSpec _elem1231; - for (int _i1232 = 0; _i1232 < _list1230.size; ++_i1232) + org.apache.thrift.protocol.TList _list1254 = iprot.readListBegin(); + struct.success = new ArrayList(_list1254.size); + PartitionSpec _elem1255; + for (int _i1256 = 0; _i1256 < _list1254.size; ++_i1256) { - _elem1231 = new PartitionSpec(); - _elem1231.read(iprot); - struct.success.add(_elem1231); + _elem1255 = new PartitionSpec(); + _elem1255.read(iprot); + struct.success.add(_elem1255); } iprot.readListEnd(); } @@ -109554,9 +109686,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_psp oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (PartitionSpec _iter1233 : struct.success) + for (PartitionSpec _iter1257 : struct.success) { - _iter1233.write(oprot); + _iter1257.write(oprot); } oprot.writeListEnd(); } @@ -109603,9 +109735,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_pspe if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (PartitionSpec _iter1234 : struct.success) + for (PartitionSpec _iter1258 : struct.success) { - _iter1234.write(oprot); + _iter1258.write(oprot); } } } @@ -109623,14 +109755,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_pspec BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1235 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1235.size); - PartitionSpec _elem1236; - for (int _i1237 = 0; _i1237 < _list1235.size; ++_i1237) + org.apache.thrift.protocol.TList _list1259 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1259.size); + PartitionSpec _elem1260; + for (int _i1261 = 0; _i1261 < _list1259.size; ++_i1261) { - _elem1236 = new PartitionSpec(); - _elem1236.read(iprot); - struct.success.add(_elem1236); + _elem1260 = new PartitionSpec(); + _elem1260.read(iprot); + struct.success.add(_elem1260); } } struct.setSuccessIsSet(true); @@ -110690,13 +110822,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partition_names case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1238 = iprot.readListBegin(); - struct.success = new ArrayList(_list1238.size); - String _elem1239; - for (int _i1240 = 0; _i1240 < _list1238.size; ++_i1240) + org.apache.thrift.protocol.TList _list1262 = iprot.readListBegin(); + struct.success = new ArrayList(_list1262.size); + String _elem1263; + for (int _i1264 = 0; _i1264 < _list1262.size; ++_i1264) { - _elem1239 = iprot.readString(); - struct.success.add(_elem1239); + _elem1263 = iprot.readString(); + struct.success.add(_elem1263); } iprot.readListEnd(); } @@ -110740,9 +110872,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partition_name oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1241 : struct.success) + for (String _iter1265 : struct.success) { - oprot.writeString(_iter1241); + oprot.writeString(_iter1265); } oprot.writeListEnd(); } @@ -110789,9 +110921,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partition_names if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1242 : struct.success) + for (String _iter1266 : struct.success) { - oprot.writeString(_iter1242); + oprot.writeString(_iter1266); } } } @@ -110809,13 +110941,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partition_names_ BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1243 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1243.size); - String _elem1244; - for (int _i1245 = 0; _i1245 < _list1243.size; ++_i1245) + org.apache.thrift.protocol.TList _list1267 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1267.size); + String _elem1268; + for (int _i1269 = 0; _i1269 < _list1267.size; ++_i1269) { - _elem1244 = iprot.readString(); - struct.success.add(_elem1244); + _elem1268 = iprot.readString(); + struct.success.add(_elem1268); } } struct.setSuccessIsSet(true); @@ -112346,13 +112478,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_ps_a case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1246 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1246.size); - String _elem1247; - for (int _i1248 = 0; _i1248 < _list1246.size; ++_i1248) + org.apache.thrift.protocol.TList _list1270 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1270.size); + String _elem1271; + for (int _i1272 = 0; _i1272 < _list1270.size; ++_i1272) { - _elem1247 = iprot.readString(); - struct.part_vals.add(_elem1247); + _elem1271 = iprot.readString(); + struct.part_vals.add(_elem1271); } iprot.readListEnd(); } @@ -112396,9 +112528,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_ps_ oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1249 : struct.part_vals) + for (String _iter1273 : struct.part_vals) { - oprot.writeString(_iter1249); + oprot.writeString(_iter1273); } oprot.writeListEnd(); } @@ -112447,9 +112579,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_a if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1250 : struct.part_vals) + for (String _iter1274 : struct.part_vals) { - oprot.writeString(_iter1250); + oprot.writeString(_iter1274); } } } @@ -112472,13 +112604,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_ar } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1251 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1251.size); - String _elem1252; - for (int _i1253 = 0; _i1253 < _list1251.size; ++_i1253) + org.apache.thrift.protocol.TList _list1275 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1275.size); + String _elem1276; + for (int _i1277 = 0; _i1277 < _list1275.size; ++_i1277) { - _elem1252 = iprot.readString(); - struct.part_vals.add(_elem1252); + _elem1276 = iprot.readString(); + struct.part_vals.add(_elem1276); } } struct.setPart_valsIsSet(true); @@ -112969,14 +113101,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_ps_r case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1254 = iprot.readListBegin(); - struct.success = new ArrayList(_list1254.size); - Partition _elem1255; - for (int _i1256 = 0; _i1256 < _list1254.size; ++_i1256) + org.apache.thrift.protocol.TList _list1278 = iprot.readListBegin(); + struct.success = new ArrayList(_list1278.size); + Partition _elem1279; + for (int _i1280 = 0; _i1280 < _list1278.size; ++_i1280) { - _elem1255 = new Partition(); - _elem1255.read(iprot); - struct.success.add(_elem1255); + _elem1279 = new Partition(); + _elem1279.read(iprot); + struct.success.add(_elem1279); } iprot.readListEnd(); } @@ -113020,9 +113152,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_ps_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1257 : struct.success) + for (Partition _iter1281 : struct.success) { - _iter1257.write(oprot); + _iter1281.write(oprot); } oprot.writeListEnd(); } @@ -113069,9 +113201,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_r if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1258 : struct.success) + for (Partition _iter1282 : struct.success) { - _iter1258.write(oprot); + _iter1282.write(oprot); } } } @@ -113089,14 +113221,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_re BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1259 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1259.size); - Partition _elem1260; - for (int _i1261 = 0; _i1261 < _list1259.size; ++_i1261) + org.apache.thrift.protocol.TList _list1283 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1283.size); + Partition _elem1284; + for (int _i1285 = 0; _i1285 < _list1283.size; ++_i1285) { - _elem1260 = new Partition(); - _elem1260.read(iprot); - struct.success.add(_elem1260); + _elem1284 = new Partition(); + _elem1284.read(iprot); + struct.success.add(_elem1284); } } struct.setSuccessIsSet(true); @@ -113868,13 +114000,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_ps_w case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1262 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1262.size); - String _elem1263; - for (int _i1264 = 0; _i1264 < _list1262.size; ++_i1264) + org.apache.thrift.protocol.TList _list1286 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1286.size); + String _elem1287; + for (int _i1288 = 0; _i1288 < _list1286.size; ++_i1288) { - _elem1263 = iprot.readString(); - struct.part_vals.add(_elem1263); + _elem1287 = iprot.readString(); + struct.part_vals.add(_elem1287); } iprot.readListEnd(); } @@ -113902,13 +114034,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_ps_w case 6: // GROUP_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1265 = iprot.readListBegin(); - struct.group_names = new ArrayList(_list1265.size); - String _elem1266; - for (int _i1267 = 0; _i1267 < _list1265.size; ++_i1267) + org.apache.thrift.protocol.TList _list1289 = iprot.readListBegin(); + struct.group_names = new ArrayList(_list1289.size); + String _elem1290; + for (int _i1291 = 0; _i1291 < _list1289.size; ++_i1291) { - _elem1266 = iprot.readString(); - struct.group_names.add(_elem1266); + _elem1290 = iprot.readString(); + struct.group_names.add(_elem1290); } iprot.readListEnd(); } @@ -113944,9 +114076,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_ps_ oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1268 : struct.part_vals) + for (String _iter1292 : struct.part_vals) { - oprot.writeString(_iter1268); + oprot.writeString(_iter1292); } oprot.writeListEnd(); } @@ -113964,9 +114096,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_ps_ oprot.writeFieldBegin(GROUP_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.group_names.size())); - for (String _iter1269 : struct.group_names) + for (String _iter1293 : struct.group_names) { - oprot.writeString(_iter1269); + oprot.writeString(_iter1293); } oprot.writeListEnd(); } @@ -114018,9 +114150,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_w if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1270 : struct.part_vals) + for (String _iter1294 : struct.part_vals) { - oprot.writeString(_iter1270); + oprot.writeString(_iter1294); } } } @@ -114033,9 +114165,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_w if (struct.isSetGroup_names()) { { oprot.writeI32(struct.group_names.size()); - for (String _iter1271 : struct.group_names) + for (String _iter1295 : struct.group_names) { - oprot.writeString(_iter1271); + oprot.writeString(_iter1295); } } } @@ -114055,13 +114187,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_wi } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1272 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1272.size); - String _elem1273; - for (int _i1274 = 0; _i1274 < _list1272.size; ++_i1274) + org.apache.thrift.protocol.TList _list1296 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1296.size); + String _elem1297; + for (int _i1298 = 0; _i1298 < _list1296.size; ++_i1298) { - _elem1273 = iprot.readString(); - struct.part_vals.add(_elem1273); + _elem1297 = iprot.readString(); + struct.part_vals.add(_elem1297); } } struct.setPart_valsIsSet(true); @@ -114076,13 +114208,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_wi } if (incoming.get(5)) { { - org.apache.thrift.protocol.TList _list1275 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.group_names = new ArrayList(_list1275.size); - String _elem1276; - for (int _i1277 = 0; _i1277 < _list1275.size; ++_i1277) + org.apache.thrift.protocol.TList _list1299 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.group_names = new ArrayList(_list1299.size); + String _elem1300; + for (int _i1301 = 0; _i1301 < _list1299.size; ++_i1301) { - _elem1276 = iprot.readString(); - struct.group_names.add(_elem1276); + _elem1300 = iprot.readString(); + struct.group_names.add(_elem1300); } } struct.setGroup_namesIsSet(true); @@ -114569,14 +114701,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_ps_w case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1278 = iprot.readListBegin(); - struct.success = new ArrayList(_list1278.size); - Partition _elem1279; - for (int _i1280 = 0; _i1280 < _list1278.size; ++_i1280) + org.apache.thrift.protocol.TList _list1302 = iprot.readListBegin(); + struct.success = new ArrayList(_list1302.size); + Partition _elem1303; + for (int _i1304 = 0; _i1304 < _list1302.size; ++_i1304) { - _elem1279 = new Partition(); - _elem1279.read(iprot); - struct.success.add(_elem1279); + _elem1303 = new Partition(); + _elem1303.read(iprot); + struct.success.add(_elem1303); } iprot.readListEnd(); } @@ -114620,9 +114752,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_ps_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1281 : struct.success) + for (Partition _iter1305 : struct.success) { - _iter1281.write(oprot); + _iter1305.write(oprot); } oprot.writeListEnd(); } @@ -114669,9 +114801,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_w if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1282 : struct.success) + for (Partition _iter1306 : struct.success) { - _iter1282.write(oprot); + _iter1306.write(oprot); } } } @@ -114689,14 +114821,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_ps_wi BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1283 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1283.size); - Partition _elem1284; - for (int _i1285 = 0; _i1285 < _list1283.size; ++_i1285) + org.apache.thrift.protocol.TList _list1307 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1307.size); + Partition _elem1308; + for (int _i1309 = 0; _i1309 < _list1307.size; ++_i1309) { - _elem1284 = new Partition(); - _elem1284.read(iprot); - struct.success.add(_elem1284); + _elem1308 = new Partition(); + _elem1308.read(iprot); + struct.success.add(_elem1308); } } struct.setSuccessIsSet(true); @@ -115289,13 +115421,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partition_names case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1286 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1286.size); - String _elem1287; - for (int _i1288 = 0; _i1288 < _list1286.size; ++_i1288) + org.apache.thrift.protocol.TList _list1310 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1310.size); + String _elem1311; + for (int _i1312 = 0; _i1312 < _list1310.size; ++_i1312) { - _elem1287 = iprot.readString(); - struct.part_vals.add(_elem1287); + _elem1311 = iprot.readString(); + struct.part_vals.add(_elem1311); } iprot.readListEnd(); } @@ -115339,9 +115471,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partition_name oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1289 : struct.part_vals) + for (String _iter1313 : struct.part_vals) { - oprot.writeString(_iter1289); + oprot.writeString(_iter1313); } oprot.writeListEnd(); } @@ -115390,9 +115522,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partition_names if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1290 : struct.part_vals) + for (String _iter1314 : struct.part_vals) { - oprot.writeString(_iter1290); + oprot.writeString(_iter1314); } } } @@ -115415,13 +115547,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partition_names_ } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1291 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1291.size); - String _elem1292; - for (int _i1293 = 0; _i1293 < _list1291.size; ++_i1293) + org.apache.thrift.protocol.TList _list1315 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1315.size); + String _elem1316; + for (int _i1317 = 0; _i1317 < _list1315.size; ++_i1317) { - _elem1292 = iprot.readString(); - struct.part_vals.add(_elem1292); + _elem1316 = iprot.readString(); + struct.part_vals.add(_elem1316); } } struct.setPart_valsIsSet(true); @@ -115909,13 +116041,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partition_names case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1294 = iprot.readListBegin(); - struct.success = new ArrayList(_list1294.size); - String _elem1295; - for (int _i1296 = 0; _i1296 < _list1294.size; ++_i1296) + org.apache.thrift.protocol.TList _list1318 = iprot.readListBegin(); + struct.success = new ArrayList(_list1318.size); + String _elem1319; + for (int _i1320 = 0; _i1320 < _list1318.size; ++_i1320) { - _elem1295 = iprot.readString(); - struct.success.add(_elem1295); + _elem1319 = iprot.readString(); + struct.success.add(_elem1319); } iprot.readListEnd(); } @@ -115959,9 +116091,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partition_name oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1297 : struct.success) + for (String _iter1321 : struct.success) { - oprot.writeString(_iter1297); + oprot.writeString(_iter1321); } oprot.writeListEnd(); } @@ -116008,9 +116140,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partition_names if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1298 : struct.success) + for (String _iter1322 : struct.success) { - oprot.writeString(_iter1298); + oprot.writeString(_iter1322); } } } @@ -116028,13 +116160,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partition_names_ BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1299 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1299.size); - String _elem1300; - for (int _i1301 = 0; _i1301 < _list1299.size; ++_i1301) + org.apache.thrift.protocol.TList _list1323 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1323.size); + String _elem1324; + for (int _i1325 = 0; _i1325 < _list1323.size; ++_i1325) { - _elem1300 = iprot.readString(); - struct.success.add(_elem1300); + _elem1324 = iprot.readString(); + struct.success.add(_elem1324); } } struct.setSuccessIsSet(true); @@ -117201,14 +117333,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_by_f case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1302 = iprot.readListBegin(); - struct.success = new ArrayList(_list1302.size); - Partition _elem1303; - for (int _i1304 = 0; _i1304 < _list1302.size; ++_i1304) + org.apache.thrift.protocol.TList _list1326 = iprot.readListBegin(); + struct.success = new ArrayList(_list1326.size); + Partition _elem1327; + for (int _i1328 = 0; _i1328 < _list1326.size; ++_i1328) { - _elem1303 = new Partition(); - _elem1303.read(iprot); - struct.success.add(_elem1303); + _elem1327 = new Partition(); + _elem1327.read(iprot); + struct.success.add(_elem1327); } iprot.readListEnd(); } @@ -117252,9 +117384,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_by_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1305 : struct.success) + for (Partition _iter1329 : struct.success) { - _iter1305.write(oprot); + _iter1329.write(oprot); } oprot.writeListEnd(); } @@ -117301,9 +117433,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_by_f if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1306 : struct.success) + for (Partition _iter1330 : struct.success) { - _iter1306.write(oprot); + _iter1330.write(oprot); } } } @@ -117321,14 +117453,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_by_fi BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1307 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1307.size); - Partition _elem1308; - for (int _i1309 = 0; _i1309 < _list1307.size; ++_i1309) + org.apache.thrift.protocol.TList _list1331 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1331.size); + Partition _elem1332; + for (int _i1333 = 0; _i1333 < _list1331.size; ++_i1333) { - _elem1308 = new Partition(); - _elem1308.read(iprot); - struct.success.add(_elem1308); + _elem1332 = new Partition(); + _elem1332.read(iprot); + struct.success.add(_elem1332); } } struct.setSuccessIsSet(true); @@ -118495,14 +118627,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_part_specs_by_f case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1310 = iprot.readListBegin(); - struct.success = new ArrayList(_list1310.size); - PartitionSpec _elem1311; - for (int _i1312 = 0; _i1312 < _list1310.size; ++_i1312) + org.apache.thrift.protocol.TList _list1334 = iprot.readListBegin(); + struct.success = new ArrayList(_list1334.size); + PartitionSpec _elem1335; + for (int _i1336 = 0; _i1336 < _list1334.size; ++_i1336) { - _elem1311 = new PartitionSpec(); - _elem1311.read(iprot); - struct.success.add(_elem1311); + _elem1335 = new PartitionSpec(); + _elem1335.read(iprot); + struct.success.add(_elem1335); } iprot.readListEnd(); } @@ -118546,9 +118678,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_part_specs_by_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (PartitionSpec _iter1313 : struct.success) + for (PartitionSpec _iter1337 : struct.success) { - _iter1313.write(oprot); + _iter1337.write(oprot); } oprot.writeListEnd(); } @@ -118595,9 +118727,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_part_specs_by_f if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (PartitionSpec _iter1314 : struct.success) + for (PartitionSpec _iter1338 : struct.success) { - _iter1314.write(oprot); + _iter1338.write(oprot); } } } @@ -118615,14 +118747,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_part_specs_by_fi BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1315 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1315.size); - PartitionSpec _elem1316; - for (int _i1317 = 0; _i1317 < _list1315.size; ++_i1317) + org.apache.thrift.protocol.TList _list1339 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1339.size); + PartitionSpec _elem1340; + for (int _i1341 = 0; _i1341 < _list1339.size; ++_i1341) { - _elem1316 = new PartitionSpec(); - _elem1316.read(iprot); - struct.success.add(_elem1316); + _elem1340 = new PartitionSpec(); + _elem1340.read(iprot); + struct.success.add(_elem1340); } } struct.setSuccessIsSet(true); @@ -121206,13 +121338,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_by_n case 3: // NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1318 = iprot.readListBegin(); - struct.names = new ArrayList(_list1318.size); - String _elem1319; - for (int _i1320 = 0; _i1320 < _list1318.size; ++_i1320) + org.apache.thrift.protocol.TList _list1342 = iprot.readListBegin(); + struct.names = new ArrayList(_list1342.size); + String _elem1343; + for (int _i1344 = 0; _i1344 < _list1342.size; ++_i1344) { - _elem1319 = iprot.readString(); - struct.names.add(_elem1319); + _elem1343 = iprot.readString(); + struct.names.add(_elem1343); } iprot.readListEnd(); } @@ -121248,9 +121380,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_by_ oprot.writeFieldBegin(NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.names.size())); - for (String _iter1321 : struct.names) + for (String _iter1345 : struct.names) { - oprot.writeString(_iter1321); + oprot.writeString(_iter1345); } oprot.writeListEnd(); } @@ -121293,9 +121425,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_by_n if (struct.isSetNames()) { { oprot.writeI32(struct.names.size()); - for (String _iter1322 : struct.names) + for (String _iter1346 : struct.names) { - oprot.writeString(_iter1322); + oprot.writeString(_iter1346); } } } @@ -121315,13 +121447,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_by_na } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1323 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.names = new ArrayList(_list1323.size); - String _elem1324; - for (int _i1325 = 0; _i1325 < _list1323.size; ++_i1325) + org.apache.thrift.protocol.TList _list1347 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.names = new ArrayList(_list1347.size); + String _elem1348; + for (int _i1349 = 0; _i1349 < _list1347.size; ++_i1349) { - _elem1324 = iprot.readString(); - struct.names.add(_elem1324); + _elem1348 = iprot.readString(); + struct.names.add(_elem1348); } } struct.setNamesIsSet(true); @@ -121808,14 +121940,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_partitions_by_n case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1326 = iprot.readListBegin(); - struct.success = new ArrayList(_list1326.size); - Partition _elem1327; - for (int _i1328 = 0; _i1328 < _list1326.size; ++_i1328) + org.apache.thrift.protocol.TList _list1350 = iprot.readListBegin(); + struct.success = new ArrayList(_list1350.size); + Partition _elem1351; + for (int _i1352 = 0; _i1352 < _list1350.size; ++_i1352) { - _elem1327 = new Partition(); - _elem1327.read(iprot); - struct.success.add(_elem1327); + _elem1351 = new Partition(); + _elem1351.read(iprot); + struct.success.add(_elem1351); } iprot.readListEnd(); } @@ -121859,9 +121991,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_partitions_by_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Partition _iter1329 : struct.success) + for (Partition _iter1353 : struct.success) { - _iter1329.write(oprot); + _iter1353.write(oprot); } oprot.writeListEnd(); } @@ -121908,9 +122040,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_partitions_by_n if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Partition _iter1330 : struct.success) + for (Partition _iter1354 : struct.success) { - _iter1330.write(oprot); + _iter1354.write(oprot); } } } @@ -121928,14 +122060,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_partitions_by_na BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1331 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1331.size); - Partition _elem1332; - for (int _i1333 = 0; _i1333 < _list1331.size; ++_i1333) + org.apache.thrift.protocol.TList _list1355 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1355.size); + Partition _elem1356; + for (int _i1357 = 0; _i1357 < _list1355.size; ++_i1357) { - _elem1332 = new Partition(); - _elem1332.read(iprot); - struct.success.add(_elem1332); + _elem1356 = new Partition(); + _elem1356.read(iprot); + struct.success.add(_elem1356); } } struct.setSuccessIsSet(true); @@ -123485,14 +123617,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, alter_partitions_ar case 3: // NEW_PARTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1334 = iprot.readListBegin(); - struct.new_parts = new ArrayList(_list1334.size); - Partition _elem1335; - for (int _i1336 = 0; _i1336 < _list1334.size; ++_i1336) + org.apache.thrift.protocol.TList _list1358 = iprot.readListBegin(); + struct.new_parts = new ArrayList(_list1358.size); + Partition _elem1359; + for (int _i1360 = 0; _i1360 < _list1358.size; ++_i1360) { - _elem1335 = new Partition(); - _elem1335.read(iprot); - struct.new_parts.add(_elem1335); + _elem1359 = new Partition(); + _elem1359.read(iprot); + struct.new_parts.add(_elem1359); } iprot.readListEnd(); } @@ -123528,9 +123660,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, alter_partitions_a oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size())); - for (Partition _iter1337 : struct.new_parts) + for (Partition _iter1361 : struct.new_parts) { - _iter1337.write(oprot); + _iter1361.write(oprot); } oprot.writeListEnd(); } @@ -123573,9 +123705,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, alter_partitions_ar if (struct.isSetNew_parts()) { { oprot.writeI32(struct.new_parts.size()); - for (Partition _iter1338 : struct.new_parts) + for (Partition _iter1362 : struct.new_parts) { - _iter1338.write(oprot); + _iter1362.write(oprot); } } } @@ -123595,14 +123727,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, alter_partitions_arg } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1339 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.new_parts = new ArrayList(_list1339.size); - Partition _elem1340; - for (int _i1341 = 0; _i1341 < _list1339.size; ++_i1341) + org.apache.thrift.protocol.TList _list1363 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.new_parts = new ArrayList(_list1363.size); + Partition _elem1364; + for (int _i1365 = 0; _i1365 < _list1363.size; ++_i1365) { - _elem1340 = new Partition(); - _elem1340.read(iprot); - struct.new_parts.add(_elem1340); + _elem1364 = new Partition(); + _elem1364.read(iprot); + struct.new_parts.add(_elem1364); } } struct.setNew_partsIsSet(true); @@ -124655,14 +124787,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, alter_partitions_wi case 3: // NEW_PARTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1342 = iprot.readListBegin(); - struct.new_parts = new ArrayList(_list1342.size); - Partition _elem1343; - for (int _i1344 = 0; _i1344 < _list1342.size; ++_i1344) + org.apache.thrift.protocol.TList _list1366 = iprot.readListBegin(); + struct.new_parts = new ArrayList(_list1366.size); + Partition _elem1367; + for (int _i1368 = 0; _i1368 < _list1366.size; ++_i1368) { - _elem1343 = new Partition(); - _elem1343.read(iprot); - struct.new_parts.add(_elem1343); + _elem1367 = new Partition(); + _elem1367.read(iprot); + struct.new_parts.add(_elem1367); } iprot.readListEnd(); } @@ -124707,9 +124839,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, alter_partitions_w oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size())); - for (Partition _iter1345 : struct.new_parts) + for (Partition _iter1369 : struct.new_parts) { - _iter1345.write(oprot); + _iter1369.write(oprot); } oprot.writeListEnd(); } @@ -124760,9 +124892,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, alter_partitions_wi if (struct.isSetNew_parts()) { { oprot.writeI32(struct.new_parts.size()); - for (Partition _iter1346 : struct.new_parts) + for (Partition _iter1370 : struct.new_parts) { - _iter1346.write(oprot); + _iter1370.write(oprot); } } } @@ -124785,14 +124917,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, alter_partitions_wit } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1347 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.new_parts = new ArrayList(_list1347.size); - Partition _elem1348; - for (int _i1349 = 0; _i1349 < _list1347.size; ++_i1349) + org.apache.thrift.protocol.TList _list1371 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.new_parts = new ArrayList(_list1371.size); + Partition _elem1372; + for (int _i1373 = 0; _i1373 < _list1371.size; ++_i1373) { - _elem1348 = new Partition(); - _elem1348.read(iprot); - struct.new_parts.add(_elem1348); + _elem1372 = new Partition(); + _elem1372.read(iprot); + struct.new_parts.add(_elem1372); } } struct.setNew_partsIsSet(true); @@ -126993,13 +127125,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, rename_partition_ar case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1350 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1350.size); - String _elem1351; - for (int _i1352 = 0; _i1352 < _list1350.size; ++_i1352) + org.apache.thrift.protocol.TList _list1374 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1374.size); + String _elem1375; + for (int _i1376 = 0; _i1376 < _list1374.size; ++_i1376) { - _elem1351 = iprot.readString(); - struct.part_vals.add(_elem1351); + _elem1375 = iprot.readString(); + struct.part_vals.add(_elem1375); } iprot.readListEnd(); } @@ -127044,9 +127176,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, rename_partition_a oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1353 : struct.part_vals) + for (String _iter1377 : struct.part_vals) { - oprot.writeString(_iter1353); + oprot.writeString(_iter1377); } oprot.writeListEnd(); } @@ -127097,9 +127229,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, rename_partition_ar if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1354 : struct.part_vals) + for (String _iter1378 : struct.part_vals) { - oprot.writeString(_iter1354); + oprot.writeString(_iter1378); } } } @@ -127122,13 +127254,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, rename_partition_arg } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1355 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1355.size); - String _elem1356; - for (int _i1357 = 0; _i1357 < _list1355.size; ++_i1357) + org.apache.thrift.protocol.TList _list1379 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1379.size); + String _elem1380; + for (int _i1381 = 0; _i1381 < _list1379.size; ++_i1381) { - _elem1356 = iprot.readString(); - struct.part_vals.add(_elem1356); + _elem1380 = iprot.readString(); + struct.part_vals.add(_elem1380); } } struct.setPart_valsIsSet(true); @@ -128002,13 +128134,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, partition_name_has_ case 1: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1358 = iprot.readListBegin(); - struct.part_vals = new ArrayList(_list1358.size); - String _elem1359; - for (int _i1360 = 0; _i1360 < _list1358.size; ++_i1360) + org.apache.thrift.protocol.TList _list1382 = iprot.readListBegin(); + struct.part_vals = new ArrayList(_list1382.size); + String _elem1383; + for (int _i1384 = 0; _i1384 < _list1382.size; ++_i1384) { - _elem1359 = iprot.readString(); - struct.part_vals.add(_elem1359); + _elem1383 = iprot.readString(); + struct.part_vals.add(_elem1383); } iprot.readListEnd(); } @@ -128042,9 +128174,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, partition_name_has oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter1361 : struct.part_vals) + for (String _iter1385 : struct.part_vals) { - oprot.writeString(_iter1361); + oprot.writeString(_iter1385); } oprot.writeListEnd(); } @@ -128081,9 +128213,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, partition_name_has_ if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (String _iter1362 : struct.part_vals) + for (String _iter1386 : struct.part_vals) { - oprot.writeString(_iter1362); + oprot.writeString(_iter1386); } } } @@ -128098,13 +128230,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, partition_name_has_v BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1363 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList(_list1363.size); - String _elem1364; - for (int _i1365 = 0; _i1365 < _list1363.size; ++_i1365) + org.apache.thrift.protocol.TList _list1387 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new ArrayList(_list1387.size); + String _elem1388; + for (int _i1389 = 0; _i1389 < _list1387.size; ++_i1389) { - _elem1364 = iprot.readString(); - struct.part_vals.add(_elem1364); + _elem1388 = iprot.readString(); + struct.part_vals.add(_elem1388); } } struct.setPart_valsIsSet(true); @@ -130259,13 +130391,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, partition_name_to_v case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1366 = iprot.readListBegin(); - struct.success = new ArrayList(_list1366.size); - String _elem1367; - for (int _i1368 = 0; _i1368 < _list1366.size; ++_i1368) + org.apache.thrift.protocol.TList _list1390 = iprot.readListBegin(); + struct.success = new ArrayList(_list1390.size); + String _elem1391; + for (int _i1392 = 0; _i1392 < _list1390.size; ++_i1392) { - _elem1367 = iprot.readString(); - struct.success.add(_elem1367); + _elem1391 = iprot.readString(); + struct.success.add(_elem1391); } iprot.readListEnd(); } @@ -130300,9 +130432,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, partition_name_to_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1369 : struct.success) + for (String _iter1393 : struct.success) { - oprot.writeString(_iter1369); + oprot.writeString(_iter1393); } oprot.writeListEnd(); } @@ -130341,9 +130473,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, partition_name_to_v if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1370 : struct.success) + for (String _iter1394 : struct.success) { - oprot.writeString(_iter1370); + oprot.writeString(_iter1394); } } } @@ -130358,13 +130490,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, partition_name_to_va BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1371 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1371.size); - String _elem1372; - for (int _i1373 = 0; _i1373 < _list1371.size; ++_i1373) + org.apache.thrift.protocol.TList _list1395 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1395.size); + String _elem1396; + for (int _i1397 = 0; _i1397 < _list1395.size; ++_i1397) { - _elem1372 = iprot.readString(); - struct.success.add(_elem1372); + _elem1396 = iprot.readString(); + struct.success.add(_elem1396); } } struct.setSuccessIsSet(true); @@ -131127,15 +131259,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, partition_name_to_s case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map1374 = iprot.readMapBegin(); - struct.success = new HashMap(2*_map1374.size); - String _key1375; - String _val1376; - for (int _i1377 = 0; _i1377 < _map1374.size; ++_i1377) + org.apache.thrift.protocol.TMap _map1398 = iprot.readMapBegin(); + struct.success = new HashMap(2*_map1398.size); + String _key1399; + String _val1400; + for (int _i1401 = 0; _i1401 < _map1398.size; ++_i1401) { - _key1375 = iprot.readString(); - _val1376 = iprot.readString(); - struct.success.put(_key1375, _val1376); + _key1399 = iprot.readString(); + _val1400 = iprot.readString(); + struct.success.put(_key1399, _val1400); } iprot.readMapEnd(); } @@ -131170,10 +131302,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, partition_name_to_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (Map.Entry _iter1378 : struct.success.entrySet()) + for (Map.Entry _iter1402 : struct.success.entrySet()) { - oprot.writeString(_iter1378.getKey()); - oprot.writeString(_iter1378.getValue()); + oprot.writeString(_iter1402.getKey()); + oprot.writeString(_iter1402.getValue()); } oprot.writeMapEnd(); } @@ -131212,10 +131344,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, partition_name_to_s if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Map.Entry _iter1379 : struct.success.entrySet()) + for (Map.Entry _iter1403 : struct.success.entrySet()) { - oprot.writeString(_iter1379.getKey()); - oprot.writeString(_iter1379.getValue()); + oprot.writeString(_iter1403.getKey()); + oprot.writeString(_iter1403.getValue()); } } } @@ -131230,15 +131362,15 @@ public void read(org.apache.thrift.protocol.TProtocol prot, partition_name_to_sp BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TMap _map1380 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new HashMap(2*_map1380.size); - String _key1381; - String _val1382; - for (int _i1383 = 0; _i1383 < _map1380.size; ++_i1383) + org.apache.thrift.protocol.TMap _map1404 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new HashMap(2*_map1404.size); + String _key1405; + String _val1406; + for (int _i1407 = 0; _i1407 < _map1404.size; ++_i1407) { - _key1381 = iprot.readString(); - _val1382 = iprot.readString(); - struct.success.put(_key1381, _val1382); + _key1405 = iprot.readString(); + _val1406 = iprot.readString(); + struct.success.put(_key1405, _val1406); } } struct.setSuccessIsSet(true); @@ -131833,15 +131965,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, markPartitionForEve case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map1384 = iprot.readMapBegin(); - struct.part_vals = new HashMap(2*_map1384.size); - String _key1385; - String _val1386; - for (int _i1387 = 0; _i1387 < _map1384.size; ++_i1387) + org.apache.thrift.protocol.TMap _map1408 = iprot.readMapBegin(); + struct.part_vals = new HashMap(2*_map1408.size); + String _key1409; + String _val1410; + for (int _i1411 = 0; _i1411 < _map1408.size; ++_i1411) { - _key1385 = iprot.readString(); - _val1386 = iprot.readString(); - struct.part_vals.put(_key1385, _val1386); + _key1409 = iprot.readString(); + _val1410 = iprot.readString(); + struct.part_vals.put(_key1409, _val1410); } iprot.readMapEnd(); } @@ -131885,10 +132017,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, markPartitionForEv oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (Map.Entry _iter1388 : struct.part_vals.entrySet()) + for (Map.Entry _iter1412 : struct.part_vals.entrySet()) { - oprot.writeString(_iter1388.getKey()); - oprot.writeString(_iter1388.getValue()); + oprot.writeString(_iter1412.getKey()); + oprot.writeString(_iter1412.getValue()); } oprot.writeMapEnd(); } @@ -131939,10 +132071,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, markPartitionForEve if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (Map.Entry _iter1389 : struct.part_vals.entrySet()) + for (Map.Entry _iter1413 : struct.part_vals.entrySet()) { - oprot.writeString(_iter1389.getKey()); - oprot.writeString(_iter1389.getValue()); + oprot.writeString(_iter1413.getKey()); + oprot.writeString(_iter1413.getValue()); } } } @@ -131965,15 +132097,15 @@ public void read(org.apache.thrift.protocol.TProtocol prot, markPartitionForEven } if (incoming.get(2)) { { - org.apache.thrift.protocol.TMap _map1390 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new HashMap(2*_map1390.size); - String _key1391; - String _val1392; - for (int _i1393 = 0; _i1393 < _map1390.size; ++_i1393) + org.apache.thrift.protocol.TMap _map1414 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new HashMap(2*_map1414.size); + String _key1415; + String _val1416; + for (int _i1417 = 0; _i1417 < _map1414.size; ++_i1417) { - _key1391 = iprot.readString(); - _val1392 = iprot.readString(); - struct.part_vals.put(_key1391, _val1392); + _key1415 = iprot.readString(); + _val1416 = iprot.readString(); + struct.part_vals.put(_key1415, _val1416); } } struct.setPart_valsIsSet(true); @@ -133457,15 +133589,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, isPartitionMarkedFo case 3: // PART_VALS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map1394 = iprot.readMapBegin(); - struct.part_vals = new HashMap(2*_map1394.size); - String _key1395; - String _val1396; - for (int _i1397 = 0; _i1397 < _map1394.size; ++_i1397) + org.apache.thrift.protocol.TMap _map1418 = iprot.readMapBegin(); + struct.part_vals = new HashMap(2*_map1418.size); + String _key1419; + String _val1420; + for (int _i1421 = 0; _i1421 < _map1418.size; ++_i1421) { - _key1395 = iprot.readString(); - _val1396 = iprot.readString(); - struct.part_vals.put(_key1395, _val1396); + _key1419 = iprot.readString(); + _val1420 = iprot.readString(); + struct.part_vals.put(_key1419, _val1420); } iprot.readMapEnd(); } @@ -133509,10 +133641,10 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, isPartitionMarkedF oprot.writeFieldBegin(PART_VALS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (Map.Entry _iter1398 : struct.part_vals.entrySet()) + for (Map.Entry _iter1422 : struct.part_vals.entrySet()) { - oprot.writeString(_iter1398.getKey()); - oprot.writeString(_iter1398.getValue()); + oprot.writeString(_iter1422.getKey()); + oprot.writeString(_iter1422.getValue()); } oprot.writeMapEnd(); } @@ -133563,10 +133695,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, isPartitionMarkedFo if (struct.isSetPart_vals()) { { oprot.writeI32(struct.part_vals.size()); - for (Map.Entry _iter1399 : struct.part_vals.entrySet()) + for (Map.Entry _iter1423 : struct.part_vals.entrySet()) { - oprot.writeString(_iter1399.getKey()); - oprot.writeString(_iter1399.getValue()); + oprot.writeString(_iter1423.getKey()); + oprot.writeString(_iter1423.getValue()); } } } @@ -133589,15 +133721,15 @@ public void read(org.apache.thrift.protocol.TProtocol prot, isPartitionMarkedFor } if (incoming.get(2)) { { - org.apache.thrift.protocol.TMap _map1400 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new HashMap(2*_map1400.size); - String _key1401; - String _val1402; - for (int _i1403 = 0; _i1403 < _map1400.size; ++_i1403) + org.apache.thrift.protocol.TMap _map1424 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.part_vals = new HashMap(2*_map1424.size); + String _key1425; + String _val1426; + for (int _i1427 = 0; _i1427 < _map1424.size; ++_i1427) { - _key1401 = iprot.readString(); - _val1402 = iprot.readString(); - struct.part_vals.put(_key1401, _val1402); + _key1425 = iprot.readString(); + _val1426 = iprot.readString(); + struct.part_vals.put(_key1425, _val1426); } } struct.setPart_valsIsSet(true); @@ -155953,13 +156085,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_functions_resul case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1404 = iprot.readListBegin(); - struct.success = new ArrayList(_list1404.size); - String _elem1405; - for (int _i1406 = 0; _i1406 < _list1404.size; ++_i1406) + org.apache.thrift.protocol.TList _list1428 = iprot.readListBegin(); + struct.success = new ArrayList(_list1428.size); + String _elem1429; + for (int _i1430 = 0; _i1430 < _list1428.size; ++_i1430) { - _elem1405 = iprot.readString(); - struct.success.add(_elem1405); + _elem1429 = iprot.readString(); + struct.success.add(_elem1429); } iprot.readListEnd(); } @@ -155994,9 +156126,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_functions_resu oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1407 : struct.success) + for (String _iter1431 : struct.success) { - oprot.writeString(_iter1407); + oprot.writeString(_iter1431); } oprot.writeListEnd(); } @@ -156035,9 +156167,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_functions_resul if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1408 : struct.success) + for (String _iter1432 : struct.success) { - oprot.writeString(_iter1408); + oprot.writeString(_iter1432); } } } @@ -156052,13 +156184,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_functions_result BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1409 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1409.size); - String _elem1410; - for (int _i1411 = 0; _i1411 < _list1409.size; ++_i1411) + org.apache.thrift.protocol.TList _list1433 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1433.size); + String _elem1434; + for (int _i1435 = 0; _i1435 < _list1433.size; ++_i1435) { - _elem1410 = iprot.readString(); - struct.success.add(_elem1410); + _elem1434 = iprot.readString(); + struct.success.add(_elem1434); } } struct.setSuccessIsSet(true); @@ -160113,13 +160245,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_role_names_resu case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1412 = iprot.readListBegin(); - struct.success = new ArrayList(_list1412.size); - String _elem1413; - for (int _i1414 = 0; _i1414 < _list1412.size; ++_i1414) + org.apache.thrift.protocol.TList _list1436 = iprot.readListBegin(); + struct.success = new ArrayList(_list1436.size); + String _elem1437; + for (int _i1438 = 0; _i1438 < _list1436.size; ++_i1438) { - _elem1413 = iprot.readString(); - struct.success.add(_elem1413); + _elem1437 = iprot.readString(); + struct.success.add(_elem1437); } iprot.readListEnd(); } @@ -160154,9 +160286,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_role_names_res oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1415 : struct.success) + for (String _iter1439 : struct.success) { - oprot.writeString(_iter1415); + oprot.writeString(_iter1439); } oprot.writeListEnd(); } @@ -160195,9 +160327,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_role_names_resu if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1416 : struct.success) + for (String _iter1440 : struct.success) { - oprot.writeString(_iter1416); + oprot.writeString(_iter1440); } } } @@ -160212,13 +160344,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_role_names_resul BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1417 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1417.size); - String _elem1418; - for (int _i1419 = 0; _i1419 < _list1417.size; ++_i1419) + org.apache.thrift.protocol.TList _list1441 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1441.size); + String _elem1442; + for (int _i1443 = 0; _i1443 < _list1441.size; ++_i1443) { - _elem1418 = iprot.readString(); - struct.success.add(_elem1418); + _elem1442 = iprot.readString(); + struct.success.add(_elem1442); } } struct.setSuccessIsSet(true); @@ -163509,14 +163641,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, list_roles_result s case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1420 = iprot.readListBegin(); - struct.success = new ArrayList(_list1420.size); - Role _elem1421; - for (int _i1422 = 0; _i1422 < _list1420.size; ++_i1422) + org.apache.thrift.protocol.TList _list1444 = iprot.readListBegin(); + struct.success = new ArrayList(_list1444.size); + Role _elem1445; + for (int _i1446 = 0; _i1446 < _list1444.size; ++_i1446) { - _elem1421 = new Role(); - _elem1421.read(iprot); - struct.success.add(_elem1421); + _elem1445 = new Role(); + _elem1445.read(iprot); + struct.success.add(_elem1445); } iprot.readListEnd(); } @@ -163551,9 +163683,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, list_roles_result oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Role _iter1423 : struct.success) + for (Role _iter1447 : struct.success) { - _iter1423.write(oprot); + _iter1447.write(oprot); } oprot.writeListEnd(); } @@ -163592,9 +163724,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, list_roles_result s if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Role _iter1424 : struct.success) + for (Role _iter1448 : struct.success) { - _iter1424.write(oprot); + _iter1448.write(oprot); } } } @@ -163609,14 +163741,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, list_roles_result st BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1425 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1425.size); - Role _elem1426; - for (int _i1427 = 0; _i1427 < _list1425.size; ++_i1427) + org.apache.thrift.protocol.TList _list1449 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1449.size); + Role _elem1450; + for (int _i1451 = 0; _i1451 < _list1449.size; ++_i1451) { - _elem1426 = new Role(); - _elem1426.read(iprot); - struct.success.add(_elem1426); + _elem1450 = new Role(); + _elem1450.read(iprot); + struct.success.add(_elem1450); } } struct.setSuccessIsSet(true); @@ -166621,13 +166753,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_privilege_set_a case 3: // GROUP_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1428 = iprot.readListBegin(); - struct.group_names = new ArrayList(_list1428.size); - String _elem1429; - for (int _i1430 = 0; _i1430 < _list1428.size; ++_i1430) + org.apache.thrift.protocol.TList _list1452 = iprot.readListBegin(); + struct.group_names = new ArrayList(_list1452.size); + String _elem1453; + for (int _i1454 = 0; _i1454 < _list1452.size; ++_i1454) { - _elem1429 = iprot.readString(); - struct.group_names.add(_elem1429); + _elem1453 = iprot.readString(); + struct.group_names.add(_elem1453); } iprot.readListEnd(); } @@ -166663,9 +166795,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_privilege_set_ oprot.writeFieldBegin(GROUP_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.group_names.size())); - for (String _iter1431 : struct.group_names) + for (String _iter1455 : struct.group_names) { - oprot.writeString(_iter1431); + oprot.writeString(_iter1455); } oprot.writeListEnd(); } @@ -166708,9 +166840,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_privilege_set_a if (struct.isSetGroup_names()) { { oprot.writeI32(struct.group_names.size()); - for (String _iter1432 : struct.group_names) + for (String _iter1456 : struct.group_names) { - oprot.writeString(_iter1432); + oprot.writeString(_iter1456); } } } @@ -166731,13 +166863,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_privilege_set_ar } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list1433 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.group_names = new ArrayList(_list1433.size); - String _elem1434; - for (int _i1435 = 0; _i1435 < _list1433.size; ++_i1435) + org.apache.thrift.protocol.TList _list1457 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.group_names = new ArrayList(_list1457.size); + String _elem1458; + for (int _i1459 = 0; _i1459 < _list1457.size; ++_i1459) { - _elem1434 = iprot.readString(); - struct.group_names.add(_elem1434); + _elem1458 = iprot.readString(); + struct.group_names.add(_elem1458); } } struct.setGroup_namesIsSet(true); @@ -168195,14 +168327,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, list_privileges_res case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1436 = iprot.readListBegin(); - struct.success = new ArrayList(_list1436.size); - HiveObjectPrivilege _elem1437; - for (int _i1438 = 0; _i1438 < _list1436.size; ++_i1438) + org.apache.thrift.protocol.TList _list1460 = iprot.readListBegin(); + struct.success = new ArrayList(_list1460.size); + HiveObjectPrivilege _elem1461; + for (int _i1462 = 0; _i1462 < _list1460.size; ++_i1462) { - _elem1437 = new HiveObjectPrivilege(); - _elem1437.read(iprot); - struct.success.add(_elem1437); + _elem1461 = new HiveObjectPrivilege(); + _elem1461.read(iprot); + struct.success.add(_elem1461); } iprot.readListEnd(); } @@ -168237,9 +168369,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, list_privileges_re oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (HiveObjectPrivilege _iter1439 : struct.success) + for (HiveObjectPrivilege _iter1463 : struct.success) { - _iter1439.write(oprot); + _iter1463.write(oprot); } oprot.writeListEnd(); } @@ -168278,9 +168410,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, list_privileges_res if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (HiveObjectPrivilege _iter1440 : struct.success) + for (HiveObjectPrivilege _iter1464 : struct.success) { - _iter1440.write(oprot); + _iter1464.write(oprot); } } } @@ -168295,14 +168427,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, list_privileges_resu BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1441 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1441.size); - HiveObjectPrivilege _elem1442; - for (int _i1443 = 0; _i1443 < _list1441.size; ++_i1443) + org.apache.thrift.protocol.TList _list1465 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1465.size); + HiveObjectPrivilege _elem1466; + for (int _i1467 = 0; _i1467 < _list1465.size; ++_i1467) { - _elem1442 = new HiveObjectPrivilege(); - _elem1442.read(iprot); - struct.success.add(_elem1442); + _elem1466 = new HiveObjectPrivilege(); + _elem1466.read(iprot); + struct.success.add(_elem1466); } } struct.setSuccessIsSet(true); @@ -172249,13 +172381,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, set_ugi_args struct case 2: // GROUP_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1444 = iprot.readListBegin(); - struct.group_names = new ArrayList(_list1444.size); - String _elem1445; - for (int _i1446 = 0; _i1446 < _list1444.size; ++_i1446) + org.apache.thrift.protocol.TList _list1468 = iprot.readListBegin(); + struct.group_names = new ArrayList(_list1468.size); + String _elem1469; + for (int _i1470 = 0; _i1470 < _list1468.size; ++_i1470) { - _elem1445 = iprot.readString(); - struct.group_names.add(_elem1445); + _elem1469 = iprot.readString(); + struct.group_names.add(_elem1469); } iprot.readListEnd(); } @@ -172286,9 +172418,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, set_ugi_args struc oprot.writeFieldBegin(GROUP_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.group_names.size())); - for (String _iter1447 : struct.group_names) + for (String _iter1471 : struct.group_names) { - oprot.writeString(_iter1447); + oprot.writeString(_iter1471); } oprot.writeListEnd(); } @@ -172325,9 +172457,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, set_ugi_args struct if (struct.isSetGroup_names()) { { oprot.writeI32(struct.group_names.size()); - for (String _iter1448 : struct.group_names) + for (String _iter1472 : struct.group_names) { - oprot.writeString(_iter1448); + oprot.writeString(_iter1472); } } } @@ -172343,13 +172475,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, set_ugi_args struct) } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list1449 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.group_names = new ArrayList(_list1449.size); - String _elem1450; - for (int _i1451 = 0; _i1451 < _list1449.size; ++_i1451) + org.apache.thrift.protocol.TList _list1473 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.group_names = new ArrayList(_list1473.size); + String _elem1474; + for (int _i1475 = 0; _i1475 < _list1473.size; ++_i1475) { - _elem1450 = iprot.readString(); - struct.group_names.add(_elem1450); + _elem1474 = iprot.readString(); + struct.group_names.add(_elem1474); } } struct.setGroup_namesIsSet(true); @@ -172752,13 +172884,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, set_ugi_result stru case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1452 = iprot.readListBegin(); - struct.success = new ArrayList(_list1452.size); - String _elem1453; - for (int _i1454 = 0; _i1454 < _list1452.size; ++_i1454) + org.apache.thrift.protocol.TList _list1476 = iprot.readListBegin(); + struct.success = new ArrayList(_list1476.size); + String _elem1477; + for (int _i1478 = 0; _i1478 < _list1476.size; ++_i1478) { - _elem1453 = iprot.readString(); - struct.success.add(_elem1453); + _elem1477 = iprot.readString(); + struct.success.add(_elem1477); } iprot.readListEnd(); } @@ -172793,9 +172925,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, set_ugi_result str oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1455 : struct.success) + for (String _iter1479 : struct.success) { - oprot.writeString(_iter1455); + oprot.writeString(_iter1479); } oprot.writeListEnd(); } @@ -172834,9 +172966,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, set_ugi_result stru if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1456 : struct.success) + for (String _iter1480 : struct.success) { - oprot.writeString(_iter1456); + oprot.writeString(_iter1480); } } } @@ -172851,13 +172983,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, set_ugi_result struc BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1457 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1457.size); - String _elem1458; - for (int _i1459 = 0; _i1459 < _list1457.size; ++_i1459) + org.apache.thrift.protocol.TList _list1481 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1481.size); + String _elem1482; + for (int _i1483 = 0; _i1483 < _list1481.size; ++_i1483) { - _elem1458 = iprot.readString(); - struct.success.add(_elem1458); + _elem1482 = iprot.readString(); + struct.success.add(_elem1482); } } struct.setSuccessIsSet(true); @@ -178148,13 +178280,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_all_token_ident case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1460 = iprot.readListBegin(); - struct.success = new ArrayList(_list1460.size); - String _elem1461; - for (int _i1462 = 0; _i1462 < _list1460.size; ++_i1462) + org.apache.thrift.protocol.TList _list1484 = iprot.readListBegin(); + struct.success = new ArrayList(_list1484.size); + String _elem1485; + for (int _i1486 = 0; _i1486 < _list1484.size; ++_i1486) { - _elem1461 = iprot.readString(); - struct.success.add(_elem1461); + _elem1485 = iprot.readString(); + struct.success.add(_elem1485); } iprot.readListEnd(); } @@ -178180,9 +178312,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_all_token_iden oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1463 : struct.success) + for (String _iter1487 : struct.success) { - oprot.writeString(_iter1463); + oprot.writeString(_iter1487); } oprot.writeListEnd(); } @@ -178213,9 +178345,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_all_token_ident if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1464 : struct.success) + for (String _iter1488 : struct.success) { - oprot.writeString(_iter1464); + oprot.writeString(_iter1488); } } } @@ -178227,13 +178359,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_all_token_identi BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1465 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1465.size); - String _elem1466; - for (int _i1467 = 0; _i1467 < _list1465.size; ++_i1467) + org.apache.thrift.protocol.TList _list1489 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1489.size); + String _elem1490; + for (int _i1491 = 0; _i1491 < _list1489.size; ++_i1491) { - _elem1466 = iprot.readString(); - struct.success.add(_elem1466); + _elem1490 = iprot.readString(); + struct.success.add(_elem1490); } } struct.setSuccessIsSet(true); @@ -181263,13 +181395,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_master_keys_res case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1468 = iprot.readListBegin(); - struct.success = new ArrayList(_list1468.size); - String _elem1469; - for (int _i1470 = 0; _i1470 < _list1468.size; ++_i1470) + org.apache.thrift.protocol.TList _list1492 = iprot.readListBegin(); + struct.success = new ArrayList(_list1492.size); + String _elem1493; + for (int _i1494 = 0; _i1494 < _list1492.size; ++_i1494) { - _elem1469 = iprot.readString(); - struct.success.add(_elem1469); + _elem1493 = iprot.readString(); + struct.success.add(_elem1493); } iprot.readListEnd(); } @@ -181295,9 +181427,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_master_keys_re oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter1471 : struct.success) + for (String _iter1495 : struct.success) { - oprot.writeString(_iter1471); + oprot.writeString(_iter1495); } oprot.writeListEnd(); } @@ -181328,9 +181460,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_master_keys_res if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter1472 : struct.success) + for (String _iter1496 : struct.success) { - oprot.writeString(_iter1472); + oprot.writeString(_iter1496); } } } @@ -181342,13 +181474,13 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_master_keys_resu BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1473 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList(_list1473.size); - String _elem1474; - for (int _i1475 = 0; _i1475 < _list1473.size; ++_i1475) + org.apache.thrift.protocol.TList _list1497 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList(_list1497.size); + String _elem1498; + for (int _i1499 = 0; _i1499 < _list1497.size; ++_i1499) { - _elem1474 = iprot.readString(); - struct.success.add(_elem1474); + _elem1498 = iprot.readString(); + struct.success.add(_elem1498); } } struct.setSuccessIsSet(true); @@ -199113,33 +199245,578 @@ public String getFieldName() { static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); metaDataMap = Collections.unmodifiableMap(tmpMap); - org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(flushCache_args.class, metaDataMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(flushCache_args.class, metaDataMap); + } + + public flushCache_args() { + } + + /** + * Performs a deep copy on other. + */ + public flushCache_args(flushCache_args other) { + } + + public flushCache_args deepCopy() { + return new flushCache_args(this); + } + + @Override + public void clear() { + } + + public void setFieldValue(_Fields field, Object value) { + switch (field) { + } + } + + public Object getFieldValue(_Fields field) { + switch (field) { + } + throw new IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new IllegalArgumentException(); + } + + switch (field) { + } + throw new IllegalStateException(); + } + + @Override + public boolean equals(Object that) { + if (that == null) + return false; + if (that instanceof flushCache_args) + return this.equals((flushCache_args)that); + return false; + } + + public boolean equals(flushCache_args that) { + if (that == null) + return false; + + return true; + } + + @Override + public int hashCode() { + List list = new ArrayList(); + + return list.hashCode(); + } + + @Override + public int compareTo(flushCache_args other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + return 0; + } + + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + schemes.get(iprot.getScheme()).getScheme().read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + schemes.get(oprot.getScheme()).getScheme().write(oprot, this); + } + + @Override + public String toString() { + StringBuilder sb = new StringBuilder("flushCache_args("); + boolean first = true; + + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException { + try { + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class flushCache_argsStandardSchemeFactory implements SchemeFactory { + public flushCache_argsStandardScheme getScheme() { + return new flushCache_argsStandardScheme(); + } + } + + private static class flushCache_argsStandardScheme extends StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_args struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, flushCache_args struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class flushCache_argsTupleSchemeFactory implements SchemeFactory { + public flushCache_argsTupleScheme getScheme() { + return new flushCache_argsTupleScheme(); + } + } + + private static class flushCache_argsTupleScheme extends TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, flushCache_args struct) throws org.apache.thrift.TException { + TTupleProtocol oprot = (TTupleProtocol) prot; + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, flushCache_args struct) throws org.apache.thrift.TException { + TTupleProtocol iprot = (TTupleProtocol) prot; + } + } + + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class flushCache_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("flushCache_result"); + + + private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); + static { + schemes.put(StandardScheme.class, new flushCache_resultStandardSchemeFactory()); + schemes.put(TupleScheme.class, new flushCache_resultTupleSchemeFactory()); + } + + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { +; + + private static final Map byName = new HashMap(); + + static { + for (_Fields field : EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + public static _Fields findByName(String name) { + return byName.get(name); + } + + private final short _thriftId; + private final String _fieldName; + + _Fields(short thriftId, String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public String getFieldName() { + return _fieldName; + } + } + public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + metaDataMap = Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(flushCache_result.class, metaDataMap); + } + + public flushCache_result() { + } + + /** + * Performs a deep copy on other. + */ + public flushCache_result(flushCache_result other) { + } + + public flushCache_result deepCopy() { + return new flushCache_result(this); + } + + @Override + public void clear() { + } + + public void setFieldValue(_Fields field, Object value) { + switch (field) { + } + } + + public Object getFieldValue(_Fields field) { + switch (field) { + } + throw new IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new IllegalArgumentException(); + } + + switch (field) { + } + throw new IllegalStateException(); + } + + @Override + public boolean equals(Object that) { + if (that == null) + return false; + if (that instanceof flushCache_result) + return this.equals((flushCache_result)that); + return false; + } + + public boolean equals(flushCache_result that) { + if (that == null) + return false; + + return true; + } + + @Override + public int hashCode() { + List list = new ArrayList(); + + return list.hashCode(); + } + + @Override + public int compareTo(flushCache_result other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + return 0; + } + + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + schemes.get(iprot.getScheme()).getScheme().read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + schemes.get(oprot.getScheme()).getScheme().write(oprot, this); + } + + @Override + public String toString() { + StringBuilder sb = new StringBuilder("flushCache_result("); + boolean first = true; + + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException { + try { + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class flushCache_resultStandardSchemeFactory implements SchemeFactory { + public flushCache_resultStandardScheme getScheme() { + return new flushCache_resultStandardScheme(); + } + } + + private static class flushCache_resultStandardScheme extends StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_result struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, flushCache_result struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class flushCache_resultTupleSchemeFactory implements SchemeFactory { + public flushCache_resultTupleScheme getScheme() { + return new flushCache_resultTupleScheme(); + } + } + + private static class flushCache_resultTupleScheme extends TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, flushCache_result struct) throws org.apache.thrift.TException { + TTupleProtocol oprot = (TTupleProtocol) prot; + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, flushCache_result struct) throws org.apache.thrift.TException { + TTupleProtocol iprot = (TTupleProtocol) prot; + } + } + + } + + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_write_notification_log_args implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_write_notification_log_args"); + + private static final org.apache.thrift.protocol.TField RQST_FIELD_DESC = new org.apache.thrift.protocol.TField("rqst", org.apache.thrift.protocol.TType.STRUCT, (short)-1); + + private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); + static { + schemes.put(StandardScheme.class, new add_write_notification_log_argsStandardSchemeFactory()); + schemes.put(TupleScheme.class, new add_write_notification_log_argsTupleSchemeFactory()); + } + + private WriteNotificationLogRequest rqst; // required + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + RQST((short)-1, "rqst"); + + private static final Map byName = new HashMap(); + + static { + for (_Fields field : EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case -1: // RQST + return RQST; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + public static _Fields findByName(String name) { + return byName.get(name); + } + + private final short _thriftId; + private final String _fieldName; + + _Fields(short thriftId, String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.RQST, new org.apache.thrift.meta_data.FieldMetaData("rqst", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, WriteNotificationLogRequest.class))); + metaDataMap = Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(add_write_notification_log_args.class, metaDataMap); } - public flushCache_args() { + public add_write_notification_log_args() { + } + + public add_write_notification_log_args( + WriteNotificationLogRequest rqst) + { + this(); + this.rqst = rqst; } /** * Performs a deep copy on other. */ - public flushCache_args(flushCache_args other) { + public add_write_notification_log_args(add_write_notification_log_args other) { + if (other.isSetRqst()) { + this.rqst = new WriteNotificationLogRequest(other.rqst); + } } - public flushCache_args deepCopy() { - return new flushCache_args(this); + public add_write_notification_log_args deepCopy() { + return new add_write_notification_log_args(this); } @Override public void clear() { + this.rqst = null; + } + + public WriteNotificationLogRequest getRqst() { + return this.rqst; + } + + public void setRqst(WriteNotificationLogRequest rqst) { + this.rqst = rqst; + } + + public void unsetRqst() { + this.rqst = null; + } + + /** Returns true if field rqst is set (has been assigned a value) and false otherwise */ + public boolean isSetRqst() { + return this.rqst != null; + } + + public void setRqstIsSet(boolean value) { + if (!value) { + this.rqst = null; + } } public void setFieldValue(_Fields field, Object value) { switch (field) { + case RQST: + if (value == null) { + unsetRqst(); + } else { + setRqst((WriteNotificationLogRequest)value); + } + break; + } } public Object getFieldValue(_Fields field) { switch (field) { + case RQST: + return getRqst(); + } throw new IllegalStateException(); } @@ -199151,6 +199828,8 @@ public boolean isSet(_Fields field) { } switch (field) { + case RQST: + return isSetRqst(); } throw new IllegalStateException(); } @@ -199159,15 +199838,24 @@ public boolean isSet(_Fields field) { public boolean equals(Object that) { if (that == null) return false; - if (that instanceof flushCache_args) - return this.equals((flushCache_args)that); + if (that instanceof add_write_notification_log_args) + return this.equals((add_write_notification_log_args)that); return false; } - public boolean equals(flushCache_args that) { + public boolean equals(add_write_notification_log_args that) { if (that == null) return false; + boolean this_present_rqst = true && this.isSetRqst(); + boolean that_present_rqst = true && that.isSetRqst(); + if (this_present_rqst || that_present_rqst) { + if (!(this_present_rqst && that_present_rqst)) + return false; + if (!this.rqst.equals(that.rqst)) + return false; + } + return true; } @@ -199175,17 +199863,32 @@ public boolean equals(flushCache_args that) { public int hashCode() { List list = new ArrayList(); + boolean present_rqst = true && (isSetRqst()); + list.add(present_rqst); + if (present_rqst) + list.add(rqst); + return list.hashCode(); } @Override - public int compareTo(flushCache_args other) { + public int compareTo(add_write_notification_log_args other) { if (!getClass().equals(other.getClass())) { return getClass().getName().compareTo(other.getClass().getName()); } int lastComparison = 0; + lastComparison = Boolean.valueOf(isSetRqst()).compareTo(other.isSetRqst()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetRqst()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.rqst, other.rqst); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -199203,9 +199906,16 @@ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache. @Override public String toString() { - StringBuilder sb = new StringBuilder("flushCache_args("); + StringBuilder sb = new StringBuilder("add_write_notification_log_args("); boolean first = true; + sb.append("rqst:"); + if (this.rqst == null) { + sb.append("null"); + } else { + sb.append(this.rqst); + } + first = false; sb.append(")"); return sb.toString(); } @@ -199213,6 +199923,9 @@ public String toString() { public void validate() throws org.apache.thrift.TException { // check for required fields // check for sub-struct validity + if (rqst != null) { + rqst.validate(); + } } private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { @@ -199231,15 +199944,15 @@ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException } } - private static class flushCache_argsStandardSchemeFactory implements SchemeFactory { - public flushCache_argsStandardScheme getScheme() { - return new flushCache_argsStandardScheme(); + private static class add_write_notification_log_argsStandardSchemeFactory implements SchemeFactory { + public add_write_notification_log_argsStandardScheme getScheme() { + return new add_write_notification_log_argsStandardScheme(); } } - private static class flushCache_argsStandardScheme extends StandardScheme { + private static class add_write_notification_log_argsStandardScheme extends StandardScheme { - public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_args struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol iprot, add_write_notification_log_args struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TField schemeField; iprot.readStructBegin(); while (true) @@ -199249,6 +199962,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_args str break; } switch (schemeField.id) { + case -1: // RQST + if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) { + struct.rqst = new WriteNotificationLogRequest(); + struct.rqst.read(iprot); + struct.setRqstIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -199258,51 +199980,72 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_args str struct.validate(); } - public void write(org.apache.thrift.protocol.TProtocol oprot, flushCache_args struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol oprot, add_write_notification_log_args struct) throws org.apache.thrift.TException { struct.validate(); oprot.writeStructBegin(STRUCT_DESC); + if (struct.rqst != null) { + oprot.writeFieldBegin(RQST_FIELD_DESC); + struct.rqst.write(oprot); + oprot.writeFieldEnd(); + } oprot.writeFieldStop(); oprot.writeStructEnd(); } } - private static class flushCache_argsTupleSchemeFactory implements SchemeFactory { - public flushCache_argsTupleScheme getScheme() { - return new flushCache_argsTupleScheme(); + private static class add_write_notification_log_argsTupleSchemeFactory implements SchemeFactory { + public add_write_notification_log_argsTupleScheme getScheme() { + return new add_write_notification_log_argsTupleScheme(); } } - private static class flushCache_argsTupleScheme extends TupleScheme { + private static class add_write_notification_log_argsTupleScheme extends TupleScheme { @Override - public void write(org.apache.thrift.protocol.TProtocol prot, flushCache_args struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol prot, add_write_notification_log_args struct) throws org.apache.thrift.TException { TTupleProtocol oprot = (TTupleProtocol) prot; + BitSet optionals = new BitSet(); + if (struct.isSetRqst()) { + optionals.set(0); + } + oprot.writeBitSet(optionals, 1); + if (struct.isSetRqst()) { + struct.rqst.write(oprot); + } } @Override - public void read(org.apache.thrift.protocol.TProtocol prot, flushCache_args struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol prot, add_write_notification_log_args struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; + BitSet incoming = iprot.readBitSet(1); + if (incoming.get(0)) { + struct.rqst = new WriteNotificationLogRequest(); + struct.rqst.read(iprot); + struct.setRqstIsSet(true); + } } } } - @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class flushCache_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { - private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("flushCache_result"); + @org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public static class add_write_notification_log_result implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_write_notification_log_result"); + private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRUCT, (short)0); private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); static { - schemes.put(StandardScheme.class, new flushCache_resultStandardSchemeFactory()); - schemes.put(TupleScheme.class, new flushCache_resultTupleSchemeFactory()); + schemes.put(StandardScheme.class, new add_write_notification_log_resultStandardSchemeFactory()); + schemes.put(TupleScheme.class, new add_write_notification_log_resultTupleSchemeFactory()); } + private WriteNotificationLogResponse success; // required /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { -; + SUCCESS((short)0, "success"); private static final Map byName = new HashMap(); @@ -199317,6 +200060,8 @@ public void read(org.apache.thrift.protocol.TProtocol prot, flushCache_args stru */ public static _Fields findByThriftId(int fieldId) { switch(fieldId) { + case 0: // SUCCESS + return SUCCESS; default: return null; } @@ -199355,37 +200100,86 @@ public String getFieldName() { return _fieldName; } } + + // isset id assignments public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT, + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, WriteNotificationLogResponse.class))); metaDataMap = Collections.unmodifiableMap(tmpMap); - org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(flushCache_result.class, metaDataMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(add_write_notification_log_result.class, metaDataMap); } - public flushCache_result() { + public add_write_notification_log_result() { + } + + public add_write_notification_log_result( + WriteNotificationLogResponse success) + { + this(); + this.success = success; } /** * Performs a deep copy on other. */ - public flushCache_result(flushCache_result other) { + public add_write_notification_log_result(add_write_notification_log_result other) { + if (other.isSetSuccess()) { + this.success = new WriteNotificationLogResponse(other.success); + } } - public flushCache_result deepCopy() { - return new flushCache_result(this); + public add_write_notification_log_result deepCopy() { + return new add_write_notification_log_result(this); } @Override public void clear() { + this.success = null; + } + + public WriteNotificationLogResponse getSuccess() { + return this.success; + } + + public void setSuccess(WriteNotificationLogResponse success) { + this.success = success; + } + + public void unsetSuccess() { + this.success = null; + } + + /** Returns true if field success is set (has been assigned a value) and false otherwise */ + public boolean isSetSuccess() { + return this.success != null; + } + + public void setSuccessIsSet(boolean value) { + if (!value) { + this.success = null; + } } public void setFieldValue(_Fields field, Object value) { switch (field) { + case SUCCESS: + if (value == null) { + unsetSuccess(); + } else { + setSuccess((WriteNotificationLogResponse)value); + } + break; + } } public Object getFieldValue(_Fields field) { switch (field) { + case SUCCESS: + return getSuccess(); + } throw new IllegalStateException(); } @@ -199397,6 +200191,8 @@ public boolean isSet(_Fields field) { } switch (field) { + case SUCCESS: + return isSetSuccess(); } throw new IllegalStateException(); } @@ -199405,15 +200201,24 @@ public boolean isSet(_Fields field) { public boolean equals(Object that) { if (that == null) return false; - if (that instanceof flushCache_result) - return this.equals((flushCache_result)that); + if (that instanceof add_write_notification_log_result) + return this.equals((add_write_notification_log_result)that); return false; } - public boolean equals(flushCache_result that) { + public boolean equals(add_write_notification_log_result that) { if (that == null) return false; + boolean this_present_success = true && this.isSetSuccess(); + boolean that_present_success = true && that.isSetSuccess(); + if (this_present_success || that_present_success) { + if (!(this_present_success && that_present_success)) + return false; + if (!this.success.equals(that.success)) + return false; + } + return true; } @@ -199421,17 +200226,32 @@ public boolean equals(flushCache_result that) { public int hashCode() { List list = new ArrayList(); + boolean present_success = true && (isSetSuccess()); + list.add(present_success); + if (present_success) + list.add(success); + return list.hashCode(); } @Override - public int compareTo(flushCache_result other) { + public int compareTo(add_write_notification_log_result other) { if (!getClass().equals(other.getClass())) { return getClass().getName().compareTo(other.getClass().getName()); } int lastComparison = 0; + lastComparison = Boolean.valueOf(isSetSuccess()).compareTo(other.isSetSuccess()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetSuccess()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.success, other.success); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -199449,9 +200269,16 @@ public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache. @Override public String toString() { - StringBuilder sb = new StringBuilder("flushCache_result("); + StringBuilder sb = new StringBuilder("add_write_notification_log_result("); boolean first = true; + sb.append("success:"); + if (this.success == null) { + sb.append("null"); + } else { + sb.append(this.success); + } + first = false; sb.append(")"); return sb.toString(); } @@ -199459,6 +200286,9 @@ public String toString() { public void validate() throws org.apache.thrift.TException { // check for required fields // check for sub-struct validity + if (success != null) { + success.validate(); + } } private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { @@ -199477,15 +200307,15 @@ private void readObject(java.io.ObjectInputStream in) throws java.io.IOException } } - private static class flushCache_resultStandardSchemeFactory implements SchemeFactory { - public flushCache_resultStandardScheme getScheme() { - return new flushCache_resultStandardScheme(); + private static class add_write_notification_log_resultStandardSchemeFactory implements SchemeFactory { + public add_write_notification_log_resultStandardScheme getScheme() { + return new add_write_notification_log_resultStandardScheme(); } } - private static class flushCache_resultStandardScheme extends StandardScheme { + private static class add_write_notification_log_resultStandardScheme extends StandardScheme { - public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_result struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol iprot, add_write_notification_log_result struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TField schemeField; iprot.readStructBegin(); while (true) @@ -199495,6 +200325,15 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_result s break; } switch (schemeField.id) { + case 0: // SUCCESS + if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) { + struct.success = new WriteNotificationLogResponse(); + struct.success.read(iprot); + struct.setSuccessIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -199504,32 +200343,51 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, flushCache_result s struct.validate(); } - public void write(org.apache.thrift.protocol.TProtocol oprot, flushCache_result struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol oprot, add_write_notification_log_result struct) throws org.apache.thrift.TException { struct.validate(); oprot.writeStructBegin(STRUCT_DESC); + if (struct.success != null) { + oprot.writeFieldBegin(SUCCESS_FIELD_DESC); + struct.success.write(oprot); + oprot.writeFieldEnd(); + } oprot.writeFieldStop(); oprot.writeStructEnd(); } } - private static class flushCache_resultTupleSchemeFactory implements SchemeFactory { - public flushCache_resultTupleScheme getScheme() { - return new flushCache_resultTupleScheme(); + private static class add_write_notification_log_resultTupleSchemeFactory implements SchemeFactory { + public add_write_notification_log_resultTupleScheme getScheme() { + return new add_write_notification_log_resultTupleScheme(); } } - private static class flushCache_resultTupleScheme extends TupleScheme { + private static class add_write_notification_log_resultTupleScheme extends TupleScheme { @Override - public void write(org.apache.thrift.protocol.TProtocol prot, flushCache_result struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol prot, add_write_notification_log_result struct) throws org.apache.thrift.TException { TTupleProtocol oprot = (TTupleProtocol) prot; + BitSet optionals = new BitSet(); + if (struct.isSetSuccess()) { + optionals.set(0); + } + oprot.writeBitSet(optionals, 1); + if (struct.isSetSuccess()) { + struct.success.write(oprot); + } } @Override - public void read(org.apache.thrift.protocol.TProtocol prot, flushCache_result struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol prot, add_write_notification_log_result struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; + BitSet incoming = iprot.readBitSet(1); + if (incoming.get(0)) { + struct.success = new WriteNotificationLogResponse(); + struct.success.read(iprot); + struct.setSuccessIsSet(true); + } } } @@ -229531,14 +230389,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_schema_all_vers case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1476 = iprot.readListBegin(); - struct.success = new ArrayList(_list1476.size); - SchemaVersion _elem1477; - for (int _i1478 = 0; _i1478 < _list1476.size; ++_i1478) + org.apache.thrift.protocol.TList _list1500 = iprot.readListBegin(); + struct.success = new ArrayList(_list1500.size); + SchemaVersion _elem1501; + for (int _i1502 = 0; _i1502 < _list1500.size; ++_i1502) { - _elem1477 = new SchemaVersion(); - _elem1477.read(iprot); - struct.success.add(_elem1477); + _elem1501 = new SchemaVersion(); + _elem1501.read(iprot); + struct.success.add(_elem1501); } iprot.readListEnd(); } @@ -229582,9 +230440,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_schema_all_ver oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (SchemaVersion _iter1479 : struct.success) + for (SchemaVersion _iter1503 : struct.success) { - _iter1479.write(oprot); + _iter1503.write(oprot); } oprot.writeListEnd(); } @@ -229631,9 +230489,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_schema_all_vers if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (SchemaVersion _iter1480 : struct.success) + for (SchemaVersion _iter1504 : struct.success) { - _iter1480.write(oprot); + _iter1504.write(oprot); } } } @@ -229651,14 +230509,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_schema_all_versi BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1481 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1481.size); - SchemaVersion _elem1482; - for (int _i1483 = 0; _i1483 < _list1481.size; ++_i1483) + org.apache.thrift.protocol.TList _list1505 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1505.size); + SchemaVersion _elem1506; + for (int _i1507 = 0; _i1507 < _list1505.size; ++_i1507) { - _elem1482 = new SchemaVersion(); - _elem1482.read(iprot); - struct.success.add(_elem1482); + _elem1506 = new SchemaVersion(); + _elem1506.read(iprot); + struct.success.add(_elem1506); } } struct.setSuccessIsSet(true); @@ -238201,14 +239059,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, get_runtime_stats_r case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list1484 = iprot.readListBegin(); - struct.success = new ArrayList(_list1484.size); - RuntimeStat _elem1485; - for (int _i1486 = 0; _i1486 < _list1484.size; ++_i1486) + org.apache.thrift.protocol.TList _list1508 = iprot.readListBegin(); + struct.success = new ArrayList(_list1508.size); + RuntimeStat _elem1509; + for (int _i1510 = 0; _i1510 < _list1508.size; ++_i1510) { - _elem1485 = new RuntimeStat(); - _elem1485.read(iprot); - struct.success.add(_elem1485); + _elem1509 = new RuntimeStat(); + _elem1509.read(iprot); + struct.success.add(_elem1509); } iprot.readListEnd(); } @@ -238243,9 +239101,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, get_runtime_stats_ oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (RuntimeStat _iter1487 : struct.success) + for (RuntimeStat _iter1511 : struct.success) { - _iter1487.write(oprot); + _iter1511.write(oprot); } oprot.writeListEnd(); } @@ -238284,9 +239142,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, get_runtime_stats_r if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (RuntimeStat _iter1488 : struct.success) + for (RuntimeStat _iter1512 : struct.success) { - _iter1488.write(oprot); + _iter1512.write(oprot); } } } @@ -238301,14 +239159,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, get_runtime_stats_re BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list1489 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList(_list1489.size); - RuntimeStat _elem1490; - for (int _i1491 = 0; _i1491 < _list1489.size; ++_i1491) + org.apache.thrift.protocol.TList _list1513 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList(_list1513.size); + RuntimeStat _elem1514; + for (int _i1515 = 0; _i1515 < _list1513.size; ++_i1515) { - _elem1490 = new RuntimeStat(); - _elem1490.read(iprot); - struct.success.add(_elem1490); + _elem1514 = new RuntimeStat(); + _elem1514.read(iprot); + struct.success.add(_elem1514); } } struct.setSuccessIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMFullResourcePlan.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMFullResourcePlan.java index f4e30f037e..f0c308d7a3 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMFullResourcePlan.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMFullResourcePlan.java @@ -755,14 +755,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMFullResourcePlan case 2: // POOLS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list840 = iprot.readListBegin(); - struct.pools = new ArrayList(_list840.size); - WMPool _elem841; - for (int _i842 = 0; _i842 < _list840.size; ++_i842) + org.apache.thrift.protocol.TList _list864 = iprot.readListBegin(); + struct.pools = new ArrayList(_list864.size); + WMPool _elem865; + for (int _i866 = 0; _i866 < _list864.size; ++_i866) { - _elem841 = new WMPool(); - _elem841.read(iprot); - struct.pools.add(_elem841); + _elem865 = new WMPool(); + _elem865.read(iprot); + struct.pools.add(_elem865); } iprot.readListEnd(); } @@ -774,14 +774,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMFullResourcePlan case 3: // MAPPINGS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list843 = iprot.readListBegin(); - struct.mappings = new ArrayList(_list843.size); - WMMapping _elem844; - for (int _i845 = 0; _i845 < _list843.size; ++_i845) + org.apache.thrift.protocol.TList _list867 = iprot.readListBegin(); + struct.mappings = new ArrayList(_list867.size); + WMMapping _elem868; + for (int _i869 = 0; _i869 < _list867.size; ++_i869) { - _elem844 = new WMMapping(); - _elem844.read(iprot); - struct.mappings.add(_elem844); + _elem868 = new WMMapping(); + _elem868.read(iprot); + struct.mappings.add(_elem868); } iprot.readListEnd(); } @@ -793,14 +793,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMFullResourcePlan case 4: // TRIGGERS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list846 = iprot.readListBegin(); - struct.triggers = new ArrayList(_list846.size); - WMTrigger _elem847; - for (int _i848 = 0; _i848 < _list846.size; ++_i848) + org.apache.thrift.protocol.TList _list870 = iprot.readListBegin(); + struct.triggers = new ArrayList(_list870.size); + WMTrigger _elem871; + for (int _i872 = 0; _i872 < _list870.size; ++_i872) { - _elem847 = new WMTrigger(); - _elem847.read(iprot); - struct.triggers.add(_elem847); + _elem871 = new WMTrigger(); + _elem871.read(iprot); + struct.triggers.add(_elem871); } iprot.readListEnd(); } @@ -812,14 +812,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMFullResourcePlan case 5: // POOL_TRIGGERS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list849 = iprot.readListBegin(); - struct.poolTriggers = new ArrayList(_list849.size); - WMPoolTrigger _elem850; - for (int _i851 = 0; _i851 < _list849.size; ++_i851) + org.apache.thrift.protocol.TList _list873 = iprot.readListBegin(); + struct.poolTriggers = new ArrayList(_list873.size); + WMPoolTrigger _elem874; + for (int _i875 = 0; _i875 < _list873.size; ++_i875) { - _elem850 = new WMPoolTrigger(); - _elem850.read(iprot); - struct.poolTriggers.add(_elem850); + _elem874 = new WMPoolTrigger(); + _elem874.read(iprot); + struct.poolTriggers.add(_elem874); } iprot.readListEnd(); } @@ -850,9 +850,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMFullResourcePlan oprot.writeFieldBegin(POOLS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.pools.size())); - for (WMPool _iter852 : struct.pools) + for (WMPool _iter876 : struct.pools) { - _iter852.write(oprot); + _iter876.write(oprot); } oprot.writeListEnd(); } @@ -863,9 +863,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMFullResourcePlan oprot.writeFieldBegin(MAPPINGS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.mappings.size())); - for (WMMapping _iter853 : struct.mappings) + for (WMMapping _iter877 : struct.mappings) { - _iter853.write(oprot); + _iter877.write(oprot); } oprot.writeListEnd(); } @@ -877,9 +877,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMFullResourcePlan oprot.writeFieldBegin(TRIGGERS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.triggers.size())); - for (WMTrigger _iter854 : struct.triggers) + for (WMTrigger _iter878 : struct.triggers) { - _iter854.write(oprot); + _iter878.write(oprot); } oprot.writeListEnd(); } @@ -891,9 +891,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMFullResourcePlan oprot.writeFieldBegin(POOL_TRIGGERS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.poolTriggers.size())); - for (WMPoolTrigger _iter855 : struct.poolTriggers) + for (WMPoolTrigger _iter879 : struct.poolTriggers) { - _iter855.write(oprot); + _iter879.write(oprot); } oprot.writeListEnd(); } @@ -920,9 +920,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, WMFullResourcePlan struct.plan.write(oprot); { oprot.writeI32(struct.pools.size()); - for (WMPool _iter856 : struct.pools) + for (WMPool _iter880 : struct.pools) { - _iter856.write(oprot); + _iter880.write(oprot); } } BitSet optionals = new BitSet(); @@ -939,27 +939,27 @@ public void write(org.apache.thrift.protocol.TProtocol prot, WMFullResourcePlan if (struct.isSetMappings()) { { oprot.writeI32(struct.mappings.size()); - for (WMMapping _iter857 : struct.mappings) + for (WMMapping _iter881 : struct.mappings) { - _iter857.write(oprot); + _iter881.write(oprot); } } } if (struct.isSetTriggers()) { { oprot.writeI32(struct.triggers.size()); - for (WMTrigger _iter858 : struct.triggers) + for (WMTrigger _iter882 : struct.triggers) { - _iter858.write(oprot); + _iter882.write(oprot); } } } if (struct.isSetPoolTriggers()) { { oprot.writeI32(struct.poolTriggers.size()); - for (WMPoolTrigger _iter859 : struct.poolTriggers) + for (WMPoolTrigger _iter883 : struct.poolTriggers) { - _iter859.write(oprot); + _iter883.write(oprot); } } } @@ -972,56 +972,56 @@ public void read(org.apache.thrift.protocol.TProtocol prot, WMFullResourcePlan s struct.plan.read(iprot); struct.setPlanIsSet(true); { - org.apache.thrift.protocol.TList _list860 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.pools = new ArrayList(_list860.size); - WMPool _elem861; - for (int _i862 = 0; _i862 < _list860.size; ++_i862) + org.apache.thrift.protocol.TList _list884 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.pools = new ArrayList(_list884.size); + WMPool _elem885; + for (int _i886 = 0; _i886 < _list884.size; ++_i886) { - _elem861 = new WMPool(); - _elem861.read(iprot); - struct.pools.add(_elem861); + _elem885 = new WMPool(); + _elem885.read(iprot); + struct.pools.add(_elem885); } } struct.setPoolsIsSet(true); BitSet incoming = iprot.readBitSet(3); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list863 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.mappings = new ArrayList(_list863.size); - WMMapping _elem864; - for (int _i865 = 0; _i865 < _list863.size; ++_i865) + org.apache.thrift.protocol.TList _list887 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.mappings = new ArrayList(_list887.size); + WMMapping _elem888; + for (int _i889 = 0; _i889 < _list887.size; ++_i889) { - _elem864 = new WMMapping(); - _elem864.read(iprot); - struct.mappings.add(_elem864); + _elem888 = new WMMapping(); + _elem888.read(iprot); + struct.mappings.add(_elem888); } } struct.setMappingsIsSet(true); } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list866 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.triggers = new ArrayList(_list866.size); - WMTrigger _elem867; - for (int _i868 = 0; _i868 < _list866.size; ++_i868) + org.apache.thrift.protocol.TList _list890 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.triggers = new ArrayList(_list890.size); + WMTrigger _elem891; + for (int _i892 = 0; _i892 < _list890.size; ++_i892) { - _elem867 = new WMTrigger(); - _elem867.read(iprot); - struct.triggers.add(_elem867); + _elem891 = new WMTrigger(); + _elem891.read(iprot); + struct.triggers.add(_elem891); } } struct.setTriggersIsSet(true); } if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list869 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.poolTriggers = new ArrayList(_list869.size); - WMPoolTrigger _elem870; - for (int _i871 = 0; _i871 < _list869.size; ++_i871) + org.apache.thrift.protocol.TList _list893 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.poolTriggers = new ArrayList(_list893.size); + WMPoolTrigger _elem894; + for (int _i895 = 0; _i895 < _list893.size; ++_i895) { - _elem870 = new WMPoolTrigger(); - _elem870.read(iprot); - struct.poolTriggers.add(_elem870); + _elem894 = new WMPoolTrigger(); + _elem894.read(iprot); + struct.poolTriggers.add(_elem894); } } struct.setPoolTriggersIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetAllResourcePlanResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetAllResourcePlanResponse.java index ba81ce9684..6eed84b222 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetAllResourcePlanResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetAllResourcePlanResponse.java @@ -346,14 +346,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMGetAllResourcePla case 1: // RESOURCE_PLANS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list872 = iprot.readListBegin(); - struct.resourcePlans = new ArrayList(_list872.size); - WMResourcePlan _elem873; - for (int _i874 = 0; _i874 < _list872.size; ++_i874) + org.apache.thrift.protocol.TList _list896 = iprot.readListBegin(); + struct.resourcePlans = new ArrayList(_list896.size); + WMResourcePlan _elem897; + for (int _i898 = 0; _i898 < _list896.size; ++_i898) { - _elem873 = new WMResourcePlan(); - _elem873.read(iprot); - struct.resourcePlans.add(_elem873); + _elem897 = new WMResourcePlan(); + _elem897.read(iprot); + struct.resourcePlans.add(_elem897); } iprot.readListEnd(); } @@ -380,9 +380,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMGetAllResourcePl oprot.writeFieldBegin(RESOURCE_PLANS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.resourcePlans.size())); - for (WMResourcePlan _iter875 : struct.resourcePlans) + for (WMResourcePlan _iter899 : struct.resourcePlans) { - _iter875.write(oprot); + _iter899.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, WMGetAllResourcePla if (struct.isSetResourcePlans()) { { oprot.writeI32(struct.resourcePlans.size()); - for (WMResourcePlan _iter876 : struct.resourcePlans) + for (WMResourcePlan _iter900 : struct.resourcePlans) { - _iter876.write(oprot); + _iter900.write(oprot); } } } @@ -428,14 +428,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, WMGetAllResourcePlan BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list877 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.resourcePlans = new ArrayList(_list877.size); - WMResourcePlan _elem878; - for (int _i879 = 0; _i879 < _list877.size; ++_i879) + org.apache.thrift.protocol.TList _list901 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.resourcePlans = new ArrayList(_list901.size); + WMResourcePlan _elem902; + for (int _i903 = 0; _i903 < _list901.size; ++_i903) { - _elem878 = new WMResourcePlan(); - _elem878.read(iprot); - struct.resourcePlans.add(_elem878); + _elem902 = new WMResourcePlan(); + _elem902.read(iprot); + struct.resourcePlans.add(_elem902); } } struct.setResourcePlansIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetTriggersForResourePlanResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetTriggersForResourePlanResponse.java index 10ed67ce2b..53ea5d56f3 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetTriggersForResourePlanResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMGetTriggersForResourePlanResponse.java @@ -346,14 +346,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMGetTriggersForRes case 1: // TRIGGERS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list896 = iprot.readListBegin(); - struct.triggers = new ArrayList(_list896.size); - WMTrigger _elem897; - for (int _i898 = 0; _i898 < _list896.size; ++_i898) + org.apache.thrift.protocol.TList _list920 = iprot.readListBegin(); + struct.triggers = new ArrayList(_list920.size); + WMTrigger _elem921; + for (int _i922 = 0; _i922 < _list920.size; ++_i922) { - _elem897 = new WMTrigger(); - _elem897.read(iprot); - struct.triggers.add(_elem897); + _elem921 = new WMTrigger(); + _elem921.read(iprot); + struct.triggers.add(_elem921); } iprot.readListEnd(); } @@ -380,9 +380,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMGetTriggersForRe oprot.writeFieldBegin(TRIGGERS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.triggers.size())); - for (WMTrigger _iter899 : struct.triggers) + for (WMTrigger _iter923 : struct.triggers) { - _iter899.write(oprot); + _iter923.write(oprot); } oprot.writeListEnd(); } @@ -414,9 +414,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, WMGetTriggersForRes if (struct.isSetTriggers()) { { oprot.writeI32(struct.triggers.size()); - for (WMTrigger _iter900 : struct.triggers) + for (WMTrigger _iter924 : struct.triggers) { - _iter900.write(oprot); + _iter924.write(oprot); } } } @@ -428,14 +428,14 @@ public void read(org.apache.thrift.protocol.TProtocol prot, WMGetTriggersForReso BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list901 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.triggers = new ArrayList(_list901.size); - WMTrigger _elem902; - for (int _i903 = 0; _i903 < _list901.size; ++_i903) + org.apache.thrift.protocol.TList _list925 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.triggers = new ArrayList(_list925.size); + WMTrigger _elem926; + for (int _i927 = 0; _i927 < _list925.size; ++_i927) { - _elem902 = new WMTrigger(); - _elem902.read(iprot); - struct.triggers.add(_elem902); + _elem926 = new WMTrigger(); + _elem926.read(iprot); + struct.triggers.add(_elem926); } } struct.setTriggersIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMValidateResourcePlanResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMValidateResourcePlanResponse.java index 86d7d5c694..0dd8a5e191 100644 --- a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMValidateResourcePlanResponse.java +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WMValidateResourcePlanResponse.java @@ -441,13 +441,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMValidateResourceP case 1: // ERRORS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list880 = iprot.readListBegin(); - struct.errors = new ArrayList(_list880.size); - String _elem881; - for (int _i882 = 0; _i882 < _list880.size; ++_i882) + org.apache.thrift.protocol.TList _list904 = iprot.readListBegin(); + struct.errors = new ArrayList(_list904.size); + String _elem905; + for (int _i906 = 0; _i906 < _list904.size; ++_i906) { - _elem881 = iprot.readString(); - struct.errors.add(_elem881); + _elem905 = iprot.readString(); + struct.errors.add(_elem905); } iprot.readListEnd(); } @@ -459,13 +459,13 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, WMValidateResourceP case 2: // WARNINGS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list883 = iprot.readListBegin(); - struct.warnings = new ArrayList(_list883.size); - String _elem884; - for (int _i885 = 0; _i885 < _list883.size; ++_i885) + org.apache.thrift.protocol.TList _list907 = iprot.readListBegin(); + struct.warnings = new ArrayList(_list907.size); + String _elem908; + for (int _i909 = 0; _i909 < _list907.size; ++_i909) { - _elem884 = iprot.readString(); - struct.warnings.add(_elem884); + _elem908 = iprot.readString(); + struct.warnings.add(_elem908); } iprot.readListEnd(); } @@ -492,9 +492,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMValidateResource oprot.writeFieldBegin(ERRORS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.errors.size())); - for (String _iter886 : struct.errors) + for (String _iter910 : struct.errors) { - oprot.writeString(_iter886); + oprot.writeString(_iter910); } oprot.writeListEnd(); } @@ -506,9 +506,9 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, WMValidateResource oprot.writeFieldBegin(WARNINGS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.warnings.size())); - for (String _iter887 : struct.warnings) + for (String _iter911 : struct.warnings) { - oprot.writeString(_iter887); + oprot.writeString(_iter911); } oprot.writeListEnd(); } @@ -543,18 +543,18 @@ public void write(org.apache.thrift.protocol.TProtocol prot, WMValidateResourceP if (struct.isSetErrors()) { { oprot.writeI32(struct.errors.size()); - for (String _iter888 : struct.errors) + for (String _iter912 : struct.errors) { - oprot.writeString(_iter888); + oprot.writeString(_iter912); } } } if (struct.isSetWarnings()) { { oprot.writeI32(struct.warnings.size()); - for (String _iter889 : struct.warnings) + for (String _iter913 : struct.warnings) { - oprot.writeString(_iter889); + oprot.writeString(_iter913); } } } @@ -566,26 +566,26 @@ public void read(org.apache.thrift.protocol.TProtocol prot, WMValidateResourcePl BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list890 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.errors = new ArrayList(_list890.size); - String _elem891; - for (int _i892 = 0; _i892 < _list890.size; ++_i892) + org.apache.thrift.protocol.TList _list914 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.errors = new ArrayList(_list914.size); + String _elem915; + for (int _i916 = 0; _i916 < _list914.size; ++_i916) { - _elem891 = iprot.readString(); - struct.errors.add(_elem891); + _elem915 = iprot.readString(); + struct.errors.add(_elem915); } } struct.setErrorsIsSet(true); } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list893 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.warnings = new ArrayList(_list893.size); - String _elem894; - for (int _i895 = 0; _i895 < _list893.size; ++_i895) + org.apache.thrift.protocol.TList _list917 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.warnings = new ArrayList(_list917.size); + String _elem918; + for (int _i919 = 0; _i919 < _list917.size; ++_i919) { - _elem894 = iprot.readString(); - struct.warnings.add(_elem894); + _elem918 = iprot.readString(); + struct.warnings.add(_elem918); } } struct.setWarningsIsSet(true); diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteEventInfo.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteEventInfo.java new file mode 100644 index 0000000000..22f26095be --- /dev/null +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteEventInfo.java @@ -0,0 +1,1012 @@ +/** + * Autogenerated by Thrift Compiler (0.9.3) + * + * DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING + * @generated + */ +package org.apache.hadoop.hive.metastore.api; + +import org.apache.thrift.scheme.IScheme; +import org.apache.thrift.scheme.SchemeFactory; +import org.apache.thrift.scheme.StandardScheme; + +import org.apache.thrift.scheme.TupleScheme; +import org.apache.thrift.protocol.TTupleProtocol; +import org.apache.thrift.protocol.TProtocolException; +import org.apache.thrift.EncodingUtils; +import org.apache.thrift.TException; +import org.apache.thrift.async.AsyncMethodCallback; +import org.apache.thrift.server.AbstractNonblockingServer.*; +import java.util.List; +import java.util.ArrayList; +import java.util.Map; +import java.util.HashMap; +import java.util.EnumMap; +import java.util.Set; +import java.util.HashSet; +import java.util.EnumSet; +import java.util.Collections; +import java.util.BitSet; +import java.nio.ByteBuffer; +import java.util.Arrays; +import javax.annotation.Generated; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +@SuppressWarnings({"cast", "rawtypes", "serial", "unchecked"}) +@Generated(value = "Autogenerated by Thrift Compiler (0.9.3)") +@org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public class WriteEventInfo implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("WriteEventInfo"); + + private static final org.apache.thrift.protocol.TField WRITE_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("writeId", org.apache.thrift.protocol.TType.I64, (short)1); + private static final org.apache.thrift.protocol.TField DATABASE_FIELD_DESC = new org.apache.thrift.protocol.TField("database", org.apache.thrift.protocol.TType.STRING, (short)2); + private static final org.apache.thrift.protocol.TField TABLE_FIELD_DESC = new org.apache.thrift.protocol.TField("table", org.apache.thrift.protocol.TType.STRING, (short)3); + private static final org.apache.thrift.protocol.TField FILES_FIELD_DESC = new org.apache.thrift.protocol.TField("files", org.apache.thrift.protocol.TType.STRING, (short)4); + private static final org.apache.thrift.protocol.TField PARTITION_FIELD_DESC = new org.apache.thrift.protocol.TField("partition", org.apache.thrift.protocol.TType.STRING, (short)5); + private static final org.apache.thrift.protocol.TField TABLE_OBJ_FIELD_DESC = new org.apache.thrift.protocol.TField("tableObj", org.apache.thrift.protocol.TType.STRING, (short)6); + private static final org.apache.thrift.protocol.TField PARTITION_OBJ_FIELD_DESC = new org.apache.thrift.protocol.TField("partitionObj", org.apache.thrift.protocol.TType.STRING, (short)7); + + private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); + static { + schemes.put(StandardScheme.class, new WriteEventInfoStandardSchemeFactory()); + schemes.put(TupleScheme.class, new WriteEventInfoTupleSchemeFactory()); + } + + private long writeId; // required + private String database; // required + private String table; // required + private String files; // required + private String partition; // optional + private String tableObj; // optional + private String partitionObj; // optional + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + WRITE_ID((short)1, "writeId"), + DATABASE((short)2, "database"), + TABLE((short)3, "table"), + FILES((short)4, "files"), + PARTITION((short)5, "partition"), + TABLE_OBJ((short)6, "tableObj"), + PARTITION_OBJ((short)7, "partitionObj"); + + private static final Map byName = new HashMap(); + + static { + for (_Fields field : EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 1: // WRITE_ID + return WRITE_ID; + case 2: // DATABASE + return DATABASE; + case 3: // TABLE + return TABLE; + case 4: // FILES + return FILES; + case 5: // PARTITION + return PARTITION; + case 6: // TABLE_OBJ + return TABLE_OBJ; + case 7: // PARTITION_OBJ + return PARTITION_OBJ; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + public static _Fields findByName(String name) { + return byName.get(name); + } + + private final short _thriftId; + private final String _fieldName; + + _Fields(short thriftId, String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + private static final int __WRITEID_ISSET_ID = 0; + private byte __isset_bitfield = 0; + private static final _Fields optionals[] = {_Fields.PARTITION,_Fields.TABLE_OBJ,_Fields.PARTITION_OBJ}; + public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.WRITE_ID, new org.apache.thrift.meta_data.FieldMetaData("writeId", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.DATABASE, new org.apache.thrift.meta_data.FieldMetaData("database", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TABLE, new org.apache.thrift.meta_data.FieldMetaData("table", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.FILES, new org.apache.thrift.meta_data.FieldMetaData("files", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.PARTITION, new org.apache.thrift.meta_data.FieldMetaData("partition", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TABLE_OBJ, new org.apache.thrift.meta_data.FieldMetaData("tableObj", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.PARTITION_OBJ, new org.apache.thrift.meta_data.FieldMetaData("partitionObj", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + metaDataMap = Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(WriteEventInfo.class, metaDataMap); + } + + public WriteEventInfo() { + } + + public WriteEventInfo( + long writeId, + String database, + String table, + String files) + { + this(); + this.writeId = writeId; + setWriteIdIsSet(true); + this.database = database; + this.table = table; + this.files = files; + } + + /** + * Performs a deep copy on other. + */ + public WriteEventInfo(WriteEventInfo other) { + __isset_bitfield = other.__isset_bitfield; + this.writeId = other.writeId; + if (other.isSetDatabase()) { + this.database = other.database; + } + if (other.isSetTable()) { + this.table = other.table; + } + if (other.isSetFiles()) { + this.files = other.files; + } + if (other.isSetPartition()) { + this.partition = other.partition; + } + if (other.isSetTableObj()) { + this.tableObj = other.tableObj; + } + if (other.isSetPartitionObj()) { + this.partitionObj = other.partitionObj; + } + } + + public WriteEventInfo deepCopy() { + return new WriteEventInfo(this); + } + + @Override + public void clear() { + setWriteIdIsSet(false); + this.writeId = 0; + this.database = null; + this.table = null; + this.files = null; + this.partition = null; + this.tableObj = null; + this.partitionObj = null; + } + + public long getWriteId() { + return this.writeId; + } + + public void setWriteId(long writeId) { + this.writeId = writeId; + setWriteIdIsSet(true); + } + + public void unsetWriteId() { + __isset_bitfield = EncodingUtils.clearBit(__isset_bitfield, __WRITEID_ISSET_ID); + } + + /** Returns true if field writeId is set (has been assigned a value) and false otherwise */ + public boolean isSetWriteId() { + return EncodingUtils.testBit(__isset_bitfield, __WRITEID_ISSET_ID); + } + + public void setWriteIdIsSet(boolean value) { + __isset_bitfield = EncodingUtils.setBit(__isset_bitfield, __WRITEID_ISSET_ID, value); + } + + public String getDatabase() { + return this.database; + } + + public void setDatabase(String database) { + this.database = database; + } + + public void unsetDatabase() { + this.database = null; + } + + /** Returns true if field database is set (has been assigned a value) and false otherwise */ + public boolean isSetDatabase() { + return this.database != null; + } + + public void setDatabaseIsSet(boolean value) { + if (!value) { + this.database = null; + } + } + + public String getTable() { + return this.table; + } + + public void setTable(String table) { + this.table = table; + } + + public void unsetTable() { + this.table = null; + } + + /** Returns true if field table is set (has been assigned a value) and false otherwise */ + public boolean isSetTable() { + return this.table != null; + } + + public void setTableIsSet(boolean value) { + if (!value) { + this.table = null; + } + } + + public String getFiles() { + return this.files; + } + + public void setFiles(String files) { + this.files = files; + } + + public void unsetFiles() { + this.files = null; + } + + /** Returns true if field files is set (has been assigned a value) and false otherwise */ + public boolean isSetFiles() { + return this.files != null; + } + + public void setFilesIsSet(boolean value) { + if (!value) { + this.files = null; + } + } + + public String getPartition() { + return this.partition; + } + + public void setPartition(String partition) { + this.partition = partition; + } + + public void unsetPartition() { + this.partition = null; + } + + /** Returns true if field partition is set (has been assigned a value) and false otherwise */ + public boolean isSetPartition() { + return this.partition != null; + } + + public void setPartitionIsSet(boolean value) { + if (!value) { + this.partition = null; + } + } + + public String getTableObj() { + return this.tableObj; + } + + public void setTableObj(String tableObj) { + this.tableObj = tableObj; + } + + public void unsetTableObj() { + this.tableObj = null; + } + + /** Returns true if field tableObj is set (has been assigned a value) and false otherwise */ + public boolean isSetTableObj() { + return this.tableObj != null; + } + + public void setTableObjIsSet(boolean value) { + if (!value) { + this.tableObj = null; + } + } + + public String getPartitionObj() { + return this.partitionObj; + } + + public void setPartitionObj(String partitionObj) { + this.partitionObj = partitionObj; + } + + public void unsetPartitionObj() { + this.partitionObj = null; + } + + /** Returns true if field partitionObj is set (has been assigned a value) and false otherwise */ + public boolean isSetPartitionObj() { + return this.partitionObj != null; + } + + public void setPartitionObjIsSet(boolean value) { + if (!value) { + this.partitionObj = null; + } + } + + public void setFieldValue(_Fields field, Object value) { + switch (field) { + case WRITE_ID: + if (value == null) { + unsetWriteId(); + } else { + setWriteId((Long)value); + } + break; + + case DATABASE: + if (value == null) { + unsetDatabase(); + } else { + setDatabase((String)value); + } + break; + + case TABLE: + if (value == null) { + unsetTable(); + } else { + setTable((String)value); + } + break; + + case FILES: + if (value == null) { + unsetFiles(); + } else { + setFiles((String)value); + } + break; + + case PARTITION: + if (value == null) { + unsetPartition(); + } else { + setPartition((String)value); + } + break; + + case TABLE_OBJ: + if (value == null) { + unsetTableObj(); + } else { + setTableObj((String)value); + } + break; + + case PARTITION_OBJ: + if (value == null) { + unsetPartitionObj(); + } else { + setPartitionObj((String)value); + } + break; + + } + } + + public Object getFieldValue(_Fields field) { + switch (field) { + case WRITE_ID: + return getWriteId(); + + case DATABASE: + return getDatabase(); + + case TABLE: + return getTable(); + + case FILES: + return getFiles(); + + case PARTITION: + return getPartition(); + + case TABLE_OBJ: + return getTableObj(); + + case PARTITION_OBJ: + return getPartitionObj(); + + } + throw new IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new IllegalArgumentException(); + } + + switch (field) { + case WRITE_ID: + return isSetWriteId(); + case DATABASE: + return isSetDatabase(); + case TABLE: + return isSetTable(); + case FILES: + return isSetFiles(); + case PARTITION: + return isSetPartition(); + case TABLE_OBJ: + return isSetTableObj(); + case PARTITION_OBJ: + return isSetPartitionObj(); + } + throw new IllegalStateException(); + } + + @Override + public boolean equals(Object that) { + if (that == null) + return false; + if (that instanceof WriteEventInfo) + return this.equals((WriteEventInfo)that); + return false; + } + + public boolean equals(WriteEventInfo that) { + if (that == null) + return false; + + boolean this_present_writeId = true; + boolean that_present_writeId = true; + if (this_present_writeId || that_present_writeId) { + if (!(this_present_writeId && that_present_writeId)) + return false; + if (this.writeId != that.writeId) + return false; + } + + boolean this_present_database = true && this.isSetDatabase(); + boolean that_present_database = true && that.isSetDatabase(); + if (this_present_database || that_present_database) { + if (!(this_present_database && that_present_database)) + return false; + if (!this.database.equals(that.database)) + return false; + } + + boolean this_present_table = true && this.isSetTable(); + boolean that_present_table = true && that.isSetTable(); + if (this_present_table || that_present_table) { + if (!(this_present_table && that_present_table)) + return false; + if (!this.table.equals(that.table)) + return false; + } + + boolean this_present_files = true && this.isSetFiles(); + boolean that_present_files = true && that.isSetFiles(); + if (this_present_files || that_present_files) { + if (!(this_present_files && that_present_files)) + return false; + if (!this.files.equals(that.files)) + return false; + } + + boolean this_present_partition = true && this.isSetPartition(); + boolean that_present_partition = true && that.isSetPartition(); + if (this_present_partition || that_present_partition) { + if (!(this_present_partition && that_present_partition)) + return false; + if (!this.partition.equals(that.partition)) + return false; + } + + boolean this_present_tableObj = true && this.isSetTableObj(); + boolean that_present_tableObj = true && that.isSetTableObj(); + if (this_present_tableObj || that_present_tableObj) { + if (!(this_present_tableObj && that_present_tableObj)) + return false; + if (!this.tableObj.equals(that.tableObj)) + return false; + } + + boolean this_present_partitionObj = true && this.isSetPartitionObj(); + boolean that_present_partitionObj = true && that.isSetPartitionObj(); + if (this_present_partitionObj || that_present_partitionObj) { + if (!(this_present_partitionObj && that_present_partitionObj)) + return false; + if (!this.partitionObj.equals(that.partitionObj)) + return false; + } + + return true; + } + + @Override + public int hashCode() { + List list = new ArrayList(); + + boolean present_writeId = true; + list.add(present_writeId); + if (present_writeId) + list.add(writeId); + + boolean present_database = true && (isSetDatabase()); + list.add(present_database); + if (present_database) + list.add(database); + + boolean present_table = true && (isSetTable()); + list.add(present_table); + if (present_table) + list.add(table); + + boolean present_files = true && (isSetFiles()); + list.add(present_files); + if (present_files) + list.add(files); + + boolean present_partition = true && (isSetPartition()); + list.add(present_partition); + if (present_partition) + list.add(partition); + + boolean present_tableObj = true && (isSetTableObj()); + list.add(present_tableObj); + if (present_tableObj) + list.add(tableObj); + + boolean present_partitionObj = true && (isSetPartitionObj()); + list.add(present_partitionObj); + if (present_partitionObj) + list.add(partitionObj); + + return list.hashCode(); + } + + @Override + public int compareTo(WriteEventInfo other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = Boolean.valueOf(isSetWriteId()).compareTo(other.isSetWriteId()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetWriteId()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.writeId, other.writeId); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetDatabase()).compareTo(other.isSetDatabase()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetDatabase()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.database, other.database); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetTable()).compareTo(other.isSetTable()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTable()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.table, other.table); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetFiles()).compareTo(other.isSetFiles()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetFiles()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.files, other.files); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetPartition()).compareTo(other.isSetPartition()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetPartition()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.partition, other.partition); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetTableObj()).compareTo(other.isSetTableObj()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTableObj()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tableObj, other.tableObj); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetPartitionObj()).compareTo(other.isSetPartitionObj()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetPartitionObj()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.partitionObj, other.partitionObj); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + schemes.get(iprot.getScheme()).getScheme().read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + schemes.get(oprot.getScheme()).getScheme().write(oprot, this); + } + + @Override + public String toString() { + StringBuilder sb = new StringBuilder("WriteEventInfo("); + boolean first = true; + + sb.append("writeId:"); + sb.append(this.writeId); + first = false; + if (!first) sb.append(", "); + sb.append("database:"); + if (this.database == null) { + sb.append("null"); + } else { + sb.append(this.database); + } + first = false; + if (!first) sb.append(", "); + sb.append("table:"); + if (this.table == null) { + sb.append("null"); + } else { + sb.append(this.table); + } + first = false; + if (!first) sb.append(", "); + sb.append("files:"); + if (this.files == null) { + sb.append("null"); + } else { + sb.append(this.files); + } + first = false; + if (isSetPartition()) { + if (!first) sb.append(", "); + sb.append("partition:"); + if (this.partition == null) { + sb.append("null"); + } else { + sb.append(this.partition); + } + first = false; + } + if (isSetTableObj()) { + if (!first) sb.append(", "); + sb.append("tableObj:"); + if (this.tableObj == null) { + sb.append("null"); + } else { + sb.append(this.tableObj); + } + first = false; + } + if (isSetPartitionObj()) { + if (!first) sb.append(", "); + sb.append("partitionObj:"); + if (this.partitionObj == null) { + sb.append("null"); + } else { + sb.append(this.partitionObj); + } + first = false; + } + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + if (!isSetWriteId()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'writeId' is unset! Struct:" + toString()); + } + + if (!isSetDatabase()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'database' is unset! Struct:" + toString()); + } + + if (!isSetTable()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'table' is unset! Struct:" + toString()); + } + + if (!isSetFiles()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'files' is unset! Struct:" + toString()); + } + + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException { + try { + // it doesn't seem like you should have to do this, but java serialization is wacky, and doesn't call the default constructor. + __isset_bitfield = 0; + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class WriteEventInfoStandardSchemeFactory implements SchemeFactory { + public WriteEventInfoStandardScheme getScheme() { + return new WriteEventInfoStandardScheme(); + } + } + + private static class WriteEventInfoStandardScheme extends StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, WriteEventInfo struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 1: // WRITE_ID + if (schemeField.type == org.apache.thrift.protocol.TType.I64) { + struct.writeId = iprot.readI64(); + struct.setWriteIdIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 2: // DATABASE + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.database = iprot.readString(); + struct.setDatabaseIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 3: // TABLE + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.table = iprot.readString(); + struct.setTableIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 4: // FILES + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.files = iprot.readString(); + struct.setFilesIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 5: // PARTITION + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.partition = iprot.readString(); + struct.setPartitionIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 6: // TABLE_OBJ + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.tableObj = iprot.readString(); + struct.setTableObjIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 7: // PARTITION_OBJ + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.partitionObj = iprot.readString(); + struct.setPartitionObjIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, WriteEventInfo struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + oprot.writeFieldBegin(WRITE_ID_FIELD_DESC); + oprot.writeI64(struct.writeId); + oprot.writeFieldEnd(); + if (struct.database != null) { + oprot.writeFieldBegin(DATABASE_FIELD_DESC); + oprot.writeString(struct.database); + oprot.writeFieldEnd(); + } + if (struct.table != null) { + oprot.writeFieldBegin(TABLE_FIELD_DESC); + oprot.writeString(struct.table); + oprot.writeFieldEnd(); + } + if (struct.files != null) { + oprot.writeFieldBegin(FILES_FIELD_DESC); + oprot.writeString(struct.files); + oprot.writeFieldEnd(); + } + if (struct.partition != null) { + if (struct.isSetPartition()) { + oprot.writeFieldBegin(PARTITION_FIELD_DESC); + oprot.writeString(struct.partition); + oprot.writeFieldEnd(); + } + } + if (struct.tableObj != null) { + if (struct.isSetTableObj()) { + oprot.writeFieldBegin(TABLE_OBJ_FIELD_DESC); + oprot.writeString(struct.tableObj); + oprot.writeFieldEnd(); + } + } + if (struct.partitionObj != null) { + if (struct.isSetPartitionObj()) { + oprot.writeFieldBegin(PARTITION_OBJ_FIELD_DESC); + oprot.writeString(struct.partitionObj); + oprot.writeFieldEnd(); + } + } + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class WriteEventInfoTupleSchemeFactory implements SchemeFactory { + public WriteEventInfoTupleScheme getScheme() { + return new WriteEventInfoTupleScheme(); + } + } + + private static class WriteEventInfoTupleScheme extends TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, WriteEventInfo struct) throws org.apache.thrift.TException { + TTupleProtocol oprot = (TTupleProtocol) prot; + oprot.writeI64(struct.writeId); + oprot.writeString(struct.database); + oprot.writeString(struct.table); + oprot.writeString(struct.files); + BitSet optionals = new BitSet(); + if (struct.isSetPartition()) { + optionals.set(0); + } + if (struct.isSetTableObj()) { + optionals.set(1); + } + if (struct.isSetPartitionObj()) { + optionals.set(2); + } + oprot.writeBitSet(optionals, 3); + if (struct.isSetPartition()) { + oprot.writeString(struct.partition); + } + if (struct.isSetTableObj()) { + oprot.writeString(struct.tableObj); + } + if (struct.isSetPartitionObj()) { + oprot.writeString(struct.partitionObj); + } + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, WriteEventInfo struct) throws org.apache.thrift.TException { + TTupleProtocol iprot = (TTupleProtocol) prot; + struct.writeId = iprot.readI64(); + struct.setWriteIdIsSet(true); + struct.database = iprot.readString(); + struct.setDatabaseIsSet(true); + struct.table = iprot.readString(); + struct.setTableIsSet(true); + struct.files = iprot.readString(); + struct.setFilesIsSet(true); + BitSet incoming = iprot.readBitSet(3); + if (incoming.get(0)) { + struct.partition = iprot.readString(); + struct.setPartitionIsSet(true); + } + if (incoming.get(1)) { + struct.tableObj = iprot.readString(); + struct.setTableObjIsSet(true); + } + if (incoming.get(2)) { + struct.partitionObj = iprot.readString(); + struct.setPartitionObjIsSet(true); + } + } + } + +} + diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteNotificationLogRequest.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteNotificationLogRequest.java new file mode 100644 index 0000000000..5758820ada --- /dev/null +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteNotificationLogRequest.java @@ -0,0 +1,949 @@ +/** + * Autogenerated by Thrift Compiler (0.9.3) + * + * DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING + * @generated + */ +package org.apache.hadoop.hive.metastore.api; + +import org.apache.thrift.scheme.IScheme; +import org.apache.thrift.scheme.SchemeFactory; +import org.apache.thrift.scheme.StandardScheme; + +import org.apache.thrift.scheme.TupleScheme; +import org.apache.thrift.protocol.TTupleProtocol; +import org.apache.thrift.protocol.TProtocolException; +import org.apache.thrift.EncodingUtils; +import org.apache.thrift.TException; +import org.apache.thrift.async.AsyncMethodCallback; +import org.apache.thrift.server.AbstractNonblockingServer.*; +import java.util.List; +import java.util.ArrayList; +import java.util.Map; +import java.util.HashMap; +import java.util.EnumMap; +import java.util.Set; +import java.util.HashSet; +import java.util.EnumSet; +import java.util.Collections; +import java.util.BitSet; +import java.nio.ByteBuffer; +import java.util.Arrays; +import javax.annotation.Generated; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +@SuppressWarnings({"cast", "rawtypes", "serial", "unchecked"}) +@Generated(value = "Autogenerated by Thrift Compiler (0.9.3)") +@org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public class WriteNotificationLogRequest implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("WriteNotificationLogRequest"); + + private static final org.apache.thrift.protocol.TField TXN_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("txnId", org.apache.thrift.protocol.TType.I64, (short)1); + private static final org.apache.thrift.protocol.TField WRITE_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("writeId", org.apache.thrift.protocol.TType.I64, (short)2); + private static final org.apache.thrift.protocol.TField DB_FIELD_DESC = new org.apache.thrift.protocol.TField("db", org.apache.thrift.protocol.TType.STRING, (short)3); + private static final org.apache.thrift.protocol.TField TABLE_FIELD_DESC = new org.apache.thrift.protocol.TField("table", org.apache.thrift.protocol.TType.STRING, (short)4); + private static final org.apache.thrift.protocol.TField FILE_INFO_FIELD_DESC = new org.apache.thrift.protocol.TField("fileInfo", org.apache.thrift.protocol.TType.STRUCT, (short)5); + private static final org.apache.thrift.protocol.TField PARTITION_VALS_FIELD_DESC = new org.apache.thrift.protocol.TField("partitionVals", org.apache.thrift.protocol.TType.LIST, (short)6); + + private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); + static { + schemes.put(StandardScheme.class, new WriteNotificationLogRequestStandardSchemeFactory()); + schemes.put(TupleScheme.class, new WriteNotificationLogRequestTupleSchemeFactory()); + } + + private long txnId; // required + private long writeId; // required + private String db; // required + private String table; // required + private InsertEventRequestData fileInfo; // required + private List partitionVals; // optional + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { + TXN_ID((short)1, "txnId"), + WRITE_ID((short)2, "writeId"), + DB((short)3, "db"), + TABLE((short)4, "table"), + FILE_INFO((short)5, "fileInfo"), + PARTITION_VALS((short)6, "partitionVals"); + + private static final Map byName = new HashMap(); + + static { + for (_Fields field : EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + case 1: // TXN_ID + return TXN_ID; + case 2: // WRITE_ID + return WRITE_ID; + case 3: // DB + return DB; + case 4: // TABLE + return TABLE; + case 5: // FILE_INFO + return FILE_INFO; + case 6: // PARTITION_VALS + return PARTITION_VALS; + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + public static _Fields findByName(String name) { + return byName.get(name); + } + + private final short _thriftId; + private final String _fieldName; + + _Fields(short thriftId, String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public String getFieldName() { + return _fieldName; + } + } + + // isset id assignments + private static final int __TXNID_ISSET_ID = 0; + private static final int __WRITEID_ISSET_ID = 1; + private byte __isset_bitfield = 0; + private static final _Fields optionals[] = {_Fields.PARTITION_VALS}; + public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + tmpMap.put(_Fields.TXN_ID, new org.apache.thrift.meta_data.FieldMetaData("txnId", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.WRITE_ID, new org.apache.thrift.meta_data.FieldMetaData("writeId", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.DB, new org.apache.thrift.meta_data.FieldMetaData("db", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.TABLE, new org.apache.thrift.meta_data.FieldMetaData("table", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); + tmpMap.put(_Fields.FILE_INFO, new org.apache.thrift.meta_data.FieldMetaData("fileInfo", org.apache.thrift.TFieldRequirementType.REQUIRED, + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, InsertEventRequestData.class))); + tmpMap.put(_Fields.PARTITION_VALS, new org.apache.thrift.meta_data.FieldMetaData("partitionVals", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.ListMetaData(org.apache.thrift.protocol.TType.LIST, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)))); + metaDataMap = Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(WriteNotificationLogRequest.class, metaDataMap); + } + + public WriteNotificationLogRequest() { + } + + public WriteNotificationLogRequest( + long txnId, + long writeId, + String db, + String table, + InsertEventRequestData fileInfo) + { + this(); + this.txnId = txnId; + setTxnIdIsSet(true); + this.writeId = writeId; + setWriteIdIsSet(true); + this.db = db; + this.table = table; + this.fileInfo = fileInfo; + } + + /** + * Performs a deep copy on other. + */ + public WriteNotificationLogRequest(WriteNotificationLogRequest other) { + __isset_bitfield = other.__isset_bitfield; + this.txnId = other.txnId; + this.writeId = other.writeId; + if (other.isSetDb()) { + this.db = other.db; + } + if (other.isSetTable()) { + this.table = other.table; + } + if (other.isSetFileInfo()) { + this.fileInfo = new InsertEventRequestData(other.fileInfo); + } + if (other.isSetPartitionVals()) { + List __this__partitionVals = new ArrayList(other.partitionVals); + this.partitionVals = __this__partitionVals; + } + } + + public WriteNotificationLogRequest deepCopy() { + return new WriteNotificationLogRequest(this); + } + + @Override + public void clear() { + setTxnIdIsSet(false); + this.txnId = 0; + setWriteIdIsSet(false); + this.writeId = 0; + this.db = null; + this.table = null; + this.fileInfo = null; + this.partitionVals = null; + } + + public long getTxnId() { + return this.txnId; + } + + public void setTxnId(long txnId) { + this.txnId = txnId; + setTxnIdIsSet(true); + } + + public void unsetTxnId() { + __isset_bitfield = EncodingUtils.clearBit(__isset_bitfield, __TXNID_ISSET_ID); + } + + /** Returns true if field txnId is set (has been assigned a value) and false otherwise */ + public boolean isSetTxnId() { + return EncodingUtils.testBit(__isset_bitfield, __TXNID_ISSET_ID); + } + + public void setTxnIdIsSet(boolean value) { + __isset_bitfield = EncodingUtils.setBit(__isset_bitfield, __TXNID_ISSET_ID, value); + } + + public long getWriteId() { + return this.writeId; + } + + public void setWriteId(long writeId) { + this.writeId = writeId; + setWriteIdIsSet(true); + } + + public void unsetWriteId() { + __isset_bitfield = EncodingUtils.clearBit(__isset_bitfield, __WRITEID_ISSET_ID); + } + + /** Returns true if field writeId is set (has been assigned a value) and false otherwise */ + public boolean isSetWriteId() { + return EncodingUtils.testBit(__isset_bitfield, __WRITEID_ISSET_ID); + } + + public void setWriteIdIsSet(boolean value) { + __isset_bitfield = EncodingUtils.setBit(__isset_bitfield, __WRITEID_ISSET_ID, value); + } + + public String getDb() { + return this.db; + } + + public void setDb(String db) { + this.db = db; + } + + public void unsetDb() { + this.db = null; + } + + /** Returns true if field db is set (has been assigned a value) and false otherwise */ + public boolean isSetDb() { + return this.db != null; + } + + public void setDbIsSet(boolean value) { + if (!value) { + this.db = null; + } + } + + public String getTable() { + return this.table; + } + + public void setTable(String table) { + this.table = table; + } + + public void unsetTable() { + this.table = null; + } + + /** Returns true if field table is set (has been assigned a value) and false otherwise */ + public boolean isSetTable() { + return this.table != null; + } + + public void setTableIsSet(boolean value) { + if (!value) { + this.table = null; + } + } + + public InsertEventRequestData getFileInfo() { + return this.fileInfo; + } + + public void setFileInfo(InsertEventRequestData fileInfo) { + this.fileInfo = fileInfo; + } + + public void unsetFileInfo() { + this.fileInfo = null; + } + + /** Returns true if field fileInfo is set (has been assigned a value) and false otherwise */ + public boolean isSetFileInfo() { + return this.fileInfo != null; + } + + public void setFileInfoIsSet(boolean value) { + if (!value) { + this.fileInfo = null; + } + } + + public int getPartitionValsSize() { + return (this.partitionVals == null) ? 0 : this.partitionVals.size(); + } + + public java.util.Iterator getPartitionValsIterator() { + return (this.partitionVals == null) ? null : this.partitionVals.iterator(); + } + + public void addToPartitionVals(String elem) { + if (this.partitionVals == null) { + this.partitionVals = new ArrayList(); + } + this.partitionVals.add(elem); + } + + public List getPartitionVals() { + return this.partitionVals; + } + + public void setPartitionVals(List partitionVals) { + this.partitionVals = partitionVals; + } + + public void unsetPartitionVals() { + this.partitionVals = null; + } + + /** Returns true if field partitionVals is set (has been assigned a value) and false otherwise */ + public boolean isSetPartitionVals() { + return this.partitionVals != null; + } + + public void setPartitionValsIsSet(boolean value) { + if (!value) { + this.partitionVals = null; + } + } + + public void setFieldValue(_Fields field, Object value) { + switch (field) { + case TXN_ID: + if (value == null) { + unsetTxnId(); + } else { + setTxnId((Long)value); + } + break; + + case WRITE_ID: + if (value == null) { + unsetWriteId(); + } else { + setWriteId((Long)value); + } + break; + + case DB: + if (value == null) { + unsetDb(); + } else { + setDb((String)value); + } + break; + + case TABLE: + if (value == null) { + unsetTable(); + } else { + setTable((String)value); + } + break; + + case FILE_INFO: + if (value == null) { + unsetFileInfo(); + } else { + setFileInfo((InsertEventRequestData)value); + } + break; + + case PARTITION_VALS: + if (value == null) { + unsetPartitionVals(); + } else { + setPartitionVals((List)value); + } + break; + + } + } + + public Object getFieldValue(_Fields field) { + switch (field) { + case TXN_ID: + return getTxnId(); + + case WRITE_ID: + return getWriteId(); + + case DB: + return getDb(); + + case TABLE: + return getTable(); + + case FILE_INFO: + return getFileInfo(); + + case PARTITION_VALS: + return getPartitionVals(); + + } + throw new IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new IllegalArgumentException(); + } + + switch (field) { + case TXN_ID: + return isSetTxnId(); + case WRITE_ID: + return isSetWriteId(); + case DB: + return isSetDb(); + case TABLE: + return isSetTable(); + case FILE_INFO: + return isSetFileInfo(); + case PARTITION_VALS: + return isSetPartitionVals(); + } + throw new IllegalStateException(); + } + + @Override + public boolean equals(Object that) { + if (that == null) + return false; + if (that instanceof WriteNotificationLogRequest) + return this.equals((WriteNotificationLogRequest)that); + return false; + } + + public boolean equals(WriteNotificationLogRequest that) { + if (that == null) + return false; + + boolean this_present_txnId = true; + boolean that_present_txnId = true; + if (this_present_txnId || that_present_txnId) { + if (!(this_present_txnId && that_present_txnId)) + return false; + if (this.txnId != that.txnId) + return false; + } + + boolean this_present_writeId = true; + boolean that_present_writeId = true; + if (this_present_writeId || that_present_writeId) { + if (!(this_present_writeId && that_present_writeId)) + return false; + if (this.writeId != that.writeId) + return false; + } + + boolean this_present_db = true && this.isSetDb(); + boolean that_present_db = true && that.isSetDb(); + if (this_present_db || that_present_db) { + if (!(this_present_db && that_present_db)) + return false; + if (!this.db.equals(that.db)) + return false; + } + + boolean this_present_table = true && this.isSetTable(); + boolean that_present_table = true && that.isSetTable(); + if (this_present_table || that_present_table) { + if (!(this_present_table && that_present_table)) + return false; + if (!this.table.equals(that.table)) + return false; + } + + boolean this_present_fileInfo = true && this.isSetFileInfo(); + boolean that_present_fileInfo = true && that.isSetFileInfo(); + if (this_present_fileInfo || that_present_fileInfo) { + if (!(this_present_fileInfo && that_present_fileInfo)) + return false; + if (!this.fileInfo.equals(that.fileInfo)) + return false; + } + + boolean this_present_partitionVals = true && this.isSetPartitionVals(); + boolean that_present_partitionVals = true && that.isSetPartitionVals(); + if (this_present_partitionVals || that_present_partitionVals) { + if (!(this_present_partitionVals && that_present_partitionVals)) + return false; + if (!this.partitionVals.equals(that.partitionVals)) + return false; + } + + return true; + } + + @Override + public int hashCode() { + List list = new ArrayList(); + + boolean present_txnId = true; + list.add(present_txnId); + if (present_txnId) + list.add(txnId); + + boolean present_writeId = true; + list.add(present_writeId); + if (present_writeId) + list.add(writeId); + + boolean present_db = true && (isSetDb()); + list.add(present_db); + if (present_db) + list.add(db); + + boolean present_table = true && (isSetTable()); + list.add(present_table); + if (present_table) + list.add(table); + + boolean present_fileInfo = true && (isSetFileInfo()); + list.add(present_fileInfo); + if (present_fileInfo) + list.add(fileInfo); + + boolean present_partitionVals = true && (isSetPartitionVals()); + list.add(present_partitionVals); + if (present_partitionVals) + list.add(partitionVals); + + return list.hashCode(); + } + + @Override + public int compareTo(WriteNotificationLogRequest other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + lastComparison = Boolean.valueOf(isSetTxnId()).compareTo(other.isSetTxnId()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTxnId()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.txnId, other.txnId); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetWriteId()).compareTo(other.isSetWriteId()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetWriteId()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.writeId, other.writeId); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetDb()).compareTo(other.isSetDb()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetDb()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.db, other.db); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetTable()).compareTo(other.isSetTable()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetTable()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.table, other.table); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetFileInfo()).compareTo(other.isSetFileInfo()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetFileInfo()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.fileInfo, other.fileInfo); + if (lastComparison != 0) { + return lastComparison; + } + } + lastComparison = Boolean.valueOf(isSetPartitionVals()).compareTo(other.isSetPartitionVals()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetPartitionVals()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.partitionVals, other.partitionVals); + if (lastComparison != 0) { + return lastComparison; + } + } + return 0; + } + + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + schemes.get(iprot.getScheme()).getScheme().read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + schemes.get(oprot.getScheme()).getScheme().write(oprot, this); + } + + @Override + public String toString() { + StringBuilder sb = new StringBuilder("WriteNotificationLogRequest("); + boolean first = true; + + sb.append("txnId:"); + sb.append(this.txnId); + first = false; + if (!first) sb.append(", "); + sb.append("writeId:"); + sb.append(this.writeId); + first = false; + if (!first) sb.append(", "); + sb.append("db:"); + if (this.db == null) { + sb.append("null"); + } else { + sb.append(this.db); + } + first = false; + if (!first) sb.append(", "); + sb.append("table:"); + if (this.table == null) { + sb.append("null"); + } else { + sb.append(this.table); + } + first = false; + if (!first) sb.append(", "); + sb.append("fileInfo:"); + if (this.fileInfo == null) { + sb.append("null"); + } else { + sb.append(this.fileInfo); + } + first = false; + if (isSetPartitionVals()) { + if (!first) sb.append(", "); + sb.append("partitionVals:"); + if (this.partitionVals == null) { + sb.append("null"); + } else { + sb.append(this.partitionVals); + } + first = false; + } + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + if (!isSetTxnId()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'txnId' is unset! Struct:" + toString()); + } + + if (!isSetWriteId()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'writeId' is unset! Struct:" + toString()); + } + + if (!isSetDb()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'db' is unset! Struct:" + toString()); + } + + if (!isSetTable()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'table' is unset! Struct:" + toString()); + } + + if (!isSetFileInfo()) { + throw new org.apache.thrift.protocol.TProtocolException("Required field 'fileInfo' is unset! Struct:" + toString()); + } + + // check for sub-struct validity + if (fileInfo != null) { + fileInfo.validate(); + } + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException { + try { + // it doesn't seem like you should have to do this, but java serialization is wacky, and doesn't call the default constructor. + __isset_bitfield = 0; + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class WriteNotificationLogRequestStandardSchemeFactory implements SchemeFactory { + public WriteNotificationLogRequestStandardScheme getScheme() { + return new WriteNotificationLogRequestStandardScheme(); + } + } + + private static class WriteNotificationLogRequestStandardScheme extends StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, WriteNotificationLogRequest struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + case 1: // TXN_ID + if (schemeField.type == org.apache.thrift.protocol.TType.I64) { + struct.txnId = iprot.readI64(); + struct.setTxnIdIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 2: // WRITE_ID + if (schemeField.type == org.apache.thrift.protocol.TType.I64) { + struct.writeId = iprot.readI64(); + struct.setWriteIdIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 3: // DB + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.db = iprot.readString(); + struct.setDbIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 4: // TABLE + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.table = iprot.readString(); + struct.setTableIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 5: // FILE_INFO + if (schemeField.type == org.apache.thrift.protocol.TType.STRUCT) { + struct.fileInfo = new InsertEventRequestData(); + struct.fileInfo.read(iprot); + struct.setFileInfoIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + case 6: // PARTITION_VALS + if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { + { + org.apache.thrift.protocol.TList _list764 = iprot.readListBegin(); + struct.partitionVals = new ArrayList(_list764.size); + String _elem765; + for (int _i766 = 0; _i766 < _list764.size; ++_i766) + { + _elem765 = iprot.readString(); + struct.partitionVals.add(_elem765); + } + iprot.readListEnd(); + } + struct.setPartitionValsIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, WriteNotificationLogRequest struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + oprot.writeFieldBegin(TXN_ID_FIELD_DESC); + oprot.writeI64(struct.txnId); + oprot.writeFieldEnd(); + oprot.writeFieldBegin(WRITE_ID_FIELD_DESC); + oprot.writeI64(struct.writeId); + oprot.writeFieldEnd(); + if (struct.db != null) { + oprot.writeFieldBegin(DB_FIELD_DESC); + oprot.writeString(struct.db); + oprot.writeFieldEnd(); + } + if (struct.table != null) { + oprot.writeFieldBegin(TABLE_FIELD_DESC); + oprot.writeString(struct.table); + oprot.writeFieldEnd(); + } + if (struct.fileInfo != null) { + oprot.writeFieldBegin(FILE_INFO_FIELD_DESC); + struct.fileInfo.write(oprot); + oprot.writeFieldEnd(); + } + if (struct.partitionVals != null) { + if (struct.isSetPartitionVals()) { + oprot.writeFieldBegin(PARTITION_VALS_FIELD_DESC); + { + oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.partitionVals.size())); + for (String _iter767 : struct.partitionVals) + { + oprot.writeString(_iter767); + } + oprot.writeListEnd(); + } + oprot.writeFieldEnd(); + } + } + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class WriteNotificationLogRequestTupleSchemeFactory implements SchemeFactory { + public WriteNotificationLogRequestTupleScheme getScheme() { + return new WriteNotificationLogRequestTupleScheme(); + } + } + + private static class WriteNotificationLogRequestTupleScheme extends TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, WriteNotificationLogRequest struct) throws org.apache.thrift.TException { + TTupleProtocol oprot = (TTupleProtocol) prot; + oprot.writeI64(struct.txnId); + oprot.writeI64(struct.writeId); + oprot.writeString(struct.db); + oprot.writeString(struct.table); + struct.fileInfo.write(oprot); + BitSet optionals = new BitSet(); + if (struct.isSetPartitionVals()) { + optionals.set(0); + } + oprot.writeBitSet(optionals, 1); + if (struct.isSetPartitionVals()) { + { + oprot.writeI32(struct.partitionVals.size()); + for (String _iter768 : struct.partitionVals) + { + oprot.writeString(_iter768); + } + } + } + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, WriteNotificationLogRequest struct) throws org.apache.thrift.TException { + TTupleProtocol iprot = (TTupleProtocol) prot; + struct.txnId = iprot.readI64(); + struct.setTxnIdIsSet(true); + struct.writeId = iprot.readI64(); + struct.setWriteIdIsSet(true); + struct.db = iprot.readString(); + struct.setDbIsSet(true); + struct.table = iprot.readString(); + struct.setTableIsSet(true); + struct.fileInfo = new InsertEventRequestData(); + struct.fileInfo.read(iprot); + struct.setFileInfoIsSet(true); + BitSet incoming = iprot.readBitSet(1); + if (incoming.get(0)) { + { + org.apache.thrift.protocol.TList _list769 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.partitionVals = new ArrayList(_list769.size); + String _elem770; + for (int _i771 = 0; _i771 < _list769.size; ++_i771) + { + _elem770 = iprot.readString(); + struct.partitionVals.add(_elem770); + } + } + struct.setPartitionValsIsSet(true); + } + } + } + +} + diff --git a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteNotificationLogResponse.java b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteNotificationLogResponse.java new file mode 100644 index 0000000000..fab4da2af2 --- /dev/null +++ b/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/WriteNotificationLogResponse.java @@ -0,0 +1,283 @@ +/** + * Autogenerated by Thrift Compiler (0.9.3) + * + * DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING + * @generated + */ +package org.apache.hadoop.hive.metastore.api; + +import org.apache.thrift.scheme.IScheme; +import org.apache.thrift.scheme.SchemeFactory; +import org.apache.thrift.scheme.StandardScheme; + +import org.apache.thrift.scheme.TupleScheme; +import org.apache.thrift.protocol.TTupleProtocol; +import org.apache.thrift.protocol.TProtocolException; +import org.apache.thrift.EncodingUtils; +import org.apache.thrift.TException; +import org.apache.thrift.async.AsyncMethodCallback; +import org.apache.thrift.server.AbstractNonblockingServer.*; +import java.util.List; +import java.util.ArrayList; +import java.util.Map; +import java.util.HashMap; +import java.util.EnumMap; +import java.util.Set; +import java.util.HashSet; +import java.util.EnumSet; +import java.util.Collections; +import java.util.BitSet; +import java.nio.ByteBuffer; +import java.util.Arrays; +import javax.annotation.Generated; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +@SuppressWarnings({"cast", "rawtypes", "serial", "unchecked"}) +@Generated(value = "Autogenerated by Thrift Compiler (0.9.3)") +@org.apache.hadoop.classification.InterfaceAudience.Public @org.apache.hadoop.classification.InterfaceStability.Stable public class WriteNotificationLogResponse implements org.apache.thrift.TBase, java.io.Serializable, Cloneable, Comparable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("WriteNotificationLogResponse"); + + + private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); + static { + schemes.put(StandardScheme.class, new WriteNotificationLogResponseStandardSchemeFactory()); + schemes.put(TupleScheme.class, new WriteNotificationLogResponseTupleSchemeFactory()); + } + + + /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ + public enum _Fields implements org.apache.thrift.TFieldIdEnum { +; + + private static final Map byName = new HashMap(); + + static { + for (_Fields field : EnumSet.allOf(_Fields.class)) { + byName.put(field.getFieldName(), field); + } + } + + /** + * Find the _Fields constant that matches fieldId, or null if its not found. + */ + public static _Fields findByThriftId(int fieldId) { + switch(fieldId) { + default: + return null; + } + } + + /** + * Find the _Fields constant that matches fieldId, throwing an exception + * if it is not found. + */ + public static _Fields findByThriftIdOrThrow(int fieldId) { + _Fields fields = findByThriftId(fieldId); + if (fields == null) throw new IllegalArgumentException("Field " + fieldId + " doesn't exist!"); + return fields; + } + + /** + * Find the _Fields constant that matches name, or null if its not found. + */ + public static _Fields findByName(String name) { + return byName.get(name); + } + + private final short _thriftId; + private final String _fieldName; + + _Fields(short thriftId, String fieldName) { + _thriftId = thriftId; + _fieldName = fieldName; + } + + public short getThriftFieldId() { + return _thriftId; + } + + public String getFieldName() { + return _fieldName; + } + } + public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; + static { + Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); + metaDataMap = Collections.unmodifiableMap(tmpMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(WriteNotificationLogResponse.class, metaDataMap); + } + + public WriteNotificationLogResponse() { + } + + /** + * Performs a deep copy on other. + */ + public WriteNotificationLogResponse(WriteNotificationLogResponse other) { + } + + public WriteNotificationLogResponse deepCopy() { + return new WriteNotificationLogResponse(this); + } + + @Override + public void clear() { + } + + public void setFieldValue(_Fields field, Object value) { + switch (field) { + } + } + + public Object getFieldValue(_Fields field) { + switch (field) { + } + throw new IllegalStateException(); + } + + /** Returns true if field corresponding to fieldID is set (has been assigned a value) and false otherwise */ + public boolean isSet(_Fields field) { + if (field == null) { + throw new IllegalArgumentException(); + } + + switch (field) { + } + throw new IllegalStateException(); + } + + @Override + public boolean equals(Object that) { + if (that == null) + return false; + if (that instanceof WriteNotificationLogResponse) + return this.equals((WriteNotificationLogResponse)that); + return false; + } + + public boolean equals(WriteNotificationLogResponse that) { + if (that == null) + return false; + + return true; + } + + @Override + public int hashCode() { + List list = new ArrayList(); + + return list.hashCode(); + } + + @Override + public int compareTo(WriteNotificationLogResponse other) { + if (!getClass().equals(other.getClass())) { + return getClass().getName().compareTo(other.getClass().getName()); + } + + int lastComparison = 0; + + return 0; + } + + public _Fields fieldForId(int fieldId) { + return _Fields.findByThriftId(fieldId); + } + + public void read(org.apache.thrift.protocol.TProtocol iprot) throws org.apache.thrift.TException { + schemes.get(iprot.getScheme()).getScheme().read(iprot, this); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot) throws org.apache.thrift.TException { + schemes.get(oprot.getScheme()).getScheme().write(oprot, this); + } + + @Override + public String toString() { + StringBuilder sb = new StringBuilder("WriteNotificationLogResponse("); + boolean first = true; + + sb.append(")"); + return sb.toString(); + } + + public void validate() throws org.apache.thrift.TException { + // check for required fields + // check for sub-struct validity + } + + private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException { + try { + write(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(out))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, ClassNotFoundException { + try { + read(new org.apache.thrift.protocol.TCompactProtocol(new org.apache.thrift.transport.TIOStreamTransport(in))); + } catch (org.apache.thrift.TException te) { + throw new java.io.IOException(te); + } + } + + private static class WriteNotificationLogResponseStandardSchemeFactory implements SchemeFactory { + public WriteNotificationLogResponseStandardScheme getScheme() { + return new WriteNotificationLogResponseStandardScheme(); + } + } + + private static class WriteNotificationLogResponseStandardScheme extends StandardScheme { + + public void read(org.apache.thrift.protocol.TProtocol iprot, WriteNotificationLogResponse struct) throws org.apache.thrift.TException { + org.apache.thrift.protocol.TField schemeField; + iprot.readStructBegin(); + while (true) + { + schemeField = iprot.readFieldBegin(); + if (schemeField.type == org.apache.thrift.protocol.TType.STOP) { + break; + } + switch (schemeField.id) { + default: + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + iprot.readFieldEnd(); + } + iprot.readStructEnd(); + struct.validate(); + } + + public void write(org.apache.thrift.protocol.TProtocol oprot, WriteNotificationLogResponse struct) throws org.apache.thrift.TException { + struct.validate(); + + oprot.writeStructBegin(STRUCT_DESC); + oprot.writeFieldStop(); + oprot.writeStructEnd(); + } + + } + + private static class WriteNotificationLogResponseTupleSchemeFactory implements SchemeFactory { + public WriteNotificationLogResponseTupleScheme getScheme() { + return new WriteNotificationLogResponseTupleScheme(); + } + } + + private static class WriteNotificationLogResponseTupleScheme extends TupleScheme { + + @Override + public void write(org.apache.thrift.protocol.TProtocol prot, WriteNotificationLogResponse struct) throws org.apache.thrift.TException { + TTupleProtocol oprot = (TTupleProtocol) prot; + } + + @Override + public void read(org.apache.thrift.protocol.TProtocol prot, WriteNotificationLogResponse struct) throws org.apache.thrift.TException { + TTupleProtocol iprot = (TTupleProtocol) prot; + } + } + +} + diff --git a/standalone-metastore/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore.php b/standalone-metastore/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore.php index 632b82c396..29e787ba6e 100644 --- a/standalone-metastore/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore.php +++ b/standalone-metastore/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore.php @@ -1270,6 +1270,11 @@ interface ThriftHiveMetastoreIf extends \FacebookServiceIf { /** */ public function flushCache(); + /** + * @param \metastore\WriteNotificationLogRequest $rqst + * @return \metastore\WriteNotificationLogResponse + */ + public function add_write_notification_log(\metastore\WriteNotificationLogRequest $rqst); /** * @param \metastore\CmRecycleRequest $request * @return \metastore\CmRecycleResponse @@ -10933,6 +10938,57 @@ class ThriftHiveMetastoreClient extends \FacebookServiceClient implements \metas return; } + public function add_write_notification_log(\metastore\WriteNotificationLogRequest $rqst) + { + $this->send_add_write_notification_log($rqst); + return $this->recv_add_write_notification_log(); + } + + public function send_add_write_notification_log(\metastore\WriteNotificationLogRequest $rqst) + { + $args = new \metastore\ThriftHiveMetastore_add_write_notification_log_args(); + $args->rqst = $rqst; + $bin_accel = ($this->output_ instanceof TBinaryProtocolAccelerated) && function_exists('thrift_protocol_write_binary'); + if ($bin_accel) + { + thrift_protocol_write_binary($this->output_, 'add_write_notification_log', TMessageType::CALL, $args, $this->seqid_, $this->output_->isStrictWrite()); + } + else + { + $this->output_->writeMessageBegin('add_write_notification_log', TMessageType::CALL, $this->seqid_); + $args->write($this->output_); + $this->output_->writeMessageEnd(); + $this->output_->getTransport()->flush(); + } + } + + public function recv_add_write_notification_log() + { + $bin_accel = ($this->input_ instanceof TBinaryProtocolAccelerated) && function_exists('thrift_protocol_read_binary'); + if ($bin_accel) $result = thrift_protocol_read_binary($this->input_, '\metastore\ThriftHiveMetastore_add_write_notification_log_result', $this->input_->isStrictRead()); + else + { + $rseqid = 0; + $fname = null; + $mtype = 0; + + $this->input_->readMessageBegin($fname, $mtype, $rseqid); + if ($mtype == TMessageType::EXCEPTION) { + $x = new TApplicationException(); + $x->read($this->input_); + $this->input_->readMessageEnd(); + throw $x; + } + $result = new \metastore\ThriftHiveMetastore_add_write_notification_log_result(); + $result->read($this->input_); + $this->input_->readMessageEnd(); + } + if ($result->success !== null) { + return $result->success; + } + throw new \Exception("add_write_notification_log failed: unknown result"); + } + public function cm_recycle(\metastore\CmRecycleRequest $request) { $this->send_cm_recycle($request); @@ -15440,14 +15496,14 @@ class ThriftHiveMetastore_get_databases_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size812 = 0; - $_etype815 = 0; - $xfer += $input->readListBegin($_etype815, $_size812); - for ($_i816 = 0; $_i816 < $_size812; ++$_i816) + $_size833 = 0; + $_etype836 = 0; + $xfer += $input->readListBegin($_etype836, $_size833); + for ($_i837 = 0; $_i837 < $_size833; ++$_i837) { - $elem817 = null; - $xfer += $input->readString($elem817); - $this->success []= $elem817; + $elem838 = null; + $xfer += $input->readString($elem838); + $this->success []= $elem838; } $xfer += $input->readListEnd(); } else { @@ -15483,9 +15539,9 @@ class ThriftHiveMetastore_get_databases_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter818) + foreach ($this->success as $iter839) { - $xfer += $output->writeString($iter818); + $xfer += $output->writeString($iter839); } } $output->writeListEnd(); @@ -15616,14 +15672,14 @@ class ThriftHiveMetastore_get_all_databases_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size819 = 0; - $_etype822 = 0; - $xfer += $input->readListBegin($_etype822, $_size819); - for ($_i823 = 0; $_i823 < $_size819; ++$_i823) + $_size840 = 0; + $_etype843 = 0; + $xfer += $input->readListBegin($_etype843, $_size840); + for ($_i844 = 0; $_i844 < $_size840; ++$_i844) { - $elem824 = null; - $xfer += $input->readString($elem824); - $this->success []= $elem824; + $elem845 = null; + $xfer += $input->readString($elem845); + $this->success []= $elem845; } $xfer += $input->readListEnd(); } else { @@ -15659,9 +15715,9 @@ class ThriftHiveMetastore_get_all_databases_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter825) + foreach ($this->success as $iter846) { - $xfer += $output->writeString($iter825); + $xfer += $output->writeString($iter846); } } $output->writeListEnd(); @@ -16662,18 +16718,18 @@ class ThriftHiveMetastore_get_type_all_result { case 0: if ($ftype == TType::MAP) { $this->success = array(); - $_size826 = 0; - $_ktype827 = 0; - $_vtype828 = 0; - $xfer += $input->readMapBegin($_ktype827, $_vtype828, $_size826); - for ($_i830 = 0; $_i830 < $_size826; ++$_i830) + $_size847 = 0; + $_ktype848 = 0; + $_vtype849 = 0; + $xfer += $input->readMapBegin($_ktype848, $_vtype849, $_size847); + for ($_i851 = 0; $_i851 < $_size847; ++$_i851) { - $key831 = ''; - $val832 = new \metastore\Type(); - $xfer += $input->readString($key831); - $val832 = new \metastore\Type(); - $xfer += $val832->read($input); - $this->success[$key831] = $val832; + $key852 = ''; + $val853 = new \metastore\Type(); + $xfer += $input->readString($key852); + $val853 = new \metastore\Type(); + $xfer += $val853->read($input); + $this->success[$key852] = $val853; } $xfer += $input->readMapEnd(); } else { @@ -16709,10 +16765,10 @@ class ThriftHiveMetastore_get_type_all_result { { $output->writeMapBegin(TType::STRING, TType::STRUCT, count($this->success)); { - foreach ($this->success as $kiter833 => $viter834) + foreach ($this->success as $kiter854 => $viter855) { - $xfer += $output->writeString($kiter833); - $xfer += $viter834->write($output); + $xfer += $output->writeString($kiter854); + $xfer += $viter855->write($output); } } $output->writeMapEnd(); @@ -16916,15 +16972,15 @@ class ThriftHiveMetastore_get_fields_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size835 = 0; - $_etype838 = 0; - $xfer += $input->readListBegin($_etype838, $_size835); - for ($_i839 = 0; $_i839 < $_size835; ++$_i839) + $_size856 = 0; + $_etype859 = 0; + $xfer += $input->readListBegin($_etype859, $_size856); + for ($_i860 = 0; $_i860 < $_size856; ++$_i860) { - $elem840 = null; - $elem840 = new \metastore\FieldSchema(); - $xfer += $elem840->read($input); - $this->success []= $elem840; + $elem861 = null; + $elem861 = new \metastore\FieldSchema(); + $xfer += $elem861->read($input); + $this->success []= $elem861; } $xfer += $input->readListEnd(); } else { @@ -16976,9 +17032,9 @@ class ThriftHiveMetastore_get_fields_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter841) + foreach ($this->success as $iter862) { - $xfer += $iter841->write($output); + $xfer += $iter862->write($output); } } $output->writeListEnd(); @@ -17220,15 +17276,15 @@ class ThriftHiveMetastore_get_fields_with_environment_context_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size842 = 0; - $_etype845 = 0; - $xfer += $input->readListBegin($_etype845, $_size842); - for ($_i846 = 0; $_i846 < $_size842; ++$_i846) + $_size863 = 0; + $_etype866 = 0; + $xfer += $input->readListBegin($_etype866, $_size863); + for ($_i867 = 0; $_i867 < $_size863; ++$_i867) { - $elem847 = null; - $elem847 = new \metastore\FieldSchema(); - $xfer += $elem847->read($input); - $this->success []= $elem847; + $elem868 = null; + $elem868 = new \metastore\FieldSchema(); + $xfer += $elem868->read($input); + $this->success []= $elem868; } $xfer += $input->readListEnd(); } else { @@ -17280,9 +17336,9 @@ class ThriftHiveMetastore_get_fields_with_environment_context_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter848) + foreach ($this->success as $iter869) { - $xfer += $iter848->write($output); + $xfer += $iter869->write($output); } } $output->writeListEnd(); @@ -17496,15 +17552,15 @@ class ThriftHiveMetastore_get_schema_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size849 = 0; - $_etype852 = 0; - $xfer += $input->readListBegin($_etype852, $_size849); - for ($_i853 = 0; $_i853 < $_size849; ++$_i853) + $_size870 = 0; + $_etype873 = 0; + $xfer += $input->readListBegin($_etype873, $_size870); + for ($_i874 = 0; $_i874 < $_size870; ++$_i874) { - $elem854 = null; - $elem854 = new \metastore\FieldSchema(); - $xfer += $elem854->read($input); - $this->success []= $elem854; + $elem875 = null; + $elem875 = new \metastore\FieldSchema(); + $xfer += $elem875->read($input); + $this->success []= $elem875; } $xfer += $input->readListEnd(); } else { @@ -17556,9 +17612,9 @@ class ThriftHiveMetastore_get_schema_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter855) + foreach ($this->success as $iter876) { - $xfer += $iter855->write($output); + $xfer += $iter876->write($output); } } $output->writeListEnd(); @@ -17800,15 +17856,15 @@ class ThriftHiveMetastore_get_schema_with_environment_context_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size856 = 0; - $_etype859 = 0; - $xfer += $input->readListBegin($_etype859, $_size856); - for ($_i860 = 0; $_i860 < $_size856; ++$_i860) + $_size877 = 0; + $_etype880 = 0; + $xfer += $input->readListBegin($_etype880, $_size877); + for ($_i881 = 0; $_i881 < $_size877; ++$_i881) { - $elem861 = null; - $elem861 = new \metastore\FieldSchema(); - $xfer += $elem861->read($input); - $this->success []= $elem861; + $elem882 = null; + $elem882 = new \metastore\FieldSchema(); + $xfer += $elem882->read($input); + $this->success []= $elem882; } $xfer += $input->readListEnd(); } else { @@ -17860,9 +17916,9 @@ class ThriftHiveMetastore_get_schema_with_environment_context_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter862) + foreach ($this->success as $iter883) { - $xfer += $iter862->write($output); + $xfer += $iter883->write($output); } } $output->writeListEnd(); @@ -18534,15 +18590,15 @@ class ThriftHiveMetastore_create_table_with_constraints_args { case 2: if ($ftype == TType::LST) { $this->primaryKeys = array(); - $_size863 = 0; - $_etype866 = 0; - $xfer += $input->readListBegin($_etype866, $_size863); - for ($_i867 = 0; $_i867 < $_size863; ++$_i867) + $_size884 = 0; + $_etype887 = 0; + $xfer += $input->readListBegin($_etype887, $_size884); + for ($_i888 = 0; $_i888 < $_size884; ++$_i888) { - $elem868 = null; - $elem868 = new \metastore\SQLPrimaryKey(); - $xfer += $elem868->read($input); - $this->primaryKeys []= $elem868; + $elem889 = null; + $elem889 = new \metastore\SQLPrimaryKey(); + $xfer += $elem889->read($input); + $this->primaryKeys []= $elem889; } $xfer += $input->readListEnd(); } else { @@ -18552,15 +18608,15 @@ class ThriftHiveMetastore_create_table_with_constraints_args { case 3: if ($ftype == TType::LST) { $this->foreignKeys = array(); - $_size869 = 0; - $_etype872 = 0; - $xfer += $input->readListBegin($_etype872, $_size869); - for ($_i873 = 0; $_i873 < $_size869; ++$_i873) + $_size890 = 0; + $_etype893 = 0; + $xfer += $input->readListBegin($_etype893, $_size890); + for ($_i894 = 0; $_i894 < $_size890; ++$_i894) { - $elem874 = null; - $elem874 = new \metastore\SQLForeignKey(); - $xfer += $elem874->read($input); - $this->foreignKeys []= $elem874; + $elem895 = null; + $elem895 = new \metastore\SQLForeignKey(); + $xfer += $elem895->read($input); + $this->foreignKeys []= $elem895; } $xfer += $input->readListEnd(); } else { @@ -18570,15 +18626,15 @@ class ThriftHiveMetastore_create_table_with_constraints_args { case 4: if ($ftype == TType::LST) { $this->uniqueConstraints = array(); - $_size875 = 0; - $_etype878 = 0; - $xfer += $input->readListBegin($_etype878, $_size875); - for ($_i879 = 0; $_i879 < $_size875; ++$_i879) + $_size896 = 0; + $_etype899 = 0; + $xfer += $input->readListBegin($_etype899, $_size896); + for ($_i900 = 0; $_i900 < $_size896; ++$_i900) { - $elem880 = null; - $elem880 = new \metastore\SQLUniqueConstraint(); - $xfer += $elem880->read($input); - $this->uniqueConstraints []= $elem880; + $elem901 = null; + $elem901 = new \metastore\SQLUniqueConstraint(); + $xfer += $elem901->read($input); + $this->uniqueConstraints []= $elem901; } $xfer += $input->readListEnd(); } else { @@ -18588,15 +18644,15 @@ class ThriftHiveMetastore_create_table_with_constraints_args { case 5: if ($ftype == TType::LST) { $this->notNullConstraints = array(); - $_size881 = 0; - $_etype884 = 0; - $xfer += $input->readListBegin($_etype884, $_size881); - for ($_i885 = 0; $_i885 < $_size881; ++$_i885) + $_size902 = 0; + $_etype905 = 0; + $xfer += $input->readListBegin($_etype905, $_size902); + for ($_i906 = 0; $_i906 < $_size902; ++$_i906) { - $elem886 = null; - $elem886 = new \metastore\SQLNotNullConstraint(); - $xfer += $elem886->read($input); - $this->notNullConstraints []= $elem886; + $elem907 = null; + $elem907 = new \metastore\SQLNotNullConstraint(); + $xfer += $elem907->read($input); + $this->notNullConstraints []= $elem907; } $xfer += $input->readListEnd(); } else { @@ -18606,15 +18662,15 @@ class ThriftHiveMetastore_create_table_with_constraints_args { case 6: if ($ftype == TType::LST) { $this->defaultConstraints = array(); - $_size887 = 0; - $_etype890 = 0; - $xfer += $input->readListBegin($_etype890, $_size887); - for ($_i891 = 0; $_i891 < $_size887; ++$_i891) + $_size908 = 0; + $_etype911 = 0; + $xfer += $input->readListBegin($_etype911, $_size908); + for ($_i912 = 0; $_i912 < $_size908; ++$_i912) { - $elem892 = null; - $elem892 = new \metastore\SQLDefaultConstraint(); - $xfer += $elem892->read($input); - $this->defaultConstraints []= $elem892; + $elem913 = null; + $elem913 = new \metastore\SQLDefaultConstraint(); + $xfer += $elem913->read($input); + $this->defaultConstraints []= $elem913; } $xfer += $input->readListEnd(); } else { @@ -18624,15 +18680,15 @@ class ThriftHiveMetastore_create_table_with_constraints_args { case 7: if ($ftype == TType::LST) { $this->checkConstraints = array(); - $_size893 = 0; - $_etype896 = 0; - $xfer += $input->readListBegin($_etype896, $_size893); - for ($_i897 = 0; $_i897 < $_size893; ++$_i897) + $_size914 = 0; + $_etype917 = 0; + $xfer += $input->readListBegin($_etype917, $_size914); + for ($_i918 = 0; $_i918 < $_size914; ++$_i918) { - $elem898 = null; - $elem898 = new \metastore\SQLCheckConstraint(); - $xfer += $elem898->read($input); - $this->checkConstraints []= $elem898; + $elem919 = null; + $elem919 = new \metastore\SQLCheckConstraint(); + $xfer += $elem919->read($input); + $this->checkConstraints []= $elem919; } $xfer += $input->readListEnd(); } else { @@ -18668,9 +18724,9 @@ class ThriftHiveMetastore_create_table_with_constraints_args { { $output->writeListBegin(TType::STRUCT, count($this->primaryKeys)); { - foreach ($this->primaryKeys as $iter899) + foreach ($this->primaryKeys as $iter920) { - $xfer += $iter899->write($output); + $xfer += $iter920->write($output); } } $output->writeListEnd(); @@ -18685,9 +18741,9 @@ class ThriftHiveMetastore_create_table_with_constraints_args { { $output->writeListBegin(TType::STRUCT, count($this->foreignKeys)); { - foreach ($this->foreignKeys as $iter900) + foreach ($this->foreignKeys as $iter921) { - $xfer += $iter900->write($output); + $xfer += $iter921->write($output); } } $output->writeListEnd(); @@ -18702,9 +18758,9 @@ class ThriftHiveMetastore_create_table_with_constraints_args { { $output->writeListBegin(TType::STRUCT, count($this->uniqueConstraints)); { - foreach ($this->uniqueConstraints as $iter901) + foreach ($this->uniqueConstraints as $iter922) { - $xfer += $iter901->write($output); + $xfer += $iter922->write($output); } } $output->writeListEnd(); @@ -18719,9 +18775,9 @@ class ThriftHiveMetastore_create_table_with_constraints_args { { $output->writeListBegin(TType::STRUCT, count($this->notNullConstraints)); { - foreach ($this->notNullConstraints as $iter902) + foreach ($this->notNullConstraints as $iter923) { - $xfer += $iter902->write($output); + $xfer += $iter923->write($output); } } $output->writeListEnd(); @@ -18736,9 +18792,9 @@ class ThriftHiveMetastore_create_table_with_constraints_args { { $output->writeListBegin(TType::STRUCT, count($this->defaultConstraints)); { - foreach ($this->defaultConstraints as $iter903) + foreach ($this->defaultConstraints as $iter924) { - $xfer += $iter903->write($output); + $xfer += $iter924->write($output); } } $output->writeListEnd(); @@ -18753,9 +18809,9 @@ class ThriftHiveMetastore_create_table_with_constraints_args { { $output->writeListBegin(TType::STRUCT, count($this->checkConstraints)); { - foreach ($this->checkConstraints as $iter904) + foreach ($this->checkConstraints as $iter925) { - $xfer += $iter904->write($output); + $xfer += $iter925->write($output); } } $output->writeListEnd(); @@ -20755,14 +20811,14 @@ class ThriftHiveMetastore_truncate_table_args { case 3: if ($ftype == TType::LST) { $this->partNames = array(); - $_size905 = 0; - $_etype908 = 0; - $xfer += $input->readListBegin($_etype908, $_size905); - for ($_i909 = 0; $_i909 < $_size905; ++$_i909) + $_size926 = 0; + $_etype929 = 0; + $xfer += $input->readListBegin($_etype929, $_size926); + for ($_i930 = 0; $_i930 < $_size926; ++$_i930) { - $elem910 = null; - $xfer += $input->readString($elem910); - $this->partNames []= $elem910; + $elem931 = null; + $xfer += $input->readString($elem931); + $this->partNames []= $elem931; } $xfer += $input->readListEnd(); } else { @@ -20800,9 +20856,9 @@ class ThriftHiveMetastore_truncate_table_args { { $output->writeListBegin(TType::STRING, count($this->partNames)); { - foreach ($this->partNames as $iter911) + foreach ($this->partNames as $iter932) { - $xfer += $output->writeString($iter911); + $xfer += $output->writeString($iter932); } } $output->writeListEnd(); @@ -21053,14 +21109,14 @@ class ThriftHiveMetastore_get_tables_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size912 = 0; - $_etype915 = 0; - $xfer += $input->readListBegin($_etype915, $_size912); - for ($_i916 = 0; $_i916 < $_size912; ++$_i916) + $_size933 = 0; + $_etype936 = 0; + $xfer += $input->readListBegin($_etype936, $_size933); + for ($_i937 = 0; $_i937 < $_size933; ++$_i937) { - $elem917 = null; - $xfer += $input->readString($elem917); - $this->success []= $elem917; + $elem938 = null; + $xfer += $input->readString($elem938); + $this->success []= $elem938; } $xfer += $input->readListEnd(); } else { @@ -21096,9 +21152,9 @@ class ThriftHiveMetastore_get_tables_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter918) + foreach ($this->success as $iter939) { - $xfer += $output->writeString($iter918); + $xfer += $output->writeString($iter939); } } $output->writeListEnd(); @@ -21300,14 +21356,14 @@ class ThriftHiveMetastore_get_tables_by_type_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size919 = 0; - $_etype922 = 0; - $xfer += $input->readListBegin($_etype922, $_size919); - for ($_i923 = 0; $_i923 < $_size919; ++$_i923) + $_size940 = 0; + $_etype943 = 0; + $xfer += $input->readListBegin($_etype943, $_size940); + for ($_i944 = 0; $_i944 < $_size940; ++$_i944) { - $elem924 = null; - $xfer += $input->readString($elem924); - $this->success []= $elem924; + $elem945 = null; + $xfer += $input->readString($elem945); + $this->success []= $elem945; } $xfer += $input->readListEnd(); } else { @@ -21343,9 +21399,9 @@ class ThriftHiveMetastore_get_tables_by_type_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter925) + foreach ($this->success as $iter946) { - $xfer += $output->writeString($iter925); + $xfer += $output->writeString($iter946); } } $output->writeListEnd(); @@ -21501,14 +21557,14 @@ class ThriftHiveMetastore_get_materialized_views_for_rewriting_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size926 = 0; - $_etype929 = 0; - $xfer += $input->readListBegin($_etype929, $_size926); - for ($_i930 = 0; $_i930 < $_size926; ++$_i930) + $_size947 = 0; + $_etype950 = 0; + $xfer += $input->readListBegin($_etype950, $_size947); + for ($_i951 = 0; $_i951 < $_size947; ++$_i951) { - $elem931 = null; - $xfer += $input->readString($elem931); - $this->success []= $elem931; + $elem952 = null; + $xfer += $input->readString($elem952); + $this->success []= $elem952; } $xfer += $input->readListEnd(); } else { @@ -21544,9 +21600,9 @@ class ThriftHiveMetastore_get_materialized_views_for_rewriting_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter932) + foreach ($this->success as $iter953) { - $xfer += $output->writeString($iter932); + $xfer += $output->writeString($iter953); } } $output->writeListEnd(); @@ -21651,14 +21707,14 @@ class ThriftHiveMetastore_get_table_meta_args { case 3: if ($ftype == TType::LST) { $this->tbl_types = array(); - $_size933 = 0; - $_etype936 = 0; - $xfer += $input->readListBegin($_etype936, $_size933); - for ($_i937 = 0; $_i937 < $_size933; ++$_i937) + $_size954 = 0; + $_etype957 = 0; + $xfer += $input->readListBegin($_etype957, $_size954); + for ($_i958 = 0; $_i958 < $_size954; ++$_i958) { - $elem938 = null; - $xfer += $input->readString($elem938); - $this->tbl_types []= $elem938; + $elem959 = null; + $xfer += $input->readString($elem959); + $this->tbl_types []= $elem959; } $xfer += $input->readListEnd(); } else { @@ -21696,9 +21752,9 @@ class ThriftHiveMetastore_get_table_meta_args { { $output->writeListBegin(TType::STRING, count($this->tbl_types)); { - foreach ($this->tbl_types as $iter939) + foreach ($this->tbl_types as $iter960) { - $xfer += $output->writeString($iter939); + $xfer += $output->writeString($iter960); } } $output->writeListEnd(); @@ -21775,15 +21831,15 @@ class ThriftHiveMetastore_get_table_meta_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size940 = 0; - $_etype943 = 0; - $xfer += $input->readListBegin($_etype943, $_size940); - for ($_i944 = 0; $_i944 < $_size940; ++$_i944) + $_size961 = 0; + $_etype964 = 0; + $xfer += $input->readListBegin($_etype964, $_size961); + for ($_i965 = 0; $_i965 < $_size961; ++$_i965) { - $elem945 = null; - $elem945 = new \metastore\TableMeta(); - $xfer += $elem945->read($input); - $this->success []= $elem945; + $elem966 = null; + $elem966 = new \metastore\TableMeta(); + $xfer += $elem966->read($input); + $this->success []= $elem966; } $xfer += $input->readListEnd(); } else { @@ -21819,9 +21875,9 @@ class ThriftHiveMetastore_get_table_meta_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter946) + foreach ($this->success as $iter967) { - $xfer += $iter946->write($output); + $xfer += $iter967->write($output); } } $output->writeListEnd(); @@ -21977,14 +22033,14 @@ class ThriftHiveMetastore_get_all_tables_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size947 = 0; - $_etype950 = 0; - $xfer += $input->readListBegin($_etype950, $_size947); - for ($_i951 = 0; $_i951 < $_size947; ++$_i951) + $_size968 = 0; + $_etype971 = 0; + $xfer += $input->readListBegin($_etype971, $_size968); + for ($_i972 = 0; $_i972 < $_size968; ++$_i972) { - $elem952 = null; - $xfer += $input->readString($elem952); - $this->success []= $elem952; + $elem973 = null; + $xfer += $input->readString($elem973); + $this->success []= $elem973; } $xfer += $input->readListEnd(); } else { @@ -22020,9 +22076,9 @@ class ThriftHiveMetastore_get_all_tables_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter953) + foreach ($this->success as $iter974) { - $xfer += $output->writeString($iter953); + $xfer += $output->writeString($iter974); } } $output->writeListEnd(); @@ -22337,14 +22393,14 @@ class ThriftHiveMetastore_get_table_objects_by_name_args { case 2: if ($ftype == TType::LST) { $this->tbl_names = array(); - $_size954 = 0; - $_etype957 = 0; - $xfer += $input->readListBegin($_etype957, $_size954); - for ($_i958 = 0; $_i958 < $_size954; ++$_i958) + $_size975 = 0; + $_etype978 = 0; + $xfer += $input->readListBegin($_etype978, $_size975); + for ($_i979 = 0; $_i979 < $_size975; ++$_i979) { - $elem959 = null; - $xfer += $input->readString($elem959); - $this->tbl_names []= $elem959; + $elem980 = null; + $xfer += $input->readString($elem980); + $this->tbl_names []= $elem980; } $xfer += $input->readListEnd(); } else { @@ -22377,9 +22433,9 @@ class ThriftHiveMetastore_get_table_objects_by_name_args { { $output->writeListBegin(TType::STRING, count($this->tbl_names)); { - foreach ($this->tbl_names as $iter960) + foreach ($this->tbl_names as $iter981) { - $xfer += $output->writeString($iter960); + $xfer += $output->writeString($iter981); } } $output->writeListEnd(); @@ -22444,15 +22500,15 @@ class ThriftHiveMetastore_get_table_objects_by_name_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size961 = 0; - $_etype964 = 0; - $xfer += $input->readListBegin($_etype964, $_size961); - for ($_i965 = 0; $_i965 < $_size961; ++$_i965) + $_size982 = 0; + $_etype985 = 0; + $xfer += $input->readListBegin($_etype985, $_size982); + for ($_i986 = 0; $_i986 < $_size982; ++$_i986) { - $elem966 = null; - $elem966 = new \metastore\Table(); - $xfer += $elem966->read($input); - $this->success []= $elem966; + $elem987 = null; + $elem987 = new \metastore\Table(); + $xfer += $elem987->read($input); + $this->success []= $elem987; } $xfer += $input->readListEnd(); } else { @@ -22480,9 +22536,9 @@ class ThriftHiveMetastore_get_table_objects_by_name_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter967) + foreach ($this->success as $iter988) { - $xfer += $iter967->write($output); + $xfer += $iter988->write($output); } } $output->writeListEnd(); @@ -23682,14 +23738,14 @@ class ThriftHiveMetastore_get_table_names_by_filter_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size968 = 0; - $_etype971 = 0; - $xfer += $input->readListBegin($_etype971, $_size968); - for ($_i972 = 0; $_i972 < $_size968; ++$_i972) + $_size989 = 0; + $_etype992 = 0; + $xfer += $input->readListBegin($_etype992, $_size989); + for ($_i993 = 0; $_i993 < $_size989; ++$_i993) { - $elem973 = null; - $xfer += $input->readString($elem973); - $this->success []= $elem973; + $elem994 = null; + $xfer += $input->readString($elem994); + $this->success []= $elem994; } $xfer += $input->readListEnd(); } else { @@ -23741,9 +23797,9 @@ class ThriftHiveMetastore_get_table_names_by_filter_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter974) + foreach ($this->success as $iter995) { - $xfer += $output->writeString($iter974); + $xfer += $output->writeString($iter995); } } $output->writeListEnd(); @@ -25056,15 +25112,15 @@ class ThriftHiveMetastore_add_partitions_args { case 1: if ($ftype == TType::LST) { $this->new_parts = array(); - $_size975 = 0; - $_etype978 = 0; - $xfer += $input->readListBegin($_etype978, $_size975); - for ($_i979 = 0; $_i979 < $_size975; ++$_i979) + $_size996 = 0; + $_etype999 = 0; + $xfer += $input->readListBegin($_etype999, $_size996); + for ($_i1000 = 0; $_i1000 < $_size996; ++$_i1000) { - $elem980 = null; - $elem980 = new \metastore\Partition(); - $xfer += $elem980->read($input); - $this->new_parts []= $elem980; + $elem1001 = null; + $elem1001 = new \metastore\Partition(); + $xfer += $elem1001->read($input); + $this->new_parts []= $elem1001; } $xfer += $input->readListEnd(); } else { @@ -25092,9 +25148,9 @@ class ThriftHiveMetastore_add_partitions_args { { $output->writeListBegin(TType::STRUCT, count($this->new_parts)); { - foreach ($this->new_parts as $iter981) + foreach ($this->new_parts as $iter1002) { - $xfer += $iter981->write($output); + $xfer += $iter1002->write($output); } } $output->writeListEnd(); @@ -25309,15 +25365,15 @@ class ThriftHiveMetastore_add_partitions_pspec_args { case 1: if ($ftype == TType::LST) { $this->new_parts = array(); - $_size982 = 0; - $_etype985 = 0; - $xfer += $input->readListBegin($_etype985, $_size982); - for ($_i986 = 0; $_i986 < $_size982; ++$_i986) + $_size1003 = 0; + $_etype1006 = 0; + $xfer += $input->readListBegin($_etype1006, $_size1003); + for ($_i1007 = 0; $_i1007 < $_size1003; ++$_i1007) { - $elem987 = null; - $elem987 = new \metastore\PartitionSpec(); - $xfer += $elem987->read($input); - $this->new_parts []= $elem987; + $elem1008 = null; + $elem1008 = new \metastore\PartitionSpec(); + $xfer += $elem1008->read($input); + $this->new_parts []= $elem1008; } $xfer += $input->readListEnd(); } else { @@ -25345,9 +25401,9 @@ class ThriftHiveMetastore_add_partitions_pspec_args { { $output->writeListBegin(TType::STRUCT, count($this->new_parts)); { - foreach ($this->new_parts as $iter988) + foreach ($this->new_parts as $iter1009) { - $xfer += $iter988->write($output); + $xfer += $iter1009->write($output); } } $output->writeListEnd(); @@ -25597,14 +25653,14 @@ class ThriftHiveMetastore_append_partition_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size989 = 0; - $_etype992 = 0; - $xfer += $input->readListBegin($_etype992, $_size989); - for ($_i993 = 0; $_i993 < $_size989; ++$_i993) + $_size1010 = 0; + $_etype1013 = 0; + $xfer += $input->readListBegin($_etype1013, $_size1010); + for ($_i1014 = 0; $_i1014 < $_size1010; ++$_i1014) { - $elem994 = null; - $xfer += $input->readString($elem994); - $this->part_vals []= $elem994; + $elem1015 = null; + $xfer += $input->readString($elem1015); + $this->part_vals []= $elem1015; } $xfer += $input->readListEnd(); } else { @@ -25642,9 +25698,9 @@ class ThriftHiveMetastore_append_partition_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter995) + foreach ($this->part_vals as $iter1016) { - $xfer += $output->writeString($iter995); + $xfer += $output->writeString($iter1016); } } $output->writeListEnd(); @@ -26146,14 +26202,14 @@ class ThriftHiveMetastore_append_partition_with_environment_context_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size996 = 0; - $_etype999 = 0; - $xfer += $input->readListBegin($_etype999, $_size996); - for ($_i1000 = 0; $_i1000 < $_size996; ++$_i1000) + $_size1017 = 0; + $_etype1020 = 0; + $xfer += $input->readListBegin($_etype1020, $_size1017); + for ($_i1021 = 0; $_i1021 < $_size1017; ++$_i1021) { - $elem1001 = null; - $xfer += $input->readString($elem1001); - $this->part_vals []= $elem1001; + $elem1022 = null; + $xfer += $input->readString($elem1022); + $this->part_vals []= $elem1022; } $xfer += $input->readListEnd(); } else { @@ -26199,9 +26255,9 @@ class ThriftHiveMetastore_append_partition_with_environment_context_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1002) + foreach ($this->part_vals as $iter1023) { - $xfer += $output->writeString($iter1002); + $xfer += $output->writeString($iter1023); } } $output->writeListEnd(); @@ -27055,14 +27111,14 @@ class ThriftHiveMetastore_drop_partition_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1003 = 0; - $_etype1006 = 0; - $xfer += $input->readListBegin($_etype1006, $_size1003); - for ($_i1007 = 0; $_i1007 < $_size1003; ++$_i1007) + $_size1024 = 0; + $_etype1027 = 0; + $xfer += $input->readListBegin($_etype1027, $_size1024); + for ($_i1028 = 0; $_i1028 < $_size1024; ++$_i1028) { - $elem1008 = null; - $xfer += $input->readString($elem1008); - $this->part_vals []= $elem1008; + $elem1029 = null; + $xfer += $input->readString($elem1029); + $this->part_vals []= $elem1029; } $xfer += $input->readListEnd(); } else { @@ -27107,9 +27163,9 @@ class ThriftHiveMetastore_drop_partition_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1009) + foreach ($this->part_vals as $iter1030) { - $xfer += $output->writeString($iter1009); + $xfer += $output->writeString($iter1030); } } $output->writeListEnd(); @@ -27362,14 +27418,14 @@ class ThriftHiveMetastore_drop_partition_with_environment_context_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1010 = 0; - $_etype1013 = 0; - $xfer += $input->readListBegin($_etype1013, $_size1010); - for ($_i1014 = 0; $_i1014 < $_size1010; ++$_i1014) + $_size1031 = 0; + $_etype1034 = 0; + $xfer += $input->readListBegin($_etype1034, $_size1031); + for ($_i1035 = 0; $_i1035 < $_size1031; ++$_i1035) { - $elem1015 = null; - $xfer += $input->readString($elem1015); - $this->part_vals []= $elem1015; + $elem1036 = null; + $xfer += $input->readString($elem1036); + $this->part_vals []= $elem1036; } $xfer += $input->readListEnd(); } else { @@ -27422,9 +27478,9 @@ class ThriftHiveMetastore_drop_partition_with_environment_context_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1016) + foreach ($this->part_vals as $iter1037) { - $xfer += $output->writeString($iter1016); + $xfer += $output->writeString($iter1037); } } $output->writeListEnd(); @@ -28438,14 +28494,14 @@ class ThriftHiveMetastore_get_partition_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1017 = 0; - $_etype1020 = 0; - $xfer += $input->readListBegin($_etype1020, $_size1017); - for ($_i1021 = 0; $_i1021 < $_size1017; ++$_i1021) + $_size1038 = 0; + $_etype1041 = 0; + $xfer += $input->readListBegin($_etype1041, $_size1038); + for ($_i1042 = 0; $_i1042 < $_size1038; ++$_i1042) { - $elem1022 = null; - $xfer += $input->readString($elem1022); - $this->part_vals []= $elem1022; + $elem1043 = null; + $xfer += $input->readString($elem1043); + $this->part_vals []= $elem1043; } $xfer += $input->readListEnd(); } else { @@ -28483,9 +28539,9 @@ class ThriftHiveMetastore_get_partition_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1023) + foreach ($this->part_vals as $iter1044) { - $xfer += $output->writeString($iter1023); + $xfer += $output->writeString($iter1044); } } $output->writeListEnd(); @@ -28727,17 +28783,17 @@ class ThriftHiveMetastore_exchange_partition_args { case 1: if ($ftype == TType::MAP) { $this->partitionSpecs = array(); - $_size1024 = 0; - $_ktype1025 = 0; - $_vtype1026 = 0; - $xfer += $input->readMapBegin($_ktype1025, $_vtype1026, $_size1024); - for ($_i1028 = 0; $_i1028 < $_size1024; ++$_i1028) + $_size1045 = 0; + $_ktype1046 = 0; + $_vtype1047 = 0; + $xfer += $input->readMapBegin($_ktype1046, $_vtype1047, $_size1045); + for ($_i1049 = 0; $_i1049 < $_size1045; ++$_i1049) { - $key1029 = ''; - $val1030 = ''; - $xfer += $input->readString($key1029); - $xfer += $input->readString($val1030); - $this->partitionSpecs[$key1029] = $val1030; + $key1050 = ''; + $val1051 = ''; + $xfer += $input->readString($key1050); + $xfer += $input->readString($val1051); + $this->partitionSpecs[$key1050] = $val1051; } $xfer += $input->readMapEnd(); } else { @@ -28793,10 +28849,10 @@ class ThriftHiveMetastore_exchange_partition_args { { $output->writeMapBegin(TType::STRING, TType::STRING, count($this->partitionSpecs)); { - foreach ($this->partitionSpecs as $kiter1031 => $viter1032) + foreach ($this->partitionSpecs as $kiter1052 => $viter1053) { - $xfer += $output->writeString($kiter1031); - $xfer += $output->writeString($viter1032); + $xfer += $output->writeString($kiter1052); + $xfer += $output->writeString($viter1053); } } $output->writeMapEnd(); @@ -29108,17 +29164,17 @@ class ThriftHiveMetastore_exchange_partitions_args { case 1: if ($ftype == TType::MAP) { $this->partitionSpecs = array(); - $_size1033 = 0; - $_ktype1034 = 0; - $_vtype1035 = 0; - $xfer += $input->readMapBegin($_ktype1034, $_vtype1035, $_size1033); - for ($_i1037 = 0; $_i1037 < $_size1033; ++$_i1037) + $_size1054 = 0; + $_ktype1055 = 0; + $_vtype1056 = 0; + $xfer += $input->readMapBegin($_ktype1055, $_vtype1056, $_size1054); + for ($_i1058 = 0; $_i1058 < $_size1054; ++$_i1058) { - $key1038 = ''; - $val1039 = ''; - $xfer += $input->readString($key1038); - $xfer += $input->readString($val1039); - $this->partitionSpecs[$key1038] = $val1039; + $key1059 = ''; + $val1060 = ''; + $xfer += $input->readString($key1059); + $xfer += $input->readString($val1060); + $this->partitionSpecs[$key1059] = $val1060; } $xfer += $input->readMapEnd(); } else { @@ -29174,10 +29230,10 @@ class ThriftHiveMetastore_exchange_partitions_args { { $output->writeMapBegin(TType::STRING, TType::STRING, count($this->partitionSpecs)); { - foreach ($this->partitionSpecs as $kiter1040 => $viter1041) + foreach ($this->partitionSpecs as $kiter1061 => $viter1062) { - $xfer += $output->writeString($kiter1040); - $xfer += $output->writeString($viter1041); + $xfer += $output->writeString($kiter1061); + $xfer += $output->writeString($viter1062); } } $output->writeMapEnd(); @@ -29310,15 +29366,15 @@ class ThriftHiveMetastore_exchange_partitions_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1042 = 0; - $_etype1045 = 0; - $xfer += $input->readListBegin($_etype1045, $_size1042); - for ($_i1046 = 0; $_i1046 < $_size1042; ++$_i1046) + $_size1063 = 0; + $_etype1066 = 0; + $xfer += $input->readListBegin($_etype1066, $_size1063); + for ($_i1067 = 0; $_i1067 < $_size1063; ++$_i1067) { - $elem1047 = null; - $elem1047 = new \metastore\Partition(); - $xfer += $elem1047->read($input); - $this->success []= $elem1047; + $elem1068 = null; + $elem1068 = new \metastore\Partition(); + $xfer += $elem1068->read($input); + $this->success []= $elem1068; } $xfer += $input->readListEnd(); } else { @@ -29378,9 +29434,9 @@ class ThriftHiveMetastore_exchange_partitions_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1048) + foreach ($this->success as $iter1069) { - $xfer += $iter1048->write($output); + $xfer += $iter1069->write($output); } } $output->writeListEnd(); @@ -29526,14 +29582,14 @@ class ThriftHiveMetastore_get_partition_with_auth_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1049 = 0; - $_etype1052 = 0; - $xfer += $input->readListBegin($_etype1052, $_size1049); - for ($_i1053 = 0; $_i1053 < $_size1049; ++$_i1053) + $_size1070 = 0; + $_etype1073 = 0; + $xfer += $input->readListBegin($_etype1073, $_size1070); + for ($_i1074 = 0; $_i1074 < $_size1070; ++$_i1074) { - $elem1054 = null; - $xfer += $input->readString($elem1054); - $this->part_vals []= $elem1054; + $elem1075 = null; + $xfer += $input->readString($elem1075); + $this->part_vals []= $elem1075; } $xfer += $input->readListEnd(); } else { @@ -29550,14 +29606,14 @@ class ThriftHiveMetastore_get_partition_with_auth_args { case 5: if ($ftype == TType::LST) { $this->group_names = array(); - $_size1055 = 0; - $_etype1058 = 0; - $xfer += $input->readListBegin($_etype1058, $_size1055); - for ($_i1059 = 0; $_i1059 < $_size1055; ++$_i1059) + $_size1076 = 0; + $_etype1079 = 0; + $xfer += $input->readListBegin($_etype1079, $_size1076); + for ($_i1080 = 0; $_i1080 < $_size1076; ++$_i1080) { - $elem1060 = null; - $xfer += $input->readString($elem1060); - $this->group_names []= $elem1060; + $elem1081 = null; + $xfer += $input->readString($elem1081); + $this->group_names []= $elem1081; } $xfer += $input->readListEnd(); } else { @@ -29595,9 +29651,9 @@ class ThriftHiveMetastore_get_partition_with_auth_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1061) + foreach ($this->part_vals as $iter1082) { - $xfer += $output->writeString($iter1061); + $xfer += $output->writeString($iter1082); } } $output->writeListEnd(); @@ -29617,9 +29673,9 @@ class ThriftHiveMetastore_get_partition_with_auth_args { { $output->writeListBegin(TType::STRING, count($this->group_names)); { - foreach ($this->group_names as $iter1062) + foreach ($this->group_names as $iter1083) { - $xfer += $output->writeString($iter1062); + $xfer += $output->writeString($iter1083); } } $output->writeListEnd(); @@ -30210,15 +30266,15 @@ class ThriftHiveMetastore_get_partitions_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1063 = 0; - $_etype1066 = 0; - $xfer += $input->readListBegin($_etype1066, $_size1063); - for ($_i1067 = 0; $_i1067 < $_size1063; ++$_i1067) + $_size1084 = 0; + $_etype1087 = 0; + $xfer += $input->readListBegin($_etype1087, $_size1084); + for ($_i1088 = 0; $_i1088 < $_size1084; ++$_i1088) { - $elem1068 = null; - $elem1068 = new \metastore\Partition(); - $xfer += $elem1068->read($input); - $this->success []= $elem1068; + $elem1089 = null; + $elem1089 = new \metastore\Partition(); + $xfer += $elem1089->read($input); + $this->success []= $elem1089; } $xfer += $input->readListEnd(); } else { @@ -30262,9 +30318,9 @@ class ThriftHiveMetastore_get_partitions_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1069) + foreach ($this->success as $iter1090) { - $xfer += $iter1069->write($output); + $xfer += $iter1090->write($output); } } $output->writeListEnd(); @@ -30410,14 +30466,14 @@ class ThriftHiveMetastore_get_partitions_with_auth_args { case 5: if ($ftype == TType::LST) { $this->group_names = array(); - $_size1070 = 0; - $_etype1073 = 0; - $xfer += $input->readListBegin($_etype1073, $_size1070); - for ($_i1074 = 0; $_i1074 < $_size1070; ++$_i1074) + $_size1091 = 0; + $_etype1094 = 0; + $xfer += $input->readListBegin($_etype1094, $_size1091); + for ($_i1095 = 0; $_i1095 < $_size1091; ++$_i1095) { - $elem1075 = null; - $xfer += $input->readString($elem1075); - $this->group_names []= $elem1075; + $elem1096 = null; + $xfer += $input->readString($elem1096); + $this->group_names []= $elem1096; } $xfer += $input->readListEnd(); } else { @@ -30465,9 +30521,9 @@ class ThriftHiveMetastore_get_partitions_with_auth_args { { $output->writeListBegin(TType::STRING, count($this->group_names)); { - foreach ($this->group_names as $iter1076) + foreach ($this->group_names as $iter1097) { - $xfer += $output->writeString($iter1076); + $xfer += $output->writeString($iter1097); } } $output->writeListEnd(); @@ -30556,15 +30612,15 @@ class ThriftHiveMetastore_get_partitions_with_auth_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1077 = 0; - $_etype1080 = 0; - $xfer += $input->readListBegin($_etype1080, $_size1077); - for ($_i1081 = 0; $_i1081 < $_size1077; ++$_i1081) + $_size1098 = 0; + $_etype1101 = 0; + $xfer += $input->readListBegin($_etype1101, $_size1098); + for ($_i1102 = 0; $_i1102 < $_size1098; ++$_i1102) { - $elem1082 = null; - $elem1082 = new \metastore\Partition(); - $xfer += $elem1082->read($input); - $this->success []= $elem1082; + $elem1103 = null; + $elem1103 = new \metastore\Partition(); + $xfer += $elem1103->read($input); + $this->success []= $elem1103; } $xfer += $input->readListEnd(); } else { @@ -30608,9 +30664,9 @@ class ThriftHiveMetastore_get_partitions_with_auth_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1083) + foreach ($this->success as $iter1104) { - $xfer += $iter1083->write($output); + $xfer += $iter1104->write($output); } } $output->writeListEnd(); @@ -30830,15 +30886,15 @@ class ThriftHiveMetastore_get_partitions_pspec_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1084 = 0; - $_etype1087 = 0; - $xfer += $input->readListBegin($_etype1087, $_size1084); - for ($_i1088 = 0; $_i1088 < $_size1084; ++$_i1088) + $_size1105 = 0; + $_etype1108 = 0; + $xfer += $input->readListBegin($_etype1108, $_size1105); + for ($_i1109 = 0; $_i1109 < $_size1105; ++$_i1109) { - $elem1089 = null; - $elem1089 = new \metastore\PartitionSpec(); - $xfer += $elem1089->read($input); - $this->success []= $elem1089; + $elem1110 = null; + $elem1110 = new \metastore\PartitionSpec(); + $xfer += $elem1110->read($input); + $this->success []= $elem1110; } $xfer += $input->readListEnd(); } else { @@ -30882,9 +30938,9 @@ class ThriftHiveMetastore_get_partitions_pspec_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1090) + foreach ($this->success as $iter1111) { - $xfer += $iter1090->write($output); + $xfer += $iter1111->write($output); } } $output->writeListEnd(); @@ -31103,14 +31159,14 @@ class ThriftHiveMetastore_get_partition_names_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1091 = 0; - $_etype1094 = 0; - $xfer += $input->readListBegin($_etype1094, $_size1091); - for ($_i1095 = 0; $_i1095 < $_size1091; ++$_i1095) + $_size1112 = 0; + $_etype1115 = 0; + $xfer += $input->readListBegin($_etype1115, $_size1112); + for ($_i1116 = 0; $_i1116 < $_size1112; ++$_i1116) { - $elem1096 = null; - $xfer += $input->readString($elem1096); - $this->success []= $elem1096; + $elem1117 = null; + $xfer += $input->readString($elem1117); + $this->success []= $elem1117; } $xfer += $input->readListEnd(); } else { @@ -31154,9 +31210,9 @@ class ThriftHiveMetastore_get_partition_names_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1097) + foreach ($this->success as $iter1118) { - $xfer += $output->writeString($iter1097); + $xfer += $output->writeString($iter1118); } } $output->writeListEnd(); @@ -31487,14 +31543,14 @@ class ThriftHiveMetastore_get_partitions_ps_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1098 = 0; - $_etype1101 = 0; - $xfer += $input->readListBegin($_etype1101, $_size1098); - for ($_i1102 = 0; $_i1102 < $_size1098; ++$_i1102) + $_size1119 = 0; + $_etype1122 = 0; + $xfer += $input->readListBegin($_etype1122, $_size1119); + for ($_i1123 = 0; $_i1123 < $_size1119; ++$_i1123) { - $elem1103 = null; - $xfer += $input->readString($elem1103); - $this->part_vals []= $elem1103; + $elem1124 = null; + $xfer += $input->readString($elem1124); + $this->part_vals []= $elem1124; } $xfer += $input->readListEnd(); } else { @@ -31539,9 +31595,9 @@ class ThriftHiveMetastore_get_partitions_ps_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1104) + foreach ($this->part_vals as $iter1125) { - $xfer += $output->writeString($iter1104); + $xfer += $output->writeString($iter1125); } } $output->writeListEnd(); @@ -31635,15 +31691,15 @@ class ThriftHiveMetastore_get_partitions_ps_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1105 = 0; - $_etype1108 = 0; - $xfer += $input->readListBegin($_etype1108, $_size1105); - for ($_i1109 = 0; $_i1109 < $_size1105; ++$_i1109) + $_size1126 = 0; + $_etype1129 = 0; + $xfer += $input->readListBegin($_etype1129, $_size1126); + for ($_i1130 = 0; $_i1130 < $_size1126; ++$_i1130) { - $elem1110 = null; - $elem1110 = new \metastore\Partition(); - $xfer += $elem1110->read($input); - $this->success []= $elem1110; + $elem1131 = null; + $elem1131 = new \metastore\Partition(); + $xfer += $elem1131->read($input); + $this->success []= $elem1131; } $xfer += $input->readListEnd(); } else { @@ -31687,9 +31743,9 @@ class ThriftHiveMetastore_get_partitions_ps_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1111) + foreach ($this->success as $iter1132) { - $xfer += $iter1111->write($output); + $xfer += $iter1132->write($output); } } $output->writeListEnd(); @@ -31836,14 +31892,14 @@ class ThriftHiveMetastore_get_partitions_ps_with_auth_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1112 = 0; - $_etype1115 = 0; - $xfer += $input->readListBegin($_etype1115, $_size1112); - for ($_i1116 = 0; $_i1116 < $_size1112; ++$_i1116) + $_size1133 = 0; + $_etype1136 = 0; + $xfer += $input->readListBegin($_etype1136, $_size1133); + for ($_i1137 = 0; $_i1137 < $_size1133; ++$_i1137) { - $elem1117 = null; - $xfer += $input->readString($elem1117); - $this->part_vals []= $elem1117; + $elem1138 = null; + $xfer += $input->readString($elem1138); + $this->part_vals []= $elem1138; } $xfer += $input->readListEnd(); } else { @@ -31867,14 +31923,14 @@ class ThriftHiveMetastore_get_partitions_ps_with_auth_args { case 6: if ($ftype == TType::LST) { $this->group_names = array(); - $_size1118 = 0; - $_etype1121 = 0; - $xfer += $input->readListBegin($_etype1121, $_size1118); - for ($_i1122 = 0; $_i1122 < $_size1118; ++$_i1122) + $_size1139 = 0; + $_etype1142 = 0; + $xfer += $input->readListBegin($_etype1142, $_size1139); + for ($_i1143 = 0; $_i1143 < $_size1139; ++$_i1143) { - $elem1123 = null; - $xfer += $input->readString($elem1123); - $this->group_names []= $elem1123; + $elem1144 = null; + $xfer += $input->readString($elem1144); + $this->group_names []= $elem1144; } $xfer += $input->readListEnd(); } else { @@ -31912,9 +31968,9 @@ class ThriftHiveMetastore_get_partitions_ps_with_auth_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1124) + foreach ($this->part_vals as $iter1145) { - $xfer += $output->writeString($iter1124); + $xfer += $output->writeString($iter1145); } } $output->writeListEnd(); @@ -31939,9 +31995,9 @@ class ThriftHiveMetastore_get_partitions_ps_with_auth_args { { $output->writeListBegin(TType::STRING, count($this->group_names)); { - foreach ($this->group_names as $iter1125) + foreach ($this->group_names as $iter1146) { - $xfer += $output->writeString($iter1125); + $xfer += $output->writeString($iter1146); } } $output->writeListEnd(); @@ -32030,15 +32086,15 @@ class ThriftHiveMetastore_get_partitions_ps_with_auth_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1126 = 0; - $_etype1129 = 0; - $xfer += $input->readListBegin($_etype1129, $_size1126); - for ($_i1130 = 0; $_i1130 < $_size1126; ++$_i1130) + $_size1147 = 0; + $_etype1150 = 0; + $xfer += $input->readListBegin($_etype1150, $_size1147); + for ($_i1151 = 0; $_i1151 < $_size1147; ++$_i1151) { - $elem1131 = null; - $elem1131 = new \metastore\Partition(); - $xfer += $elem1131->read($input); - $this->success []= $elem1131; + $elem1152 = null; + $elem1152 = new \metastore\Partition(); + $xfer += $elem1152->read($input); + $this->success []= $elem1152; } $xfer += $input->readListEnd(); } else { @@ -32082,9 +32138,9 @@ class ThriftHiveMetastore_get_partitions_ps_with_auth_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1132) + foreach ($this->success as $iter1153) { - $xfer += $iter1132->write($output); + $xfer += $iter1153->write($output); } } $output->writeListEnd(); @@ -32205,14 +32261,14 @@ class ThriftHiveMetastore_get_partition_names_ps_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1133 = 0; - $_etype1136 = 0; - $xfer += $input->readListBegin($_etype1136, $_size1133); - for ($_i1137 = 0; $_i1137 < $_size1133; ++$_i1137) + $_size1154 = 0; + $_etype1157 = 0; + $xfer += $input->readListBegin($_etype1157, $_size1154); + for ($_i1158 = 0; $_i1158 < $_size1154; ++$_i1158) { - $elem1138 = null; - $xfer += $input->readString($elem1138); - $this->part_vals []= $elem1138; + $elem1159 = null; + $xfer += $input->readString($elem1159); + $this->part_vals []= $elem1159; } $xfer += $input->readListEnd(); } else { @@ -32257,9 +32313,9 @@ class ThriftHiveMetastore_get_partition_names_ps_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1139) + foreach ($this->part_vals as $iter1160) { - $xfer += $output->writeString($iter1139); + $xfer += $output->writeString($iter1160); } } $output->writeListEnd(); @@ -32352,14 +32408,14 @@ class ThriftHiveMetastore_get_partition_names_ps_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1140 = 0; - $_etype1143 = 0; - $xfer += $input->readListBegin($_etype1143, $_size1140); - for ($_i1144 = 0; $_i1144 < $_size1140; ++$_i1144) + $_size1161 = 0; + $_etype1164 = 0; + $xfer += $input->readListBegin($_etype1164, $_size1161); + for ($_i1165 = 0; $_i1165 < $_size1161; ++$_i1165) { - $elem1145 = null; - $xfer += $input->readString($elem1145); - $this->success []= $elem1145; + $elem1166 = null; + $xfer += $input->readString($elem1166); + $this->success []= $elem1166; } $xfer += $input->readListEnd(); } else { @@ -32403,9 +32459,9 @@ class ThriftHiveMetastore_get_partition_names_ps_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1146) + foreach ($this->success as $iter1167) { - $xfer += $output->writeString($iter1146); + $xfer += $output->writeString($iter1167); } } $output->writeListEnd(); @@ -32648,15 +32704,15 @@ class ThriftHiveMetastore_get_partitions_by_filter_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1147 = 0; - $_etype1150 = 0; - $xfer += $input->readListBegin($_etype1150, $_size1147); - for ($_i1151 = 0; $_i1151 < $_size1147; ++$_i1151) + $_size1168 = 0; + $_etype1171 = 0; + $xfer += $input->readListBegin($_etype1171, $_size1168); + for ($_i1172 = 0; $_i1172 < $_size1168; ++$_i1172) { - $elem1152 = null; - $elem1152 = new \metastore\Partition(); - $xfer += $elem1152->read($input); - $this->success []= $elem1152; + $elem1173 = null; + $elem1173 = new \metastore\Partition(); + $xfer += $elem1173->read($input); + $this->success []= $elem1173; } $xfer += $input->readListEnd(); } else { @@ -32700,9 +32756,9 @@ class ThriftHiveMetastore_get_partitions_by_filter_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1153) + foreach ($this->success as $iter1174) { - $xfer += $iter1153->write($output); + $xfer += $iter1174->write($output); } } $output->writeListEnd(); @@ -32945,15 +33001,15 @@ class ThriftHiveMetastore_get_part_specs_by_filter_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1154 = 0; - $_etype1157 = 0; - $xfer += $input->readListBegin($_etype1157, $_size1154); - for ($_i1158 = 0; $_i1158 < $_size1154; ++$_i1158) + $_size1175 = 0; + $_etype1178 = 0; + $xfer += $input->readListBegin($_etype1178, $_size1175); + for ($_i1179 = 0; $_i1179 < $_size1175; ++$_i1179) { - $elem1159 = null; - $elem1159 = new \metastore\PartitionSpec(); - $xfer += $elem1159->read($input); - $this->success []= $elem1159; + $elem1180 = null; + $elem1180 = new \metastore\PartitionSpec(); + $xfer += $elem1180->read($input); + $this->success []= $elem1180; } $xfer += $input->readListEnd(); } else { @@ -32997,9 +33053,9 @@ class ThriftHiveMetastore_get_part_specs_by_filter_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1160) + foreach ($this->success as $iter1181) { - $xfer += $iter1160->write($output); + $xfer += $iter1181->write($output); } } $output->writeListEnd(); @@ -33565,14 +33621,14 @@ class ThriftHiveMetastore_get_partitions_by_names_args { case 3: if ($ftype == TType::LST) { $this->names = array(); - $_size1161 = 0; - $_etype1164 = 0; - $xfer += $input->readListBegin($_etype1164, $_size1161); - for ($_i1165 = 0; $_i1165 < $_size1161; ++$_i1165) + $_size1182 = 0; + $_etype1185 = 0; + $xfer += $input->readListBegin($_etype1185, $_size1182); + for ($_i1186 = 0; $_i1186 < $_size1182; ++$_i1186) { - $elem1166 = null; - $xfer += $input->readString($elem1166); - $this->names []= $elem1166; + $elem1187 = null; + $xfer += $input->readString($elem1187); + $this->names []= $elem1187; } $xfer += $input->readListEnd(); } else { @@ -33610,9 +33666,9 @@ class ThriftHiveMetastore_get_partitions_by_names_args { { $output->writeListBegin(TType::STRING, count($this->names)); { - foreach ($this->names as $iter1167) + foreach ($this->names as $iter1188) { - $xfer += $output->writeString($iter1167); + $xfer += $output->writeString($iter1188); } } $output->writeListEnd(); @@ -33701,15 +33757,15 @@ class ThriftHiveMetastore_get_partitions_by_names_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1168 = 0; - $_etype1171 = 0; - $xfer += $input->readListBegin($_etype1171, $_size1168); - for ($_i1172 = 0; $_i1172 < $_size1168; ++$_i1172) + $_size1189 = 0; + $_etype1192 = 0; + $xfer += $input->readListBegin($_etype1192, $_size1189); + for ($_i1193 = 0; $_i1193 < $_size1189; ++$_i1193) { - $elem1173 = null; - $elem1173 = new \metastore\Partition(); - $xfer += $elem1173->read($input); - $this->success []= $elem1173; + $elem1194 = null; + $elem1194 = new \metastore\Partition(); + $xfer += $elem1194->read($input); + $this->success []= $elem1194; } $xfer += $input->readListEnd(); } else { @@ -33753,9 +33809,9 @@ class ThriftHiveMetastore_get_partitions_by_names_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1174) + foreach ($this->success as $iter1195) { - $xfer += $iter1174->write($output); + $xfer += $iter1195->write($output); } } $output->writeListEnd(); @@ -34094,15 +34150,15 @@ class ThriftHiveMetastore_alter_partitions_args { case 3: if ($ftype == TType::LST) { $this->new_parts = array(); - $_size1175 = 0; - $_etype1178 = 0; - $xfer += $input->readListBegin($_etype1178, $_size1175); - for ($_i1179 = 0; $_i1179 < $_size1175; ++$_i1179) + $_size1196 = 0; + $_etype1199 = 0; + $xfer += $input->readListBegin($_etype1199, $_size1196); + for ($_i1200 = 0; $_i1200 < $_size1196; ++$_i1200) { - $elem1180 = null; - $elem1180 = new \metastore\Partition(); - $xfer += $elem1180->read($input); - $this->new_parts []= $elem1180; + $elem1201 = null; + $elem1201 = new \metastore\Partition(); + $xfer += $elem1201->read($input); + $this->new_parts []= $elem1201; } $xfer += $input->readListEnd(); } else { @@ -34140,9 +34196,9 @@ class ThriftHiveMetastore_alter_partitions_args { { $output->writeListBegin(TType::STRUCT, count($this->new_parts)); { - foreach ($this->new_parts as $iter1181) + foreach ($this->new_parts as $iter1202) { - $xfer += $iter1181->write($output); + $xfer += $iter1202->write($output); } } $output->writeListEnd(); @@ -34357,15 +34413,15 @@ class ThriftHiveMetastore_alter_partitions_with_environment_context_args { case 3: if ($ftype == TType::LST) { $this->new_parts = array(); - $_size1182 = 0; - $_etype1185 = 0; - $xfer += $input->readListBegin($_etype1185, $_size1182); - for ($_i1186 = 0; $_i1186 < $_size1182; ++$_i1186) + $_size1203 = 0; + $_etype1206 = 0; + $xfer += $input->readListBegin($_etype1206, $_size1203); + for ($_i1207 = 0; $_i1207 < $_size1203; ++$_i1207) { - $elem1187 = null; - $elem1187 = new \metastore\Partition(); - $xfer += $elem1187->read($input); - $this->new_parts []= $elem1187; + $elem1208 = null; + $elem1208 = new \metastore\Partition(); + $xfer += $elem1208->read($input); + $this->new_parts []= $elem1208; } $xfer += $input->readListEnd(); } else { @@ -34411,9 +34467,9 @@ class ThriftHiveMetastore_alter_partitions_with_environment_context_args { { $output->writeListBegin(TType::STRUCT, count($this->new_parts)); { - foreach ($this->new_parts as $iter1188) + foreach ($this->new_parts as $iter1209) { - $xfer += $iter1188->write($output); + $xfer += $iter1209->write($output); } } $output->writeListEnd(); @@ -34891,14 +34947,14 @@ class ThriftHiveMetastore_rename_partition_args { case 3: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1189 = 0; - $_etype1192 = 0; - $xfer += $input->readListBegin($_etype1192, $_size1189); - for ($_i1193 = 0; $_i1193 < $_size1189; ++$_i1193) + $_size1210 = 0; + $_etype1213 = 0; + $xfer += $input->readListBegin($_etype1213, $_size1210); + for ($_i1214 = 0; $_i1214 < $_size1210; ++$_i1214) { - $elem1194 = null; - $xfer += $input->readString($elem1194); - $this->part_vals []= $elem1194; + $elem1215 = null; + $xfer += $input->readString($elem1215); + $this->part_vals []= $elem1215; } $xfer += $input->readListEnd(); } else { @@ -34944,9 +35000,9 @@ class ThriftHiveMetastore_rename_partition_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1195) + foreach ($this->part_vals as $iter1216) { - $xfer += $output->writeString($iter1195); + $xfer += $output->writeString($iter1216); } } $output->writeListEnd(); @@ -35131,14 +35187,14 @@ class ThriftHiveMetastore_partition_name_has_valid_characters_args { case 1: if ($ftype == TType::LST) { $this->part_vals = array(); - $_size1196 = 0; - $_etype1199 = 0; - $xfer += $input->readListBegin($_etype1199, $_size1196); - for ($_i1200 = 0; $_i1200 < $_size1196; ++$_i1200) + $_size1217 = 0; + $_etype1220 = 0; + $xfer += $input->readListBegin($_etype1220, $_size1217); + for ($_i1221 = 0; $_i1221 < $_size1217; ++$_i1221) { - $elem1201 = null; - $xfer += $input->readString($elem1201); - $this->part_vals []= $elem1201; + $elem1222 = null; + $xfer += $input->readString($elem1222); + $this->part_vals []= $elem1222; } $xfer += $input->readListEnd(); } else { @@ -35173,9 +35229,9 @@ class ThriftHiveMetastore_partition_name_has_valid_characters_args { { $output->writeListBegin(TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $iter1202) + foreach ($this->part_vals as $iter1223) { - $xfer += $output->writeString($iter1202); + $xfer += $output->writeString($iter1223); } } $output->writeListEnd(); @@ -35629,14 +35685,14 @@ class ThriftHiveMetastore_partition_name_to_vals_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1203 = 0; - $_etype1206 = 0; - $xfer += $input->readListBegin($_etype1206, $_size1203); - for ($_i1207 = 0; $_i1207 < $_size1203; ++$_i1207) + $_size1224 = 0; + $_etype1227 = 0; + $xfer += $input->readListBegin($_etype1227, $_size1224); + for ($_i1228 = 0; $_i1228 < $_size1224; ++$_i1228) { - $elem1208 = null; - $xfer += $input->readString($elem1208); - $this->success []= $elem1208; + $elem1229 = null; + $xfer += $input->readString($elem1229); + $this->success []= $elem1229; } $xfer += $input->readListEnd(); } else { @@ -35672,9 +35728,9 @@ class ThriftHiveMetastore_partition_name_to_vals_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1209) + foreach ($this->success as $iter1230) { - $xfer += $output->writeString($iter1209); + $xfer += $output->writeString($iter1230); } } $output->writeListEnd(); @@ -35834,17 +35890,17 @@ class ThriftHiveMetastore_partition_name_to_spec_result { case 0: if ($ftype == TType::MAP) { $this->success = array(); - $_size1210 = 0; - $_ktype1211 = 0; - $_vtype1212 = 0; - $xfer += $input->readMapBegin($_ktype1211, $_vtype1212, $_size1210); - for ($_i1214 = 0; $_i1214 < $_size1210; ++$_i1214) + $_size1231 = 0; + $_ktype1232 = 0; + $_vtype1233 = 0; + $xfer += $input->readMapBegin($_ktype1232, $_vtype1233, $_size1231); + for ($_i1235 = 0; $_i1235 < $_size1231; ++$_i1235) { - $key1215 = ''; - $val1216 = ''; - $xfer += $input->readString($key1215); - $xfer += $input->readString($val1216); - $this->success[$key1215] = $val1216; + $key1236 = ''; + $val1237 = ''; + $xfer += $input->readString($key1236); + $xfer += $input->readString($val1237); + $this->success[$key1236] = $val1237; } $xfer += $input->readMapEnd(); } else { @@ -35880,10 +35936,10 @@ class ThriftHiveMetastore_partition_name_to_spec_result { { $output->writeMapBegin(TType::STRING, TType::STRING, count($this->success)); { - foreach ($this->success as $kiter1217 => $viter1218) + foreach ($this->success as $kiter1238 => $viter1239) { - $xfer += $output->writeString($kiter1217); - $xfer += $output->writeString($viter1218); + $xfer += $output->writeString($kiter1238); + $xfer += $output->writeString($viter1239); } } $output->writeMapEnd(); @@ -36003,17 +36059,17 @@ class ThriftHiveMetastore_markPartitionForEvent_args { case 3: if ($ftype == TType::MAP) { $this->part_vals = array(); - $_size1219 = 0; - $_ktype1220 = 0; - $_vtype1221 = 0; - $xfer += $input->readMapBegin($_ktype1220, $_vtype1221, $_size1219); - for ($_i1223 = 0; $_i1223 < $_size1219; ++$_i1223) + $_size1240 = 0; + $_ktype1241 = 0; + $_vtype1242 = 0; + $xfer += $input->readMapBegin($_ktype1241, $_vtype1242, $_size1240); + for ($_i1244 = 0; $_i1244 < $_size1240; ++$_i1244) { - $key1224 = ''; - $val1225 = ''; - $xfer += $input->readString($key1224); - $xfer += $input->readString($val1225); - $this->part_vals[$key1224] = $val1225; + $key1245 = ''; + $val1246 = ''; + $xfer += $input->readString($key1245); + $xfer += $input->readString($val1246); + $this->part_vals[$key1245] = $val1246; } $xfer += $input->readMapEnd(); } else { @@ -36058,10 +36114,10 @@ class ThriftHiveMetastore_markPartitionForEvent_args { { $output->writeMapBegin(TType::STRING, TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $kiter1226 => $viter1227) + foreach ($this->part_vals as $kiter1247 => $viter1248) { - $xfer += $output->writeString($kiter1226); - $xfer += $output->writeString($viter1227); + $xfer += $output->writeString($kiter1247); + $xfer += $output->writeString($viter1248); } } $output->writeMapEnd(); @@ -36383,17 +36439,17 @@ class ThriftHiveMetastore_isPartitionMarkedForEvent_args { case 3: if ($ftype == TType::MAP) { $this->part_vals = array(); - $_size1228 = 0; - $_ktype1229 = 0; - $_vtype1230 = 0; - $xfer += $input->readMapBegin($_ktype1229, $_vtype1230, $_size1228); - for ($_i1232 = 0; $_i1232 < $_size1228; ++$_i1232) + $_size1249 = 0; + $_ktype1250 = 0; + $_vtype1251 = 0; + $xfer += $input->readMapBegin($_ktype1250, $_vtype1251, $_size1249); + for ($_i1253 = 0; $_i1253 < $_size1249; ++$_i1253) { - $key1233 = ''; - $val1234 = ''; - $xfer += $input->readString($key1233); - $xfer += $input->readString($val1234); - $this->part_vals[$key1233] = $val1234; + $key1254 = ''; + $val1255 = ''; + $xfer += $input->readString($key1254); + $xfer += $input->readString($val1255); + $this->part_vals[$key1254] = $val1255; } $xfer += $input->readMapEnd(); } else { @@ -36438,10 +36494,10 @@ class ThriftHiveMetastore_isPartitionMarkedForEvent_args { { $output->writeMapBegin(TType::STRING, TType::STRING, count($this->part_vals)); { - foreach ($this->part_vals as $kiter1235 => $viter1236) + foreach ($this->part_vals as $kiter1256 => $viter1257) { - $xfer += $output->writeString($kiter1235); - $xfer += $output->writeString($viter1236); + $xfer += $output->writeString($kiter1256); + $xfer += $output->writeString($viter1257); } } $output->writeMapEnd(); @@ -41400,14 +41456,14 @@ class ThriftHiveMetastore_get_functions_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1237 = 0; - $_etype1240 = 0; - $xfer += $input->readListBegin($_etype1240, $_size1237); - for ($_i1241 = 0; $_i1241 < $_size1237; ++$_i1241) + $_size1258 = 0; + $_etype1261 = 0; + $xfer += $input->readListBegin($_etype1261, $_size1258); + for ($_i1262 = 0; $_i1262 < $_size1258; ++$_i1262) { - $elem1242 = null; - $xfer += $input->readString($elem1242); - $this->success []= $elem1242; + $elem1263 = null; + $xfer += $input->readString($elem1263); + $this->success []= $elem1263; } $xfer += $input->readListEnd(); } else { @@ -41443,9 +41499,9 @@ class ThriftHiveMetastore_get_functions_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1243) + foreach ($this->success as $iter1264) { - $xfer += $output->writeString($iter1243); + $xfer += $output->writeString($iter1264); } } $output->writeListEnd(); @@ -42314,14 +42370,14 @@ class ThriftHiveMetastore_get_role_names_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1244 = 0; - $_etype1247 = 0; - $xfer += $input->readListBegin($_etype1247, $_size1244); - for ($_i1248 = 0; $_i1248 < $_size1244; ++$_i1248) + $_size1265 = 0; + $_etype1268 = 0; + $xfer += $input->readListBegin($_etype1268, $_size1265); + for ($_i1269 = 0; $_i1269 < $_size1265; ++$_i1269) { - $elem1249 = null; - $xfer += $input->readString($elem1249); - $this->success []= $elem1249; + $elem1270 = null; + $xfer += $input->readString($elem1270); + $this->success []= $elem1270; } $xfer += $input->readListEnd(); } else { @@ -42357,9 +42413,9 @@ class ThriftHiveMetastore_get_role_names_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1250) + foreach ($this->success as $iter1271) { - $xfer += $output->writeString($iter1250); + $xfer += $output->writeString($iter1271); } } $output->writeListEnd(); @@ -43050,15 +43106,15 @@ class ThriftHiveMetastore_list_roles_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1251 = 0; - $_etype1254 = 0; - $xfer += $input->readListBegin($_etype1254, $_size1251); - for ($_i1255 = 0; $_i1255 < $_size1251; ++$_i1255) + $_size1272 = 0; + $_etype1275 = 0; + $xfer += $input->readListBegin($_etype1275, $_size1272); + for ($_i1276 = 0; $_i1276 < $_size1272; ++$_i1276) { - $elem1256 = null; - $elem1256 = new \metastore\Role(); - $xfer += $elem1256->read($input); - $this->success []= $elem1256; + $elem1277 = null; + $elem1277 = new \metastore\Role(); + $xfer += $elem1277->read($input); + $this->success []= $elem1277; } $xfer += $input->readListEnd(); } else { @@ -43094,9 +43150,9 @@ class ThriftHiveMetastore_list_roles_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1257) + foreach ($this->success as $iter1278) { - $xfer += $iter1257->write($output); + $xfer += $iter1278->write($output); } } $output->writeListEnd(); @@ -43758,14 +43814,14 @@ class ThriftHiveMetastore_get_privilege_set_args { case 3: if ($ftype == TType::LST) { $this->group_names = array(); - $_size1258 = 0; - $_etype1261 = 0; - $xfer += $input->readListBegin($_etype1261, $_size1258); - for ($_i1262 = 0; $_i1262 < $_size1258; ++$_i1262) + $_size1279 = 0; + $_etype1282 = 0; + $xfer += $input->readListBegin($_etype1282, $_size1279); + for ($_i1283 = 0; $_i1283 < $_size1279; ++$_i1283) { - $elem1263 = null; - $xfer += $input->readString($elem1263); - $this->group_names []= $elem1263; + $elem1284 = null; + $xfer += $input->readString($elem1284); + $this->group_names []= $elem1284; } $xfer += $input->readListEnd(); } else { @@ -43806,9 +43862,9 @@ class ThriftHiveMetastore_get_privilege_set_args { { $output->writeListBegin(TType::STRING, count($this->group_names)); { - foreach ($this->group_names as $iter1264) + foreach ($this->group_names as $iter1285) { - $xfer += $output->writeString($iter1264); + $xfer += $output->writeString($iter1285); } } $output->writeListEnd(); @@ -44116,15 +44172,15 @@ class ThriftHiveMetastore_list_privileges_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1265 = 0; - $_etype1268 = 0; - $xfer += $input->readListBegin($_etype1268, $_size1265); - for ($_i1269 = 0; $_i1269 < $_size1265; ++$_i1269) + $_size1286 = 0; + $_etype1289 = 0; + $xfer += $input->readListBegin($_etype1289, $_size1286); + for ($_i1290 = 0; $_i1290 < $_size1286; ++$_i1290) { - $elem1270 = null; - $elem1270 = new \metastore\HiveObjectPrivilege(); - $xfer += $elem1270->read($input); - $this->success []= $elem1270; + $elem1291 = null; + $elem1291 = new \metastore\HiveObjectPrivilege(); + $xfer += $elem1291->read($input); + $this->success []= $elem1291; } $xfer += $input->readListEnd(); } else { @@ -44160,9 +44216,9 @@ class ThriftHiveMetastore_list_privileges_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1271) + foreach ($this->success as $iter1292) { - $xfer += $iter1271->write($output); + $xfer += $iter1292->write($output); } } $output->writeListEnd(); @@ -45030,14 +45086,14 @@ class ThriftHiveMetastore_set_ugi_args { case 2: if ($ftype == TType::LST) { $this->group_names = array(); - $_size1272 = 0; - $_etype1275 = 0; - $xfer += $input->readListBegin($_etype1275, $_size1272); - for ($_i1276 = 0; $_i1276 < $_size1272; ++$_i1276) + $_size1293 = 0; + $_etype1296 = 0; + $xfer += $input->readListBegin($_etype1296, $_size1293); + for ($_i1297 = 0; $_i1297 < $_size1293; ++$_i1297) { - $elem1277 = null; - $xfer += $input->readString($elem1277); - $this->group_names []= $elem1277; + $elem1298 = null; + $xfer += $input->readString($elem1298); + $this->group_names []= $elem1298; } $xfer += $input->readListEnd(); } else { @@ -45070,9 +45126,9 @@ class ThriftHiveMetastore_set_ugi_args { { $output->writeListBegin(TType::STRING, count($this->group_names)); { - foreach ($this->group_names as $iter1278) + foreach ($this->group_names as $iter1299) { - $xfer += $output->writeString($iter1278); + $xfer += $output->writeString($iter1299); } } $output->writeListEnd(); @@ -45148,14 +45204,14 @@ class ThriftHiveMetastore_set_ugi_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1279 = 0; - $_etype1282 = 0; - $xfer += $input->readListBegin($_etype1282, $_size1279); - for ($_i1283 = 0; $_i1283 < $_size1279; ++$_i1283) + $_size1300 = 0; + $_etype1303 = 0; + $xfer += $input->readListBegin($_etype1303, $_size1300); + for ($_i1304 = 0; $_i1304 < $_size1300; ++$_i1304) { - $elem1284 = null; - $xfer += $input->readString($elem1284); - $this->success []= $elem1284; + $elem1305 = null; + $xfer += $input->readString($elem1305); + $this->success []= $elem1305; } $xfer += $input->readListEnd(); } else { @@ -45191,9 +45247,9 @@ class ThriftHiveMetastore_set_ugi_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1285) + foreach ($this->success as $iter1306) { - $xfer += $output->writeString($iter1285); + $xfer += $output->writeString($iter1306); } } $output->writeListEnd(); @@ -46310,14 +46366,14 @@ class ThriftHiveMetastore_get_all_token_identifiers_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1286 = 0; - $_etype1289 = 0; - $xfer += $input->readListBegin($_etype1289, $_size1286); - for ($_i1290 = 0; $_i1290 < $_size1286; ++$_i1290) + $_size1307 = 0; + $_etype1310 = 0; + $xfer += $input->readListBegin($_etype1310, $_size1307); + for ($_i1311 = 0; $_i1311 < $_size1307; ++$_i1311) { - $elem1291 = null; - $xfer += $input->readString($elem1291); - $this->success []= $elem1291; + $elem1312 = null; + $xfer += $input->readString($elem1312); + $this->success []= $elem1312; } $xfer += $input->readListEnd(); } else { @@ -46345,9 +46401,9 @@ class ThriftHiveMetastore_get_all_token_identifiers_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1292) + foreach ($this->success as $iter1313) { - $xfer += $output->writeString($iter1292); + $xfer += $output->writeString($iter1313); } } $output->writeListEnd(); @@ -46986,14 +47042,14 @@ class ThriftHiveMetastore_get_master_keys_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1293 = 0; - $_etype1296 = 0; - $xfer += $input->readListBegin($_etype1296, $_size1293); - for ($_i1297 = 0; $_i1297 < $_size1293; ++$_i1297) + $_size1314 = 0; + $_etype1317 = 0; + $xfer += $input->readListBegin($_etype1317, $_size1314); + for ($_i1318 = 0; $_i1318 < $_size1314; ++$_i1318) { - $elem1298 = null; - $xfer += $input->readString($elem1298); - $this->success []= $elem1298; + $elem1319 = null; + $xfer += $input->readString($elem1319); + $this->success []= $elem1319; } $xfer += $input->readListEnd(); } else { @@ -47021,9 +47077,9 @@ class ThriftHiveMetastore_get_master_keys_result { { $output->writeListBegin(TType::STRING, count($this->success)); { - foreach ($this->success as $iter1299) + foreach ($this->success as $iter1320) { - $xfer += $output->writeString($iter1299); + $xfer += $output->writeString($iter1320); } } $output->writeListEnd(); @@ -51024,6 +51080,166 @@ class ThriftHiveMetastore_flushCache_result { } +class ThriftHiveMetastore_add_write_notification_log_args { + static $_TSPEC; + + /** + * @var \metastore\WriteNotificationLogRequest + */ + public $rqst = null; + + public function __construct($vals=null) { + if (!isset(self::$_TSPEC)) { + self::$_TSPEC = array( + -1 => array( + 'var' => 'rqst', + 'type' => TType::STRUCT, + 'class' => '\metastore\WriteNotificationLogRequest', + ), + ); + } + if (is_array($vals)) { + if (isset($vals['rqst'])) { + $this->rqst = $vals['rqst']; + } + } + } + + public function getName() { + return 'ThriftHiveMetastore_add_write_notification_log_args'; + } + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) + { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) + { + case -1: + if ($ftype == TType::STRUCT) { + $this->rqst = new \metastore\WriteNotificationLogRequest(); + $xfer += $this->rqst->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) { + $xfer = 0; + $xfer += $output->writeStructBegin('ThriftHiveMetastore_add_write_notification_log_args'); + if ($this->rqst !== null) { + if (!is_object($this->rqst)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('rqst', TType::STRUCT, -1); + $xfer += $this->rqst->write($output); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } + +} + +class ThriftHiveMetastore_add_write_notification_log_result { + static $_TSPEC; + + /** + * @var \metastore\WriteNotificationLogResponse + */ + public $success = null; + + public function __construct($vals=null) { + if (!isset(self::$_TSPEC)) { + self::$_TSPEC = array( + 0 => array( + 'var' => 'success', + 'type' => TType::STRUCT, + 'class' => '\metastore\WriteNotificationLogResponse', + ), + ); + } + if (is_array($vals)) { + if (isset($vals['success'])) { + $this->success = $vals['success']; + } + } + } + + public function getName() { + return 'ThriftHiveMetastore_add_write_notification_log_result'; + } + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) + { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) + { + case 0: + if ($ftype == TType::STRUCT) { + $this->success = new \metastore\WriteNotificationLogResponse(); + $xfer += $this->success->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) { + $xfer = 0; + $xfer += $output->writeStructBegin('ThriftHiveMetastore_add_write_notification_log_result'); + if ($this->success !== null) { + if (!is_object($this->success)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('success', TType::STRUCT, 0); + $xfer += $this->success->write($output); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } + +} + class ThriftHiveMetastore_cm_recycle_args { static $_TSPEC; @@ -57692,15 +57908,15 @@ class ThriftHiveMetastore_get_schema_all_versions_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1300 = 0; - $_etype1303 = 0; - $xfer += $input->readListBegin($_etype1303, $_size1300); - for ($_i1304 = 0; $_i1304 < $_size1300; ++$_i1304) + $_size1321 = 0; + $_etype1324 = 0; + $xfer += $input->readListBegin($_etype1324, $_size1321); + for ($_i1325 = 0; $_i1325 < $_size1321; ++$_i1325) { - $elem1305 = null; - $elem1305 = new \metastore\SchemaVersion(); - $xfer += $elem1305->read($input); - $this->success []= $elem1305; + $elem1326 = null; + $elem1326 = new \metastore\SchemaVersion(); + $xfer += $elem1326->read($input); + $this->success []= $elem1326; } $xfer += $input->readListEnd(); } else { @@ -57744,9 +57960,9 @@ class ThriftHiveMetastore_get_schema_all_versions_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1306) + foreach ($this->success as $iter1327) { - $xfer += $iter1306->write($output); + $xfer += $iter1327->write($output); } } $output->writeListEnd(); @@ -59615,15 +59831,15 @@ class ThriftHiveMetastore_get_runtime_stats_result { case 0: if ($ftype == TType::LST) { $this->success = array(); - $_size1307 = 0; - $_etype1310 = 0; - $xfer += $input->readListBegin($_etype1310, $_size1307); - for ($_i1311 = 0; $_i1311 < $_size1307; ++$_i1311) + $_size1328 = 0; + $_etype1331 = 0; + $xfer += $input->readListBegin($_etype1331, $_size1328); + for ($_i1332 = 0; $_i1332 < $_size1328; ++$_i1332) { - $elem1312 = null; - $elem1312 = new \metastore\RuntimeStat(); - $xfer += $elem1312->read($input); - $this->success []= $elem1312; + $elem1333 = null; + $elem1333 = new \metastore\RuntimeStat(); + $xfer += $elem1333->read($input); + $this->success []= $elem1333; } $xfer += $input->readListEnd(); } else { @@ -59659,9 +59875,9 @@ class ThriftHiveMetastore_get_runtime_stats_result { { $output->writeListBegin(TType::STRUCT, count($this->success)); { - foreach ($this->success as $iter1313) + foreach ($this->success as $iter1334) { - $xfer += $iter1313->write($output); + $xfer += $iter1334->write($output); } } $output->writeListEnd(); diff --git a/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php b/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php index 84f7e3320c..ae12471afe 100644 --- a/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php +++ b/standalone-metastore/src/gen/thrift/gen-php/metastore/Types.php @@ -16558,6 +16558,10 @@ class CommitTxnRequest { * @var string */ public $replPolicy = null; + /** + * @var \metastore\WriteEventInfo[] + */ + public $writeEventInfos = null; public function __construct($vals=null) { if (!isset(self::$_TSPEC)) { @@ -16570,6 +16574,15 @@ class CommitTxnRequest { 'var' => 'replPolicy', 'type' => TType::STRING, ), + 3 => array( + 'var' => 'writeEventInfos', + 'type' => TType::LST, + 'etype' => TType::STRUCT, + 'elem' => array( + 'type' => TType::STRUCT, + 'class' => '\metastore\WriteEventInfo', + ), + ), ); } if (is_array($vals)) { @@ -16579,6 +16592,9 @@ class CommitTxnRequest { if (isset($vals['replPolicy'])) { $this->replPolicy = $vals['replPolicy']; } + if (isset($vals['writeEventInfos'])) { + $this->writeEventInfos = $vals['writeEventInfos']; + } } } @@ -16615,6 +16631,24 @@ class CommitTxnRequest { $xfer += $input->skip($ftype); } break; + case 3: + if ($ftype == TType::LST) { + $this->writeEventInfos = array(); + $_size523 = 0; + $_etype526 = 0; + $xfer += $input->readListBegin($_etype526, $_size523); + for ($_i527 = 0; $_i527 < $_size523; ++$_i527) + { + $elem528 = null; + $elem528 = new \metastore\WriteEventInfo(); + $xfer += $elem528->read($input); + $this->writeEventInfos []= $elem528; + } + $xfer += $input->readListEnd(); + } else { + $xfer += $input->skip($ftype); + } + break; default: $xfer += $input->skip($ftype); break; @@ -16638,6 +16672,236 @@ class CommitTxnRequest { $xfer += $output->writeString($this->replPolicy); $xfer += $output->writeFieldEnd(); } + if ($this->writeEventInfos !== null) { + if (!is_array($this->writeEventInfos)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('writeEventInfos', TType::LST, 3); + { + $output->writeListBegin(TType::STRUCT, count($this->writeEventInfos)); + { + foreach ($this->writeEventInfos as $iter529) + { + $xfer += $iter529->write($output); + } + } + $output->writeListEnd(); + } + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } + +} + +class WriteEventInfo { + static $_TSPEC; + + /** + * @var int + */ + public $writeId = null; + /** + * @var string + */ + public $database = null; + /** + * @var string + */ + public $table = null; + /** + * @var string + */ + public $files = null; + /** + * @var string + */ + public $partition = null; + /** + * @var string + */ + public $tableObj = null; + /** + * @var string + */ + public $partitionObj = null; + + public function __construct($vals=null) { + if (!isset(self::$_TSPEC)) { + self::$_TSPEC = array( + 1 => array( + 'var' => 'writeId', + 'type' => TType::I64, + ), + 2 => array( + 'var' => 'database', + 'type' => TType::STRING, + ), + 3 => array( + 'var' => 'table', + 'type' => TType::STRING, + ), + 4 => array( + 'var' => 'files', + 'type' => TType::STRING, + ), + 5 => array( + 'var' => 'partition', + 'type' => TType::STRING, + ), + 6 => array( + 'var' => 'tableObj', + 'type' => TType::STRING, + ), + 7 => array( + 'var' => 'partitionObj', + 'type' => TType::STRING, + ), + ); + } + if (is_array($vals)) { + if (isset($vals['writeId'])) { + $this->writeId = $vals['writeId']; + } + if (isset($vals['database'])) { + $this->database = $vals['database']; + } + if (isset($vals['table'])) { + $this->table = $vals['table']; + } + if (isset($vals['files'])) { + $this->files = $vals['files']; + } + if (isset($vals['partition'])) { + $this->partition = $vals['partition']; + } + if (isset($vals['tableObj'])) { + $this->tableObj = $vals['tableObj']; + } + if (isset($vals['partitionObj'])) { + $this->partitionObj = $vals['partitionObj']; + } + } + } + + public function getName() { + return 'WriteEventInfo'; + } + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) + { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) + { + case 1: + if ($ftype == TType::I64) { + $xfer += $input->readI64($this->writeId); + } else { + $xfer += $input->skip($ftype); + } + break; + case 2: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->database); + } else { + $xfer += $input->skip($ftype); + } + break; + case 3: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->table); + } else { + $xfer += $input->skip($ftype); + } + break; + case 4: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->files); + } else { + $xfer += $input->skip($ftype); + } + break; + case 5: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->partition); + } else { + $xfer += $input->skip($ftype); + } + break; + case 6: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->tableObj); + } else { + $xfer += $input->skip($ftype); + } + break; + case 7: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->partitionObj); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) { + $xfer = 0; + $xfer += $output->writeStructBegin('WriteEventInfo'); + if ($this->writeId !== null) { + $xfer += $output->writeFieldBegin('writeId', TType::I64, 1); + $xfer += $output->writeI64($this->writeId); + $xfer += $output->writeFieldEnd(); + } + if ($this->database !== null) { + $xfer += $output->writeFieldBegin('database', TType::STRING, 2); + $xfer += $output->writeString($this->database); + $xfer += $output->writeFieldEnd(); + } + if ($this->table !== null) { + $xfer += $output->writeFieldBegin('table', TType::STRING, 3); + $xfer += $output->writeString($this->table); + $xfer += $output->writeFieldEnd(); + } + if ($this->files !== null) { + $xfer += $output->writeFieldBegin('files', TType::STRING, 4); + $xfer += $output->writeString($this->files); + $xfer += $output->writeFieldEnd(); + } + if ($this->partition !== null) { + $xfer += $output->writeFieldBegin('partition', TType::STRING, 5); + $xfer += $output->writeString($this->partition); + $xfer += $output->writeFieldEnd(); + } + if ($this->tableObj !== null) { + $xfer += $output->writeFieldBegin('tableObj', TType::STRING, 6); + $xfer += $output->writeString($this->tableObj); + $xfer += $output->writeFieldEnd(); + } + if ($this->partitionObj !== null) { + $xfer += $output->writeFieldBegin('partitionObj', TType::STRING, 7); + $xfer += $output->writeString($this->partitionObj); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; @@ -16785,14 +17049,14 @@ class ReplTblWriteIdStateRequest { case 6: if ($ftype == TType::LST) { $this->partNames = array(); - $_size523 = 0; - $_etype526 = 0; - $xfer += $input->readListBegin($_etype526, $_size523); - for ($_i527 = 0; $_i527 < $_size523; ++$_i527) + $_size530 = 0; + $_etype533 = 0; + $xfer += $input->readListBegin($_etype533, $_size530); + for ($_i534 = 0; $_i534 < $_size530; ++$_i534) { - $elem528 = null; - $xfer += $input->readString($elem528); - $this->partNames []= $elem528; + $elem535 = null; + $xfer += $input->readString($elem535); + $this->partNames []= $elem535; } $xfer += $input->readListEnd(); } else { @@ -16845,9 +17109,9 @@ class ReplTblWriteIdStateRequest { { $output->writeListBegin(TType::STRING, count($this->partNames)); { - foreach ($this->partNames as $iter529) + foreach ($this->partNames as $iter536) { - $xfer += $output->writeString($iter529); + $xfer += $output->writeString($iter536); } } $output->writeListEnd(); @@ -16922,14 +17186,14 @@ class GetValidWriteIdsRequest { case 1: if ($ftype == TType::LST) { $this->fullTableNames = array(); - $_size530 = 0; - $_etype533 = 0; - $xfer += $input->readListBegin($_etype533, $_size530); - for ($_i534 = 0; $_i534 < $_size530; ++$_i534) + $_size537 = 0; + $_etype540 = 0; + $xfer += $input->readListBegin($_etype540, $_size537); + for ($_i541 = 0; $_i541 < $_size537; ++$_i541) { - $elem535 = null; - $xfer += $input->readString($elem535); - $this->fullTableNames []= $elem535; + $elem542 = null; + $xfer += $input->readString($elem542); + $this->fullTableNames []= $elem542; } $xfer += $input->readListEnd(); } else { @@ -16964,9 +17228,9 @@ class GetValidWriteIdsRequest { { $output->writeListBegin(TType::STRING, count($this->fullTableNames)); { - foreach ($this->fullTableNames as $iter536) + foreach ($this->fullTableNames as $iter543) { - $xfer += $output->writeString($iter536); + $xfer += $output->writeString($iter543); } } $output->writeListEnd(); @@ -17093,14 +17357,14 @@ class TableValidWriteIds { case 3: if ($ftype == TType::LST) { $this->invalidWriteIds = array(); - $_size537 = 0; - $_etype540 = 0; - $xfer += $input->readListBegin($_etype540, $_size537); - for ($_i541 = 0; $_i541 < $_size537; ++$_i541) + $_size544 = 0; + $_etype547 = 0; + $xfer += $input->readListBegin($_etype547, $_size544); + for ($_i548 = 0; $_i548 < $_size544; ++$_i548) { - $elem542 = null; - $xfer += $input->readI64($elem542); - $this->invalidWriteIds []= $elem542; + $elem549 = null; + $xfer += $input->readI64($elem549); + $this->invalidWriteIds []= $elem549; } $xfer += $input->readListEnd(); } else { @@ -17152,9 +17416,9 @@ class TableValidWriteIds { { $output->writeListBegin(TType::I64, count($this->invalidWriteIds)); { - foreach ($this->invalidWriteIds as $iter543) + foreach ($this->invalidWriteIds as $iter550) { - $xfer += $output->writeI64($iter543); + $xfer += $output->writeI64($iter550); } } $output->writeListEnd(); @@ -17229,15 +17493,15 @@ class GetValidWriteIdsResponse { case 1: if ($ftype == TType::LST) { $this->tblValidWriteIds = array(); - $_size544 = 0; - $_etype547 = 0; - $xfer += $input->readListBegin($_etype547, $_size544); - for ($_i548 = 0; $_i548 < $_size544; ++$_i548) + $_size551 = 0; + $_etype554 = 0; + $xfer += $input->readListBegin($_etype554, $_size551); + for ($_i555 = 0; $_i555 < $_size551; ++$_i555) { - $elem549 = null; - $elem549 = new \metastore\TableValidWriteIds(); - $xfer += $elem549->read($input); - $this->tblValidWriteIds []= $elem549; + $elem556 = null; + $elem556 = new \metastore\TableValidWriteIds(); + $xfer += $elem556->read($input); + $this->tblValidWriteIds []= $elem556; } $xfer += $input->readListEnd(); } else { @@ -17265,9 +17529,9 @@ class GetValidWriteIdsResponse { { $output->writeListBegin(TType::STRUCT, count($this->tblValidWriteIds)); { - foreach ($this->tblValidWriteIds as $iter550) + foreach ($this->tblValidWriteIds as $iter557) { - $xfer += $iter550->write($output); + $xfer += $iter557->write($output); } } $output->writeListEnd(); @@ -17394,14 +17658,14 @@ class AllocateTableWriteIdsRequest { case 3: if ($ftype == TType::LST) { $this->txnIds = array(); - $_size551 = 0; - $_etype554 = 0; - $xfer += $input->readListBegin($_etype554, $_size551); - for ($_i555 = 0; $_i555 < $_size551; ++$_i555) + $_size558 = 0; + $_etype561 = 0; + $xfer += $input->readListBegin($_etype561, $_size558); + for ($_i562 = 0; $_i562 < $_size558; ++$_i562) { - $elem556 = null; - $xfer += $input->readI64($elem556); - $this->txnIds []= $elem556; + $elem563 = null; + $xfer += $input->readI64($elem563); + $this->txnIds []= $elem563; } $xfer += $input->readListEnd(); } else { @@ -17418,15 +17682,15 @@ class AllocateTableWriteIdsRequest { case 5: if ($ftype == TType::LST) { $this->srcTxnToWriteIdList = array(); - $_size557 = 0; - $_etype560 = 0; - $xfer += $input->readListBegin($_etype560, $_size557); - for ($_i561 = 0; $_i561 < $_size557; ++$_i561) + $_size564 = 0; + $_etype567 = 0; + $xfer += $input->readListBegin($_etype567, $_size564); + for ($_i568 = 0; $_i568 < $_size564; ++$_i568) { - $elem562 = null; - $elem562 = new \metastore\TxnToWriteId(); - $xfer += $elem562->read($input); - $this->srcTxnToWriteIdList []= $elem562; + $elem569 = null; + $elem569 = new \metastore\TxnToWriteId(); + $xfer += $elem569->read($input); + $this->srcTxnToWriteIdList []= $elem569; } $xfer += $input->readListEnd(); } else { @@ -17464,9 +17728,9 @@ class AllocateTableWriteIdsRequest { { $output->writeListBegin(TType::I64, count($this->txnIds)); { - foreach ($this->txnIds as $iter563) + foreach ($this->txnIds as $iter570) { - $xfer += $output->writeI64($iter563); + $xfer += $output->writeI64($iter570); } } $output->writeListEnd(); @@ -17486,9 +17750,9 @@ class AllocateTableWriteIdsRequest { { $output->writeListBegin(TType::STRUCT, count($this->srcTxnToWriteIdList)); { - foreach ($this->srcTxnToWriteIdList as $iter564) + foreach ($this->srcTxnToWriteIdList as $iter571) { - $xfer += $iter564->write($output); + $xfer += $iter571->write($output); } } $output->writeListEnd(); @@ -17651,16 +17915,16 @@ class AllocateTableWriteIdsResponse { case 1: if ($ftype == TType::LST) { $this->txnToWriteIds = array(); - $_size565 = 0; - $_etype568 = 0; - $xfer += $input->readListBegin($_etype568, $_size565); - for ($_i569 = 0; $_i569 < $_size565; ++$_i569) + $_size572 = 0; + $_etype575 = 0; + $xfer += $input->readListBegin($_etype575, $_size572); + for ($_i576 = 0; $_i576 < $_size572; ++$_i576) { - $elem570 = null; - $elem570 = new \metastore\TxnToWriteId(); - $xfer += $elem570->read($input); - $this->txnToWriteIds []= $elem570; - } + $elem577 = null; + $elem577 = new \metastore\TxnToWriteId(); + $xfer += $elem577->read($input); + $this->txnToWriteIds []= $elem577; + } $xfer += $input->readListEnd(); } else { $xfer += $input->skip($ftype); @@ -17687,9 +17951,9 @@ class AllocateTableWriteIdsResponse { { $output->writeListBegin(TType::STRUCT, count($this->txnToWriteIds)); { - foreach ($this->txnToWriteIds as $iter571) + foreach ($this->txnToWriteIds as $iter578) { - $xfer += $iter571->write($output); + $xfer += $iter578->write($output); } } $output->writeListEnd(); @@ -18034,15 +18298,15 @@ class LockRequest { case 1: if ($ftype == TType::LST) { $this->component = array(); - $_size572 = 0; - $_etype575 = 0; - $xfer += $input->readListBegin($_etype575, $_size572); - for ($_i576 = 0; $_i576 < $_size572; ++$_i576) + $_size579 = 0; + $_etype582 = 0; + $xfer += $input->readListBegin($_etype582, $_size579); + for ($_i583 = 0; $_i583 < $_size579; ++$_i583) { - $elem577 = null; - $elem577 = new \metastore\LockComponent(); - $xfer += $elem577->read($input); - $this->component []= $elem577; + $elem584 = null; + $elem584 = new \metastore\LockComponent(); + $xfer += $elem584->read($input); + $this->component []= $elem584; } $xfer += $input->readListEnd(); } else { @@ -18098,9 +18362,9 @@ class LockRequest { { $output->writeListBegin(TType::STRUCT, count($this->component)); { - foreach ($this->component as $iter578) + foreach ($this->component as $iter585) { - $xfer += $iter578->write($output); + $xfer += $iter585->write($output); } } $output->writeListEnd(); @@ -19043,15 +19307,15 @@ class ShowLocksResponse { case 1: if ($ftype == TType::LST) { $this->locks = array(); - $_size579 = 0; - $_etype582 = 0; - $xfer += $input->readListBegin($_etype582, $_size579); - for ($_i583 = 0; $_i583 < $_size579; ++$_i583) + $_size586 = 0; + $_etype589 = 0; + $xfer += $input->readListBegin($_etype589, $_size586); + for ($_i590 = 0; $_i590 < $_size586; ++$_i590) { - $elem584 = null; - $elem584 = new \metastore\ShowLocksResponseElement(); - $xfer += $elem584->read($input); - $this->locks []= $elem584; + $elem591 = null; + $elem591 = new \metastore\ShowLocksResponseElement(); + $xfer += $elem591->read($input); + $this->locks []= $elem591; } $xfer += $input->readListEnd(); } else { @@ -19079,9 +19343,9 @@ class ShowLocksResponse { { $output->writeListBegin(TType::STRUCT, count($this->locks)); { - foreach ($this->locks as $iter585) + foreach ($this->locks as $iter592) { - $xfer += $iter585->write($output); + $xfer += $iter592->write($output); } } $output->writeListEnd(); @@ -19356,17 +19620,17 @@ class HeartbeatTxnRangeResponse { case 1: if ($ftype == TType::SET) { $this->aborted = array(); - $_size586 = 0; - $_etype589 = 0; - $xfer += $input->readSetBegin($_etype589, $_size586); - for ($_i590 = 0; $_i590 < $_size586; ++$_i590) + $_size593 = 0; + $_etype596 = 0; + $xfer += $input->readSetBegin($_etype596, $_size593); + for ($_i597 = 0; $_i597 < $_size593; ++$_i597) { - $elem591 = null; - $xfer += $input->readI64($elem591); - if (is_scalar($elem591)) { - $this->aborted[$elem591] = true; + $elem598 = null; + $xfer += $input->readI64($elem598); + if (is_scalar($elem598)) { + $this->aborted[$elem598] = true; } else { - $this->aborted []= $elem591; + $this->aborted []= $elem598; } } $xfer += $input->readSetEnd(); @@ -19377,17 +19641,17 @@ class HeartbeatTxnRangeResponse { case 2: if ($ftype == TType::SET) { $this->nosuch = array(); - $_size592 = 0; - $_etype595 = 0; - $xfer += $input->readSetBegin($_etype595, $_size592); - for ($_i596 = 0; $_i596 < $_size592; ++$_i596) + $_size599 = 0; + $_etype602 = 0; + $xfer += $input->readSetBegin($_etype602, $_size599); + for ($_i603 = 0; $_i603 < $_size599; ++$_i603) { - $elem597 = null; - $xfer += $input->readI64($elem597); - if (is_scalar($elem597)) { - $this->nosuch[$elem597] = true; + $elem604 = null; + $xfer += $input->readI64($elem604); + if (is_scalar($elem604)) { + $this->nosuch[$elem604] = true; } else { - $this->nosuch []= $elem597; + $this->nosuch []= $elem604; } } $xfer += $input->readSetEnd(); @@ -19416,12 +19680,12 @@ class HeartbeatTxnRangeResponse { { $output->writeSetBegin(TType::I64, count($this->aborted)); { - foreach ($this->aborted as $iter598 => $iter599) + foreach ($this->aborted as $iter605 => $iter606) { - if (is_scalar($iter599)) { - $xfer += $output->writeI64($iter598); + if (is_scalar($iter606)) { + $xfer += $output->writeI64($iter605); } else { - $xfer += $output->writeI64($iter599); + $xfer += $output->writeI64($iter606); } } } @@ -19437,12 +19701,12 @@ class HeartbeatTxnRangeResponse { { $output->writeSetBegin(TType::I64, count($this->nosuch)); { - foreach ($this->nosuch as $iter600 => $iter601) + foreach ($this->nosuch as $iter607 => $iter608) { - if (is_scalar($iter601)) { - $xfer += $output->writeI64($iter600); + if (is_scalar($iter608)) { + $xfer += $output->writeI64($iter607); } else { - $xfer += $output->writeI64($iter601); + $xfer += $output->writeI64($iter608); } } } @@ -19601,17 +19865,17 @@ class CompactionRequest { case 6: if ($ftype == TType::MAP) { $this->properties = array(); - $_size602 = 0; - $_ktype603 = 0; - $_vtype604 = 0; - $xfer += $input->readMapBegin($_ktype603, $_vtype604, $_size602); - for ($_i606 = 0; $_i606 < $_size602; ++$_i606) + $_size609 = 0; + $_ktype610 = 0; + $_vtype611 = 0; + $xfer += $input->readMapBegin($_ktype610, $_vtype611, $_size609); + for ($_i613 = 0; $_i613 < $_size609; ++$_i613) { - $key607 = ''; - $val608 = ''; - $xfer += $input->readString($key607); - $xfer += $input->readString($val608); - $this->properties[$key607] = $val608; + $key614 = ''; + $val615 = ''; + $xfer += $input->readString($key614); + $xfer += $input->readString($val615); + $this->properties[$key614] = $val615; } $xfer += $input->readMapEnd(); } else { @@ -19664,10 +19928,10 @@ class CompactionRequest { { $output->writeMapBegin(TType::STRING, TType::STRING, count($this->properties)); { - foreach ($this->properties as $kiter609 => $viter610) + foreach ($this->properties as $kiter616 => $viter617) { - $xfer += $output->writeString($kiter609); - $xfer += $output->writeString($viter610); + $xfer += $output->writeString($kiter616); + $xfer += $output->writeString($viter617); } } $output->writeMapEnd(); @@ -20254,15 +20518,15 @@ class ShowCompactResponse { case 1: if ($ftype == TType::LST) { $this->compacts = array(); - $_size611 = 0; - $_etype614 = 0; - $xfer += $input->readListBegin($_etype614, $_size611); - for ($_i615 = 0; $_i615 < $_size611; ++$_i615) + $_size618 = 0; + $_etype621 = 0; + $xfer += $input->readListBegin($_etype621, $_size618); + for ($_i622 = 0; $_i622 < $_size618; ++$_i622) { - $elem616 = null; - $elem616 = new \metastore\ShowCompactResponseElement(); - $xfer += $elem616->read($input); - $this->compacts []= $elem616; + $elem623 = null; + $elem623 = new \metastore\ShowCompactResponseElement(); + $xfer += $elem623->read($input); + $this->compacts []= $elem623; } $xfer += $input->readListEnd(); } else { @@ -20290,9 +20554,9 @@ class ShowCompactResponse { { $output->writeListBegin(TType::STRUCT, count($this->compacts)); { - foreach ($this->compacts as $iter617) + foreach ($this->compacts as $iter624) { - $xfer += $iter617->write($output); + $xfer += $iter624->write($output); } } $output->writeListEnd(); @@ -20439,14 +20703,14 @@ class AddDynamicPartitions { case 5: if ($ftype == TType::LST) { $this->partitionnames = array(); - $_size618 = 0; - $_etype621 = 0; - $xfer += $input->readListBegin($_etype621, $_size618); - for ($_i622 = 0; $_i622 < $_size618; ++$_i622) + $_size625 = 0; + $_etype628 = 0; + $xfer += $input->readListBegin($_etype628, $_size625); + for ($_i629 = 0; $_i629 < $_size625; ++$_i629) { - $elem623 = null; - $xfer += $input->readString($elem623); - $this->partitionnames []= $elem623; + $elem630 = null; + $xfer += $input->readString($elem630); + $this->partitionnames []= $elem630; } $xfer += $input->readListEnd(); } else { @@ -20501,9 +20765,9 @@ class AddDynamicPartitions { { $output->writeListBegin(TType::STRING, count($this->partitionnames)); { - foreach ($this->partitionnames as $iter624) + foreach ($this->partitionnames as $iter631) { - $xfer += $output->writeString($iter624); + $xfer += $output->writeString($iter631); } } $output->writeListEnd(); @@ -20838,17 +21102,17 @@ class CreationMetadata { case 4: if ($ftype == TType::SET) { $this->tablesUsed = array(); - $_size625 = 0; - $_etype628 = 0; - $xfer += $input->readSetBegin($_etype628, $_size625); - for ($_i629 = 0; $_i629 < $_size625; ++$_i629) + $_size632 = 0; + $_etype635 = 0; + $xfer += $input->readSetBegin($_etype635, $_size632); + for ($_i636 = 0; $_i636 < $_size632; ++$_i636) { - $elem630 = null; - $xfer += $input->readString($elem630); - if (is_scalar($elem630)) { - $this->tablesUsed[$elem630] = true; + $elem637 = null; + $xfer += $input->readString($elem637); + if (is_scalar($elem637)) { + $this->tablesUsed[$elem637] = true; } else { - $this->tablesUsed []= $elem630; + $this->tablesUsed []= $elem637; } } $xfer += $input->readSetEnd(); @@ -20906,12 +21170,12 @@ class CreationMetadata { { $output->writeSetBegin(TType::STRING, count($this->tablesUsed)); { - foreach ($this->tablesUsed as $iter631 => $iter632) + foreach ($this->tablesUsed as $iter638 => $iter639) { - if (is_scalar($iter632)) { - $xfer += $output->writeString($iter631); + if (is_scalar($iter639)) { + $xfer += $output->writeString($iter638); } else { - $xfer += $output->writeString($iter632); + $xfer += $output->writeString($iter639); } } } @@ -21321,15 +21585,15 @@ class NotificationEventResponse { case 1: if ($ftype == TType::LST) { $this->events = array(); - $_size633 = 0; - $_etype636 = 0; - $xfer += $input->readListBegin($_etype636, $_size633); - for ($_i637 = 0; $_i637 < $_size633; ++$_i637) + $_size640 = 0; + $_etype643 = 0; + $xfer += $input->readListBegin($_etype643, $_size640); + for ($_i644 = 0; $_i644 < $_size640; ++$_i644) { - $elem638 = null; - $elem638 = new \metastore\NotificationEvent(); - $xfer += $elem638->read($input); - $this->events []= $elem638; + $elem645 = null; + $elem645 = new \metastore\NotificationEvent(); + $xfer += $elem645->read($input); + $this->events []= $elem645; } $xfer += $input->readListEnd(); } else { @@ -21357,9 +21621,9 @@ class NotificationEventResponse { { $output->writeListBegin(TType::STRUCT, count($this->events)); { - foreach ($this->events as $iter639) + foreach ($this->events as $iter646) { - $xfer += $iter639->write($output); + $xfer += $iter646->write($output); } } $output->writeListEnd(); @@ -21659,6 +21923,10 @@ class InsertEventRequestData { * @var string[] */ public $filesAddedChecksum = null; + /** + * @var string[] + */ + public $subDirectoryList = null; public function __construct($vals=null) { if (!isset(self::$_TSPEC)) { @@ -21683,6 +21951,14 @@ class InsertEventRequestData { 'type' => TType::STRING, ), ), + 4 => array( + 'var' => 'subDirectoryList', + 'type' => TType::LST, + 'etype' => TType::STRING, + 'elem' => array( + 'type' => TType::STRING, + ), + ), ); } if (is_array($vals)) { @@ -21695,6 +21971,9 @@ class InsertEventRequestData { if (isset($vals['filesAddedChecksum'])) { $this->filesAddedChecksum = $vals['filesAddedChecksum']; } + if (isset($vals['subDirectoryList'])) { + $this->subDirectoryList = $vals['subDirectoryList']; + } } } @@ -21727,14 +22006,14 @@ class InsertEventRequestData { case 2: if ($ftype == TType::LST) { $this->filesAdded = array(); - $_size640 = 0; - $_etype643 = 0; - $xfer += $input->readListBegin($_etype643, $_size640); - for ($_i644 = 0; $_i644 < $_size640; ++$_i644) + $_size647 = 0; + $_etype650 = 0; + $xfer += $input->readListBegin($_etype650, $_size647); + for ($_i651 = 0; $_i651 < $_size647; ++$_i651) { - $elem645 = null; - $xfer += $input->readString($elem645); - $this->filesAdded []= $elem645; + $elem652 = null; + $xfer += $input->readString($elem652); + $this->filesAdded []= $elem652; } $xfer += $input->readListEnd(); } else { @@ -21744,14 +22023,31 @@ class InsertEventRequestData { case 3: if ($ftype == TType::LST) { $this->filesAddedChecksum = array(); - $_size646 = 0; - $_etype649 = 0; - $xfer += $input->readListBegin($_etype649, $_size646); - for ($_i650 = 0; $_i650 < $_size646; ++$_i650) + $_size653 = 0; + $_etype656 = 0; + $xfer += $input->readListBegin($_etype656, $_size653); + for ($_i657 = 0; $_i657 < $_size653; ++$_i657) + { + $elem658 = null; + $xfer += $input->readString($elem658); + $this->filesAddedChecksum []= $elem658; + } + $xfer += $input->readListEnd(); + } else { + $xfer += $input->skip($ftype); + } + break; + case 4: + if ($ftype == TType::LST) { + $this->subDirectoryList = array(); + $_size659 = 0; + $_etype662 = 0; + $xfer += $input->readListBegin($_etype662, $_size659); + for ($_i663 = 0; $_i663 < $_size659; ++$_i663) { - $elem651 = null; - $xfer += $input->readString($elem651); - $this->filesAddedChecksum []= $elem651; + $elem664 = null; + $xfer += $input->readString($elem664); + $this->subDirectoryList []= $elem664; } $xfer += $input->readListEnd(); } else { @@ -21784,32 +22080,350 @@ class InsertEventRequestData { { $output->writeListBegin(TType::STRING, count($this->filesAdded)); { - foreach ($this->filesAdded as $iter652) + foreach ($this->filesAdded as $iter665) + { + $xfer += $output->writeString($iter665); + } + } + $output->writeListEnd(); + } + $xfer += $output->writeFieldEnd(); + } + if ($this->filesAddedChecksum !== null) { + if (!is_array($this->filesAddedChecksum)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('filesAddedChecksum', TType::LST, 3); + { + $output->writeListBegin(TType::STRING, count($this->filesAddedChecksum)); + { + foreach ($this->filesAddedChecksum as $iter666) + { + $xfer += $output->writeString($iter666); + } + } + $output->writeListEnd(); + } + $xfer += $output->writeFieldEnd(); + } + if ($this->subDirectoryList !== null) { + if (!is_array($this->subDirectoryList)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('subDirectoryList', TType::LST, 4); + { + $output->writeListBegin(TType::STRING, count($this->subDirectoryList)); + { + foreach ($this->subDirectoryList as $iter667) { - $xfer += $output->writeString($iter652); + $xfer += $output->writeString($iter667); } } $output->writeListEnd(); } $xfer += $output->writeFieldEnd(); } - if ($this->filesAddedChecksum !== null) { - if (!is_array($this->filesAddedChecksum)) { + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } + +} + +class FireEventRequestData { + static $_TSPEC; + + /** + * @var \metastore\InsertEventRequestData + */ + public $insertData = null; + + public function __construct($vals=null) { + if (!isset(self::$_TSPEC)) { + self::$_TSPEC = array( + 1 => array( + 'var' => 'insertData', + 'type' => TType::STRUCT, + 'class' => '\metastore\InsertEventRequestData', + ), + ); + } + if (is_array($vals)) { + if (isset($vals['insertData'])) { + $this->insertData = $vals['insertData']; + } + } + } + + public function getName() { + return 'FireEventRequestData'; + } + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) + { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) + { + case 1: + if ($ftype == TType::STRUCT) { + $this->insertData = new \metastore\InsertEventRequestData(); + $xfer += $this->insertData->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) { + $xfer = 0; + $xfer += $output->writeStructBegin('FireEventRequestData'); + if ($this->insertData !== null) { + if (!is_object($this->insertData)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('insertData', TType::STRUCT, 1); + $xfer += $this->insertData->write($output); + $xfer += $output->writeFieldEnd(); + } + $xfer += $output->writeFieldStop(); + $xfer += $output->writeStructEnd(); + return $xfer; + } + +} + +class FireEventRequest { + static $_TSPEC; + + /** + * @var bool + */ + public $successful = null; + /** + * @var \metastore\FireEventRequestData + */ + public $data = null; + /** + * @var string + */ + public $dbName = null; + /** + * @var string + */ + public $tableName = null; + /** + * @var string[] + */ + public $partitionVals = null; + /** + * @var string + */ + public $catName = null; + + public function __construct($vals=null) { + if (!isset(self::$_TSPEC)) { + self::$_TSPEC = array( + 1 => array( + 'var' => 'successful', + 'type' => TType::BOOL, + ), + 2 => array( + 'var' => 'data', + 'type' => TType::STRUCT, + 'class' => '\metastore\FireEventRequestData', + ), + 3 => array( + 'var' => 'dbName', + 'type' => TType::STRING, + ), + 4 => array( + 'var' => 'tableName', + 'type' => TType::STRING, + ), + 5 => array( + 'var' => 'partitionVals', + 'type' => TType::LST, + 'etype' => TType::STRING, + 'elem' => array( + 'type' => TType::STRING, + ), + ), + 6 => array( + 'var' => 'catName', + 'type' => TType::STRING, + ), + ); + } + if (is_array($vals)) { + if (isset($vals['successful'])) { + $this->successful = $vals['successful']; + } + if (isset($vals['data'])) { + $this->data = $vals['data']; + } + if (isset($vals['dbName'])) { + $this->dbName = $vals['dbName']; + } + if (isset($vals['tableName'])) { + $this->tableName = $vals['tableName']; + } + if (isset($vals['partitionVals'])) { + $this->partitionVals = $vals['partitionVals']; + } + if (isset($vals['catName'])) { + $this->catName = $vals['catName']; + } + } + } + + public function getName() { + return 'FireEventRequest'; + } + + public function read($input) + { + $xfer = 0; + $fname = null; + $ftype = 0; + $fid = 0; + $xfer += $input->readStructBegin($fname); + while (true) + { + $xfer += $input->readFieldBegin($fname, $ftype, $fid); + if ($ftype == TType::STOP) { + break; + } + switch ($fid) + { + case 1: + if ($ftype == TType::BOOL) { + $xfer += $input->readBool($this->successful); + } else { + $xfer += $input->skip($ftype); + } + break; + case 2: + if ($ftype == TType::STRUCT) { + $this->data = new \metastore\FireEventRequestData(); + $xfer += $this->data->read($input); + } else { + $xfer += $input->skip($ftype); + } + break; + case 3: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->dbName); + } else { + $xfer += $input->skip($ftype); + } + break; + case 4: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->tableName); + } else { + $xfer += $input->skip($ftype); + } + break; + case 5: + if ($ftype == TType::LST) { + $this->partitionVals = array(); + $_size668 = 0; + $_etype671 = 0; + $xfer += $input->readListBegin($_etype671, $_size668); + for ($_i672 = 0; $_i672 < $_size668; ++$_i672) + { + $elem673 = null; + $xfer += $input->readString($elem673); + $this->partitionVals []= $elem673; + } + $xfer += $input->readListEnd(); + } else { + $xfer += $input->skip($ftype); + } + break; + case 6: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->catName); + } else { + $xfer += $input->skip($ftype); + } + break; + default: + $xfer += $input->skip($ftype); + break; + } + $xfer += $input->readFieldEnd(); + } + $xfer += $input->readStructEnd(); + return $xfer; + } + + public function write($output) { + $xfer = 0; + $xfer += $output->writeStructBegin('FireEventRequest'); + if ($this->successful !== null) { + $xfer += $output->writeFieldBegin('successful', TType::BOOL, 1); + $xfer += $output->writeBool($this->successful); + $xfer += $output->writeFieldEnd(); + } + if ($this->data !== null) { + if (!is_object($this->data)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('data', TType::STRUCT, 2); + $xfer += $this->data->write($output); + $xfer += $output->writeFieldEnd(); + } + if ($this->dbName !== null) { + $xfer += $output->writeFieldBegin('dbName', TType::STRING, 3); + $xfer += $output->writeString($this->dbName); + $xfer += $output->writeFieldEnd(); + } + if ($this->tableName !== null) { + $xfer += $output->writeFieldBegin('tableName', TType::STRING, 4); + $xfer += $output->writeString($this->tableName); + $xfer += $output->writeFieldEnd(); + } + if ($this->partitionVals !== null) { + if (!is_array($this->partitionVals)) { throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); } - $xfer += $output->writeFieldBegin('filesAddedChecksum', TType::LST, 3); + $xfer += $output->writeFieldBegin('partitionVals', TType::LST, 5); { - $output->writeListBegin(TType::STRING, count($this->filesAddedChecksum)); + $output->writeListBegin(TType::STRING, count($this->partitionVals)); { - foreach ($this->filesAddedChecksum as $iter653) + foreach ($this->partitionVals as $iter674) { - $xfer += $output->writeString($iter653); + $xfer += $output->writeString($iter674); } } $output->writeListEnd(); } $xfer += $output->writeFieldEnd(); } + if ($this->catName !== null) { + $xfer += $output->writeFieldBegin('catName', TType::STRING, 6); + $xfer += $output->writeString($this->catName); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; @@ -21817,33 +22431,19 @@ class InsertEventRequestData { } -class FireEventRequestData { +class FireEventResponse { static $_TSPEC; - /** - * @var \metastore\InsertEventRequestData - */ - public $insertData = null; - public function __construct($vals=null) { + public function __construct() { if (!isset(self::$_TSPEC)) { self::$_TSPEC = array( - 1 => array( - 'var' => 'insertData', - 'type' => TType::STRUCT, - 'class' => '\metastore\InsertEventRequestData', - ), ); } - if (is_array($vals)) { - if (isset($vals['insertData'])) { - $this->insertData = $vals['insertData']; - } - } } public function getName() { - return 'FireEventRequestData'; + return 'FireEventResponse'; } public function read($input) @@ -21861,14 +22461,6 @@ class FireEventRequestData { } switch ($fid) { - case 1: - if ($ftype == TType::STRUCT) { - $this->insertData = new \metastore\InsertEventRequestData(); - $xfer += $this->insertData->read($input); - } else { - $xfer += $input->skip($ftype); - } - break; default: $xfer += $input->skip($ftype); break; @@ -21881,15 +22473,7 @@ class FireEventRequestData { public function write($output) { $xfer = 0; - $xfer += $output->writeStructBegin('FireEventRequestData'); - if ($this->insertData !== null) { - if (!is_object($this->insertData)) { - throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); - } - $xfer += $output->writeFieldBegin('insertData', TType::STRUCT, 1); - $xfer += $this->insertData->write($output); - $xfer += $output->writeFieldEnd(); - } + $xfer += $output->writeStructBegin('FireEventResponse'); $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; @@ -21897,55 +22481,59 @@ class FireEventRequestData { } -class FireEventRequest { +class WriteNotificationLogRequest { static $_TSPEC; /** - * @var bool + * @var int */ - public $successful = null; + public $txnId = null; /** - * @var \metastore\FireEventRequestData + * @var int */ - public $data = null; + public $writeId = null; /** * @var string */ - public $dbName = null; + public $db = null; /** * @var string */ - public $tableName = null; + public $table = null; /** - * @var string[] + * @var \metastore\InsertEventRequestData */ - public $partitionVals = null; + public $fileInfo = null; /** - * @var string + * @var string[] */ - public $catName = null; + public $partitionVals = null; public function __construct($vals=null) { if (!isset(self::$_TSPEC)) { self::$_TSPEC = array( 1 => array( - 'var' => 'successful', - 'type' => TType::BOOL, + 'var' => 'txnId', + 'type' => TType::I64, ), 2 => array( - 'var' => 'data', - 'type' => TType::STRUCT, - 'class' => '\metastore\FireEventRequestData', + 'var' => 'writeId', + 'type' => TType::I64, ), 3 => array( - 'var' => 'dbName', + 'var' => 'db', 'type' => TType::STRING, ), 4 => array( - 'var' => 'tableName', + 'var' => 'table', 'type' => TType::STRING, ), 5 => array( + 'var' => 'fileInfo', + 'type' => TType::STRUCT, + 'class' => '\metastore\InsertEventRequestData', + ), + 6 => array( 'var' => 'partitionVals', 'type' => TType::LST, 'etype' => TType::STRING, @@ -21953,36 +22541,32 @@ class FireEventRequest { 'type' => TType::STRING, ), ), - 6 => array( - 'var' => 'catName', - 'type' => TType::STRING, - ), ); } if (is_array($vals)) { - if (isset($vals['successful'])) { - $this->successful = $vals['successful']; + if (isset($vals['txnId'])) { + $this->txnId = $vals['txnId']; } - if (isset($vals['data'])) { - $this->data = $vals['data']; + if (isset($vals['writeId'])) { + $this->writeId = $vals['writeId']; } - if (isset($vals['dbName'])) { - $this->dbName = $vals['dbName']; + if (isset($vals['db'])) { + $this->db = $vals['db']; } - if (isset($vals['tableName'])) { - $this->tableName = $vals['tableName']; + if (isset($vals['table'])) { + $this->table = $vals['table']; + } + if (isset($vals['fileInfo'])) { + $this->fileInfo = $vals['fileInfo']; } if (isset($vals['partitionVals'])) { $this->partitionVals = $vals['partitionVals']; } - if (isset($vals['catName'])) { - $this->catName = $vals['catName']; - } } } public function getName() { - return 'FireEventRequest'; + return 'WriteNotificationLogRequest'; } public function read($input) @@ -22001,54 +22585,54 @@ class FireEventRequest { switch ($fid) { case 1: - if ($ftype == TType::BOOL) { - $xfer += $input->readBool($this->successful); + if ($ftype == TType::I64) { + $xfer += $input->readI64($this->txnId); } else { $xfer += $input->skip($ftype); } break; case 2: - if ($ftype == TType::STRUCT) { - $this->data = new \metastore\FireEventRequestData(); - $xfer += $this->data->read($input); + if ($ftype == TType::I64) { + $xfer += $input->readI64($this->writeId); } else { $xfer += $input->skip($ftype); } break; case 3: if ($ftype == TType::STRING) { - $xfer += $input->readString($this->dbName); + $xfer += $input->readString($this->db); } else { $xfer += $input->skip($ftype); } break; case 4: if ($ftype == TType::STRING) { - $xfer += $input->readString($this->tableName); + $xfer += $input->readString($this->table); } else { $xfer += $input->skip($ftype); } break; case 5: - if ($ftype == TType::LST) { - $this->partitionVals = array(); - $_size654 = 0; - $_etype657 = 0; - $xfer += $input->readListBegin($_etype657, $_size654); - for ($_i658 = 0; $_i658 < $_size654; ++$_i658) - { - $elem659 = null; - $xfer += $input->readString($elem659); - $this->partitionVals []= $elem659; - } - $xfer += $input->readListEnd(); + if ($ftype == TType::STRUCT) { + $this->fileInfo = new \metastore\InsertEventRequestData(); + $xfer += $this->fileInfo->read($input); } else { $xfer += $input->skip($ftype); } break; case 6: - if ($ftype == TType::STRING) { - $xfer += $input->readString($this->catName); + if ($ftype == TType::LST) { + $this->partitionVals = array(); + $_size675 = 0; + $_etype678 = 0; + $xfer += $input->readListBegin($_etype678, $_size675); + for ($_i679 = 0; $_i679 < $_size675; ++$_i679) + { + $elem680 = null; + $xfer += $input->readString($elem680); + $this->partitionVals []= $elem680; + } + $xfer += $input->readListEnd(); } else { $xfer += $input->skip($ftype); } @@ -22065,52 +22649,52 @@ class FireEventRequest { public function write($output) { $xfer = 0; - $xfer += $output->writeStructBegin('FireEventRequest'); - if ($this->successful !== null) { - $xfer += $output->writeFieldBegin('successful', TType::BOOL, 1); - $xfer += $output->writeBool($this->successful); + $xfer += $output->writeStructBegin('WriteNotificationLogRequest'); + if ($this->txnId !== null) { + $xfer += $output->writeFieldBegin('txnId', TType::I64, 1); + $xfer += $output->writeI64($this->txnId); $xfer += $output->writeFieldEnd(); } - if ($this->data !== null) { - if (!is_object($this->data)) { - throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); - } - $xfer += $output->writeFieldBegin('data', TType::STRUCT, 2); - $xfer += $this->data->write($output); + if ($this->writeId !== null) { + $xfer += $output->writeFieldBegin('writeId', TType::I64, 2); + $xfer += $output->writeI64($this->writeId); $xfer += $output->writeFieldEnd(); } - if ($this->dbName !== null) { - $xfer += $output->writeFieldBegin('dbName', TType::STRING, 3); - $xfer += $output->writeString($this->dbName); + if ($this->db !== null) { + $xfer += $output->writeFieldBegin('db', TType::STRING, 3); + $xfer += $output->writeString($this->db); $xfer += $output->writeFieldEnd(); } - if ($this->tableName !== null) { - $xfer += $output->writeFieldBegin('tableName', TType::STRING, 4); - $xfer += $output->writeString($this->tableName); + if ($this->table !== null) { + $xfer += $output->writeFieldBegin('table', TType::STRING, 4); + $xfer += $output->writeString($this->table); + $xfer += $output->writeFieldEnd(); + } + if ($this->fileInfo !== null) { + if (!is_object($this->fileInfo)) { + throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); + } + $xfer += $output->writeFieldBegin('fileInfo', TType::STRUCT, 5); + $xfer += $this->fileInfo->write($output); $xfer += $output->writeFieldEnd(); } if ($this->partitionVals !== null) { if (!is_array($this->partitionVals)) { throw new TProtocolException('Bad type in structure.', TProtocolException::INVALID_DATA); } - $xfer += $output->writeFieldBegin('partitionVals', TType::LST, 5); + $xfer += $output->writeFieldBegin('partitionVals', TType::LST, 6); { $output->writeListBegin(TType::STRING, count($this->partitionVals)); { - foreach ($this->partitionVals as $iter660) + foreach ($this->partitionVals as $iter681) { - $xfer += $output->writeString($iter660); + $xfer += $output->writeString($iter681); } } $output->writeListEnd(); } $xfer += $output->writeFieldEnd(); } - if ($this->catName !== null) { - $xfer += $output->writeFieldBegin('catName', TType::STRING, 6); - $xfer += $output->writeString($this->catName); - $xfer += $output->writeFieldEnd(); - } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; @@ -22118,7 +22702,7 @@ class FireEventRequest { } -class FireEventResponse { +class WriteNotificationLogResponse { static $_TSPEC; @@ -22130,7 +22714,7 @@ class FireEventResponse { } public function getName() { - return 'FireEventResponse'; + return 'WriteNotificationLogResponse'; } public function read($input) @@ -22160,7 +22744,7 @@ class FireEventResponse { public function write($output) { $xfer = 0; - $xfer += $output->writeStructBegin('FireEventResponse'); + $xfer += $output->writeStructBegin('WriteNotificationLogResponse'); $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; @@ -22332,18 +22916,18 @@ class GetFileMetadataByExprResult { case 1: if ($ftype == TType::MAP) { $this->metadata = array(); - $_size661 = 0; - $_ktype662 = 0; - $_vtype663 = 0; - $xfer += $input->readMapBegin($_ktype662, $_vtype663, $_size661); - for ($_i665 = 0; $_i665 < $_size661; ++$_i665) + $_size682 = 0; + $_ktype683 = 0; + $_vtype684 = 0; + $xfer += $input->readMapBegin($_ktype683, $_vtype684, $_size682); + for ($_i686 = 0; $_i686 < $_size682; ++$_i686) { - $key666 = 0; - $val667 = new \metastore\MetadataPpdResult(); - $xfer += $input->readI64($key666); - $val667 = new \metastore\MetadataPpdResult(); - $xfer += $val667->read($input); - $this->metadata[$key666] = $val667; + $key687 = 0; + $val688 = new \metastore\MetadataPpdResult(); + $xfer += $input->readI64($key687); + $val688 = new \metastore\MetadataPpdResult(); + $xfer += $val688->read($input); + $this->metadata[$key687] = $val688; } $xfer += $input->readMapEnd(); } else { @@ -22378,10 +22962,10 @@ class GetFileMetadataByExprResult { { $output->writeMapBegin(TType::I64, TType::STRUCT, count($this->metadata)); { - foreach ($this->metadata as $kiter668 => $viter669) + foreach ($this->metadata as $kiter689 => $viter690) { - $xfer += $output->writeI64($kiter668); - $xfer += $viter669->write($output); + $xfer += $output->writeI64($kiter689); + $xfer += $viter690->write($output); } } $output->writeMapEnd(); @@ -22483,14 +23067,14 @@ class GetFileMetadataByExprRequest { case 1: if ($ftype == TType::LST) { $this->fileIds = array(); - $_size670 = 0; - $_etype673 = 0; - $xfer += $input->readListBegin($_etype673, $_size670); - for ($_i674 = 0; $_i674 < $_size670; ++$_i674) + $_size691 = 0; + $_etype694 = 0; + $xfer += $input->readListBegin($_etype694, $_size691); + for ($_i695 = 0; $_i695 < $_size691; ++$_i695) { - $elem675 = null; - $xfer += $input->readI64($elem675); - $this->fileIds []= $elem675; + $elem696 = null; + $xfer += $input->readI64($elem696); + $this->fileIds []= $elem696; } $xfer += $input->readListEnd(); } else { @@ -22539,9 +23123,9 @@ class GetFileMetadataByExprRequest { { $output->writeListBegin(TType::I64, count($this->fileIds)); { - foreach ($this->fileIds as $iter676) + foreach ($this->fileIds as $iter697) { - $xfer += $output->writeI64($iter676); + $xfer += $output->writeI64($iter697); } } $output->writeListEnd(); @@ -22635,17 +23219,17 @@ class GetFileMetadataResult { case 1: if ($ftype == TType::MAP) { $this->metadata = array(); - $_size677 = 0; - $_ktype678 = 0; - $_vtype679 = 0; - $xfer += $input->readMapBegin($_ktype678, $_vtype679, $_size677); - for ($_i681 = 0; $_i681 < $_size677; ++$_i681) + $_size698 = 0; + $_ktype699 = 0; + $_vtype700 = 0; + $xfer += $input->readMapBegin($_ktype699, $_vtype700, $_size698); + for ($_i702 = 0; $_i702 < $_size698; ++$_i702) { - $key682 = 0; - $val683 = ''; - $xfer += $input->readI64($key682); - $xfer += $input->readString($val683); - $this->metadata[$key682] = $val683; + $key703 = 0; + $val704 = ''; + $xfer += $input->readI64($key703); + $xfer += $input->readString($val704); + $this->metadata[$key703] = $val704; } $xfer += $input->readMapEnd(); } else { @@ -22680,10 +23264,10 @@ class GetFileMetadataResult { { $output->writeMapBegin(TType::I64, TType::STRING, count($this->metadata)); { - foreach ($this->metadata as $kiter684 => $viter685) + foreach ($this->metadata as $kiter705 => $viter706) { - $xfer += $output->writeI64($kiter684); - $xfer += $output->writeString($viter685); + $xfer += $output->writeI64($kiter705); + $xfer += $output->writeString($viter706); } } $output->writeMapEnd(); @@ -22752,14 +23336,14 @@ class GetFileMetadataRequest { case 1: if ($ftype == TType::LST) { $this->fileIds = array(); - $_size686 = 0; - $_etype689 = 0; - $xfer += $input->readListBegin($_etype689, $_size686); - for ($_i690 = 0; $_i690 < $_size686; ++$_i690) + $_size707 = 0; + $_etype710 = 0; + $xfer += $input->readListBegin($_etype710, $_size707); + for ($_i711 = 0; $_i711 < $_size707; ++$_i711) { - $elem691 = null; - $xfer += $input->readI64($elem691); - $this->fileIds []= $elem691; + $elem712 = null; + $xfer += $input->readI64($elem712); + $this->fileIds []= $elem712; } $xfer += $input->readListEnd(); } else { @@ -22787,9 +23371,9 @@ class GetFileMetadataRequest { { $output->writeListBegin(TType::I64, count($this->fileIds)); { - foreach ($this->fileIds as $iter692) + foreach ($this->fileIds as $iter713) { - $xfer += $output->writeI64($iter692); + $xfer += $output->writeI64($iter713); } } $output->writeListEnd(); @@ -22929,14 +23513,14 @@ class PutFileMetadataRequest { case 1: if ($ftype == TType::LST) { $this->fileIds = array(); - $_size693 = 0; - $_etype696 = 0; - $xfer += $input->readListBegin($_etype696, $_size693); - for ($_i697 = 0; $_i697 < $_size693; ++$_i697) + $_size714 = 0; + $_etype717 = 0; + $xfer += $input->readListBegin($_etype717, $_size714); + for ($_i718 = 0; $_i718 < $_size714; ++$_i718) { - $elem698 = null; - $xfer += $input->readI64($elem698); - $this->fileIds []= $elem698; + $elem719 = null; + $xfer += $input->readI64($elem719); + $this->fileIds []= $elem719; } $xfer += $input->readListEnd(); } else { @@ -22946,14 +23530,14 @@ class PutFileMetadataRequest { case 2: if ($ftype == TType::LST) { $this->metadata = array(); - $_size699 = 0; - $_etype702 = 0; - $xfer += $input->readListBegin($_etype702, $_size699); - for ($_i703 = 0; $_i703 < $_size699; ++$_i703) + $_size720 = 0; + $_etype723 = 0; + $xfer += $input->readListBegin($_etype723, $_size720); + for ($_i724 = 0; $_i724 < $_size720; ++$_i724) { - $elem704 = null; - $xfer += $input->readString($elem704); - $this->metadata []= $elem704; + $elem725 = null; + $xfer += $input->readString($elem725); + $this->metadata []= $elem725; } $xfer += $input->readListEnd(); } else { @@ -22988,9 +23572,9 @@ class PutFileMetadataRequest { { $output->writeListBegin(TType::I64, count($this->fileIds)); { - foreach ($this->fileIds as $iter705) + foreach ($this->fileIds as $iter726) { - $xfer += $output->writeI64($iter705); + $xfer += $output->writeI64($iter726); } } $output->writeListEnd(); @@ -23005,9 +23589,9 @@ class PutFileMetadataRequest { { $output->writeListBegin(TType::STRING, count($this->metadata)); { - foreach ($this->metadata as $iter706) + foreach ($this->metadata as $iter727) { - $xfer += $output->writeString($iter706); + $xfer += $output->writeString($iter727); } } $output->writeListEnd(); @@ -23126,14 +23710,14 @@ class ClearFileMetadataRequest { case 1: if ($ftype == TType::LST) { $this->fileIds = array(); - $_size707 = 0; - $_etype710 = 0; - $xfer += $input->readListBegin($_etype710, $_size707); - for ($_i711 = 0; $_i711 < $_size707; ++$_i711) + $_size728 = 0; + $_etype731 = 0; + $xfer += $input->readListBegin($_etype731, $_size728); + for ($_i732 = 0; $_i732 < $_size728; ++$_i732) { - $elem712 = null; - $xfer += $input->readI64($elem712); - $this->fileIds []= $elem712; + $elem733 = null; + $xfer += $input->readI64($elem733); + $this->fileIds []= $elem733; } $xfer += $input->readListEnd(); } else { @@ -23161,9 +23745,9 @@ class ClearFileMetadataRequest { { $output->writeListBegin(TType::I64, count($this->fileIds)); { - foreach ($this->fileIds as $iter713) + foreach ($this->fileIds as $iter734) { - $xfer += $output->writeI64($iter713); + $xfer += $output->writeI64($iter734); } } $output->writeListEnd(); @@ -23447,15 +24031,15 @@ class GetAllFunctionsResponse { case 1: if ($ftype == TType::LST) { $this->functions = array(); - $_size714 = 0; - $_etype717 = 0; - $xfer += $input->readListBegin($_etype717, $_size714); - for ($_i718 = 0; $_i718 < $_size714; ++$_i718) + $_size735 = 0; + $_etype738 = 0; + $xfer += $input->readListBegin($_etype738, $_size735); + for ($_i739 = 0; $_i739 < $_size735; ++$_i739) { - $elem719 = null; - $elem719 = new \metastore\Function(); - $xfer += $elem719->read($input); - $this->functions []= $elem719; + $elem740 = null; + $elem740 = new \metastore\Function(); + $xfer += $elem740->read($input); + $this->functions []= $elem740; } $xfer += $input->readListEnd(); } else { @@ -23483,9 +24067,9 @@ class GetAllFunctionsResponse { { $output->writeListBegin(TType::STRUCT, count($this->functions)); { - foreach ($this->functions as $iter720) + foreach ($this->functions as $iter741) { - $xfer += $iter720->write($output); + $xfer += $iter741->write($output); } } $output->writeListEnd(); @@ -23549,14 +24133,14 @@ class ClientCapabilities { case 1: if ($ftype == TType::LST) { $this->values = array(); - $_size721 = 0; - $_etype724 = 0; - $xfer += $input->readListBegin($_etype724, $_size721); - for ($_i725 = 0; $_i725 < $_size721; ++$_i725) + $_size742 = 0; + $_etype745 = 0; + $xfer += $input->readListBegin($_etype745, $_size742); + for ($_i746 = 0; $_i746 < $_size742; ++$_i746) { - $elem726 = null; - $xfer += $input->readI32($elem726); - $this->values []= $elem726; + $elem747 = null; + $xfer += $input->readI32($elem747); + $this->values []= $elem747; } $xfer += $input->readListEnd(); } else { @@ -23584,9 +24168,9 @@ class ClientCapabilities { { $output->writeListBegin(TType::I32, count($this->values)); { - foreach ($this->values as $iter727) + foreach ($this->values as $iter748) { - $xfer += $output->writeI32($iter727); + $xfer += $output->writeI32($iter748); } } $output->writeListEnd(); @@ -23920,14 +24504,14 @@ class GetTablesRequest { case 2: if ($ftype == TType::LST) { $this->tblNames = array(); - $_size728 = 0; - $_etype731 = 0; - $xfer += $input->readListBegin($_etype731, $_size728); - for ($_i732 = 0; $_i732 < $_size728; ++$_i732) + $_size749 = 0; + $_etype752 = 0; + $xfer += $input->readListBegin($_etype752, $_size749); + for ($_i753 = 0; $_i753 < $_size749; ++$_i753) { - $elem733 = null; - $xfer += $input->readString($elem733); - $this->tblNames []= $elem733; + $elem754 = null; + $xfer += $input->readString($elem754); + $this->tblNames []= $elem754; } $xfer += $input->readListEnd(); } else { @@ -23975,9 +24559,9 @@ class GetTablesRequest { { $output->writeListBegin(TType::STRING, count($this->tblNames)); { - foreach ($this->tblNames as $iter734) + foreach ($this->tblNames as $iter755) { - $xfer += $output->writeString($iter734); + $xfer += $output->writeString($iter755); } } $output->writeListEnd(); @@ -24055,15 +24639,15 @@ class GetTablesResult { case 1: if ($ftype == TType::LST) { $this->tables = array(); - $_size735 = 0; - $_etype738 = 0; - $xfer += $input->readListBegin($_etype738, $_size735); - for ($_i739 = 0; $_i739 < $_size735; ++$_i739) + $_size756 = 0; + $_etype759 = 0; + $xfer += $input->readListBegin($_etype759, $_size756); + for ($_i760 = 0; $_i760 < $_size756; ++$_i760) { - $elem740 = null; - $elem740 = new \metastore\Table(); - $xfer += $elem740->read($input); - $this->tables []= $elem740; + $elem761 = null; + $elem761 = new \metastore\Table(); + $xfer += $elem761->read($input); + $this->tables []= $elem761; } $xfer += $input->readListEnd(); } else { @@ -24091,9 +24675,9 @@ class GetTablesResult { { $output->writeListBegin(TType::STRUCT, count($this->tables)); { - foreach ($this->tables as $iter741) + foreach ($this->tables as $iter762) { - $xfer += $iter741->write($output); + $xfer += $iter762->write($output); } } $output->writeListEnd(); @@ -25739,15 +26323,15 @@ class WMFullResourcePlan { case 2: if ($ftype == TType::LST) { $this->pools = array(); - $_size742 = 0; - $_etype745 = 0; - $xfer += $input->readListBegin($_etype745, $_size742); - for ($_i746 = 0; $_i746 < $_size742; ++$_i746) + $_size763 = 0; + $_etype766 = 0; + $xfer += $input->readListBegin($_etype766, $_size763); + for ($_i767 = 0; $_i767 < $_size763; ++$_i767) { - $elem747 = null; - $elem747 = new \metastore\WMPool(); - $xfer += $elem747->read($input); - $this->pools []= $elem747; + $elem768 = null; + $elem768 = new \metastore\WMPool(); + $xfer += $elem768->read($input); + $this->pools []= $elem768; } $xfer += $input->readListEnd(); } else { @@ -25757,15 +26341,15 @@ class WMFullResourcePlan { case 3: if ($ftype == TType::LST) { $this->mappings = array(); - $_size748 = 0; - $_etype751 = 0; - $xfer += $input->readListBegin($_etype751, $_size748); - for ($_i752 = 0; $_i752 < $_size748; ++$_i752) + $_size769 = 0; + $_etype772 = 0; + $xfer += $input->readListBegin($_etype772, $_size769); + for ($_i773 = 0; $_i773 < $_size769; ++$_i773) { - $elem753 = null; - $elem753 = new \metastore\WMMapping(); - $xfer += $elem753->read($input); - $this->mappings []= $elem753; + $elem774 = null; + $elem774 = new \metastore\WMMapping(); + $xfer += $elem774->read($input); + $this->mappings []= $elem774; } $xfer += $input->readListEnd(); } else { @@ -25775,15 +26359,15 @@ class WMFullResourcePlan { case 4: if ($ftype == TType::LST) { $this->triggers = array(); - $_size754 = 0; - $_etype757 = 0; - $xfer += $input->readListBegin($_etype757, $_size754); - for ($_i758 = 0; $_i758 < $_size754; ++$_i758) + $_size775 = 0; + $_etype778 = 0; + $xfer += $input->readListBegin($_etype778, $_size775); + for ($_i779 = 0; $_i779 < $_size775; ++$_i779) { - $elem759 = null; - $elem759 = new \metastore\WMTrigger(); - $xfer += $elem759->read($input); - $this->triggers []= $elem759; + $elem780 = null; + $elem780 = new \metastore\WMTrigger(); + $xfer += $elem780->read($input); + $this->triggers []= $elem780; } $xfer += $input->readListEnd(); } else { @@ -25793,15 +26377,15 @@ class WMFullResourcePlan { case 5: if ($ftype == TType::LST) { $this->poolTriggers = array(); - $_size760 = 0; - $_etype763 = 0; - $xfer += $input->readListBegin($_etype763, $_size760); - for ($_i764 = 0; $_i764 < $_size760; ++$_i764) + $_size781 = 0; + $_etype784 = 0; + $xfer += $input->readListBegin($_etype784, $_size781); + for ($_i785 = 0; $_i785 < $_size781; ++$_i785) { - $elem765 = null; - $elem765 = new \metastore\WMPoolTrigger(); - $xfer += $elem765->read($input); - $this->poolTriggers []= $elem765; + $elem786 = null; + $elem786 = new \metastore\WMPoolTrigger(); + $xfer += $elem786->read($input); + $this->poolTriggers []= $elem786; } $xfer += $input->readListEnd(); } else { @@ -25837,9 +26421,9 @@ class WMFullResourcePlan { { $output->writeListBegin(TType::STRUCT, count($this->pools)); { - foreach ($this->pools as $iter766) + foreach ($this->pools as $iter787) { - $xfer += $iter766->write($output); + $xfer += $iter787->write($output); } } $output->writeListEnd(); @@ -25854,9 +26438,9 @@ class WMFullResourcePlan { { $output->writeListBegin(TType::STRUCT, count($this->mappings)); { - foreach ($this->mappings as $iter767) + foreach ($this->mappings as $iter788) { - $xfer += $iter767->write($output); + $xfer += $iter788->write($output); } } $output->writeListEnd(); @@ -25871,9 +26455,9 @@ class WMFullResourcePlan { { $output->writeListBegin(TType::STRUCT, count($this->triggers)); { - foreach ($this->triggers as $iter768) + foreach ($this->triggers as $iter789) { - $xfer += $iter768->write($output); + $xfer += $iter789->write($output); } } $output->writeListEnd(); @@ -25888,9 +26472,9 @@ class WMFullResourcePlan { { $output->writeListBegin(TType::STRUCT, count($this->poolTriggers)); { - foreach ($this->poolTriggers as $iter769) + foreach ($this->poolTriggers as $iter790) { - $xfer += $iter769->write($output); + $xfer += $iter790->write($output); } } $output->writeListEnd(); @@ -26443,15 +27027,15 @@ class WMGetAllResourcePlanResponse { case 1: if ($ftype == TType::LST) { $this->resourcePlans = array(); - $_size770 = 0; - $_etype773 = 0; - $xfer += $input->readListBegin($_etype773, $_size770); - for ($_i774 = 0; $_i774 < $_size770; ++$_i774) + $_size791 = 0; + $_etype794 = 0; + $xfer += $input->readListBegin($_etype794, $_size791); + for ($_i795 = 0; $_i795 < $_size791; ++$_i795) { - $elem775 = null; - $elem775 = new \metastore\WMResourcePlan(); - $xfer += $elem775->read($input); - $this->resourcePlans []= $elem775; + $elem796 = null; + $elem796 = new \metastore\WMResourcePlan(); + $xfer += $elem796->read($input); + $this->resourcePlans []= $elem796; } $xfer += $input->readListEnd(); } else { @@ -26479,9 +27063,9 @@ class WMGetAllResourcePlanResponse { { $output->writeListBegin(TType::STRUCT, count($this->resourcePlans)); { - foreach ($this->resourcePlans as $iter776) + foreach ($this->resourcePlans as $iter797) { - $xfer += $iter776->write($output); + $xfer += $iter797->write($output); } } $output->writeListEnd(); @@ -26887,14 +27471,14 @@ class WMValidateResourcePlanResponse { case 1: if ($ftype == TType::LST) { $this->errors = array(); - $_size777 = 0; - $_etype780 = 0; - $xfer += $input->readListBegin($_etype780, $_size777); - for ($_i781 = 0; $_i781 < $_size777; ++$_i781) + $_size798 = 0; + $_etype801 = 0; + $xfer += $input->readListBegin($_etype801, $_size798); + for ($_i802 = 0; $_i802 < $_size798; ++$_i802) { - $elem782 = null; - $xfer += $input->readString($elem782); - $this->errors []= $elem782; + $elem803 = null; + $xfer += $input->readString($elem803); + $this->errors []= $elem803; } $xfer += $input->readListEnd(); } else { @@ -26904,14 +27488,14 @@ class WMValidateResourcePlanResponse { case 2: if ($ftype == TType::LST) { $this->warnings = array(); - $_size783 = 0; - $_etype786 = 0; - $xfer += $input->readListBegin($_etype786, $_size783); - for ($_i787 = 0; $_i787 < $_size783; ++$_i787) + $_size804 = 0; + $_etype807 = 0; + $xfer += $input->readListBegin($_etype807, $_size804); + for ($_i808 = 0; $_i808 < $_size804; ++$_i808) { - $elem788 = null; - $xfer += $input->readString($elem788); - $this->warnings []= $elem788; + $elem809 = null; + $xfer += $input->readString($elem809); + $this->warnings []= $elem809; } $xfer += $input->readListEnd(); } else { @@ -26939,9 +27523,9 @@ class WMValidateResourcePlanResponse { { $output->writeListBegin(TType::STRING, count($this->errors)); { - foreach ($this->errors as $iter789) + foreach ($this->errors as $iter810) { - $xfer += $output->writeString($iter789); + $xfer += $output->writeString($iter810); } } $output->writeListEnd(); @@ -26956,9 +27540,9 @@ class WMValidateResourcePlanResponse { { $output->writeListBegin(TType::STRING, count($this->warnings)); { - foreach ($this->warnings as $iter790) + foreach ($this->warnings as $iter811) { - $xfer += $output->writeString($iter790); + $xfer += $output->writeString($iter811); } } $output->writeListEnd(); @@ -27631,15 +28215,15 @@ class WMGetTriggersForResourePlanResponse { case 1: if ($ftype == TType::LST) { $this->triggers = array(); - $_size791 = 0; - $_etype794 = 0; - $xfer += $input->readListBegin($_etype794, $_size791); - for ($_i795 = 0; $_i795 < $_size791; ++$_i795) + $_size812 = 0; + $_etype815 = 0; + $xfer += $input->readListBegin($_etype815, $_size812); + for ($_i816 = 0; $_i816 < $_size812; ++$_i816) { - $elem796 = null; - $elem796 = new \metastore\WMTrigger(); - $xfer += $elem796->read($input); - $this->triggers []= $elem796; + $elem817 = null; + $elem817 = new \metastore\WMTrigger(); + $xfer += $elem817->read($input); + $this->triggers []= $elem817; } $xfer += $input->readListEnd(); } else { @@ -27667,9 +28251,9 @@ class WMGetTriggersForResourePlanResponse { { $output->writeListBegin(TType::STRUCT, count($this->triggers)); { - foreach ($this->triggers as $iter797) + foreach ($this->triggers as $iter818) { - $xfer += $iter797->write($output); + $xfer += $iter818->write($output); } } $output->writeListEnd(); @@ -29253,15 +29837,15 @@ class SchemaVersion { case 4: if ($ftype == TType::LST) { $this->cols = array(); - $_size798 = 0; - $_etype801 = 0; - $xfer += $input->readListBegin($_etype801, $_size798); - for ($_i802 = 0; $_i802 < $_size798; ++$_i802) + $_size819 = 0; + $_etype822 = 0; + $xfer += $input->readListBegin($_etype822, $_size819); + for ($_i823 = 0; $_i823 < $_size819; ++$_i823) { - $elem803 = null; - $elem803 = new \metastore\FieldSchema(); - $xfer += $elem803->read($input); - $this->cols []= $elem803; + $elem824 = null; + $elem824 = new \metastore\FieldSchema(); + $xfer += $elem824->read($input); + $this->cols []= $elem824; } $xfer += $input->readListEnd(); } else { @@ -29350,9 +29934,9 @@ class SchemaVersion { { $output->writeListBegin(TType::STRUCT, count($this->cols)); { - foreach ($this->cols as $iter804) + foreach ($this->cols as $iter825) { - $xfer += $iter804->write($output); + $xfer += $iter825->write($output); } } $output->writeListEnd(); @@ -29674,15 +30258,15 @@ class FindSchemasByColsResp { case 1: if ($ftype == TType::LST) { $this->schemaVersions = array(); - $_size805 = 0; - $_etype808 = 0; - $xfer += $input->readListBegin($_etype808, $_size805); - for ($_i809 = 0; $_i809 < $_size805; ++$_i809) + $_size826 = 0; + $_etype829 = 0; + $xfer += $input->readListBegin($_etype829, $_size826); + for ($_i830 = 0; $_i830 < $_size826; ++$_i830) { - $elem810 = null; - $elem810 = new \metastore\SchemaVersionDescriptor(); - $xfer += $elem810->read($input); - $this->schemaVersions []= $elem810; + $elem831 = null; + $elem831 = new \metastore\SchemaVersionDescriptor(); + $xfer += $elem831->read($input); + $this->schemaVersions []= $elem831; } $xfer += $input->readListEnd(); } else { @@ -29710,9 +30294,9 @@ class FindSchemasByColsResp { { $output->writeListBegin(TType::STRUCT, count($this->schemaVersions)); { - foreach ($this->schemaVersions as $iter811) + foreach ($this->schemaVersions as $iter832) { - $xfer += $iter811->write($output); + $xfer += $iter832->write($output); } } $output->writeListEnd(); diff --git a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote index 6c98efe93d..e5d943d5d8 100755 --- a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote +++ b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote @@ -189,6 +189,7 @@ if len(sys.argv) <= 1 or sys.argv[1] == '--help': print(' NotificationEventsCountResponse get_notification_events_count(NotificationEventsCountRequest rqst)') print(' FireEventResponse fire_listener_event(FireEventRequest rqst)') print(' void flushCache()') + print(' WriteNotificationLogResponse add_write_notification_log(WriteNotificationLogRequest rqst)') print(' CmRecycleResponse cm_recycle(CmRecycleRequest request)') print(' GetFileMetadataByExprResult get_file_metadata_by_expr(GetFileMetadataByExprRequest req)') print(' GetFileMetadataResult get_file_metadata(GetFileMetadataRequest req)') @@ -1290,6 +1291,12 @@ elif cmd == 'flushCache': sys.exit(1) pp.pprint(client.flushCache()) +elif cmd == 'add_write_notification_log': + if len(args) != 1: + print('add_write_notification_log requires 1 args') + sys.exit(1) + pp.pprint(client.add_write_notification_log(eval(args[0]),)) + elif cmd == 'cm_recycle': if len(args) != 1: print('cm_recycle requires 1 args') diff --git a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py index c4f3f02603..2ae6d9ab6f 100644 --- a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py +++ b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py @@ -1309,6 +1309,13 @@ def fire_listener_event(self, rqst): def flushCache(self): pass + def add_write_notification_log(self, rqst): + """ + Parameters: + - rqst + """ + pass + def cm_recycle(self, request): """ Parameters: @@ -7546,6 +7553,37 @@ def recv_flushCache(self): iprot.readMessageEnd() return + def add_write_notification_log(self, rqst): + """ + Parameters: + - rqst + """ + self.send_add_write_notification_log(rqst) + return self.recv_add_write_notification_log() + + def send_add_write_notification_log(self, rqst): + self._oprot.writeMessageBegin('add_write_notification_log', TMessageType.CALL, self._seqid) + args = add_write_notification_log_args() + args.rqst = rqst + args.write(self._oprot) + self._oprot.writeMessageEnd() + self._oprot.trans.flush() + + def recv_add_write_notification_log(self): + iprot = self._iprot + (fname, mtype, rseqid) = iprot.readMessageBegin() + if mtype == TMessageType.EXCEPTION: + x = TApplicationException() + x.read(iprot) + iprot.readMessageEnd() + raise x + result = add_write_notification_log_result() + result.read(iprot) + iprot.readMessageEnd() + if result.success is not None: + return result.success + raise TApplicationException(TApplicationException.MISSING_RESULT, "add_write_notification_log failed: unknown result") + def cm_recycle(self, request): """ Parameters: @@ -9170,6 +9208,7 @@ def __init__(self, handler): self._processMap["get_notification_events_count"] = Processor.process_get_notification_events_count self._processMap["fire_listener_event"] = Processor.process_fire_listener_event self._processMap["flushCache"] = Processor.process_flushCache + self._processMap["add_write_notification_log"] = Processor.process_add_write_notification_log self._processMap["cm_recycle"] = Processor.process_cm_recycle self._processMap["get_file_metadata_by_expr"] = Processor.process_get_file_metadata_by_expr self._processMap["get_file_metadata"] = Processor.process_get_file_metadata @@ -13287,6 +13326,25 @@ def process_flushCache(self, seqid, iprot, oprot): oprot.writeMessageEnd() oprot.trans.flush() + def process_add_write_notification_log(self, seqid, iprot, oprot): + args = add_write_notification_log_args() + args.read(iprot) + iprot.readMessageEnd() + result = add_write_notification_log_result() + try: + result.success = self._handler.add_write_notification_log(args.rqst) + msg_type = TMessageType.REPLY + except (TTransport.TTransportException, KeyboardInterrupt, SystemExit): + raise + except Exception as ex: + msg_type = TMessageType.EXCEPTION + logging.exception(ex) + result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') + oprot.writeMessageBegin("add_write_notification_log", msg_type, seqid) + result.write(oprot) + oprot.writeMessageEnd() + oprot.trans.flush() + def process_cm_recycle(self, seqid, iprot, oprot): args = cm_recycle_args() args.read(iprot) @@ -15987,10 +16045,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype812, _size809) = iprot.readListBegin() - for _i813 in xrange(_size809): - _elem814 = iprot.readString() - self.success.append(_elem814) + (_etype833, _size830) = iprot.readListBegin() + for _i834 in xrange(_size830): + _elem835 = iprot.readString() + self.success.append(_elem835) iprot.readListEnd() else: iprot.skip(ftype) @@ -16013,8 +16071,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter815 in self.success: - oprot.writeString(iter815) + for iter836 in self.success: + oprot.writeString(iter836) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -16119,10 +16177,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype819, _size816) = iprot.readListBegin() - for _i820 in xrange(_size816): - _elem821 = iprot.readString() - self.success.append(_elem821) + (_etype840, _size837) = iprot.readListBegin() + for _i841 in xrange(_size837): + _elem842 = iprot.readString() + self.success.append(_elem842) iprot.readListEnd() else: iprot.skip(ftype) @@ -16145,8 +16203,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter822 in self.success: - oprot.writeString(iter822) + for iter843 in self.success: + oprot.writeString(iter843) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -16916,12 +16974,12 @@ def read(self, iprot): if fid == 0: if ftype == TType.MAP: self.success = {} - (_ktype824, _vtype825, _size823 ) = iprot.readMapBegin() - for _i827 in xrange(_size823): - _key828 = iprot.readString() - _val829 = Type() - _val829.read(iprot) - self.success[_key828] = _val829 + (_ktype845, _vtype846, _size844 ) = iprot.readMapBegin() + for _i848 in xrange(_size844): + _key849 = iprot.readString() + _val850 = Type() + _val850.read(iprot) + self.success[_key849] = _val850 iprot.readMapEnd() else: iprot.skip(ftype) @@ -16944,9 +17002,9 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.MAP, 0) oprot.writeMapBegin(TType.STRING, TType.STRUCT, len(self.success)) - for kiter830,viter831 in self.success.items(): - oprot.writeString(kiter830) - viter831.write(oprot) + for kiter851,viter852 in self.success.items(): + oprot.writeString(kiter851) + viter852.write(oprot) oprot.writeMapEnd() oprot.writeFieldEnd() if self.o2 is not None: @@ -17089,11 +17147,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype835, _size832) = iprot.readListBegin() - for _i836 in xrange(_size832): - _elem837 = FieldSchema() - _elem837.read(iprot) - self.success.append(_elem837) + (_etype856, _size853) = iprot.readListBegin() + for _i857 in xrange(_size853): + _elem858 = FieldSchema() + _elem858.read(iprot) + self.success.append(_elem858) iprot.readListEnd() else: iprot.skip(ftype) @@ -17128,8 +17186,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter838 in self.success: - iter838.write(oprot) + for iter859 in self.success: + iter859.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -17296,11 +17354,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype842, _size839) = iprot.readListBegin() - for _i843 in xrange(_size839): - _elem844 = FieldSchema() - _elem844.read(iprot) - self.success.append(_elem844) + (_etype863, _size860) = iprot.readListBegin() + for _i864 in xrange(_size860): + _elem865 = FieldSchema() + _elem865.read(iprot) + self.success.append(_elem865) iprot.readListEnd() else: iprot.skip(ftype) @@ -17335,8 +17393,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter845 in self.success: - iter845.write(oprot) + for iter866 in self.success: + iter866.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -17489,11 +17547,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype849, _size846) = iprot.readListBegin() - for _i850 in xrange(_size846): - _elem851 = FieldSchema() - _elem851.read(iprot) - self.success.append(_elem851) + (_etype870, _size867) = iprot.readListBegin() + for _i871 in xrange(_size867): + _elem872 = FieldSchema() + _elem872.read(iprot) + self.success.append(_elem872) iprot.readListEnd() else: iprot.skip(ftype) @@ -17528,8 +17586,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter852 in self.success: - iter852.write(oprot) + for iter873 in self.success: + iter873.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -17696,11 +17754,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype856, _size853) = iprot.readListBegin() - for _i857 in xrange(_size853): - _elem858 = FieldSchema() - _elem858.read(iprot) - self.success.append(_elem858) + (_etype877, _size874) = iprot.readListBegin() + for _i878 in xrange(_size874): + _elem879 = FieldSchema() + _elem879.read(iprot) + self.success.append(_elem879) iprot.readListEnd() else: iprot.skip(ftype) @@ -17735,8 +17793,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter859 in self.success: - iter859.write(oprot) + for iter880 in self.success: + iter880.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -18189,66 +18247,66 @@ def read(self, iprot): elif fid == 2: if ftype == TType.LIST: self.primaryKeys = [] - (_etype863, _size860) = iprot.readListBegin() - for _i864 in xrange(_size860): - _elem865 = SQLPrimaryKey() - _elem865.read(iprot) - self.primaryKeys.append(_elem865) + (_etype884, _size881) = iprot.readListBegin() + for _i885 in xrange(_size881): + _elem886 = SQLPrimaryKey() + _elem886.read(iprot) + self.primaryKeys.append(_elem886) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 3: if ftype == TType.LIST: self.foreignKeys = [] - (_etype869, _size866) = iprot.readListBegin() - for _i870 in xrange(_size866): - _elem871 = SQLForeignKey() - _elem871.read(iprot) - self.foreignKeys.append(_elem871) + (_etype890, _size887) = iprot.readListBegin() + for _i891 in xrange(_size887): + _elem892 = SQLForeignKey() + _elem892.read(iprot) + self.foreignKeys.append(_elem892) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 4: if ftype == TType.LIST: self.uniqueConstraints = [] - (_etype875, _size872) = iprot.readListBegin() - for _i876 in xrange(_size872): - _elem877 = SQLUniqueConstraint() - _elem877.read(iprot) - self.uniqueConstraints.append(_elem877) + (_etype896, _size893) = iprot.readListBegin() + for _i897 in xrange(_size893): + _elem898 = SQLUniqueConstraint() + _elem898.read(iprot) + self.uniqueConstraints.append(_elem898) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 5: if ftype == TType.LIST: self.notNullConstraints = [] - (_etype881, _size878) = iprot.readListBegin() - for _i882 in xrange(_size878): - _elem883 = SQLNotNullConstraint() - _elem883.read(iprot) - self.notNullConstraints.append(_elem883) + (_etype902, _size899) = iprot.readListBegin() + for _i903 in xrange(_size899): + _elem904 = SQLNotNullConstraint() + _elem904.read(iprot) + self.notNullConstraints.append(_elem904) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 6: if ftype == TType.LIST: self.defaultConstraints = [] - (_etype887, _size884) = iprot.readListBegin() - for _i888 in xrange(_size884): - _elem889 = SQLDefaultConstraint() - _elem889.read(iprot) - self.defaultConstraints.append(_elem889) + (_etype908, _size905) = iprot.readListBegin() + for _i909 in xrange(_size905): + _elem910 = SQLDefaultConstraint() + _elem910.read(iprot) + self.defaultConstraints.append(_elem910) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 7: if ftype == TType.LIST: self.checkConstraints = [] - (_etype893, _size890) = iprot.readListBegin() - for _i894 in xrange(_size890): - _elem895 = SQLCheckConstraint() - _elem895.read(iprot) - self.checkConstraints.append(_elem895) + (_etype914, _size911) = iprot.readListBegin() + for _i915 in xrange(_size911): + _elem916 = SQLCheckConstraint() + _elem916.read(iprot) + self.checkConstraints.append(_elem916) iprot.readListEnd() else: iprot.skip(ftype) @@ -18269,43 +18327,43 @@ def write(self, oprot): if self.primaryKeys is not None: oprot.writeFieldBegin('primaryKeys', TType.LIST, 2) oprot.writeListBegin(TType.STRUCT, len(self.primaryKeys)) - for iter896 in self.primaryKeys: - iter896.write(oprot) + for iter917 in self.primaryKeys: + iter917.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.foreignKeys is not None: oprot.writeFieldBegin('foreignKeys', TType.LIST, 3) oprot.writeListBegin(TType.STRUCT, len(self.foreignKeys)) - for iter897 in self.foreignKeys: - iter897.write(oprot) + for iter918 in self.foreignKeys: + iter918.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.uniqueConstraints is not None: oprot.writeFieldBegin('uniqueConstraints', TType.LIST, 4) oprot.writeListBegin(TType.STRUCT, len(self.uniqueConstraints)) - for iter898 in self.uniqueConstraints: - iter898.write(oprot) + for iter919 in self.uniqueConstraints: + iter919.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.notNullConstraints is not None: oprot.writeFieldBegin('notNullConstraints', TType.LIST, 5) oprot.writeListBegin(TType.STRUCT, len(self.notNullConstraints)) - for iter899 in self.notNullConstraints: - iter899.write(oprot) + for iter920 in self.notNullConstraints: + iter920.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.defaultConstraints is not None: oprot.writeFieldBegin('defaultConstraints', TType.LIST, 6) oprot.writeListBegin(TType.STRUCT, len(self.defaultConstraints)) - for iter900 in self.defaultConstraints: - iter900.write(oprot) + for iter921 in self.defaultConstraints: + iter921.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.checkConstraints is not None: oprot.writeFieldBegin('checkConstraints', TType.LIST, 7) oprot.writeListBegin(TType.STRUCT, len(self.checkConstraints)) - for iter901 in self.checkConstraints: - iter901.write(oprot) + for iter922 in self.checkConstraints: + iter922.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -19865,10 +19923,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.partNames = [] - (_etype905, _size902) = iprot.readListBegin() - for _i906 in xrange(_size902): - _elem907 = iprot.readString() - self.partNames.append(_elem907) + (_etype926, _size923) = iprot.readListBegin() + for _i927 in xrange(_size923): + _elem928 = iprot.readString() + self.partNames.append(_elem928) iprot.readListEnd() else: iprot.skip(ftype) @@ -19893,8 +19951,8 @@ def write(self, oprot): if self.partNames is not None: oprot.writeFieldBegin('partNames', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.partNames)) - for iter908 in self.partNames: - oprot.writeString(iter908) + for iter929 in self.partNames: + oprot.writeString(iter929) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -20094,10 +20152,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype912, _size909) = iprot.readListBegin() - for _i913 in xrange(_size909): - _elem914 = iprot.readString() - self.success.append(_elem914) + (_etype933, _size930) = iprot.readListBegin() + for _i934 in xrange(_size930): + _elem935 = iprot.readString() + self.success.append(_elem935) iprot.readListEnd() else: iprot.skip(ftype) @@ -20120,8 +20178,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter915 in self.success: - oprot.writeString(iter915) + for iter936 in self.success: + oprot.writeString(iter936) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -20271,10 +20329,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype919, _size916) = iprot.readListBegin() - for _i920 in xrange(_size916): - _elem921 = iprot.readString() - self.success.append(_elem921) + (_etype940, _size937) = iprot.readListBegin() + for _i941 in xrange(_size937): + _elem942 = iprot.readString() + self.success.append(_elem942) iprot.readListEnd() else: iprot.skip(ftype) @@ -20297,8 +20355,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter922 in self.success: - oprot.writeString(iter922) + for iter943 in self.success: + oprot.writeString(iter943) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -20422,10 +20480,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype926, _size923) = iprot.readListBegin() - for _i927 in xrange(_size923): - _elem928 = iprot.readString() - self.success.append(_elem928) + (_etype947, _size944) = iprot.readListBegin() + for _i948 in xrange(_size944): + _elem949 = iprot.readString() + self.success.append(_elem949) iprot.readListEnd() else: iprot.skip(ftype) @@ -20448,8 +20506,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter929 in self.success: - oprot.writeString(iter929) + for iter950 in self.success: + oprot.writeString(iter950) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -20522,10 +20580,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.tbl_types = [] - (_etype933, _size930) = iprot.readListBegin() - for _i934 in xrange(_size930): - _elem935 = iprot.readString() - self.tbl_types.append(_elem935) + (_etype954, _size951) = iprot.readListBegin() + for _i955 in xrange(_size951): + _elem956 = iprot.readString() + self.tbl_types.append(_elem956) iprot.readListEnd() else: iprot.skip(ftype) @@ -20550,8 +20608,8 @@ def write(self, oprot): if self.tbl_types is not None: oprot.writeFieldBegin('tbl_types', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.tbl_types)) - for iter936 in self.tbl_types: - oprot.writeString(iter936) + for iter957 in self.tbl_types: + oprot.writeString(iter957) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -20607,11 +20665,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype940, _size937) = iprot.readListBegin() - for _i941 in xrange(_size937): - _elem942 = TableMeta() - _elem942.read(iprot) - self.success.append(_elem942) + (_etype961, _size958) = iprot.readListBegin() + for _i962 in xrange(_size958): + _elem963 = TableMeta() + _elem963.read(iprot) + self.success.append(_elem963) iprot.readListEnd() else: iprot.skip(ftype) @@ -20634,8 +20692,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter943 in self.success: - iter943.write(oprot) + for iter964 in self.success: + iter964.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -20759,10 +20817,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype947, _size944) = iprot.readListBegin() - for _i948 in xrange(_size944): - _elem949 = iprot.readString() - self.success.append(_elem949) + (_etype968, _size965) = iprot.readListBegin() + for _i969 in xrange(_size965): + _elem970 = iprot.readString() + self.success.append(_elem970) iprot.readListEnd() else: iprot.skip(ftype) @@ -20785,8 +20843,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter950 in self.success: - oprot.writeString(iter950) + for iter971 in self.success: + oprot.writeString(iter971) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -21022,10 +21080,10 @@ def read(self, iprot): elif fid == 2: if ftype == TType.LIST: self.tbl_names = [] - (_etype954, _size951) = iprot.readListBegin() - for _i955 in xrange(_size951): - _elem956 = iprot.readString() - self.tbl_names.append(_elem956) + (_etype975, _size972) = iprot.readListBegin() + for _i976 in xrange(_size972): + _elem977 = iprot.readString() + self.tbl_names.append(_elem977) iprot.readListEnd() else: iprot.skip(ftype) @@ -21046,8 +21104,8 @@ def write(self, oprot): if self.tbl_names is not None: oprot.writeFieldBegin('tbl_names', TType.LIST, 2) oprot.writeListBegin(TType.STRING, len(self.tbl_names)) - for iter957 in self.tbl_names: - oprot.writeString(iter957) + for iter978 in self.tbl_names: + oprot.writeString(iter978) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -21099,11 +21157,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype961, _size958) = iprot.readListBegin() - for _i962 in xrange(_size958): - _elem963 = Table() - _elem963.read(iprot) - self.success.append(_elem963) + (_etype982, _size979) = iprot.readListBegin() + for _i983 in xrange(_size979): + _elem984 = Table() + _elem984.read(iprot) + self.success.append(_elem984) iprot.readListEnd() else: iprot.skip(ftype) @@ -21120,8 +21178,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter964 in self.success: - iter964.write(oprot) + for iter985 in self.success: + iter985.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -21989,10 +22047,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype968, _size965) = iprot.readListBegin() - for _i969 in xrange(_size965): - _elem970 = iprot.readString() - self.success.append(_elem970) + (_etype989, _size986) = iprot.readListBegin() + for _i990 in xrange(_size986): + _elem991 = iprot.readString() + self.success.append(_elem991) iprot.readListEnd() else: iprot.skip(ftype) @@ -22027,8 +22085,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter971 in self.success: - oprot.writeString(iter971) + for iter992 in self.success: + oprot.writeString(iter992) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -22998,11 +23056,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.new_parts = [] - (_etype975, _size972) = iprot.readListBegin() - for _i976 in xrange(_size972): - _elem977 = Partition() - _elem977.read(iprot) - self.new_parts.append(_elem977) + (_etype996, _size993) = iprot.readListBegin() + for _i997 in xrange(_size993): + _elem998 = Partition() + _elem998.read(iprot) + self.new_parts.append(_elem998) iprot.readListEnd() else: iprot.skip(ftype) @@ -23019,8 +23077,8 @@ def write(self, oprot): if self.new_parts is not None: oprot.writeFieldBegin('new_parts', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.new_parts)) - for iter978 in self.new_parts: - iter978.write(oprot) + for iter999 in self.new_parts: + iter999.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -23178,11 +23236,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.new_parts = [] - (_etype982, _size979) = iprot.readListBegin() - for _i983 in xrange(_size979): - _elem984 = PartitionSpec() - _elem984.read(iprot) - self.new_parts.append(_elem984) + (_etype1003, _size1000) = iprot.readListBegin() + for _i1004 in xrange(_size1000): + _elem1005 = PartitionSpec() + _elem1005.read(iprot) + self.new_parts.append(_elem1005) iprot.readListEnd() else: iprot.skip(ftype) @@ -23199,8 +23257,8 @@ def write(self, oprot): if self.new_parts is not None: oprot.writeFieldBegin('new_parts', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.new_parts)) - for iter985 in self.new_parts: - iter985.write(oprot) + for iter1006 in self.new_parts: + iter1006.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -23374,10 +23432,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype989, _size986) = iprot.readListBegin() - for _i990 in xrange(_size986): - _elem991 = iprot.readString() - self.part_vals.append(_elem991) + (_etype1010, _size1007) = iprot.readListBegin() + for _i1011 in xrange(_size1007): + _elem1012 = iprot.readString() + self.part_vals.append(_elem1012) iprot.readListEnd() else: iprot.skip(ftype) @@ -23402,8 +23460,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter992 in self.part_vals: - oprot.writeString(iter992) + for iter1013 in self.part_vals: + oprot.writeString(iter1013) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -23756,10 +23814,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype996, _size993) = iprot.readListBegin() - for _i997 in xrange(_size993): - _elem998 = iprot.readString() - self.part_vals.append(_elem998) + (_etype1017, _size1014) = iprot.readListBegin() + for _i1018 in xrange(_size1014): + _elem1019 = iprot.readString() + self.part_vals.append(_elem1019) iprot.readListEnd() else: iprot.skip(ftype) @@ -23790,8 +23848,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter999 in self.part_vals: - oprot.writeString(iter999) + for iter1020 in self.part_vals: + oprot.writeString(iter1020) oprot.writeListEnd() oprot.writeFieldEnd() if self.environment_context is not None: @@ -24386,10 +24444,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1003, _size1000) = iprot.readListBegin() - for _i1004 in xrange(_size1000): - _elem1005 = iprot.readString() - self.part_vals.append(_elem1005) + (_etype1024, _size1021) = iprot.readListBegin() + for _i1025 in xrange(_size1021): + _elem1026 = iprot.readString() + self.part_vals.append(_elem1026) iprot.readListEnd() else: iprot.skip(ftype) @@ -24419,8 +24477,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1006 in self.part_vals: - oprot.writeString(iter1006) + for iter1027 in self.part_vals: + oprot.writeString(iter1027) oprot.writeListEnd() oprot.writeFieldEnd() if self.deleteData is not None: @@ -24593,10 +24651,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1010, _size1007) = iprot.readListBegin() - for _i1011 in xrange(_size1007): - _elem1012 = iprot.readString() - self.part_vals.append(_elem1012) + (_etype1031, _size1028) = iprot.readListBegin() + for _i1032 in xrange(_size1028): + _elem1033 = iprot.readString() + self.part_vals.append(_elem1033) iprot.readListEnd() else: iprot.skip(ftype) @@ -24632,8 +24690,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1013 in self.part_vals: - oprot.writeString(iter1013) + for iter1034 in self.part_vals: + oprot.writeString(iter1034) oprot.writeListEnd() oprot.writeFieldEnd() if self.deleteData is not None: @@ -25370,10 +25428,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1017, _size1014) = iprot.readListBegin() - for _i1018 in xrange(_size1014): - _elem1019 = iprot.readString() - self.part_vals.append(_elem1019) + (_etype1038, _size1035) = iprot.readListBegin() + for _i1039 in xrange(_size1035): + _elem1040 = iprot.readString() + self.part_vals.append(_elem1040) iprot.readListEnd() else: iprot.skip(ftype) @@ -25398,8 +25456,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1020 in self.part_vals: - oprot.writeString(iter1020) + for iter1041 in self.part_vals: + oprot.writeString(iter1041) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -25558,11 +25616,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.MAP: self.partitionSpecs = {} - (_ktype1022, _vtype1023, _size1021 ) = iprot.readMapBegin() - for _i1025 in xrange(_size1021): - _key1026 = iprot.readString() - _val1027 = iprot.readString() - self.partitionSpecs[_key1026] = _val1027 + (_ktype1043, _vtype1044, _size1042 ) = iprot.readMapBegin() + for _i1046 in xrange(_size1042): + _key1047 = iprot.readString() + _val1048 = iprot.readString() + self.partitionSpecs[_key1047] = _val1048 iprot.readMapEnd() else: iprot.skip(ftype) @@ -25599,9 +25657,9 @@ def write(self, oprot): if self.partitionSpecs is not None: oprot.writeFieldBegin('partitionSpecs', TType.MAP, 1) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.partitionSpecs)) - for kiter1028,viter1029 in self.partitionSpecs.items(): - oprot.writeString(kiter1028) - oprot.writeString(viter1029) + for kiter1049,viter1050 in self.partitionSpecs.items(): + oprot.writeString(kiter1049) + oprot.writeString(viter1050) oprot.writeMapEnd() oprot.writeFieldEnd() if self.source_db is not None: @@ -25806,11 +25864,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.MAP: self.partitionSpecs = {} - (_ktype1031, _vtype1032, _size1030 ) = iprot.readMapBegin() - for _i1034 in xrange(_size1030): - _key1035 = iprot.readString() - _val1036 = iprot.readString() - self.partitionSpecs[_key1035] = _val1036 + (_ktype1052, _vtype1053, _size1051 ) = iprot.readMapBegin() + for _i1055 in xrange(_size1051): + _key1056 = iprot.readString() + _val1057 = iprot.readString() + self.partitionSpecs[_key1056] = _val1057 iprot.readMapEnd() else: iprot.skip(ftype) @@ -25847,9 +25905,9 @@ def write(self, oprot): if self.partitionSpecs is not None: oprot.writeFieldBegin('partitionSpecs', TType.MAP, 1) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.partitionSpecs)) - for kiter1037,viter1038 in self.partitionSpecs.items(): - oprot.writeString(kiter1037) - oprot.writeString(viter1038) + for kiter1058,viter1059 in self.partitionSpecs.items(): + oprot.writeString(kiter1058) + oprot.writeString(viter1059) oprot.writeMapEnd() oprot.writeFieldEnd() if self.source_db is not None: @@ -25932,11 +25990,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1042, _size1039) = iprot.readListBegin() - for _i1043 in xrange(_size1039): - _elem1044 = Partition() - _elem1044.read(iprot) - self.success.append(_elem1044) + (_etype1063, _size1060) = iprot.readListBegin() + for _i1064 in xrange(_size1060): + _elem1065 = Partition() + _elem1065.read(iprot) + self.success.append(_elem1065) iprot.readListEnd() else: iprot.skip(ftype) @@ -25977,8 +26035,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1045 in self.success: - iter1045.write(oprot) + for iter1066 in self.success: + iter1066.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -26072,10 +26130,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1049, _size1046) = iprot.readListBegin() - for _i1050 in xrange(_size1046): - _elem1051 = iprot.readString() - self.part_vals.append(_elem1051) + (_etype1070, _size1067) = iprot.readListBegin() + for _i1071 in xrange(_size1067): + _elem1072 = iprot.readString() + self.part_vals.append(_elem1072) iprot.readListEnd() else: iprot.skip(ftype) @@ -26087,10 +26145,10 @@ def read(self, iprot): elif fid == 5: if ftype == TType.LIST: self.group_names = [] - (_etype1055, _size1052) = iprot.readListBegin() - for _i1056 in xrange(_size1052): - _elem1057 = iprot.readString() - self.group_names.append(_elem1057) + (_etype1076, _size1073) = iprot.readListBegin() + for _i1077 in xrange(_size1073): + _elem1078 = iprot.readString() + self.group_names.append(_elem1078) iprot.readListEnd() else: iprot.skip(ftype) @@ -26115,8 +26173,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1058 in self.part_vals: - oprot.writeString(iter1058) + for iter1079 in self.part_vals: + oprot.writeString(iter1079) oprot.writeListEnd() oprot.writeFieldEnd() if self.user_name is not None: @@ -26126,8 +26184,8 @@ def write(self, oprot): if self.group_names is not None: oprot.writeFieldBegin('group_names', TType.LIST, 5) oprot.writeListBegin(TType.STRING, len(self.group_names)) - for iter1059 in self.group_names: - oprot.writeString(iter1059) + for iter1080 in self.group_names: + oprot.writeString(iter1080) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -26556,11 +26614,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1063, _size1060) = iprot.readListBegin() - for _i1064 in xrange(_size1060): - _elem1065 = Partition() - _elem1065.read(iprot) - self.success.append(_elem1065) + (_etype1084, _size1081) = iprot.readListBegin() + for _i1085 in xrange(_size1081): + _elem1086 = Partition() + _elem1086.read(iprot) + self.success.append(_elem1086) iprot.readListEnd() else: iprot.skip(ftype) @@ -26589,8 +26647,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1066 in self.success: - iter1066.write(oprot) + for iter1087 in self.success: + iter1087.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -26684,10 +26742,10 @@ def read(self, iprot): elif fid == 5: if ftype == TType.LIST: self.group_names = [] - (_etype1070, _size1067) = iprot.readListBegin() - for _i1071 in xrange(_size1067): - _elem1072 = iprot.readString() - self.group_names.append(_elem1072) + (_etype1091, _size1088) = iprot.readListBegin() + for _i1092 in xrange(_size1088): + _elem1093 = iprot.readString() + self.group_names.append(_elem1093) iprot.readListEnd() else: iprot.skip(ftype) @@ -26720,8 +26778,8 @@ def write(self, oprot): if self.group_names is not None: oprot.writeFieldBegin('group_names', TType.LIST, 5) oprot.writeListBegin(TType.STRING, len(self.group_names)) - for iter1073 in self.group_names: - oprot.writeString(iter1073) + for iter1094 in self.group_names: + oprot.writeString(iter1094) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -26782,11 +26840,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1077, _size1074) = iprot.readListBegin() - for _i1078 in xrange(_size1074): - _elem1079 = Partition() - _elem1079.read(iprot) - self.success.append(_elem1079) + (_etype1098, _size1095) = iprot.readListBegin() + for _i1099 in xrange(_size1095): + _elem1100 = Partition() + _elem1100.read(iprot) + self.success.append(_elem1100) iprot.readListEnd() else: iprot.skip(ftype) @@ -26815,8 +26873,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1080 in self.success: - iter1080.write(oprot) + for iter1101 in self.success: + iter1101.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -26974,11 +27032,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1084, _size1081) = iprot.readListBegin() - for _i1085 in xrange(_size1081): - _elem1086 = PartitionSpec() - _elem1086.read(iprot) - self.success.append(_elem1086) + (_etype1105, _size1102) = iprot.readListBegin() + for _i1106 in xrange(_size1102): + _elem1107 = PartitionSpec() + _elem1107.read(iprot) + self.success.append(_elem1107) iprot.readListEnd() else: iprot.skip(ftype) @@ -27007,8 +27065,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1087 in self.success: - iter1087.write(oprot) + for iter1108 in self.success: + iter1108.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -27166,10 +27224,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1091, _size1088) = iprot.readListBegin() - for _i1092 in xrange(_size1088): - _elem1093 = iprot.readString() - self.success.append(_elem1093) + (_etype1112, _size1109) = iprot.readListBegin() + for _i1113 in xrange(_size1109): + _elem1114 = iprot.readString() + self.success.append(_elem1114) iprot.readListEnd() else: iprot.skip(ftype) @@ -27198,8 +27256,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1094 in self.success: - oprot.writeString(iter1094) + for iter1115 in self.success: + oprot.writeString(iter1115) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -27439,10 +27497,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1098, _size1095) = iprot.readListBegin() - for _i1099 in xrange(_size1095): - _elem1100 = iprot.readString() - self.part_vals.append(_elem1100) + (_etype1119, _size1116) = iprot.readListBegin() + for _i1120 in xrange(_size1116): + _elem1121 = iprot.readString() + self.part_vals.append(_elem1121) iprot.readListEnd() else: iprot.skip(ftype) @@ -27472,8 +27530,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1101 in self.part_vals: - oprot.writeString(iter1101) + for iter1122 in self.part_vals: + oprot.writeString(iter1122) oprot.writeListEnd() oprot.writeFieldEnd() if self.max_parts is not None: @@ -27537,11 +27595,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1105, _size1102) = iprot.readListBegin() - for _i1106 in xrange(_size1102): - _elem1107 = Partition() - _elem1107.read(iprot) - self.success.append(_elem1107) + (_etype1126, _size1123) = iprot.readListBegin() + for _i1127 in xrange(_size1123): + _elem1128 = Partition() + _elem1128.read(iprot) + self.success.append(_elem1128) iprot.readListEnd() else: iprot.skip(ftype) @@ -27570,8 +27628,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1108 in self.success: - iter1108.write(oprot) + for iter1129 in self.success: + iter1129.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -27658,10 +27716,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1112, _size1109) = iprot.readListBegin() - for _i1113 in xrange(_size1109): - _elem1114 = iprot.readString() - self.part_vals.append(_elem1114) + (_etype1133, _size1130) = iprot.readListBegin() + for _i1134 in xrange(_size1130): + _elem1135 = iprot.readString() + self.part_vals.append(_elem1135) iprot.readListEnd() else: iprot.skip(ftype) @@ -27678,10 +27736,10 @@ def read(self, iprot): elif fid == 6: if ftype == TType.LIST: self.group_names = [] - (_etype1118, _size1115) = iprot.readListBegin() - for _i1119 in xrange(_size1115): - _elem1120 = iprot.readString() - self.group_names.append(_elem1120) + (_etype1139, _size1136) = iprot.readListBegin() + for _i1140 in xrange(_size1136): + _elem1141 = iprot.readString() + self.group_names.append(_elem1141) iprot.readListEnd() else: iprot.skip(ftype) @@ -27706,8 +27764,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1121 in self.part_vals: - oprot.writeString(iter1121) + for iter1142 in self.part_vals: + oprot.writeString(iter1142) oprot.writeListEnd() oprot.writeFieldEnd() if self.max_parts is not None: @@ -27721,8 +27779,8 @@ def write(self, oprot): if self.group_names is not None: oprot.writeFieldBegin('group_names', TType.LIST, 6) oprot.writeListBegin(TType.STRING, len(self.group_names)) - for iter1122 in self.group_names: - oprot.writeString(iter1122) + for iter1143 in self.group_names: + oprot.writeString(iter1143) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -27784,11 +27842,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1126, _size1123) = iprot.readListBegin() - for _i1127 in xrange(_size1123): - _elem1128 = Partition() - _elem1128.read(iprot) - self.success.append(_elem1128) + (_etype1147, _size1144) = iprot.readListBegin() + for _i1148 in xrange(_size1144): + _elem1149 = Partition() + _elem1149.read(iprot) + self.success.append(_elem1149) iprot.readListEnd() else: iprot.skip(ftype) @@ -27817,8 +27875,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1129 in self.success: - iter1129.write(oprot) + for iter1150 in self.success: + iter1150.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -27899,10 +27957,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1133, _size1130) = iprot.readListBegin() - for _i1134 in xrange(_size1130): - _elem1135 = iprot.readString() - self.part_vals.append(_elem1135) + (_etype1154, _size1151) = iprot.readListBegin() + for _i1155 in xrange(_size1151): + _elem1156 = iprot.readString() + self.part_vals.append(_elem1156) iprot.readListEnd() else: iprot.skip(ftype) @@ -27932,8 +27990,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1136 in self.part_vals: - oprot.writeString(iter1136) + for iter1157 in self.part_vals: + oprot.writeString(iter1157) oprot.writeListEnd() oprot.writeFieldEnd() if self.max_parts is not None: @@ -27997,10 +28055,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1140, _size1137) = iprot.readListBegin() - for _i1141 in xrange(_size1137): - _elem1142 = iprot.readString() - self.success.append(_elem1142) + (_etype1161, _size1158) = iprot.readListBegin() + for _i1162 in xrange(_size1158): + _elem1163 = iprot.readString() + self.success.append(_elem1163) iprot.readListEnd() else: iprot.skip(ftype) @@ -28029,8 +28087,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1143 in self.success: - oprot.writeString(iter1143) + for iter1164 in self.success: + oprot.writeString(iter1164) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -28201,11 +28259,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1147, _size1144) = iprot.readListBegin() - for _i1148 in xrange(_size1144): - _elem1149 = Partition() - _elem1149.read(iprot) - self.success.append(_elem1149) + (_etype1168, _size1165) = iprot.readListBegin() + for _i1169 in xrange(_size1165): + _elem1170 = Partition() + _elem1170.read(iprot) + self.success.append(_elem1170) iprot.readListEnd() else: iprot.skip(ftype) @@ -28234,8 +28292,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1150 in self.success: - iter1150.write(oprot) + for iter1171 in self.success: + iter1171.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -28406,11 +28464,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1154, _size1151) = iprot.readListBegin() - for _i1155 in xrange(_size1151): - _elem1156 = PartitionSpec() - _elem1156.read(iprot) - self.success.append(_elem1156) + (_etype1175, _size1172) = iprot.readListBegin() + for _i1176 in xrange(_size1172): + _elem1177 = PartitionSpec() + _elem1177.read(iprot) + self.success.append(_elem1177) iprot.readListEnd() else: iprot.skip(ftype) @@ -28439,8 +28497,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1157 in self.success: - iter1157.write(oprot) + for iter1178 in self.success: + iter1178.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -28860,10 +28918,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.names = [] - (_etype1161, _size1158) = iprot.readListBegin() - for _i1162 in xrange(_size1158): - _elem1163 = iprot.readString() - self.names.append(_elem1163) + (_etype1182, _size1179) = iprot.readListBegin() + for _i1183 in xrange(_size1179): + _elem1184 = iprot.readString() + self.names.append(_elem1184) iprot.readListEnd() else: iprot.skip(ftype) @@ -28888,8 +28946,8 @@ def write(self, oprot): if self.names is not None: oprot.writeFieldBegin('names', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.names)) - for iter1164 in self.names: - oprot.writeString(iter1164) + for iter1185 in self.names: + oprot.writeString(iter1185) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -28948,11 +29006,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1168, _size1165) = iprot.readListBegin() - for _i1169 in xrange(_size1165): - _elem1170 = Partition() - _elem1170.read(iprot) - self.success.append(_elem1170) + (_etype1189, _size1186) = iprot.readListBegin() + for _i1190 in xrange(_size1186): + _elem1191 = Partition() + _elem1191.read(iprot) + self.success.append(_elem1191) iprot.readListEnd() else: iprot.skip(ftype) @@ -28981,8 +29039,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1171 in self.success: - iter1171.write(oprot) + for iter1192 in self.success: + iter1192.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -29232,11 +29290,11 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.new_parts = [] - (_etype1175, _size1172) = iprot.readListBegin() - for _i1176 in xrange(_size1172): - _elem1177 = Partition() - _elem1177.read(iprot) - self.new_parts.append(_elem1177) + (_etype1196, _size1193) = iprot.readListBegin() + for _i1197 in xrange(_size1193): + _elem1198 = Partition() + _elem1198.read(iprot) + self.new_parts.append(_elem1198) iprot.readListEnd() else: iprot.skip(ftype) @@ -29261,8 +29319,8 @@ def write(self, oprot): if self.new_parts is not None: oprot.writeFieldBegin('new_parts', TType.LIST, 3) oprot.writeListBegin(TType.STRUCT, len(self.new_parts)) - for iter1178 in self.new_parts: - iter1178.write(oprot) + for iter1199 in self.new_parts: + iter1199.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -29415,11 +29473,11 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.new_parts = [] - (_etype1182, _size1179) = iprot.readListBegin() - for _i1183 in xrange(_size1179): - _elem1184 = Partition() - _elem1184.read(iprot) - self.new_parts.append(_elem1184) + (_etype1203, _size1200) = iprot.readListBegin() + for _i1204 in xrange(_size1200): + _elem1205 = Partition() + _elem1205.read(iprot) + self.new_parts.append(_elem1205) iprot.readListEnd() else: iprot.skip(ftype) @@ -29450,8 +29508,8 @@ def write(self, oprot): if self.new_parts is not None: oprot.writeFieldBegin('new_parts', TType.LIST, 3) oprot.writeListBegin(TType.STRUCT, len(self.new_parts)) - for iter1185 in self.new_parts: - iter1185.write(oprot) + for iter1206 in self.new_parts: + iter1206.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.environment_context is not None: @@ -29795,10 +29853,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.part_vals = [] - (_etype1189, _size1186) = iprot.readListBegin() - for _i1190 in xrange(_size1186): - _elem1191 = iprot.readString() - self.part_vals.append(_elem1191) + (_etype1210, _size1207) = iprot.readListBegin() + for _i1211 in xrange(_size1207): + _elem1212 = iprot.readString() + self.part_vals.append(_elem1212) iprot.readListEnd() else: iprot.skip(ftype) @@ -29829,8 +29887,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1192 in self.part_vals: - oprot.writeString(iter1192) + for iter1213 in self.part_vals: + oprot.writeString(iter1213) oprot.writeListEnd() oprot.writeFieldEnd() if self.new_part is not None: @@ -29972,10 +30030,10 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.part_vals = [] - (_etype1196, _size1193) = iprot.readListBegin() - for _i1197 in xrange(_size1193): - _elem1198 = iprot.readString() - self.part_vals.append(_elem1198) + (_etype1217, _size1214) = iprot.readListBegin() + for _i1218 in xrange(_size1214): + _elem1219 = iprot.readString() + self.part_vals.append(_elem1219) iprot.readListEnd() else: iprot.skip(ftype) @@ -29997,8 +30055,8 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.LIST, 1) oprot.writeListBegin(TType.STRING, len(self.part_vals)) - for iter1199 in self.part_vals: - oprot.writeString(iter1199) + for iter1220 in self.part_vals: + oprot.writeString(iter1220) oprot.writeListEnd() oprot.writeFieldEnd() if self.throw_exception is not None: @@ -30356,10 +30414,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1203, _size1200) = iprot.readListBegin() - for _i1204 in xrange(_size1200): - _elem1205 = iprot.readString() - self.success.append(_elem1205) + (_etype1224, _size1221) = iprot.readListBegin() + for _i1225 in xrange(_size1221): + _elem1226 = iprot.readString() + self.success.append(_elem1226) iprot.readListEnd() else: iprot.skip(ftype) @@ -30382,8 +30440,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1206 in self.success: - oprot.writeString(iter1206) + for iter1227 in self.success: + oprot.writeString(iter1227) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -30507,11 +30565,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.MAP: self.success = {} - (_ktype1208, _vtype1209, _size1207 ) = iprot.readMapBegin() - for _i1211 in xrange(_size1207): - _key1212 = iprot.readString() - _val1213 = iprot.readString() - self.success[_key1212] = _val1213 + (_ktype1229, _vtype1230, _size1228 ) = iprot.readMapBegin() + for _i1232 in xrange(_size1228): + _key1233 = iprot.readString() + _val1234 = iprot.readString() + self.success[_key1233] = _val1234 iprot.readMapEnd() else: iprot.skip(ftype) @@ -30534,9 +30592,9 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.MAP, 0) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.success)) - for kiter1214,viter1215 in self.success.items(): - oprot.writeString(kiter1214) - oprot.writeString(viter1215) + for kiter1235,viter1236 in self.success.items(): + oprot.writeString(kiter1235) + oprot.writeString(viter1236) oprot.writeMapEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -30612,11 +30670,11 @@ def read(self, iprot): elif fid == 3: if ftype == TType.MAP: self.part_vals = {} - (_ktype1217, _vtype1218, _size1216 ) = iprot.readMapBegin() - for _i1220 in xrange(_size1216): - _key1221 = iprot.readString() - _val1222 = iprot.readString() - self.part_vals[_key1221] = _val1222 + (_ktype1238, _vtype1239, _size1237 ) = iprot.readMapBegin() + for _i1241 in xrange(_size1237): + _key1242 = iprot.readString() + _val1243 = iprot.readString() + self.part_vals[_key1242] = _val1243 iprot.readMapEnd() else: iprot.skip(ftype) @@ -30646,9 +30704,9 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.MAP, 3) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.part_vals)) - for kiter1223,viter1224 in self.part_vals.items(): - oprot.writeString(kiter1223) - oprot.writeString(viter1224) + for kiter1244,viter1245 in self.part_vals.items(): + oprot.writeString(kiter1244) + oprot.writeString(viter1245) oprot.writeMapEnd() oprot.writeFieldEnd() if self.eventType is not None: @@ -30862,11 +30920,11 @@ def read(self, iprot): elif fid == 3: if ftype == TType.MAP: self.part_vals = {} - (_ktype1226, _vtype1227, _size1225 ) = iprot.readMapBegin() - for _i1229 in xrange(_size1225): - _key1230 = iprot.readString() - _val1231 = iprot.readString() - self.part_vals[_key1230] = _val1231 + (_ktype1247, _vtype1248, _size1246 ) = iprot.readMapBegin() + for _i1250 in xrange(_size1246): + _key1251 = iprot.readString() + _val1252 = iprot.readString() + self.part_vals[_key1251] = _val1252 iprot.readMapEnd() else: iprot.skip(ftype) @@ -30896,9 +30954,9 @@ def write(self, oprot): if self.part_vals is not None: oprot.writeFieldBegin('part_vals', TType.MAP, 3) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.part_vals)) - for kiter1232,viter1233 in self.part_vals.items(): - oprot.writeString(kiter1232) - oprot.writeString(viter1233) + for kiter1253,viter1254 in self.part_vals.items(): + oprot.writeString(kiter1253) + oprot.writeString(viter1254) oprot.writeMapEnd() oprot.writeFieldEnd() if self.eventType is not None: @@ -34550,10 +34608,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1237, _size1234) = iprot.readListBegin() - for _i1238 in xrange(_size1234): - _elem1239 = iprot.readString() - self.success.append(_elem1239) + (_etype1258, _size1255) = iprot.readListBegin() + for _i1259 in xrange(_size1255): + _elem1260 = iprot.readString() + self.success.append(_elem1260) iprot.readListEnd() else: iprot.skip(ftype) @@ -34576,8 +34634,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1240 in self.success: - oprot.writeString(iter1240) + for iter1261 in self.success: + oprot.writeString(iter1261) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -35265,10 +35323,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1244, _size1241) = iprot.readListBegin() - for _i1245 in xrange(_size1241): - _elem1246 = iprot.readString() - self.success.append(_elem1246) + (_etype1265, _size1262) = iprot.readListBegin() + for _i1266 in xrange(_size1262): + _elem1267 = iprot.readString() + self.success.append(_elem1267) iprot.readListEnd() else: iprot.skip(ftype) @@ -35291,8 +35349,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1247 in self.success: - oprot.writeString(iter1247) + for iter1268 in self.success: + oprot.writeString(iter1268) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -35806,11 +35864,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1251, _size1248) = iprot.readListBegin() - for _i1252 in xrange(_size1248): - _elem1253 = Role() - _elem1253.read(iprot) - self.success.append(_elem1253) + (_etype1272, _size1269) = iprot.readListBegin() + for _i1273 in xrange(_size1269): + _elem1274 = Role() + _elem1274.read(iprot) + self.success.append(_elem1274) iprot.readListEnd() else: iprot.skip(ftype) @@ -35833,8 +35891,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1254 in self.success: - iter1254.write(oprot) + for iter1275 in self.success: + iter1275.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -36343,10 +36401,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.group_names = [] - (_etype1258, _size1255) = iprot.readListBegin() - for _i1259 in xrange(_size1255): - _elem1260 = iprot.readString() - self.group_names.append(_elem1260) + (_etype1279, _size1276) = iprot.readListBegin() + for _i1280 in xrange(_size1276): + _elem1281 = iprot.readString() + self.group_names.append(_elem1281) iprot.readListEnd() else: iprot.skip(ftype) @@ -36371,8 +36429,8 @@ def write(self, oprot): if self.group_names is not None: oprot.writeFieldBegin('group_names', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.group_names)) - for iter1261 in self.group_names: - oprot.writeString(iter1261) + for iter1282 in self.group_names: + oprot.writeString(iter1282) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -36599,11 +36657,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1265, _size1262) = iprot.readListBegin() - for _i1266 in xrange(_size1262): - _elem1267 = HiveObjectPrivilege() - _elem1267.read(iprot) - self.success.append(_elem1267) + (_etype1286, _size1283) = iprot.readListBegin() + for _i1287 in xrange(_size1283): + _elem1288 = HiveObjectPrivilege() + _elem1288.read(iprot) + self.success.append(_elem1288) iprot.readListEnd() else: iprot.skip(ftype) @@ -36626,8 +36684,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1268 in self.success: - iter1268.write(oprot) + for iter1289 in self.success: + iter1289.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -37297,10 +37355,10 @@ def read(self, iprot): elif fid == 2: if ftype == TType.LIST: self.group_names = [] - (_etype1272, _size1269) = iprot.readListBegin() - for _i1273 in xrange(_size1269): - _elem1274 = iprot.readString() - self.group_names.append(_elem1274) + (_etype1293, _size1290) = iprot.readListBegin() + for _i1294 in xrange(_size1290): + _elem1295 = iprot.readString() + self.group_names.append(_elem1295) iprot.readListEnd() else: iprot.skip(ftype) @@ -37321,8 +37379,8 @@ def write(self, oprot): if self.group_names is not None: oprot.writeFieldBegin('group_names', TType.LIST, 2) oprot.writeListBegin(TType.STRING, len(self.group_names)) - for iter1275 in self.group_names: - oprot.writeString(iter1275) + for iter1296 in self.group_names: + oprot.writeString(iter1296) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -37377,10 +37435,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1279, _size1276) = iprot.readListBegin() - for _i1280 in xrange(_size1276): - _elem1281 = iprot.readString() - self.success.append(_elem1281) + (_etype1300, _size1297) = iprot.readListBegin() + for _i1301 in xrange(_size1297): + _elem1302 = iprot.readString() + self.success.append(_elem1302) iprot.readListEnd() else: iprot.skip(ftype) @@ -37403,8 +37461,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1282 in self.success: - oprot.writeString(iter1282) + for iter1303 in self.success: + oprot.writeString(iter1303) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -38336,10 +38394,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1286, _size1283) = iprot.readListBegin() - for _i1287 in xrange(_size1283): - _elem1288 = iprot.readString() - self.success.append(_elem1288) + (_etype1307, _size1304) = iprot.readListBegin() + for _i1308 in xrange(_size1304): + _elem1309 = iprot.readString() + self.success.append(_elem1309) iprot.readListEnd() else: iprot.skip(ftype) @@ -38356,8 +38414,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1289 in self.success: - oprot.writeString(iter1289) + for iter1310 in self.success: + oprot.writeString(iter1310) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -38884,10 +38942,10 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1293, _size1290) = iprot.readListBegin() - for _i1294 in xrange(_size1290): - _elem1295 = iprot.readString() - self.success.append(_elem1295) + (_etype1314, _size1311) = iprot.readListBegin() + for _i1315 in xrange(_size1311): + _elem1316 = iprot.readString() + self.success.append(_elem1316) iprot.readListEnd() else: iprot.skip(ftype) @@ -38904,8 +38962,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRING, len(self.success)) - for iter1296 in self.success: - oprot.writeString(iter1296) + for iter1317 in self.success: + oprot.writeString(iter1317) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -42154,6 +42212,133 @@ def __eq__(self, other): def __ne__(self, other): return not (self == other) +class add_write_notification_log_args: + """ + Attributes: + - rqst + """ + + thrift_spec = None + def __init__(self, rqst=None,): + self.rqst = rqst + + def read(self, iprot): + if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: + fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec)) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == -1: + if ftype == TType.STRUCT: + self.rqst = WriteNotificationLogRequest() + self.rqst.read(iprot) + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None: + oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec))) + return + oprot.writeStructBegin('add_write_notification_log_args') + if self.rqst is not None: + oprot.writeFieldBegin('rqst', TType.STRUCT, -1) + self.rqst.write(oprot) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + + def __hash__(self): + value = 17 + value = (value * 31) ^ hash(self.rqst) + return value + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.iteritems()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) + +class add_write_notification_log_result: + """ + Attributes: + - success + """ + + thrift_spec = ( + (0, TType.STRUCT, 'success', (WriteNotificationLogResponse, WriteNotificationLogResponse.thrift_spec), None, ), # 0 + ) + + def __init__(self, success=None,): + self.success = success + + def read(self, iprot): + if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: + fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec)) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 0: + if ftype == TType.STRUCT: + self.success = WriteNotificationLogResponse() + self.success.read(iprot) + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None: + oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec))) + return + oprot.writeStructBegin('add_write_notification_log_result') + if self.success is not None: + oprot.writeFieldBegin('success', TType.STRUCT, 0) + self.success.write(oprot) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + + def __hash__(self): + value = 17 + value = (value * 31) ^ hash(self.success) + return value + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.iteritems()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) + class cm_recycle_args: """ Attributes: @@ -47185,11 +47370,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1300, _size1297) = iprot.readListBegin() - for _i1301 in xrange(_size1297): - _elem1302 = SchemaVersion() - _elem1302.read(iprot) - self.success.append(_elem1302) + (_etype1321, _size1318) = iprot.readListBegin() + for _i1322 in xrange(_size1318): + _elem1323 = SchemaVersion() + _elem1323.read(iprot) + self.success.append(_elem1323) iprot.readListEnd() else: iprot.skip(ftype) @@ -47218,8 +47403,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1303 in self.success: - iter1303.write(oprot) + for iter1324 in self.success: + iter1324.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: @@ -48694,11 +48879,11 @@ def read(self, iprot): if fid == 0: if ftype == TType.LIST: self.success = [] - (_etype1307, _size1304) = iprot.readListBegin() - for _i1308 in xrange(_size1304): - _elem1309 = RuntimeStat() - _elem1309.read(iprot) - self.success.append(_elem1309) + (_etype1328, _size1325) = iprot.readListBegin() + for _i1329 in xrange(_size1325): + _elem1330 = RuntimeStat() + _elem1330.read(iprot) + self.success.append(_elem1330) iprot.readListEnd() else: iprot.skip(ftype) @@ -48721,8 +48906,8 @@ def write(self, oprot): if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) - for iter1310 in self.success: - iter1310.write(oprot) + for iter1331 in self.success: + iter1331.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.o1 is not None: diff --git a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py index 8fba3df3b0..1285c086c9 100644 --- a/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py +++ b/standalone-metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py @@ -11562,17 +11562,20 @@ class CommitTxnRequest: Attributes: - txnid - replPolicy + - writeEventInfos """ thrift_spec = ( None, # 0 (1, TType.I64, 'txnid', None, None, ), # 1 (2, TType.STRING, 'replPolicy', None, None, ), # 2 + (3, TType.LIST, 'writeEventInfos', (TType.STRUCT,(WriteEventInfo, WriteEventInfo.thrift_spec)), None, ), # 3 ) - def __init__(self, txnid=None, replPolicy=None,): + def __init__(self, txnid=None, replPolicy=None, writeEventInfos=None,): self.txnid = txnid self.replPolicy = replPolicy + self.writeEventInfos = writeEventInfos def read(self, iprot): if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: @@ -11593,6 +11596,17 @@ def read(self, iprot): self.replPolicy = iprot.readString() else: iprot.skip(ftype) + elif fid == 3: + if ftype == TType.LIST: + self.writeEventInfos = [] + (_etype526, _size523) = iprot.readListBegin() + for _i527 in xrange(_size523): + _elem528 = WriteEventInfo() + _elem528.read(iprot) + self.writeEventInfos.append(_elem528) + iprot.readListEnd() + else: + iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() @@ -11611,6 +11625,13 @@ def write(self, oprot): oprot.writeFieldBegin('replPolicy', TType.STRING, 2) oprot.writeString(self.replPolicy) oprot.writeFieldEnd() + if self.writeEventInfos is not None: + oprot.writeFieldBegin('writeEventInfos', TType.LIST, 3) + oprot.writeListBegin(TType.STRUCT, len(self.writeEventInfos)) + for iter529 in self.writeEventInfos: + iter529.write(oprot) + oprot.writeListEnd() + oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() @@ -11624,6 +11645,158 @@ def __hash__(self): value = 17 value = (value * 31) ^ hash(self.txnid) value = (value * 31) ^ hash(self.replPolicy) + value = (value * 31) ^ hash(self.writeEventInfos) + return value + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.iteritems()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) + +class WriteEventInfo: + """ + Attributes: + - writeId + - database + - table + - files + - partition + - tableObj + - partitionObj + """ + + thrift_spec = ( + None, # 0 + (1, TType.I64, 'writeId', None, None, ), # 1 + (2, TType.STRING, 'database', None, None, ), # 2 + (3, TType.STRING, 'table', None, None, ), # 3 + (4, TType.STRING, 'files', None, None, ), # 4 + (5, TType.STRING, 'partition', None, None, ), # 5 + (6, TType.STRING, 'tableObj', None, None, ), # 6 + (7, TType.STRING, 'partitionObj', None, None, ), # 7 + ) + + def __init__(self, writeId=None, database=None, table=None, files=None, partition=None, tableObj=None, partitionObj=None,): + self.writeId = writeId + self.database = database + self.table = table + self.files = files + self.partition = partition + self.tableObj = tableObj + self.partitionObj = partitionObj + + def read(self, iprot): + if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: + fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec)) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 1: + if ftype == TType.I64: + self.writeId = iprot.readI64() + else: + iprot.skip(ftype) + elif fid == 2: + if ftype == TType.STRING: + self.database = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 3: + if ftype == TType.STRING: + self.table = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 4: + if ftype == TType.STRING: + self.files = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 5: + if ftype == TType.STRING: + self.partition = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 6: + if ftype == TType.STRING: + self.tableObj = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 7: + if ftype == TType.STRING: + self.partitionObj = iprot.readString() + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None: + oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec))) + return + oprot.writeStructBegin('WriteEventInfo') + if self.writeId is not None: + oprot.writeFieldBegin('writeId', TType.I64, 1) + oprot.writeI64(self.writeId) + oprot.writeFieldEnd() + if self.database is not None: + oprot.writeFieldBegin('database', TType.STRING, 2) + oprot.writeString(self.database) + oprot.writeFieldEnd() + if self.table is not None: + oprot.writeFieldBegin('table', TType.STRING, 3) + oprot.writeString(self.table) + oprot.writeFieldEnd() + if self.files is not None: + oprot.writeFieldBegin('files', TType.STRING, 4) + oprot.writeString(self.files) + oprot.writeFieldEnd() + if self.partition is not None: + oprot.writeFieldBegin('partition', TType.STRING, 5) + oprot.writeString(self.partition) + oprot.writeFieldEnd() + if self.tableObj is not None: + oprot.writeFieldBegin('tableObj', TType.STRING, 6) + oprot.writeString(self.tableObj) + oprot.writeFieldEnd() + if self.partitionObj is not None: + oprot.writeFieldBegin('partitionObj', TType.STRING, 7) + oprot.writeString(self.partitionObj) + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + if self.writeId is None: + raise TProtocol.TProtocolException(message='Required field writeId is unset!') + if self.database is None: + raise TProtocol.TProtocolException(message='Required field database is unset!') + if self.table is None: + raise TProtocol.TProtocolException(message='Required field table is unset!') + if self.files is None: + raise TProtocol.TProtocolException(message='Required field files is unset!') + return + + + def __hash__(self): + value = 17 + value = (value * 31) ^ hash(self.writeId) + value = (value * 31) ^ hash(self.database) + value = (value * 31) ^ hash(self.table) + value = (value * 31) ^ hash(self.files) + value = (value * 31) ^ hash(self.partition) + value = (value * 31) ^ hash(self.tableObj) + value = (value * 31) ^ hash(self.partitionObj) return value def __repr__(self): @@ -11703,10 +11876,10 @@ def read(self, iprot): elif fid == 6: if ftype == TType.LIST: self.partNames = [] - (_etype526, _size523) = iprot.readListBegin() - for _i527 in xrange(_size523): - _elem528 = iprot.readString() - self.partNames.append(_elem528) + (_etype533, _size530) = iprot.readListBegin() + for _i534 in xrange(_size530): + _elem535 = iprot.readString() + self.partNames.append(_elem535) iprot.readListEnd() else: iprot.skip(ftype) @@ -11743,8 +11916,8 @@ def write(self, oprot): if self.partNames is not None: oprot.writeFieldBegin('partNames', TType.LIST, 6) oprot.writeListBegin(TType.STRING, len(self.partNames)) - for iter529 in self.partNames: - oprot.writeString(iter529) + for iter536 in self.partNames: + oprot.writeString(iter536) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -11814,10 +11987,10 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.fullTableNames = [] - (_etype533, _size530) = iprot.readListBegin() - for _i534 in xrange(_size530): - _elem535 = iprot.readString() - self.fullTableNames.append(_elem535) + (_etype540, _size537) = iprot.readListBegin() + for _i541 in xrange(_size537): + _elem542 = iprot.readString() + self.fullTableNames.append(_elem542) iprot.readListEnd() else: iprot.skip(ftype) @@ -11839,8 +12012,8 @@ def write(self, oprot): if self.fullTableNames is not None: oprot.writeFieldBegin('fullTableNames', TType.LIST, 1) oprot.writeListBegin(TType.STRING, len(self.fullTableNames)) - for iter536 in self.fullTableNames: - oprot.writeString(iter536) + for iter543 in self.fullTableNames: + oprot.writeString(iter543) oprot.writeListEnd() oprot.writeFieldEnd() if self.validTxnList is not None: @@ -11923,10 +12096,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.invalidWriteIds = [] - (_etype540, _size537) = iprot.readListBegin() - for _i541 in xrange(_size537): - _elem542 = iprot.readI64() - self.invalidWriteIds.append(_elem542) + (_etype547, _size544) = iprot.readListBegin() + for _i548 in xrange(_size544): + _elem549 = iprot.readI64() + self.invalidWriteIds.append(_elem549) iprot.readListEnd() else: iprot.skip(ftype) @@ -11961,8 +12134,8 @@ def write(self, oprot): if self.invalidWriteIds is not None: oprot.writeFieldBegin('invalidWriteIds', TType.LIST, 3) oprot.writeListBegin(TType.I64, len(self.invalidWriteIds)) - for iter543 in self.invalidWriteIds: - oprot.writeI64(iter543) + for iter550 in self.invalidWriteIds: + oprot.writeI64(iter550) oprot.writeListEnd() oprot.writeFieldEnd() if self.minOpenWriteId is not None: @@ -12034,11 +12207,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.tblValidWriteIds = [] - (_etype547, _size544) = iprot.readListBegin() - for _i548 in xrange(_size544): - _elem549 = TableValidWriteIds() - _elem549.read(iprot) - self.tblValidWriteIds.append(_elem549) + (_etype554, _size551) = iprot.readListBegin() + for _i555 in xrange(_size551): + _elem556 = TableValidWriteIds() + _elem556.read(iprot) + self.tblValidWriteIds.append(_elem556) iprot.readListEnd() else: iprot.skip(ftype) @@ -12055,8 +12228,8 @@ def write(self, oprot): if self.tblValidWriteIds is not None: oprot.writeFieldBegin('tblValidWriteIds', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.tblValidWriteIds)) - for iter550 in self.tblValidWriteIds: - iter550.write(oprot) + for iter557 in self.tblValidWriteIds: + iter557.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -12132,10 +12305,10 @@ def read(self, iprot): elif fid == 3: if ftype == TType.LIST: self.txnIds = [] - (_etype554, _size551) = iprot.readListBegin() - for _i555 in xrange(_size551): - _elem556 = iprot.readI64() - self.txnIds.append(_elem556) + (_etype561, _size558) = iprot.readListBegin() + for _i562 in xrange(_size558): + _elem563 = iprot.readI64() + self.txnIds.append(_elem563) iprot.readListEnd() else: iprot.skip(ftype) @@ -12147,11 +12320,11 @@ def read(self, iprot): elif fid == 5: if ftype == TType.LIST: self.srcTxnToWriteIdList = [] - (_etype560, _size557) = iprot.readListBegin() - for _i561 in xrange(_size557): - _elem562 = TxnToWriteId() - _elem562.read(iprot) - self.srcTxnToWriteIdList.append(_elem562) + (_etype567, _size564) = iprot.readListBegin() + for _i568 in xrange(_size564): + _elem569 = TxnToWriteId() + _elem569.read(iprot) + self.srcTxnToWriteIdList.append(_elem569) iprot.readListEnd() else: iprot.skip(ftype) @@ -12176,8 +12349,8 @@ def write(self, oprot): if self.txnIds is not None: oprot.writeFieldBegin('txnIds', TType.LIST, 3) oprot.writeListBegin(TType.I64, len(self.txnIds)) - for iter563 in self.txnIds: - oprot.writeI64(iter563) + for iter570 in self.txnIds: + oprot.writeI64(iter570) oprot.writeListEnd() oprot.writeFieldEnd() if self.replPolicy is not None: @@ -12187,8 +12360,8 @@ def write(self, oprot): if self.srcTxnToWriteIdList is not None: oprot.writeFieldBegin('srcTxnToWriteIdList', TType.LIST, 5) oprot.writeListBegin(TType.STRUCT, len(self.srcTxnToWriteIdList)) - for iter564 in self.srcTxnToWriteIdList: - iter564.write(oprot) + for iter571 in self.srcTxnToWriteIdList: + iter571.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -12330,11 +12503,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.txnToWriteIds = [] - (_etype568, _size565) = iprot.readListBegin() - for _i569 in xrange(_size565): - _elem570 = TxnToWriteId() - _elem570.read(iprot) - self.txnToWriteIds.append(_elem570) + (_etype575, _size572) = iprot.readListBegin() + for _i576 in xrange(_size572): + _elem577 = TxnToWriteId() + _elem577.read(iprot) + self.txnToWriteIds.append(_elem577) iprot.readListEnd() else: iprot.skip(ftype) @@ -12351,8 +12524,8 @@ def write(self, oprot): if self.txnToWriteIds is not None: oprot.writeFieldBegin('txnToWriteIds', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.txnToWriteIds)) - for iter571 in self.txnToWriteIds: - iter571.write(oprot) + for iter578 in self.txnToWriteIds: + iter578.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -12580,11 +12753,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.component = [] - (_etype575, _size572) = iprot.readListBegin() - for _i576 in xrange(_size572): - _elem577 = LockComponent() - _elem577.read(iprot) - self.component.append(_elem577) + (_etype582, _size579) = iprot.readListBegin() + for _i583 in xrange(_size579): + _elem584 = LockComponent() + _elem584.read(iprot) + self.component.append(_elem584) iprot.readListEnd() else: iprot.skip(ftype) @@ -12621,8 +12794,8 @@ def write(self, oprot): if self.component is not None: oprot.writeFieldBegin('component', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.component)) - for iter578 in self.component: - iter578.write(oprot) + for iter585 in self.component: + iter585.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.txnid is not None: @@ -13320,11 +13493,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.locks = [] - (_etype582, _size579) = iprot.readListBegin() - for _i583 in xrange(_size579): - _elem584 = ShowLocksResponseElement() - _elem584.read(iprot) - self.locks.append(_elem584) + (_etype589, _size586) = iprot.readListBegin() + for _i590 in xrange(_size586): + _elem591 = ShowLocksResponseElement() + _elem591.read(iprot) + self.locks.append(_elem591) iprot.readListEnd() else: iprot.skip(ftype) @@ -13341,8 +13514,8 @@ def write(self, oprot): if self.locks is not None: oprot.writeFieldBegin('locks', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.locks)) - for iter585 in self.locks: - iter585.write(oprot) + for iter592 in self.locks: + iter592.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -13557,20 +13730,20 @@ def read(self, iprot): if fid == 1: if ftype == TType.SET: self.aborted = set() - (_etype589, _size586) = iprot.readSetBegin() - for _i590 in xrange(_size586): - _elem591 = iprot.readI64() - self.aborted.add(_elem591) + (_etype596, _size593) = iprot.readSetBegin() + for _i597 in xrange(_size593): + _elem598 = iprot.readI64() + self.aborted.add(_elem598) iprot.readSetEnd() else: iprot.skip(ftype) elif fid == 2: if ftype == TType.SET: self.nosuch = set() - (_etype595, _size592) = iprot.readSetBegin() - for _i596 in xrange(_size592): - _elem597 = iprot.readI64() - self.nosuch.add(_elem597) + (_etype602, _size599) = iprot.readSetBegin() + for _i603 in xrange(_size599): + _elem604 = iprot.readI64() + self.nosuch.add(_elem604) iprot.readSetEnd() else: iprot.skip(ftype) @@ -13587,15 +13760,15 @@ def write(self, oprot): if self.aborted is not None: oprot.writeFieldBegin('aborted', TType.SET, 1) oprot.writeSetBegin(TType.I64, len(self.aborted)) - for iter598 in self.aborted: - oprot.writeI64(iter598) + for iter605 in self.aborted: + oprot.writeI64(iter605) oprot.writeSetEnd() oprot.writeFieldEnd() if self.nosuch is not None: oprot.writeFieldBegin('nosuch', TType.SET, 2) oprot.writeSetBegin(TType.I64, len(self.nosuch)) - for iter599 in self.nosuch: - oprot.writeI64(iter599) + for iter606 in self.nosuch: + oprot.writeI64(iter606) oprot.writeSetEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -13692,11 +13865,11 @@ def read(self, iprot): elif fid == 6: if ftype == TType.MAP: self.properties = {} - (_ktype601, _vtype602, _size600 ) = iprot.readMapBegin() - for _i604 in xrange(_size600): - _key605 = iprot.readString() - _val606 = iprot.readString() - self.properties[_key605] = _val606 + (_ktype608, _vtype609, _size607 ) = iprot.readMapBegin() + for _i611 in xrange(_size607): + _key612 = iprot.readString() + _val613 = iprot.readString() + self.properties[_key612] = _val613 iprot.readMapEnd() else: iprot.skip(ftype) @@ -13733,9 +13906,9 @@ def write(self, oprot): if self.properties is not None: oprot.writeFieldBegin('properties', TType.MAP, 6) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.properties)) - for kiter607,viter608 in self.properties.items(): - oprot.writeString(kiter607) - oprot.writeString(viter608) + for kiter614,viter615 in self.properties.items(): + oprot.writeString(kiter614) + oprot.writeString(viter615) oprot.writeMapEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -14170,11 +14343,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.compacts = [] - (_etype612, _size609) = iprot.readListBegin() - for _i613 in xrange(_size609): - _elem614 = ShowCompactResponseElement() - _elem614.read(iprot) - self.compacts.append(_elem614) + (_etype619, _size616) = iprot.readListBegin() + for _i620 in xrange(_size616): + _elem621 = ShowCompactResponseElement() + _elem621.read(iprot) + self.compacts.append(_elem621) iprot.readListEnd() else: iprot.skip(ftype) @@ -14191,8 +14364,8 @@ def write(self, oprot): if self.compacts is not None: oprot.writeFieldBegin('compacts', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.compacts)) - for iter615 in self.compacts: - iter615.write(oprot) + for iter622 in self.compacts: + iter622.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -14281,10 +14454,10 @@ def read(self, iprot): elif fid == 5: if ftype == TType.LIST: self.partitionnames = [] - (_etype619, _size616) = iprot.readListBegin() - for _i620 in xrange(_size616): - _elem621 = iprot.readString() - self.partitionnames.append(_elem621) + (_etype626, _size623) = iprot.readListBegin() + for _i627 in xrange(_size623): + _elem628 = iprot.readString() + self.partitionnames.append(_elem628) iprot.readListEnd() else: iprot.skip(ftype) @@ -14322,8 +14495,8 @@ def write(self, oprot): if self.partitionnames is not None: oprot.writeFieldBegin('partitionnames', TType.LIST, 5) oprot.writeListBegin(TType.STRING, len(self.partitionnames)) - for iter622 in self.partitionnames: - oprot.writeString(iter622) + for iter629 in self.partitionnames: + oprot.writeString(iter629) oprot.writeListEnd() oprot.writeFieldEnd() if self.operationType is not None: @@ -14556,10 +14729,10 @@ def read(self, iprot): elif fid == 4: if ftype == TType.SET: self.tablesUsed = set() - (_etype626, _size623) = iprot.readSetBegin() - for _i627 in xrange(_size623): - _elem628 = iprot.readString() - self.tablesUsed.add(_elem628) + (_etype633, _size630) = iprot.readSetBegin() + for _i634 in xrange(_size630): + _elem635 = iprot.readString() + self.tablesUsed.add(_elem635) iprot.readSetEnd() else: iprot.skip(ftype) @@ -14598,8 +14771,8 @@ def write(self, oprot): if self.tablesUsed is not None: oprot.writeFieldBegin('tablesUsed', TType.SET, 4) oprot.writeSetBegin(TType.STRING, len(self.tablesUsed)) - for iter629 in self.tablesUsed: - oprot.writeString(iter629) + for iter636 in self.tablesUsed: + oprot.writeString(iter636) oprot.writeSetEnd() oprot.writeFieldEnd() if self.validTxnList is not None: @@ -14916,11 +15089,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.events = [] - (_etype633, _size630) = iprot.readListBegin() - for _i634 in xrange(_size630): - _elem635 = NotificationEvent() - _elem635.read(iprot) - self.events.append(_elem635) + (_etype640, _size637) = iprot.readListBegin() + for _i641 in xrange(_size637): + _elem642 = NotificationEvent() + _elem642.read(iprot) + self.events.append(_elem642) iprot.readListEnd() else: iprot.skip(ftype) @@ -14937,8 +15110,8 @@ def write(self, oprot): if self.events is not None: oprot.writeFieldBegin('events', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.events)) - for iter636 in self.events: - iter636.write(oprot) + for iter643 in self.events: + iter643.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -15201,6 +15374,7 @@ class InsertEventRequestData: - replace - filesAdded - filesAddedChecksum + - subDirectoryList """ thrift_spec = ( @@ -15208,12 +15382,14 @@ class InsertEventRequestData: (1, TType.BOOL, 'replace', None, None, ), # 1 (2, TType.LIST, 'filesAdded', (TType.STRING,None), None, ), # 2 (3, TType.LIST, 'filesAddedChecksum', (TType.STRING,None), None, ), # 3 + (4, TType.LIST, 'subDirectoryList', (TType.STRING,None), None, ), # 4 ) - def __init__(self, replace=None, filesAdded=None, filesAddedChecksum=None,): + def __init__(self, replace=None, filesAdded=None, filesAddedChecksum=None, subDirectoryList=None,): self.replace = replace self.filesAdded = filesAdded self.filesAddedChecksum = filesAddedChecksum + self.subDirectoryList = subDirectoryList def read(self, iprot): if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: @@ -15232,20 +15408,30 @@ def read(self, iprot): elif fid == 2: if ftype == TType.LIST: self.filesAdded = [] - (_etype640, _size637) = iprot.readListBegin() - for _i641 in xrange(_size637): - _elem642 = iprot.readString() - self.filesAdded.append(_elem642) + (_etype647, _size644) = iprot.readListBegin() + for _i648 in xrange(_size644): + _elem649 = iprot.readString() + self.filesAdded.append(_elem649) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 3: if ftype == TType.LIST: self.filesAddedChecksum = [] - (_etype646, _size643) = iprot.readListBegin() - for _i647 in xrange(_size643): - _elem648 = iprot.readString() - self.filesAddedChecksum.append(_elem648) + (_etype653, _size650) = iprot.readListBegin() + for _i654 in xrange(_size650): + _elem655 = iprot.readString() + self.filesAddedChecksum.append(_elem655) + iprot.readListEnd() + else: + iprot.skip(ftype) + elif fid == 4: + if ftype == TType.LIST: + self.subDirectoryList = [] + (_etype659, _size656) = iprot.readListBegin() + for _i660 in xrange(_size656): + _elem661 = iprot.readString() + self.subDirectoryList.append(_elem661) iprot.readListEnd() else: iprot.skip(ftype) @@ -15266,15 +15452,22 @@ def write(self, oprot): if self.filesAdded is not None: oprot.writeFieldBegin('filesAdded', TType.LIST, 2) oprot.writeListBegin(TType.STRING, len(self.filesAdded)) - for iter649 in self.filesAdded: - oprot.writeString(iter649) + for iter662 in self.filesAdded: + oprot.writeString(iter662) oprot.writeListEnd() oprot.writeFieldEnd() if self.filesAddedChecksum is not None: oprot.writeFieldBegin('filesAddedChecksum', TType.LIST, 3) oprot.writeListBegin(TType.STRING, len(self.filesAddedChecksum)) - for iter650 in self.filesAddedChecksum: - oprot.writeString(iter650) + for iter663 in self.filesAddedChecksum: + oprot.writeString(iter663) + oprot.writeListEnd() + oprot.writeFieldEnd() + if self.subDirectoryList is not None: + oprot.writeFieldBegin('subDirectoryList', TType.LIST, 4) + oprot.writeListBegin(TType.STRING, len(self.subDirectoryList)) + for iter664 in self.subDirectoryList: + oprot.writeString(iter664) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -15291,6 +15484,7 @@ def __hash__(self): value = (value * 31) ^ hash(self.replace) value = (value * 31) ^ hash(self.filesAdded) value = (value * 31) ^ hash(self.filesAddedChecksum) + value = (value * 31) ^ hash(self.subDirectoryList) return value def __repr__(self): @@ -15432,10 +15626,10 @@ def read(self, iprot): elif fid == 5: if ftype == TType.LIST: self.partitionVals = [] - (_etype654, _size651) = iprot.readListBegin() - for _i655 in xrange(_size651): - _elem656 = iprot.readString() - self.partitionVals.append(_elem656) + (_etype668, _size665) = iprot.readListBegin() + for _i669 in xrange(_size665): + _elem670 = iprot.readString() + self.partitionVals.append(_elem670) iprot.readListEnd() else: iprot.skip(ftype) @@ -15473,8 +15667,8 @@ def write(self, oprot): if self.partitionVals is not None: oprot.writeFieldBegin('partitionVals', TType.LIST, 5) oprot.writeListBegin(TType.STRING, len(self.partitionVals)) - for iter657 in self.partitionVals: - oprot.writeString(iter657) + for iter671 in self.partitionVals: + oprot.writeString(iter671) oprot.writeListEnd() oprot.writeFieldEnd() if self.catName is not None: @@ -15559,6 +15753,201 @@ def __eq__(self, other): def __ne__(self, other): return not (self == other) +class WriteNotificationLogRequest: + """ + Attributes: + - txnId + - writeId + - db + - table + - fileInfo + - partitionVals + """ + + thrift_spec = ( + None, # 0 + (1, TType.I64, 'txnId', None, None, ), # 1 + (2, TType.I64, 'writeId', None, None, ), # 2 + (3, TType.STRING, 'db', None, None, ), # 3 + (4, TType.STRING, 'table', None, None, ), # 4 + (5, TType.STRUCT, 'fileInfo', (InsertEventRequestData, InsertEventRequestData.thrift_spec), None, ), # 5 + (6, TType.LIST, 'partitionVals', (TType.STRING,None), None, ), # 6 + ) + + def __init__(self, txnId=None, writeId=None, db=None, table=None, fileInfo=None, partitionVals=None,): + self.txnId = txnId + self.writeId = writeId + self.db = db + self.table = table + self.fileInfo = fileInfo + self.partitionVals = partitionVals + + def read(self, iprot): + if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: + fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec)) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + if fid == 1: + if ftype == TType.I64: + self.txnId = iprot.readI64() + else: + iprot.skip(ftype) + elif fid == 2: + if ftype == TType.I64: + self.writeId = iprot.readI64() + else: + iprot.skip(ftype) + elif fid == 3: + if ftype == TType.STRING: + self.db = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 4: + if ftype == TType.STRING: + self.table = iprot.readString() + else: + iprot.skip(ftype) + elif fid == 5: + if ftype == TType.STRUCT: + self.fileInfo = InsertEventRequestData() + self.fileInfo.read(iprot) + else: + iprot.skip(ftype) + elif fid == 6: + if ftype == TType.LIST: + self.partitionVals = [] + (_etype675, _size672) = iprot.readListBegin() + for _i676 in xrange(_size672): + _elem677 = iprot.readString() + self.partitionVals.append(_elem677) + iprot.readListEnd() + else: + iprot.skip(ftype) + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None: + oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec))) + return + oprot.writeStructBegin('WriteNotificationLogRequest') + if self.txnId is not None: + oprot.writeFieldBegin('txnId', TType.I64, 1) + oprot.writeI64(self.txnId) + oprot.writeFieldEnd() + if self.writeId is not None: + oprot.writeFieldBegin('writeId', TType.I64, 2) + oprot.writeI64(self.writeId) + oprot.writeFieldEnd() + if self.db is not None: + oprot.writeFieldBegin('db', TType.STRING, 3) + oprot.writeString(self.db) + oprot.writeFieldEnd() + if self.table is not None: + oprot.writeFieldBegin('table', TType.STRING, 4) + oprot.writeString(self.table) + oprot.writeFieldEnd() + if self.fileInfo is not None: + oprot.writeFieldBegin('fileInfo', TType.STRUCT, 5) + self.fileInfo.write(oprot) + oprot.writeFieldEnd() + if self.partitionVals is not None: + oprot.writeFieldBegin('partitionVals', TType.LIST, 6) + oprot.writeListBegin(TType.STRING, len(self.partitionVals)) + for iter678 in self.partitionVals: + oprot.writeString(iter678) + oprot.writeListEnd() + oprot.writeFieldEnd() + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + if self.txnId is None: + raise TProtocol.TProtocolException(message='Required field txnId is unset!') + if self.writeId is None: + raise TProtocol.TProtocolException(message='Required field writeId is unset!') + if self.db is None: + raise TProtocol.TProtocolException(message='Required field db is unset!') + if self.table is None: + raise TProtocol.TProtocolException(message='Required field table is unset!') + if self.fileInfo is None: + raise TProtocol.TProtocolException(message='Required field fileInfo is unset!') + return + + + def __hash__(self): + value = 17 + value = (value * 31) ^ hash(self.txnId) + value = (value * 31) ^ hash(self.writeId) + value = (value * 31) ^ hash(self.db) + value = (value * 31) ^ hash(self.table) + value = (value * 31) ^ hash(self.fileInfo) + value = (value * 31) ^ hash(self.partitionVals) + return value + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.iteritems()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) + +class WriteNotificationLogResponse: + + thrift_spec = ( + ) + + def read(self, iprot): + if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: + fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec)) + return + iprot.readStructBegin() + while True: + (fname, ftype, fid) = iprot.readFieldBegin() + if ftype == TType.STOP: + break + else: + iprot.skip(ftype) + iprot.readFieldEnd() + iprot.readStructEnd() + + def write(self, oprot): + if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None: + oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec))) + return + oprot.writeStructBegin('WriteNotificationLogResponse') + oprot.writeFieldStop() + oprot.writeStructEnd() + + def validate(self): + return + + + def __hash__(self): + value = 17 + return value + + def __repr__(self): + L = ['%s=%r' % (key, value) + for key, value in self.__dict__.iteritems()] + return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) + + def __eq__(self, other): + return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ + + def __ne__(self, other): + return not (self == other) + class MetadataPpdResult: """ Attributes: @@ -15666,12 +16055,12 @@ def read(self, iprot): if fid == 1: if ftype == TType.MAP: self.metadata = {} - (_ktype659, _vtype660, _size658 ) = iprot.readMapBegin() - for _i662 in xrange(_size658): - _key663 = iprot.readI64() - _val664 = MetadataPpdResult() - _val664.read(iprot) - self.metadata[_key663] = _val664 + (_ktype680, _vtype681, _size679 ) = iprot.readMapBegin() + for _i683 in xrange(_size679): + _key684 = iprot.readI64() + _val685 = MetadataPpdResult() + _val685.read(iprot) + self.metadata[_key684] = _val685 iprot.readMapEnd() else: iprot.skip(ftype) @@ -15693,9 +16082,9 @@ def write(self, oprot): if self.metadata is not None: oprot.writeFieldBegin('metadata', TType.MAP, 1) oprot.writeMapBegin(TType.I64, TType.STRUCT, len(self.metadata)) - for kiter665,viter666 in self.metadata.items(): - oprot.writeI64(kiter665) - viter666.write(oprot) + for kiter686,viter687 in self.metadata.items(): + oprot.writeI64(kiter686) + viter687.write(oprot) oprot.writeMapEnd() oprot.writeFieldEnd() if self.isSupported is not None: @@ -15765,10 +16154,10 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.fileIds = [] - (_etype670, _size667) = iprot.readListBegin() - for _i671 in xrange(_size667): - _elem672 = iprot.readI64() - self.fileIds.append(_elem672) + (_etype691, _size688) = iprot.readListBegin() + for _i692 in xrange(_size688): + _elem693 = iprot.readI64() + self.fileIds.append(_elem693) iprot.readListEnd() else: iprot.skip(ftype) @@ -15800,8 +16189,8 @@ def write(self, oprot): if self.fileIds is not None: oprot.writeFieldBegin('fileIds', TType.LIST, 1) oprot.writeListBegin(TType.I64, len(self.fileIds)) - for iter673 in self.fileIds: - oprot.writeI64(iter673) + for iter694 in self.fileIds: + oprot.writeI64(iter694) oprot.writeListEnd() oprot.writeFieldEnd() if self.expr is not None: @@ -15875,11 +16264,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.MAP: self.metadata = {} - (_ktype675, _vtype676, _size674 ) = iprot.readMapBegin() - for _i678 in xrange(_size674): - _key679 = iprot.readI64() - _val680 = iprot.readString() - self.metadata[_key679] = _val680 + (_ktype696, _vtype697, _size695 ) = iprot.readMapBegin() + for _i699 in xrange(_size695): + _key700 = iprot.readI64() + _val701 = iprot.readString() + self.metadata[_key700] = _val701 iprot.readMapEnd() else: iprot.skip(ftype) @@ -15901,9 +16290,9 @@ def write(self, oprot): if self.metadata is not None: oprot.writeFieldBegin('metadata', TType.MAP, 1) oprot.writeMapBegin(TType.I64, TType.STRING, len(self.metadata)) - for kiter681,viter682 in self.metadata.items(): - oprot.writeI64(kiter681) - oprot.writeString(viter682) + for kiter702,viter703 in self.metadata.items(): + oprot.writeI64(kiter702) + oprot.writeString(viter703) oprot.writeMapEnd() oprot.writeFieldEnd() if self.isSupported is not None: @@ -15964,10 +16353,10 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.fileIds = [] - (_etype686, _size683) = iprot.readListBegin() - for _i687 in xrange(_size683): - _elem688 = iprot.readI64() - self.fileIds.append(_elem688) + (_etype707, _size704) = iprot.readListBegin() + for _i708 in xrange(_size704): + _elem709 = iprot.readI64() + self.fileIds.append(_elem709) iprot.readListEnd() else: iprot.skip(ftype) @@ -15984,8 +16373,8 @@ def write(self, oprot): if self.fileIds is not None: oprot.writeFieldBegin('fileIds', TType.LIST, 1) oprot.writeListBegin(TType.I64, len(self.fileIds)) - for iter689 in self.fileIds: - oprot.writeI64(iter689) + for iter710 in self.fileIds: + oprot.writeI64(iter710) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -16091,20 +16480,20 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.fileIds = [] - (_etype693, _size690) = iprot.readListBegin() - for _i694 in xrange(_size690): - _elem695 = iprot.readI64() - self.fileIds.append(_elem695) + (_etype714, _size711) = iprot.readListBegin() + for _i715 in xrange(_size711): + _elem716 = iprot.readI64() + self.fileIds.append(_elem716) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 2: if ftype == TType.LIST: self.metadata = [] - (_etype699, _size696) = iprot.readListBegin() - for _i700 in xrange(_size696): - _elem701 = iprot.readString() - self.metadata.append(_elem701) + (_etype720, _size717) = iprot.readListBegin() + for _i721 in xrange(_size717): + _elem722 = iprot.readString() + self.metadata.append(_elem722) iprot.readListEnd() else: iprot.skip(ftype) @@ -16126,15 +16515,15 @@ def write(self, oprot): if self.fileIds is not None: oprot.writeFieldBegin('fileIds', TType.LIST, 1) oprot.writeListBegin(TType.I64, len(self.fileIds)) - for iter702 in self.fileIds: - oprot.writeI64(iter702) + for iter723 in self.fileIds: + oprot.writeI64(iter723) oprot.writeListEnd() oprot.writeFieldEnd() if self.metadata is not None: oprot.writeFieldBegin('metadata', TType.LIST, 2) oprot.writeListBegin(TType.STRING, len(self.metadata)) - for iter703 in self.metadata: - oprot.writeString(iter703) + for iter724 in self.metadata: + oprot.writeString(iter724) oprot.writeListEnd() oprot.writeFieldEnd() if self.type is not None: @@ -16242,10 +16631,10 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.fileIds = [] - (_etype707, _size704) = iprot.readListBegin() - for _i708 in xrange(_size704): - _elem709 = iprot.readI64() - self.fileIds.append(_elem709) + (_etype728, _size725) = iprot.readListBegin() + for _i729 in xrange(_size725): + _elem730 = iprot.readI64() + self.fileIds.append(_elem730) iprot.readListEnd() else: iprot.skip(ftype) @@ -16262,8 +16651,8 @@ def write(self, oprot): if self.fileIds is not None: oprot.writeFieldBegin('fileIds', TType.LIST, 1) oprot.writeListBegin(TType.I64, len(self.fileIds)) - for iter710 in self.fileIds: - oprot.writeI64(iter710) + for iter731 in self.fileIds: + oprot.writeI64(iter731) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -16492,11 +16881,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.functions = [] - (_etype714, _size711) = iprot.readListBegin() - for _i715 in xrange(_size711): - _elem716 = Function() - _elem716.read(iprot) - self.functions.append(_elem716) + (_etype735, _size732) = iprot.readListBegin() + for _i736 in xrange(_size732): + _elem737 = Function() + _elem737.read(iprot) + self.functions.append(_elem737) iprot.readListEnd() else: iprot.skip(ftype) @@ -16513,8 +16902,8 @@ def write(self, oprot): if self.functions is not None: oprot.writeFieldBegin('functions', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.functions)) - for iter717 in self.functions: - iter717.write(oprot) + for iter738 in self.functions: + iter738.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -16566,10 +16955,10 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.values = [] - (_etype721, _size718) = iprot.readListBegin() - for _i722 in xrange(_size718): - _elem723 = iprot.readI32() - self.values.append(_elem723) + (_etype742, _size739) = iprot.readListBegin() + for _i743 in xrange(_size739): + _elem744 = iprot.readI32() + self.values.append(_elem744) iprot.readListEnd() else: iprot.skip(ftype) @@ -16586,8 +16975,8 @@ def write(self, oprot): if self.values is not None: oprot.writeFieldBegin('values', TType.LIST, 1) oprot.writeListBegin(TType.I32, len(self.values)) - for iter724 in self.values: - oprot.writeI32(iter724) + for iter745 in self.values: + oprot.writeI32(iter745) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -16832,10 +17221,10 @@ def read(self, iprot): elif fid == 2: if ftype == TType.LIST: self.tblNames = [] - (_etype728, _size725) = iprot.readListBegin() - for _i729 in xrange(_size725): - _elem730 = iprot.readString() - self.tblNames.append(_elem730) + (_etype749, _size746) = iprot.readListBegin() + for _i750 in xrange(_size746): + _elem751 = iprot.readString() + self.tblNames.append(_elem751) iprot.readListEnd() else: iprot.skip(ftype) @@ -16867,8 +17256,8 @@ def write(self, oprot): if self.tblNames is not None: oprot.writeFieldBegin('tblNames', TType.LIST, 2) oprot.writeListBegin(TType.STRING, len(self.tblNames)) - for iter731 in self.tblNames: - oprot.writeString(iter731) + for iter752 in self.tblNames: + oprot.writeString(iter752) oprot.writeListEnd() oprot.writeFieldEnd() if self.capabilities is not None: @@ -16933,11 +17322,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.tables = [] - (_etype735, _size732) = iprot.readListBegin() - for _i736 in xrange(_size732): - _elem737 = Table() - _elem737.read(iprot) - self.tables.append(_elem737) + (_etype756, _size753) = iprot.readListBegin() + for _i757 in xrange(_size753): + _elem758 = Table() + _elem758.read(iprot) + self.tables.append(_elem758) iprot.readListEnd() else: iprot.skip(ftype) @@ -16954,8 +17343,8 @@ def write(self, oprot): if self.tables is not None: oprot.writeFieldBegin('tables', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.tables)) - for iter738 in self.tables: - iter738.write(oprot) + for iter759 in self.tables: + iter759.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -18163,44 +18552,44 @@ def read(self, iprot): elif fid == 2: if ftype == TType.LIST: self.pools = [] - (_etype742, _size739) = iprot.readListBegin() - for _i743 in xrange(_size739): - _elem744 = WMPool() - _elem744.read(iprot) - self.pools.append(_elem744) + (_etype763, _size760) = iprot.readListBegin() + for _i764 in xrange(_size760): + _elem765 = WMPool() + _elem765.read(iprot) + self.pools.append(_elem765) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 3: if ftype == TType.LIST: self.mappings = [] - (_etype748, _size745) = iprot.readListBegin() - for _i749 in xrange(_size745): - _elem750 = WMMapping() - _elem750.read(iprot) - self.mappings.append(_elem750) + (_etype769, _size766) = iprot.readListBegin() + for _i770 in xrange(_size766): + _elem771 = WMMapping() + _elem771.read(iprot) + self.mappings.append(_elem771) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 4: if ftype == TType.LIST: self.triggers = [] - (_etype754, _size751) = iprot.readListBegin() - for _i755 in xrange(_size751): - _elem756 = WMTrigger() - _elem756.read(iprot) - self.triggers.append(_elem756) + (_etype775, _size772) = iprot.readListBegin() + for _i776 in xrange(_size772): + _elem777 = WMTrigger() + _elem777.read(iprot) + self.triggers.append(_elem777) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 5: if ftype == TType.LIST: self.poolTriggers = [] - (_etype760, _size757) = iprot.readListBegin() - for _i761 in xrange(_size757): - _elem762 = WMPoolTrigger() - _elem762.read(iprot) - self.poolTriggers.append(_elem762) + (_etype781, _size778) = iprot.readListBegin() + for _i782 in xrange(_size778): + _elem783 = WMPoolTrigger() + _elem783.read(iprot) + self.poolTriggers.append(_elem783) iprot.readListEnd() else: iprot.skip(ftype) @@ -18221,29 +18610,29 @@ def write(self, oprot): if self.pools is not None: oprot.writeFieldBegin('pools', TType.LIST, 2) oprot.writeListBegin(TType.STRUCT, len(self.pools)) - for iter763 in self.pools: - iter763.write(oprot) + for iter784 in self.pools: + iter784.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.mappings is not None: oprot.writeFieldBegin('mappings', TType.LIST, 3) oprot.writeListBegin(TType.STRUCT, len(self.mappings)) - for iter764 in self.mappings: - iter764.write(oprot) + for iter785 in self.mappings: + iter785.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.triggers is not None: oprot.writeFieldBegin('triggers', TType.LIST, 4) oprot.writeListBegin(TType.STRUCT, len(self.triggers)) - for iter765 in self.triggers: - iter765.write(oprot) + for iter786 in self.triggers: + iter786.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.poolTriggers is not None: oprot.writeFieldBegin('poolTriggers', TType.LIST, 5) oprot.writeListBegin(TType.STRUCT, len(self.poolTriggers)) - for iter766 in self.poolTriggers: - iter766.write(oprot) + for iter787 in self.poolTriggers: + iter787.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -18717,11 +19106,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.resourcePlans = [] - (_etype770, _size767) = iprot.readListBegin() - for _i771 in xrange(_size767): - _elem772 = WMResourcePlan() - _elem772.read(iprot) - self.resourcePlans.append(_elem772) + (_etype791, _size788) = iprot.readListBegin() + for _i792 in xrange(_size788): + _elem793 = WMResourcePlan() + _elem793.read(iprot) + self.resourcePlans.append(_elem793) iprot.readListEnd() else: iprot.skip(ftype) @@ -18738,8 +19127,8 @@ def write(self, oprot): if self.resourcePlans is not None: oprot.writeFieldBegin('resourcePlans', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.resourcePlans)) - for iter773 in self.resourcePlans: - iter773.write(oprot) + for iter794 in self.resourcePlans: + iter794.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -19043,20 +19432,20 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.errors = [] - (_etype777, _size774) = iprot.readListBegin() - for _i778 in xrange(_size774): - _elem779 = iprot.readString() - self.errors.append(_elem779) + (_etype798, _size795) = iprot.readListBegin() + for _i799 in xrange(_size795): + _elem800 = iprot.readString() + self.errors.append(_elem800) iprot.readListEnd() else: iprot.skip(ftype) elif fid == 2: if ftype == TType.LIST: self.warnings = [] - (_etype783, _size780) = iprot.readListBegin() - for _i784 in xrange(_size780): - _elem785 = iprot.readString() - self.warnings.append(_elem785) + (_etype804, _size801) = iprot.readListBegin() + for _i805 in xrange(_size801): + _elem806 = iprot.readString() + self.warnings.append(_elem806) iprot.readListEnd() else: iprot.skip(ftype) @@ -19073,15 +19462,15 @@ def write(self, oprot): if self.errors is not None: oprot.writeFieldBegin('errors', TType.LIST, 1) oprot.writeListBegin(TType.STRING, len(self.errors)) - for iter786 in self.errors: - oprot.writeString(iter786) + for iter807 in self.errors: + oprot.writeString(iter807) oprot.writeListEnd() oprot.writeFieldEnd() if self.warnings is not None: oprot.writeFieldBegin('warnings', TType.LIST, 2) oprot.writeListBegin(TType.STRING, len(self.warnings)) - for iter787 in self.warnings: - oprot.writeString(iter787) + for iter808 in self.warnings: + oprot.writeString(iter808) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -19658,11 +20047,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.triggers = [] - (_etype791, _size788) = iprot.readListBegin() - for _i792 in xrange(_size788): - _elem793 = WMTrigger() - _elem793.read(iprot) - self.triggers.append(_elem793) + (_etype812, _size809) = iprot.readListBegin() + for _i813 in xrange(_size809): + _elem814 = WMTrigger() + _elem814.read(iprot) + self.triggers.append(_elem814) iprot.readListEnd() else: iprot.skip(ftype) @@ -19679,8 +20068,8 @@ def write(self, oprot): if self.triggers is not None: oprot.writeFieldBegin('triggers', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.triggers)) - for iter794 in self.triggers: - iter794.write(oprot) + for iter815 in self.triggers: + iter815.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() @@ -20864,11 +21253,11 @@ def read(self, iprot): elif fid == 4: if ftype == TType.LIST: self.cols = [] - (_etype798, _size795) = iprot.readListBegin() - for _i799 in xrange(_size795): - _elem800 = FieldSchema() - _elem800.read(iprot) - self.cols.append(_elem800) + (_etype819, _size816) = iprot.readListBegin() + for _i820 in xrange(_size816): + _elem821 = FieldSchema() + _elem821.read(iprot) + self.cols.append(_elem821) iprot.readListEnd() else: iprot.skip(ftype) @@ -20928,8 +21317,8 @@ def write(self, oprot): if self.cols is not None: oprot.writeFieldBegin('cols', TType.LIST, 4) oprot.writeListBegin(TType.STRUCT, len(self.cols)) - for iter801 in self.cols: - iter801.write(oprot) + for iter822 in self.cols: + iter822.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() if self.state is not None: @@ -21184,11 +21573,11 @@ def read(self, iprot): if fid == 1: if ftype == TType.LIST: self.schemaVersions = [] - (_etype805, _size802) = iprot.readListBegin() - for _i806 in xrange(_size802): - _elem807 = SchemaVersionDescriptor() - _elem807.read(iprot) - self.schemaVersions.append(_elem807) + (_etype826, _size823) = iprot.readListBegin() + for _i827 in xrange(_size823): + _elem828 = SchemaVersionDescriptor() + _elem828.read(iprot) + self.schemaVersions.append(_elem828) iprot.readListEnd() else: iprot.skip(ftype) @@ -21205,8 +21594,8 @@ def write(self, oprot): if self.schemaVersions is not None: oprot.writeFieldBegin('schemaVersions', TType.LIST, 1) oprot.writeListBegin(TType.STRUCT, len(self.schemaVersions)) - for iter808 in self.schemaVersions: - iter808.write(oprot) + for iter829 in self.schemaVersions: + iter829.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() diff --git a/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb b/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb index cc77b502bb..a0fabfe781 100644 --- a/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb +++ b/standalone-metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb @@ -2582,10 +2582,12 @@ class CommitTxnRequest include ::Thrift::Struct, ::Thrift::Struct_Union TXNID = 1 REPLPOLICY = 2 + WRITEEVENTINFOS = 3 FIELDS = { TXNID => {:type => ::Thrift::Types::I64, :name => 'txnid'}, - REPLPOLICY => {:type => ::Thrift::Types::STRING, :name => 'replPolicy', :optional => true} + REPLPOLICY => {:type => ::Thrift::Types::STRING, :name => 'replPolicy', :optional => true}, + WRITEEVENTINFOS => {:type => ::Thrift::Types::LIST, :name => 'writeEventInfos', :element => {:type => ::Thrift::Types::STRUCT, :class => ::WriteEventInfo}, :optional => true} } def struct_fields; FIELDS; end @@ -2597,6 +2599,38 @@ class CommitTxnRequest ::Thrift::Struct.generate_accessors self end +class WriteEventInfo + include ::Thrift::Struct, ::Thrift::Struct_Union + WRITEID = 1 + DATABASE = 2 + TABLE = 3 + FILES = 4 + PARTITION = 5 + TABLEOBJ = 6 + PARTITIONOBJ = 7 + + FIELDS = { + WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId'}, + DATABASE => {:type => ::Thrift::Types::STRING, :name => 'database'}, + TABLE => {:type => ::Thrift::Types::STRING, :name => 'table'}, + FILES => {:type => ::Thrift::Types::STRING, :name => 'files'}, + PARTITION => {:type => ::Thrift::Types::STRING, :name => 'partition', :optional => true}, + TABLEOBJ => {:type => ::Thrift::Types::STRING, :name => 'tableObj', :optional => true}, + PARTITIONOBJ => {:type => ::Thrift::Types::STRING, :name => 'partitionObj', :optional => true} + } + + def struct_fields; FIELDS; end + + def validate + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field writeId is unset!') unless @writeId + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field database is unset!') unless @database + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field table is unset!') unless @table + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field files is unset!') unless @files + end + + ::Thrift::Struct.generate_accessors self +end + class ReplTblWriteIdStateRequest include ::Thrift::Struct, ::Thrift::Struct_Union VALIDWRITEIDLIST = 1 @@ -3397,11 +3431,13 @@ class InsertEventRequestData REPLACE = 1 FILESADDED = 2 FILESADDEDCHECKSUM = 3 + SUBDIRECTORYLIST = 4 FIELDS = { REPLACE => {:type => ::Thrift::Types::BOOL, :name => 'replace', :optional => true}, FILESADDED => {:type => ::Thrift::Types::LIST, :name => 'filesAdded', :element => {:type => ::Thrift::Types::STRING}}, - FILESADDEDCHECKSUM => {:type => ::Thrift::Types::LIST, :name => 'filesAddedChecksum', :element => {:type => ::Thrift::Types::STRING}, :optional => true} + FILESADDEDCHECKSUM => {:type => ::Thrift::Types::LIST, :name => 'filesAddedChecksum', :element => {:type => ::Thrift::Types::STRING}, :optional => true}, + SUBDIRECTORYLIST => {:type => ::Thrift::Types::LIST, :name => 'subDirectoryList', :element => {:type => ::Thrift::Types::STRING}, :optional => true} } def struct_fields; FIELDS; end @@ -3479,6 +3515,52 @@ class FireEventResponse ::Thrift::Struct.generate_accessors self end +class WriteNotificationLogRequest + include ::Thrift::Struct, ::Thrift::Struct_Union + TXNID = 1 + WRITEID = 2 + DB = 3 + TABLE = 4 + FILEINFO = 5 + PARTITIONVALS = 6 + + FIELDS = { + TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId'}, + WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId'}, + DB => {:type => ::Thrift::Types::STRING, :name => 'db'}, + TABLE => {:type => ::Thrift::Types::STRING, :name => 'table'}, + FILEINFO => {:type => ::Thrift::Types::STRUCT, :name => 'fileInfo', :class => ::InsertEventRequestData}, + PARTITIONVALS => {:type => ::Thrift::Types::LIST, :name => 'partitionVals', :element => {:type => ::Thrift::Types::STRING}, :optional => true} + } + + def struct_fields; FIELDS; end + + def validate + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field txnId is unset!') unless @txnId + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field writeId is unset!') unless @writeId + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field db is unset!') unless @db + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field table is unset!') unless @table + raise ::Thrift::ProtocolException.new(::Thrift::ProtocolException::UNKNOWN, 'Required field fileInfo is unset!') unless @fileInfo + end + + ::Thrift::Struct.generate_accessors self +end + +class WriteNotificationLogResponse + include ::Thrift::Struct, ::Thrift::Struct_Union + + FIELDS = { + + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self +end + class MetadataPpdResult include ::Thrift::Struct, ::Thrift::Struct_Union METADATA = 1 diff --git a/standalone-metastore/src/gen/thrift/gen-rb/thrift_hive_metastore.rb b/standalone-metastore/src/gen/thrift/gen-rb/thrift_hive_metastore.rb index 1e1a18f46c..5ecfbeda64 100644 --- a/standalone-metastore/src/gen/thrift/gen-rb/thrift_hive_metastore.rb +++ b/standalone-metastore/src/gen/thrift/gen-rb/thrift_hive_metastore.rb @@ -2751,6 +2751,21 @@ module ThriftHiveMetastore return end + def add_write_notification_log(rqst) + send_add_write_notification_log(rqst) + return recv_add_write_notification_log() + end + + def send_add_write_notification_log(rqst) + send_message('add_write_notification_log', Add_write_notification_log_args, :rqst => rqst) + end + + def recv_add_write_notification_log() + result = receive_message(Add_write_notification_log_result) + return result.success unless result.success.nil? + raise ::Thrift::ApplicationException.new(::Thrift::ApplicationException::MISSING_RESULT, 'add_write_notification_log failed: unknown result') + end + def cm_recycle(request) send_cm_recycle(request) return recv_cm_recycle() @@ -5520,6 +5535,13 @@ module ThriftHiveMetastore write_result(result, oprot, 'flushCache', seqid) end + def process_add_write_notification_log(seqid, iprot, oprot) + args = read_args(iprot, Add_write_notification_log_args) + result = Add_write_notification_log_result.new() + result.success = @handler.add_write_notification_log(args.rqst) + write_result(result, oprot, 'add_write_notification_log', seqid) + end + def process_cm_recycle(seqid, iprot, oprot) args = read_args(iprot, Cm_recycle_args) result = Cm_recycle_result.new() @@ -12220,6 +12242,38 @@ module ThriftHiveMetastore ::Thrift::Struct.generate_accessors self end + class Add_write_notification_log_args + include ::Thrift::Struct, ::Thrift::Struct_Union + RQST = -1 + + FIELDS = { + RQST => {:type => ::Thrift::Types::STRUCT, :name => 'rqst', :class => ::WriteNotificationLogRequest} + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self + end + + class Add_write_notification_log_result + include ::Thrift::Struct, ::Thrift::Struct_Union + SUCCESS = 0 + + FIELDS = { + SUCCESS => {:type => ::Thrift::Types::STRUCT, :name => 'success', :class => ::WriteNotificationLogResponse} + } + + def struct_fields; FIELDS; end + + def validate + end + + ::Thrift::Struct.generate_accessors self + end + class Cm_recycle_args include ::Thrift::Struct, ::Thrift::Struct_Union REQUEST = 1 diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java index 2ccaaa4389..6182dbd2e8 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java @@ -82,6 +82,7 @@ import org.apache.hadoop.hive.common.TableName; import org.apache.hadoop.hive.metastore.api.*; import org.apache.hadoop.hive.metastore.events.AddForeignKeyEvent; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.ConfVars; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.StatsUpdateMode; @@ -7184,6 +7185,55 @@ public void abort_txns(AbortTxnsRequest rqst) throws TException { @Override public void commit_txn(CommitTxnRequest rqst) throws TException { + // in replication flow, the write notification log table will be updated here. + if (rqst.isSetWriteEventInfos()) { + long targetTxnId = getTxnHandler().getTargetTxnId(rqst.getReplPolicy(), rqst.getTxnid()); + if (targetTxnId < 0) { + //looks like a retry + return; + } + for (WriteEventInfo writeEventInfo : rqst.getWriteEventInfos()) { + String[] filesAdded = ReplChangeManager.getListFromSeparatedString(writeEventInfo.getFiles()); + List partitionValue = null; + Partition ptnObj = null; + String root; + Table tbl = getTblObject(writeEventInfo.getDatabase(), writeEventInfo.getTable()); + + if (writeEventInfo.getPartition() != null && !writeEventInfo.getPartition().isEmpty()) { + partitionValue = Warehouse.getPartValuesFromPartName(writeEventInfo.getPartition()); + ptnObj = getPartitionObj(writeEventInfo.getDatabase(), writeEventInfo.getTable(), partitionValue, tbl); + root = ptnObj.getSd().getLocation(); + } else { + root = tbl.getSd().getLocation(); + } + + InsertEventRequestData insertData = new InsertEventRequestData(); + insertData.setReplace(true); + + // The files in the commit txn message during load will have files with path corresponding to source + // warehouse. Need to transform them to target warehouse using table or partition object location. + for (String file : filesAdded) { + String[] decodedPath = ReplChangeManager.decodeFileUri(file); + String name = (new Path(decodedPath[0])).getName(); + Path newPath = FileUtils.getTransformedPath(name, decodedPath[3], root); + insertData.addToFilesAdded(newPath.toUri().toString()); + insertData.addToSubDirectoryList(decodedPath[3]); + try { + insertData.addToFilesAddedChecksum(ReplChangeManager.checksumFor(newPath, newPath.getFileSystem(conf))); + } catch (IOException e) { + LOG.error("failed to get checksum for the file " + newPath + " with error: " + e.getMessage()); + throw new TException(e.getMessage()); + } + } + + WriteNotificationLogRequest wnRqst = new WriteNotificationLogRequest(targetTxnId, + writeEventInfo.getWriteId(), writeEventInfo.getDatabase(), writeEventInfo.getTable(), insertData); + if (partitionValue != null) { + wnRqst.setPartitionVals(partitionValue); + } + addTxnWriteNotificationLog(tbl, ptnObj, wnRqst); + } + } getTxnHandler().commitTxn(rqst); if (listeners != null && !listeners.isEmpty()) { MetaStoreListenerNotifier.notifyEvent(listeners, EventType.COMMIT_TXN, @@ -7213,6 +7263,42 @@ public AllocateTableWriteIdsResponse allocate_table_write_ids( return response; } + private void addTxnWriteNotificationLog(Table tableObj, Partition ptnObj, WriteNotificationLogRequest rqst) + throws MetaException { + String partition = ""; //Empty string is an invalid partition name. Can be used for non partitioned table. + if (ptnObj != null) { + partition = Warehouse.makePartName(tableObj.getPartitionKeys(), rqst.getPartitionVals()); + } + AcidWriteEvent event = new AcidWriteEvent(partition, tableObj, ptnObj, rqst); + getTxnHandler().addWriteNotificationLog(event); + if (listeners != null && !listeners.isEmpty()) { + MetaStoreListenerNotifier.notifyEvent(listeners, EventType.ACID_WRITE, event); + } + } + + private Table getTblObject(String db, String table) throws MetaException, NoSuchObjectException { + GetTableRequest req = new GetTableRequest(db, table); + req.setCapabilities(new ClientCapabilities(Lists.newArrayList(ClientCapability.TEST_CAPABILITY))); + return get_table_req(req).getTable(); + } + + private Partition getPartitionObj(String db, String table, List partitionVals, Table tableObj) + throws MetaException, NoSuchObjectException { + if (tableObj.isSetPartitionKeys() && !tableObj.getPartitionKeys().isEmpty()) { + return get_partition(db, table, partitionVals); + } + return null; + } + + @Override + public WriteNotificationLogResponse add_write_notification_log(WriteNotificationLogRequest rqst) + throws MetaException, NoSuchObjectException { + Table tableObj = getTblObject(rqst.getDb(), rqst.getTable()); + Partition ptnObj = getPartitionObj(rqst.getDb(), rqst.getTable(), rqst.getPartitionVals(), tableObj); + addTxnWriteNotificationLog(tableObj, ptnObj, rqst); + return new WriteNotificationLogResponse(); + } + @Override public LockResponse lock(LockRequest rqst) throws TException { return getTxnHandler().lock(rqst); diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java index 2a14dd4fae..57377779e8 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java @@ -2492,10 +2492,8 @@ public void commitTxn(long txnid) } @Override - public void replCommitTxn(long srcTxnId, String replPolicy) + public void replCommitTxn(CommitTxnRequest rqst) throws NoSuchTxnException, TxnAbortedException, TException { - CommitTxnRequest rqst = new CommitTxnRequest(srcTxnId); - rqst.setReplPolicy(replPolicy); client.commit_txn(rqst); } @@ -2742,6 +2740,12 @@ public FireEventResponse fireListenerEvent(FireEventRequest rqst) throws TExcept return client.fire_listener_event(rqst); } + @InterfaceAudience.LimitedPrivate({"Apache Hive, HCatalog"}) + @Override + public void addWriteNotificationLog(WriteNotificationLogRequest rqst) throws TException { + client.add_write_notification_log(rqst); + } + /** * Creates a synchronized wrapper for any {@link IMetaStoreClient}. * This may be used by multi-threaded applications until we have diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java index 234e0cf214..9661bebc13 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java @@ -40,6 +40,7 @@ import org.apache.hadoop.hive.metastore.api.CmRecycleResponse; import org.apache.hadoop.hive.metastore.api.ColumnStatistics; import org.apache.hadoop.hive.metastore.api.ColumnStatisticsObj; +import org.apache.hadoop.hive.metastore.api.CommitTxnRequest; import org.apache.hadoop.hive.metastore.api.CompactionResponse; import org.apache.hadoop.hive.metastore.api.CompactionType; import org.apache.hadoop.hive.metastore.api.ConfigValSecurityException; @@ -125,6 +126,7 @@ import org.apache.hadoop.hive.metastore.api.WMResourcePlan; import org.apache.hadoop.hive.metastore.api.WMTrigger; import org.apache.hadoop.hive.metastore.api.WMValidateResourcePlanResponse; +import org.apache.hadoop.hive.metastore.api.WriteNotificationLogRequest; import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy; import org.apache.hadoop.hive.metastore.utils.ObjectPair; import org.apache.thrift.TException; @@ -2871,8 +2873,8 @@ void commitTxn(long txnid) /** * Commit a transaction. This will also unlock any locks associated with * this transaction. - * @param srcTxnid id of transaction at source which is committed and to be replicated. - * @param replPolicy the replication policy to identify the source cluster + * @param rqst Information containing the txn info and write event information + * of transaction at source which is committed and to be replicated * @throws NoSuchTxnException if the requested transaction does not exist. * This can result fro the transaction having timed out and been deleted by * the compactor. @@ -2880,7 +2882,7 @@ void commitTxn(long txnid) * aborted. This can result from the transaction timing out. * @throws TException */ - void replCommitTxn(long srcTxnid, String replPolicy) + void replCommitTxn(CommitTxnRequest rqst) throws NoSuchTxnException, TxnAbortedException, TException; /** @@ -3193,6 +3195,14 @@ NotificationEventsCountResponse getNotificationEventsCount(NotificationEventsCou @InterfaceAudience.LimitedPrivate({"Apache Hive, HCatalog"}) FireEventResponse fireListenerEvent(FireEventRequest request) throws TException; + /** + * Add a event related to write operations in an ACID table. + * @param rqst message containing information for acid write operation. + * @throws TException + */ + @InterfaceAudience.LimitedPrivate({"Apache Hive, HCatalog"}) + void addWriteNotificationLog(WriteNotificationLogRequest rqst) throws TException; + class IncompatibleMetastoreException extends MetaException { IncompatibleMetastoreException(String message) { super(message); diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java index e0e65cf231..de226bf327 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreEventListener.java @@ -55,6 +55,7 @@ import org.apache.hadoop.hive.metastore.events.CommitTxnEvent; import org.apache.hadoop.hive.metastore.events.AbortTxnEvent; import org.apache.hadoop.hive.metastore.events.AllocWriteIdEvent; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.tools.SQLGenerator; import java.sql.Connection; @@ -282,6 +283,17 @@ public void onAllocWriteId(AllocWriteIdEvent allocWriteIdEvent, Connection dbCon throws MetaException { } + /** + * This will be called to perform acid write operation. + * @param acidWriteEvent event to be processed + * @param dbConn jdbc connection to remote meta store db. + * @param sqlGenerator helper class to generate db specific sql string. + * @throws MetaException + */ + public void onAcidWrite(AcidWriteEvent acidWriteEvent, Connection dbConn, SQLGenerator sqlGenerator) + throws MetaException { + } + @Override public Configuration getConf() { return this.conf; diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreListenerNotifier.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreListenerNotifier.java index 3cf83149a8..c296f57559 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreListenerNotifier.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/MetaStoreListenerNotifier.java @@ -54,6 +54,7 @@ import org.apache.hadoop.hive.metastore.events.CommitTxnEvent; import org.apache.hadoop.hive.metastore.events.AbortTxnEvent; import org.apache.hadoop.hive.metastore.events.AllocWriteIdEvent; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.tools.SQLGenerator; import java.sql.Connection; import java.util.List; @@ -221,6 +222,8 @@ public void notify(MetaStoreEventListener listener, ListenerEvent event) throws (listener, event) -> listener.onAbortTxn((AbortTxnEvent) event, null, null)) .put(EventType.ALLOC_WRITE_ID, (listener, event) -> listener.onAllocWriteId((AllocWriteIdEvent) event, null, null)) + .put(EventType.ACID_WRITE, + (listener, event) -> listener.onAcidWrite((AcidWriteEvent) event, null, null)) .build() ); @@ -241,6 +244,9 @@ void notify(MetaStoreEventListener listener, ListenerEvent event, Connection dbC .put(EventType.ALLOC_WRITE_ID, (listener, event, dbConn, sqlGenerator) -> listener.onAllocWriteId((AllocWriteIdEvent) event, dbConn, sqlGenerator)) + .put(EventType.ACID_WRITE, + (listener, event, dbConn, sqlGenerator) -> + listener.onAcidWrite((AcidWriteEvent) event, dbConn, sqlGenerator)) .build() ); diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java index b8eb173e9c..7a6e2bdbf2 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java @@ -159,6 +159,7 @@ import org.apache.hadoop.hive.metastore.api.WMResourcePlanStatus; import org.apache.hadoop.hive.metastore.api.WMTrigger; import org.apache.hadoop.hive.metastore.api.WMValidateResourcePlanResponse; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.ConfVars; import org.apache.hadoop.hive.metastore.datasource.DataSourceProvider; @@ -206,6 +207,7 @@ import org.apache.hadoop.hive.metastore.model.MWMResourcePlan; import org.apache.hadoop.hive.metastore.model.MWMResourcePlan.Status; import org.apache.hadoop.hive.metastore.model.MWMTrigger; +import org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog; import org.apache.hadoop.hive.metastore.parser.ExpressionTree; import org.apache.hadoop.hive.metastore.parser.ExpressionTree.FilterBuilder; import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy; @@ -9460,6 +9462,64 @@ public NotificationEventResponse getNextNotification(NotificationEventRequest rq } } + @Override + public void cleanWriteNotificationEvents(int olderThan) { + boolean commited = false; + Query query = null; + try { + openTransaction(); + long tmp = System.currentTimeMillis() / 1000 - olderThan; + int tooOld = (tmp > Integer.MAX_VALUE) ? 0 : (int) tmp; + query = pm.newQuery(MTxnWriteNotificationLog.class, "eventTime < tooOld"); + query.declareParameters("java.lang.Integer tooOld"); + Collection toBeRemoved = (Collection) query.execute(tooOld); + if (CollectionUtils.isNotEmpty(toBeRemoved)) { + pm.deletePersistentAll(toBeRemoved); + } + commited = commitTransaction(); + } finally { + rollbackAndCleanup(commited, query); + } + } + + @Override + public List getAllWriteEventInfo(long txnId, String dbName, String tableName) throws MetaException { + List writeEventInfoList = null; + boolean commited = false; + Query query = null; + try { + openTransaction(); + List parameterVals = new ArrayList<>(); + StringBuilder filterBuilder = new StringBuilder(" txnId == " + Long.toString(txnId)); + if (dbName != null && !"*".equals(dbName)) { // * means get all database, so no need to add filter + appendSimpleCondition(filterBuilder, "database", new String[]{dbName}, parameterVals); + } + if (tableName != null && !"*".equals(tableName)) { + appendSimpleCondition(filterBuilder, "table", new String[]{tableName}, parameterVals); + } + query = pm.newQuery(MTxnWriteNotificationLog.class, filterBuilder.toString()); + query.setOrdering("database,table ascending"); + List mplans = (List)query.executeWithArray( + parameterVals.toArray(new String[parameterVals.size()])); + pm.retrieveAll(mplans); + commited = commitTransaction(); + if (mplans != null && mplans.size() > 0) { + writeEventInfoList = Lists.newArrayList(); + for (MTxnWriteNotificationLog mplan : mplans) { + WriteEventInfo writeEventInfo = new WriteEventInfo(mplan.getWriteId(), mplan.getDatabase(), + mplan.getTable(), mplan.getFiles()); + writeEventInfo.setPartition(mplan.getPartition()); + writeEventInfo.setPartitionObj(mplan.getPartObject()); + writeEventInfo.setTableObj(mplan.getTableObject()); + writeEventInfoList.add(writeEventInfo); + } + } + } finally { + rollbackAndCleanup(commited, query); + } + return writeEventInfoList; + } + private void prepareQuotes() throws SQLException { if (dbType == DatabaseProduct.MYSQL) { assert pm.currentTransaction().isActive(); diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java index c429048c54..f1573f00a5 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/RawStore.java @@ -23,6 +23,7 @@ import org.apache.hadoop.hive.metastore.api.ISchemaName; import org.apache.hadoop.hive.metastore.api.SchemaVersionDescriptor; import org.apache.hadoop.hive.metastore.api.WMFullResourcePlan; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import java.lang.annotation.ElementType; import java.lang.annotation.Retention; @@ -1648,4 +1649,17 @@ void alterSchemaVersion(SchemaVersionDescriptor version, SchemaVersion newVersio Map> getPartitionColsWithStats(String catName, String dbName, String tableName) throws MetaException, NoSuchObjectException; + /** + * Remove older notification events. + * @param olderThan Remove any events older than a given number of seconds + */ + void cleanWriteNotificationEvents(int olderThan); + + /** + * Get all write events for a specific transaction . + * @param txnId get all the events done by this transaction + * @param dbName the name of db for which dump is being taken + * @param tableName the name of the table for which the dump is being taken + */ + List getAllWriteEventInfo(long txnId, String dbName, String tableName) throws MetaException; } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java index f7018c2d0b..ac1d3c87bc 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java @@ -59,6 +59,7 @@ static final String REMAIN_IN_TRASH_TAG = "user.remain-in-trash"; private static final String URI_FRAGMENT_SEPARATOR = "#"; public static final String SOURCE_OF_REPLICATION = "repl.source.for"; + private static final String TXN_WRITE_EVENT_FILE_SEPARATOR = "]"; public enum RecycleType { MOVE, @@ -472,7 +473,6 @@ static void scheduleCMClearer(Configuration conf) { } public static boolean isSourceOfReplication(Database db) { - // Can not judge, so assuming replication is not enabled. assert (db != null); String replPolicyIds = getReplPolicyIdString(db); return !StringUtils.isEmpty(replPolicyIds); @@ -490,4 +490,12 @@ public static String getReplPolicyIdString(Database db) { } return null; } + + public static String joinWithSeparator(Iterable strings) { + return org.apache.hadoop.util.StringUtils.join(TXN_WRITE_EVENT_FILE_SEPARATOR, strings); + } + + public static String[] getListFromSeparatedString(String commaSeparatedString) { + return commaSeparatedString.split("\\s*" + TXN_WRITE_EVENT_FILE_SEPARATOR + "\\s*"); + } } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java index f6286eaa5c..54c833deaa 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/cache/CachedStore.java @@ -111,6 +111,7 @@ import org.apache.hadoop.hive.metastore.api.WMFullResourcePlan; import org.apache.hadoop.hive.metastore.api.WMMapping; import org.apache.hadoop.hive.metastore.api.WMPool; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.ConfVars; import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy; @@ -2407,6 +2408,17 @@ public long getCacheUpdateCount() { return sharedCache.getUpdateCount(); } + @Override + public void cleanWriteNotificationEvents(int olderThan) { + rawStore.cleanWriteNotificationEvents(olderThan); + } + + + @Override + public List getAllWriteEventInfo(long txnId, String dbName, String tableName) throws MetaException { + return rawStore.getAllWriteEventInfo(txnId, dbName, tableName); + } + static boolean isNotInBlackList(String catName, String dbName, String tblName) { String str = TableName.getQualified(catName, dbName, tblName); for (Pattern pattern : blacklistPatterns) { diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/events/AcidWriteEvent.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/events/AcidWriteEvent.java new file mode 100644 index 0000000000..001179a3f8 --- /dev/null +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/events/AcidWriteEvent.java @@ -0,0 +1,91 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.hive.metastore.events; + +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; +import org.apache.hadoop.hive.metastore.api.Partition; +import org.apache.hadoop.hive.metastore.api.Table; +import org.apache.hadoop.hive.metastore.api.WriteNotificationLogRequest; +import org.apache.hadoop.hive.metastore.utils.StringUtils; + +import java.util.List; + +/** + * AcidWriteEvent + * Event generated for acid write operations + */ +@InterfaceAudience.Public +@InterfaceStability.Stable +public class AcidWriteEvent extends ListenerEvent { + private final WriteNotificationLogRequest writeNotificationLogRequest; + private final String partition; + private final Table tableObj; + private final Partition partitionObj; + + public AcidWriteEvent(String partition, Table tableObj, Partition partitionObj, + WriteNotificationLogRequest writeNotificationLogRequest) { + super(true, null); + this.writeNotificationLogRequest = writeNotificationLogRequest; + this.partition = partition; + this.tableObj = tableObj; + this.partitionObj = partitionObj; + } + + public Long getTxnId() { + return writeNotificationLogRequest.getTxnId(); + } + + public List getFiles() { + return writeNotificationLogRequest.getFileInfo().getFilesAdded(); + } + + public List getChecksums() { + return writeNotificationLogRequest.getFileInfo().getFilesAddedChecksum(); + } + + public String getDatabase() { + return StringUtils.normalizeIdentifier(writeNotificationLogRequest.getDb()); + } + + public String getTable() { + return StringUtils.normalizeIdentifier(writeNotificationLogRequest.getTable()); + } + + public String getPartition() { + return partition; //Don't normalize partition value, as its case sensitive. + } + + public Long getWriteId() { + return writeNotificationLogRequest.getWriteId(); + } + + public Table getTableObj() { + return tableObj; + } + + public Partition getPartitionObj() { + return partitionObj; + } + + public List getSubDirs() { + return writeNotificationLogRequest.getFileInfo().getSubDirectoryList(); + } +} + diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/AcidWriteMessage.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/AcidWriteMessage.java new file mode 100644 index 0000000000..e2c9ccf53d --- /dev/null +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/AcidWriteMessage.java @@ -0,0 +1,50 @@ +/* * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.hive.metastore.messaging; + +import org.apache.hadoop.hive.metastore.api.Partition; +import org.apache.hadoop.hive.metastore.api.Table; +import java.util.List; + +/** + * HCat message sent when an ACID write is done. + */ +public abstract class AcidWriteMessage extends EventMessage { + + protected AcidWriteMessage() { + super(EventType.ACID_WRITE); + } + + public abstract Long getTxnId(); + + public abstract String getTable(); + + public abstract Long getWriteId(); + + public abstract String getPartition(); + + public abstract List getFiles(); + + public abstract Table getTableObj() throws Exception; + + public abstract Partition getPartitionObj() throws Exception; + + public abstract String getTableObjStr(); + + public abstract String getPartitionObjStr(); +} diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/CommitTxnMessage.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/CommitTxnMessage.java index 49004f2260..9733039f06 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/CommitTxnMessage.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/CommitTxnMessage.java @@ -17,6 +17,12 @@ package org.apache.hadoop.hive.metastore.messaging; +import org.apache.hadoop.hive.metastore.api.Partition; +import org.apache.hadoop.hive.metastore.api.Table; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; + +import java.util.List; + /** * HCat message sent when an commit transaction is done. */ @@ -33,4 +39,21 @@ protected CommitTxnMessage() { */ public abstract Long getTxnId(); + public abstract List getWriteIds(); + + public abstract List getDatabases(); + + public abstract List getTables(); + + public abstract List getPartitions(); + + public abstract Table getTableObj(int idx) throws Exception; + + public abstract Partition getPartitionObj(int idx) throws Exception; + + public abstract String getFiles(int idx); + + public abstract List getFilesList(); + + public abstract void addWriteEventInfo(List writeEventInfoList); } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/EventMessage.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/EventMessage.java index 969dd7bce5..f24b419445 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/EventMessage.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/EventMessage.java @@ -60,7 +60,8 @@ COMMIT_TXN(MessageFactory.COMMIT_TXN_EVENT), ABORT_TXN(MessageFactory.ABORT_TXN_EVENT), ALLOC_WRITE_ID(MessageFactory.ALLOC_WRITE_ID_EVENT), - ALTER_CATALOG(MessageFactory.ALTER_CATALOG_EVENT); + ALTER_CATALOG(MessageFactory.ALTER_CATALOG_EVENT), + ACID_WRITE(MessageFactory.ACID_WRITE_EVENT); private String typeString; diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageDeserializer.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageDeserializer.java index ca335790ce..b701d84ac4 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageDeserializer.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageDeserializer.java @@ -70,6 +70,10 @@ public EventMessage getEventMessage(String eventTypeString, String messageBody) return getCommitTxnMessage(messageBody); case ABORT_TXN: return getAbortTxnMessage(messageBody); + case ALLOC_WRITE_ID: + return getAllocWriteIdMessage(messageBody); + case ACID_WRITE: + return getAcidWriteMessage(messageBody); default: throw new IllegalArgumentException("Unsupported event-type: " + eventTypeString); } @@ -186,6 +190,11 @@ public EventMessage getEventMessage(String eventTypeString, String messageBody) */ public abstract AllocWriteIdMessage getAllocWriteIdMessage(String messageBody); + /* + * Method to de-serialize AcidWriteMessage instance. + */ + public abstract AcidWriteMessage getAcidWriteMessage(String messageBody); + // Protection against construction. protected MessageDeserializer() {} } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageFactory.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageFactory.java index e0629ea4ab..d529147f1c 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageFactory.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/MessageFactory.java @@ -33,6 +33,7 @@ import org.apache.hadoop.hive.metastore.api.TxnToWriteId; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.conf.MetastoreConf.ConfVars; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.utils.JavaUtils; import java.util.Iterator; @@ -74,6 +75,7 @@ public static final String ABORT_TXN_EVENT = "ABORT_TXN"; public static final String ALLOC_WRITE_ID_EVENT = "ALLOC_WRITE_ID_EVENT"; public static final String ALTER_CATALOG_EVENT = "ALTER_CATALOG"; + public static final String ACID_WRITE_EVENT = "ACID_WRITE_EVENT"; private static MessageFactory instance = null; @@ -326,4 +328,14 @@ public abstract DropConstraintMessage buildDropConstraintMessage(String dbName, public abstract DropCatalogMessage buildDropCatalogMessage(Catalog catalog); public abstract AlterCatalogMessage buildAlterCatalogMessage(Catalog oldCat, Catalog newCat); + + /** + * Factory method for building acid write message + * + * + * @param acidWriteEvent information related to the acid write operation + * @param files files added by this write operation + * @return instance of AcidWriteMessage + */ + public abstract AcidWriteMessage buildAcidWriteMessage(AcidWriteEvent acidWriteEvent, Iterator files); } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONAcidWriteMessage.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONAcidWriteMessage.java new file mode 100644 index 0000000000..515a2cb6d6 --- /dev/null +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONAcidWriteMessage.java @@ -0,0 +1,150 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.hadoop.hive.metastore.messaging.json; + +import com.google.common.collect.Lists; +import org.apache.hadoop.hive.metastore.api.Partition; +import org.apache.hadoop.hive.metastore.api.Table; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; +import org.apache.hadoop.hive.metastore.messaging.AcidWriteMessage; +import org.apache.thrift.TException; +import org.codehaus.jackson.annotate.JsonProperty; +import java.util.Iterator; +import java.util.List; + +/** + * JSON implementation of AcidWriteMessage + */ +public class JSONAcidWriteMessage extends AcidWriteMessage { + + @JsonProperty + private Long txnid, writeId, timestamp; + + @JsonProperty + private String server, servicePrincipal, database, table, partition, tableObjJson, partitionObjJson; + + @JsonProperty + private List files; + + /** + * Default constructor, needed for Jackson. + */ + public JSONAcidWriteMessage() { + } + + public JSONAcidWriteMessage(String server, String servicePrincipal, Long timestamp, AcidWriteEvent acidWriteEvent, + Iterator files) { + this.timestamp = timestamp; + this.txnid = acidWriteEvent.getTxnId(); + this.server = server; + this.servicePrincipal = servicePrincipal; + this.database = acidWriteEvent.getDatabase(); + this.table = acidWriteEvent.getTable(); + this.writeId = acidWriteEvent.getWriteId(); + this.partition = acidWriteEvent.getPartition(); + try { + this.tableObjJson = JSONMessageFactory.createTableObjJson(acidWriteEvent.getTableObj()); + if (acidWriteEvent.getPartitionObj() != null) { + this.partitionObjJson = JSONMessageFactory.createPartitionObjJson(acidWriteEvent.getPartitionObj()); + } else { + this.partitionObjJson = null; + } + } catch (TException e) { + throw new IllegalArgumentException("Could not serialize JSONAcidWriteMessage : ", e); + } + this.files = Lists.newArrayList(files); + } + + @Override + public Long getTxnId() { + return txnid; + } + + @Override + public Long getTimestamp() { + return timestamp; + } + + @Override + public String getDB() { + return database; + } + + @Override + public String getServicePrincipal() { + return servicePrincipal; + } + + @Override + public String getServer() { + return server; + } + + @Override + public String getTable() { + return table; + } + + @Override + public Long getWriteId() { + return writeId; + } + + @Override + public String getPartition() { + return partition; + } + + @Override + public List getFiles() { + return files; + } + + @Override + public Table getTableObj() throws Exception { + return (tableObjJson == null) ? null : (Table) JSONMessageFactory.getTObj(tableObjJson, Table.class); + } + + @Override + public Partition getPartitionObj() throws Exception { + return ((partitionObjJson == null) ? null : + (Partition) JSONMessageFactory.getTObj(partitionObjJson, Partition.class)); + } + + @Override + public String getTableObjStr() { + return tableObjJson; + } + + @Override + public String getPartitionObjStr() { + return partitionObjJson; + } + + @Override + public String toString() { + try { + return JSONMessageDeserializer.mapper.writeValueAsString(this); + } catch (Exception exception) { + throw new IllegalArgumentException("Could not serialize: ", exception); + } + } +} + diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONCommitTxnMessage.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONCommitTxnMessage.java index 595a3d1b57..6082b8eb42 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONCommitTxnMessage.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONCommitTxnMessage.java @@ -18,9 +18,15 @@ */ package org.apache.hadoop.hive.metastore.messaging.json; +import com.google.common.collect.Lists; +import org.apache.hadoop.hive.metastore.api.Partition; +import org.apache.hadoop.hive.metastore.api.Table; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.messaging.CommitTxnMessage; import org.codehaus.jackson.annotate.JsonProperty; +import java.util.List; + /** * JSON implementation of CommitTxnMessage */ @@ -38,6 +44,12 @@ @JsonProperty private String servicePrincipal; + @JsonProperty + private List writeIds; + + @JsonProperty + private List databases, tables, partitions, tableObjs, partitionObjs, files; + /** * Default constructor, needed for Jackson. */ @@ -49,6 +61,13 @@ public JSONCommitTxnMessage(String server, String servicePrincipal, Long txnid, this.txnid = txnid; this.server = server; this.servicePrincipal = servicePrincipal; + this.databases = null; + this.tables = null; + this.writeIds = null; + this.partitions = null; + this.tableObjs = null; + this.partitionObjs = null; + this.files = null; } @Override @@ -76,6 +95,82 @@ public String getServer() { return server; } + @Override + public List getWriteIds() { + return writeIds; + } + + @Override + public List getDatabases() { + return databases; + } + + @Override + public List getTables() { + return tables; + } + + @Override + public List getPartitions() { + return partitions; + } + + @Override + public Table getTableObj(int idx) throws Exception { + return tableObjs == null ? null : (Table) JSONMessageFactory.getTObj(tableObjs.get(idx), Table.class); + } + + @Override + public Partition getPartitionObj(int idx) throws Exception { + return (partitionObjs == null ? null : (partitionObjs.get(idx) == null ? null : + (Partition)JSONMessageFactory.getTObj(partitionObjs.get(idx), Partition.class))); + } + + @Override + public String getFiles(int idx) { + return files == null ? null : files.get(idx); + } + + @Override + public List getFilesList() { + return files; + } + + @Override + public void addWriteEventInfo(List writeEventInfoList) { + if (this.databases == null) { + this.databases = Lists.newArrayList(); + } + if (this.tables == null) { + this.tables = Lists.newArrayList(); + } + if (this.writeIds == null) { + this.writeIds = Lists.newArrayList(); + } + if (this.tableObjs == null) { + this.tableObjs = Lists.newArrayList(); + } + if (this.partitions == null) { + this.partitions = Lists.newArrayList(); + } + if (this.partitionObjs == null) { + this.partitionObjs = Lists.newArrayList(); + } + if (this.files == null) { + this.files = Lists.newArrayList(); + } + + for (WriteEventInfo writeEventInfo : writeEventInfoList) { + this.databases.add(writeEventInfo.getDatabase()); + this.tables.add(writeEventInfo.getTable()); + this.writeIds.add(writeEventInfo.getWriteId()); + this.partitions.add(writeEventInfo.getPartition()); + this.tableObjs.add(writeEventInfo.getTableObj()); + this.partitionObjs.add(writeEventInfo.getPartitionObj()); + this.files.add(writeEventInfo.getFiles()); + } + } + @Override public String toString() { try { diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageDeserializer.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageDeserializer.java index f54e24d41f..be6b7513cd 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageDeserializer.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageDeserializer.java @@ -41,6 +41,7 @@ import org.apache.hadoop.hive.metastore.messaging.CommitTxnMessage; import org.apache.hadoop.hive.metastore.messaging.AbortTxnMessage; import org.apache.hadoop.hive.metastore.messaging.AllocWriteIdMessage; +import org.apache.hadoop.hive.metastore.messaging.AcidWriteMessage; import org.codehaus.jackson.map.DeserializationConfig; import org.codehaus.jackson.map.ObjectMapper; import org.codehaus.jackson.map.SerializationConfig; @@ -259,4 +260,12 @@ public AllocWriteIdMessage getAllocWriteIdMessage(String messageBody) { throw new IllegalArgumentException("Could not construct AllocWriteIdMessage", e); } } + + public AcidWriteMessage getAcidWriteMessage(String messageBody) { + try { + return mapper.readValue(messageBody, JSONAcidWriteMessage.class); + } catch (Exception e) { + throw new IllegalArgumentException("Could not construct AcidWriteMessage", e); + } + } } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageFactory.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageFactory.java index d64c3ffb80..07f51f0a8a 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageFactory.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/messaging/json/JSONMessageFactory.java @@ -39,6 +39,7 @@ import org.apache.hadoop.hive.metastore.api.SQLUniqueConstraint; import org.apache.hadoop.hive.metastore.api.Table; import org.apache.hadoop.hive.metastore.api.TxnToWriteId; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.messaging.AddForeignKeyMessage; import org.apache.hadoop.hive.metastore.messaging.AddNotNullConstraintMessage; import org.apache.hadoop.hive.metastore.messaging.AddPartitionMessage; @@ -66,6 +67,7 @@ import org.apache.hadoop.hive.metastore.messaging.CommitTxnMessage; import org.apache.hadoop.hive.metastore.messaging.AbortTxnMessage; import org.apache.hadoop.hive.metastore.messaging.AllocWriteIdMessage; +import org.apache.hadoop.hive.metastore.messaging.AcidWriteMessage; import org.apache.thrift.TBase; import org.apache.thrift.TDeserializer; import org.apache.thrift.TException; @@ -230,11 +232,17 @@ public AbortTxnMessage buildAbortTxnMessage(Long txnId) { return new JSONAbortTxnMessage(MS_SERVER_URL, MS_SERVICE_PRINCIPAL, txnId, now()); } + @Override public AllocWriteIdMessage buildAllocWriteIdMessage(List txnToWriteIdList, String dbName, String tableName) { return new JSONAllocWriteIdMessage(MS_SERVER_URL, MS_SERVICE_PRINCIPAL, txnToWriteIdList, dbName, tableName, now()); } + @Override + public AcidWriteMessage buildAcidWriteMessage(AcidWriteEvent acidWriteEvent, Iterator files) { + return new JSONAcidWriteMessage(MS_SERVER_URL, MS_SERVICE_PRINCIPAL, now(), acidWriteEvent, files); + } + private long now() { return System.currentTimeMillis() / 1000; } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTxnWriteNotificationLog.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTxnWriteNotificationLog.java new file mode 100644 index 0000000000..f5ca3862f6 --- /dev/null +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/model/MTxnWriteNotificationLog.java @@ -0,0 +1,123 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hive.metastore.model; + +/** + * MTxnWriteNotificationLog + * DN table for ACID write events. + */ +public class MTxnWriteNotificationLog { + private long txnId; + private long writeId; + private int eventTime; + private String database; + private String table; + private String partition; + private String tableObject; + private String partObject; + private String files; + + public MTxnWriteNotificationLog() { + } + + public MTxnWriteNotificationLog(long txnId, long writeId, int eventTime, String database, String table, + String partition, String tableObject, String partObject, String files) { + this.txnId = txnId; + this.writeId = writeId; + this.eventTime = eventTime; + this.database = database; + this.table = table; + this.partition = partition; + this.tableObject = tableObject; + this.partObject = partObject; + this.files = files; + } + + public long getTxnId() { + return txnId; + } + + public void setTxnId(long txnId) { + this.txnId = txnId; + } + + public long getWriteId() { + return writeId; + } + + public void setWriteId(long writeId) { + this.writeId = writeId; + } + + public int getEventTime() { + return eventTime; + } + + public void setEventTime(int eventTime) { + this.eventTime = eventTime; + } + + public String getDatabase() { + return database; + } + + public void setDatabase(String database) { + this.database = database; + } + + public String getTable() { + return table; + } + + public void setTable(String table) { + this.table = table; + } + + public String getPartition() { + return partition; + } + + public void setPartition(String partition) { + this.partition = partition; + } + + public String getTableObject() { + return tableObject; + } + + public void setTableObject(String tableObject) { + this.tableObject = tableObject; + } + + public String getPartObject() { + return partObject; + } + + public void setPartObject(String partObject) { + this.partObject = partObject; + } + + public String getFiles() { + return files; + } + + public void setFiles(String files) { + this.files = files; + } +} + diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/tools/SQLGenerator.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/tools/SQLGenerator.java index b23a6d7709..d0ac7dbe67 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/tools/SQLGenerator.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/tools/SQLGenerator.java @@ -175,4 +175,13 @@ public DatabaseProduct getDbProduct() { return dbProduct; } + // This is required for SQL executed directly. If the SQL has double quotes then some dbs tend to + // remove the escape characters and store the variable without double quote. + public String addEscapeCharacters(String s) { + if (dbProduct == DatabaseProduct.MYSQL) { + return s.replaceAll("\\\\", "\\\\\\\\"); + } + return s; + } + } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnDbUtil.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnDbUtil.java index 8764c21bd8..38cef62c7e 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnDbUtil.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnDbUtil.java @@ -253,6 +253,34 @@ public static void prepDb(Configuration conf) throws Exception { stmt.execute("INSERT INTO \"APP\".\"NOTIFICATION_SEQUENCE\" (\"NNI_ID\", \"NEXT_EVENT_ID\")" + " SELECT * FROM (VALUES (1,1)) tmp_table WHERE NOT EXISTS ( SELECT " + "\"NEXT_EVENT_ID\" FROM \"APP\".\"NOTIFICATION_SEQUENCE\")"); + + try { + stmt.execute("CREATE TABLE TXN_WRITE_NOTIFICATION_LOG (" + + "WNL_ID bigint NOT NULL," + + "WNL_TXNID bigint NOT NULL," + + "WNL_WRITEID bigint NOT NULL," + + "WNL_DATABASE varchar(128) NOT NULL," + + "WNL_TABLE varchar(128) NOT NULL," + + "WNL_PARTITION varchar(1024) NOT NULL," + + "WNL_TABLE_OBJ clob NOT NULL," + + "WNL_PARTITION_OBJ clob," + + "WNL_FILES clob," + + "WNL_EVENT_TIME integer NOT NULL," + + "PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION))" + ); + } catch (SQLException e) { + if (e.getMessage() != null && e.getMessage().contains("already exists")) { + LOG.info("TXN_WRITE_NOTIFICATION_LOG table already exist, ignoring"); + } else { + throw e; + } + } + + stmt.execute("INSERT INTO \"APP\".\"SEQUENCE_TABLE\" (\"SEQUENCE_NAME\", \"NEXT_VAL\") " + + "SELECT * FROM (VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', " + + "1)) tmp_table WHERE NOT EXISTS ( SELECT \"NEXT_VAL\" FROM \"APP\"" + + ".\"SEQUENCE_TABLE\" WHERE \"SEQUENCE_NAME\" = 'org.apache.hadoop.hive.metastore" + + ".model.MTxnWriteNotificationLog')"); } catch (SQLException e) { try { conn.rollback(); diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java index b2a22f193a..4f56eba201 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java @@ -80,6 +80,7 @@ import org.apache.hadoop.hive.metastore.events.AllocWriteIdEvent; import org.apache.hadoop.hive.metastore.events.CommitTxnEvent; import org.apache.hadoop.hive.metastore.events.OpenTxnEvent; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import org.apache.hadoop.hive.metastore.messaging.EventMessage; import org.apache.hadoop.hive.metastore.metrics.Metrics; import org.apache.hadoop.hive.metastore.metrics.MetricsConstants; @@ -700,6 +701,38 @@ public OpenTxnsResponse openTxns(OpenTxnRequest rqst) throws MetaException { } } + @Override + @RetrySemantics.Idempotent + public long getTargetTxnId(String replPolicy, long sourceTxnId) throws MetaException { + try { + Connection dbConn = null; + Statement stmt = null; + try { + lockInternal(); + dbConn = getDbConn(Connection.TRANSACTION_READ_COMMITTED); + stmt = dbConn.createStatement(); + List targetTxnIds = getTargetTxnIdList(replPolicy, Collections.singletonList(sourceTxnId), stmt); + if (targetTxnIds.isEmpty()) { + LOG.info("Txn {} not present for repl policy {}", sourceTxnId, replPolicy); + return -1; + } + assert (targetTxnIds.size() == 1); + return targetTxnIds.get(0); + } catch (SQLException e) { + LOG.debug("Going to rollback"); + rollbackDBConn(dbConn); + checkRetryable(dbConn, e, "getTargetTxnId(" + replPolicy + sourceTxnId + ")"); + throw new MetaException("Unable to get target transaction id " + + StringUtils.stringifyException(e)); + } finally { + close(null, stmt, dbConn); + unlockInternal(); + } + } catch (RetryException e) { + return getTargetTxnId(replPolicy, sourceTxnId); + } + } + @Override @RetrySemantics.Idempotent public void abortTxn(AbortTxnRequest rqst) throws NoSuchTxnException, MetaException, TxnAbortedException { @@ -893,10 +926,18 @@ public void commitTxn(CommitTxnRequest rqst) shouldNeverHappen(txnid); //dbConn is rolled back in finally{} } - String conflictSQLSuffix = "from TXN_COMPONENTS where tc_txnid=" + txnid + " and tc_operation_type IN(" + - quoteChar(OpertaionType.UPDATE.sqlConst) + "," + quoteChar(OpertaionType.DELETE.sqlConst) + ")"; - rs = stmt.executeQuery(sqlGenerator.addLimitClause(1, "tc_operation_type " + conflictSQLSuffix)); - if (rs.next()) { + + String conflictSQLSuffix = null; + if (rqst.isSetReplPolicy()) { + rs = null; + } else { + conflictSQLSuffix = "from TXN_COMPONENTS where tc_txnid=" + txnid + " and tc_operation_type IN(" + + quoteChar(OpertaionType.UPDATE.sqlConst) + "," + quoteChar(OpertaionType.DELETE.sqlConst) + ")"; + rs = stmt.executeQuery(sqlGenerator.addLimitClause(1, + "tc_operation_type " + conflictSQLSuffix)); + } + + if (rs != null && rs.next()) { isUpdateDelete = 'Y'; close(rs); //if here it means currently committing txn performed update/delete and we should check WW conflict @@ -985,45 +1026,66 @@ public void commitTxn(CommitTxnRequest rqst) * Consider: if RO txn is after a W txn, then RO's openTxns() will be mutexed with W's * commitTxn() because both do S4U on NEXT_TXN_ID and thus RO will see result of W txn. * If RO < W, then there is no reads-from relationship. + * In replication flow we don't expect any write write conflict as it should have been handled at source. */ } - // Move the record from txn_components into completed_txn_components so that the compactor - // knows where to look to compact. - String s = "insert into COMPLETED_TXN_COMPONENTS (ctc_txnid, ctc_database, " + - "ctc_table, ctc_partition, ctc_writeid, ctc_update_delete) select tc_txnid, tc_database, tc_table, " + - "tc_partition, tc_writeid, '" + isUpdateDelete + "' from TXN_COMPONENTS where tc_txnid = " + txnid; - LOG.debug("Going to execute insert <" + s + ">"); - int modCount = 0; - if ((modCount = stmt.executeUpdate(s)) < 1) { - //this can be reasonable for an empty txn START/COMMIT or read-only txn - //also an IUD with DP that didn't match any rows. - LOG.info("Expected to move at least one record from txn_components to " + - "completed_txn_components when committing txn! " + JavaUtils.txnIdToString(txnid)); + + String s; + if (!rqst.isSetReplPolicy()) { + // Move the record from txn_components into completed_txn_components so that the compactor + // knows where to look to compact. + s = "insert into COMPLETED_TXN_COMPONENTS (ctc_txnid, ctc_database, " + + "ctc_table, ctc_partition, ctc_writeid, ctc_update_delete) select tc_txnid, tc_database, tc_table, " + + "tc_partition, tc_writeid, '" + isUpdateDelete + "' from TXN_COMPONENTS where tc_txnid = " + txnid; + LOG.debug("Going to execute insert <" + s + ">"); + int modCount = 0; + if ((modCount = stmt.executeUpdate(s)) < 1) { + //this can be reasonable for an empty txn START/COMMIT or read-only txn + //also an IUD with DP that didn't match any rows. + LOG.info("Expected to move at least one record from txn_components to " + + "completed_txn_components when committing txn! " + JavaUtils.txnIdToString(txnid)); + } + } else { + if (rqst.isSetWriteEventInfos()) { + List rows = new ArrayList<>(); + for (WriteEventInfo writeEventInfo : rqst.getWriteEventInfos()) { + rows.add(txnid + "," + quoteString(writeEventInfo.getDatabase()) + "," + + quoteString(writeEventInfo.getTable()) + "," + + quoteString(writeEventInfo.getPartition()) + "," + + writeEventInfo.getWriteId() + "," + + quoteChar(isUpdateDelete)); + } + List queries = sqlGenerator.createInsertValuesStmt("COMPLETED_TXN_COMPONENTS " + + "(ctc_txnid," + " ctc_database, ctc_table, ctc_partition, ctc_writeid, ctc_update_delete)", rows); + for (String q : queries) { + LOG.debug("Going to execute insert <" + q + "> "); + stmt.execute(q); + } + } + + s = "delete from REPL_TXN_MAP where RTM_SRC_TXN_ID = " + sourceTxnId + + " and RTM_REPL_POLICY = " + quoteString(rqst.getReplPolicy()); + LOG.info("Repl going to execute <" + s + ">"); + stmt.executeUpdate(s); } + s = "delete from TXN_COMPONENTS where tc_txnid = " + txnid; LOG.debug("Going to execute update <" + s + ">"); - modCount = stmt.executeUpdate(s); + stmt.executeUpdate(s); s = "delete from HIVE_LOCKS where hl_txnid = " + txnid; LOG.debug("Going to execute update <" + s + ">"); - modCount = stmt.executeUpdate(s); + stmt.executeUpdate(s); s = "delete from TXNS where txn_id = " + txnid; LOG.debug("Going to execute update <" + s + ">"); - modCount = stmt.executeUpdate(s); + stmt.executeUpdate(s); s = "delete from MIN_HISTORY_LEVEL where mhl_txnid = " + txnid; LOG.debug("Going to execute update <" + s + ">"); - modCount = stmt.executeUpdate(s); + stmt.executeUpdate(s); LOG.info("Removed committed transaction: (" + txnid + ") from MIN_HISTORY_LEVEL"); s = "delete from MATERIALIZATION_REBUILD_LOCKS where mrl_txn_id = " + txnid; LOG.debug("Going to execute update <" + s + ">"); - modCount = stmt.executeUpdate(s); - - if (rqst.isSetReplPolicy()) { - s = "delete from REPL_TXN_MAP where RTM_SRC_TXN_ID = " + sourceTxnId + - " and RTM_REPL_POLICY = " + quoteString(rqst.getReplPolicy()); - LOG.info("Repl going to execute <" + s + ">"); - stmt.executeUpdate(s); - } + stmt.executeUpdate(s); if (transactionalListeners != null) { MetaStoreListenerNotifier.notifyEventWithDirectSql(transactionalListeners, @@ -1525,6 +1587,43 @@ public void seedWriteIdOnAcidConversion(InitializeTableWriteIdsRequest rqst) } } + @Override + @RetrySemantics.Idempotent + public void addWriteNotificationLog(AcidWriteEvent acidWriteEvent) + throws MetaException { + Connection dbConn = null; + try { + try { + //Idempotent case is handled by notify Event + lockInternal(); + dbConn = getDbConn(Connection.TRANSACTION_READ_COMMITTED); + MetaStoreListenerNotifier.notifyEventWithDirectSql(transactionalListeners, + EventMessage.EventType.ACID_WRITE, acidWriteEvent, dbConn, sqlGenerator); + LOG.debug("Going to commit"); + dbConn.commit(); + return; + } catch (SQLException e) { + LOG.debug("Going to rollback"); + rollbackDBConn(dbConn); + if (isDuplicateKeyError(e)) { + // in case of key duplicate error, retry as it might be because of race condition + if (waitForRetry("addWriteNotificationLog(" + acidWriteEvent + ")", e.getMessage())) { + throw new RetryException(); + } + retryNum = 0; + throw new MetaException(e.getMessage()); + } + checkRetryable(dbConn, e, "addWriteNotificationLog(" + acidWriteEvent + ")"); + throw new MetaException("Unable to add write notification event " + StringUtils.stringifyException(e)); + } finally{ + closeDbConn(dbConn); + unlockInternal(); + } + } catch (RetryException e) { + addWriteNotificationLog(acidWriteEvent); + } + } + @Override @RetrySemantics.SafeToRetry public void performWriteSetGC() { @@ -3214,6 +3313,22 @@ static void close(ResultSet rs, Statement stmt, Connection dbConn) { closeStmt(stmt); closeDbConn(dbConn); } + + private boolean waitForRetry(String caller, String errMsg) { + if (retryNum++ < retryLimit) { + LOG.warn("Retryable error detected in " + caller + ". Will wait " + retryInterval + + "ms and retry up to " + (retryLimit - retryNum + 1) + " times. Error: " + errMsg); + try { + Thread.sleep(retryInterval); + } catch (InterruptedException ex) { + // + } + return true; + } else { + LOG.error("Fatal error in " + caller + ". Retry limit (" + retryLimit + ") reached. Last error: " + errMsg); + } + return false; + } /** * Determine if an exception was such that it makes sense to retry. Unfortunately there is no standard way to do * this, so we have to inspect the error messages and catch the telltale signs for each @@ -3257,18 +3372,7 @@ protected void checkRetryable(Connection conn, } } else if (isRetryable(conf, e)) { //in MSSQL this means Communication Link Failure - if (retryNum++ < retryLimit) { - LOG.warn("Retryable error detected in " + caller + ". Will wait " + retryInterval + - "ms and retry up to " + (retryLimit - retryNum + 1) + " times. Error: " + getMessage(e)); - try { - Thread.sleep(retryInterval); - } catch (InterruptedException ex) { - // - } - sendRetrySignal = true; - } else { - LOG.error("Fatal error in " + caller + ". Retry limit (" + retryLimit + ") reached. Last error: " + getMessage(e)); - } + sendRetrySignal = waitForRetry(caller, e.getMessage()); } else { //make sure we know we saw an error that we don't recognize diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java index 2c47ee4a20..33f24fbba6 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnStore.java @@ -25,6 +25,7 @@ import org.apache.hadoop.hive.common.ValidWriteIdList; import org.apache.hadoop.hive.common.classification.RetrySemantics; import org.apache.hadoop.hive.metastore.api.*; +import org.apache.hadoop.hive.metastore.events.AcidWriteEvent; import java.sql.SQLException; import java.util.Iterator; @@ -88,6 +89,9 @@ @RetrySemantics.Idempotent OpenTxnsResponse openTxns(OpenTxnRequest rqst) throws MetaException; + @RetrySemantics.Idempotent + long getTargetTxnId(String replPolicy, long sourceTxnId) throws MetaException; + /** * Abort (rollback) a transaction. * @param rqst info on transaction to abort @@ -490,4 +494,11 @@ void onRename(String oldCatName, String oldDbName, String oldTabName, String old */ @RetrySemantics.Idempotent void setHadoopJobId(String hadoopJobId, long id); + + /** + * Add the ACID write event information to writeNotificationLog table. + * @param acidWriteEvent + */ + @RetrySemantics.Idempotent + void addWriteNotificationLog(AcidWriteEvent acidWriteEvent) throws MetaException; } diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java index ec9e9e2b95..565c72b229 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/utils/FileUtils.java @@ -510,4 +510,15 @@ public static Path makeQualified(Path path, Configuration conf) throws IOExcepti return new Path(scheme, authority, pathUri.getPath()); } + + public static Path getTransformedPath(String name, String subDir, String root) { + if (root != null) { + Path newPath = new Path(root); + if (subDir != null) { + newPath = new Path(newPath, subDir); + } + return new Path(newPath, name); + } + return null; + } } diff --git a/standalone-metastore/src/main/resources/package.jdo b/standalone-metastore/src/main/resources/package.jdo index 1be3e986a5..5fb548cf88 100644 --- a/standalone-metastore/src/main/resources/package.jdo +++ b/standalone-metastore/src/main/resources/package.jdo @@ -1182,6 +1182,41 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/standalone-metastore/src/main/sql/derby/hive-schema-3.2.0.derby.sql b/standalone-metastore/src/main/sql/derby/hive-schema-3.2.0.derby.sql new file mode 100644 index 0000000000..50d0815395 --- /dev/null +++ b/standalone-metastore/src/main/sql/derby/hive-schema-3.2.0.derby.sql @@ -0,0 +1,720 @@ +-- Timestamp: 2011-09-22 15:32:02.024 +-- Source database is: /home/carl/Work/repos/hive1/metastore/scripts/upgrade/derby/mdb +-- Connection URL is: jdbc:derby:/home/carl/Work/repos/hive1/metastore/scripts/upgrade/derby/mdb +-- Specified schema is: APP +-- appendLogs: false + +-- ---------------------------------------------- +-- DDL Statements for functions +-- ---------------------------------------------- + +CREATE FUNCTION "APP"."NUCLEUS_ASCII" (C CHAR(1)) RETURNS INTEGER LANGUAGE JAVA PARAMETER STYLE JAVA READS SQL DATA CALLED ON NULL INPUT EXTERNAL NAME 'org.datanucleus.store.rdbms.adapter.DerbySQLFunction.ascii' ; + +CREATE FUNCTION "APP"."NUCLEUS_MATCHES" (TEXT VARCHAR(8000),PATTERN VARCHAR(8000)) RETURNS INTEGER LANGUAGE JAVA PARAMETER STYLE JAVA READS SQL DATA CALLED ON NULL INPUT EXTERNAL NAME 'org.datanucleus.store.rdbms.adapter.DerbySQLFunction.matches' ; + +-- ---------------------------------------------- +-- DDL Statements for tables +-- ---------------------------------------------- +CREATE TABLE "APP"."DBS" ( + "DB_ID" BIGINT NOT NULL, + "DESC" VARCHAR(4000), + "DB_LOCATION_URI" VARCHAR(4000) NOT NULL, + "NAME" VARCHAR(128), + "OWNER_NAME" VARCHAR(128), + "OWNER_TYPE" VARCHAR(10), + "CTLG_NAME" VARCHAR(256) NOT NULL +); + +CREATE TABLE "APP"."TBL_PRIVS" ("TBL_GRANT_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "TBL_PRIV" VARCHAR(128), "TBL_ID" BIGINT, "AUTHORIZER" VARCHAR(128)); + +CREATE TABLE "APP"."DATABASE_PARAMS" ("DB_ID" BIGINT NOT NULL, "PARAM_KEY" VARCHAR(180) NOT NULL, "PARAM_VALUE" VARCHAR(4000)); + +CREATE TABLE "APP"."TBL_COL_PRIVS" ("TBL_COLUMN_GRANT_ID" BIGINT NOT NULL, "COLUMN_NAME" VARCHAR(767), "CREATE_TIME" INTEGER NOT NULL, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "TBL_COL_PRIV" VARCHAR(128), "TBL_ID" BIGINT, "AUTHORIZER" VARCHAR(128)); + +CREATE TABLE "APP"."SERDE_PARAMS" ("SERDE_ID" BIGINT NOT NULL, "PARAM_KEY" VARCHAR(256) NOT NULL, "PARAM_VALUE" CLOB); + +CREATE TABLE "APP"."COLUMNS_V2" ("CD_ID" BIGINT NOT NULL, "COMMENT" VARCHAR(4000), "COLUMN_NAME" VARCHAR(767) NOT NULL, "TYPE_NAME" CLOB, "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."SORT_COLS" ("SD_ID" BIGINT NOT NULL, "COLUMN_NAME" VARCHAR(767), "ORDER" INTEGER NOT NULL, "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."CDS" ("CD_ID" BIGINT NOT NULL); + +CREATE TABLE "APP"."PARTITION_KEY_VALS" ("PART_ID" BIGINT NOT NULL, "PART_KEY_VAL" VARCHAR(256), "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."DB_PRIVS" ("DB_GRANT_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "DB_ID" BIGINT, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "DB_PRIV" VARCHAR(128), "AUTHORIZER" VARCHAR(128)); + +CREATE TABLE "APP"."IDXS" ("INDEX_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "DEFERRED_REBUILD" CHAR(1) NOT NULL, "INDEX_HANDLER_CLASS" VARCHAR(4000), "INDEX_NAME" VARCHAR(128), "INDEX_TBL_ID" BIGINT, "LAST_ACCESS_TIME" INTEGER NOT NULL, "ORIG_TBL_ID" BIGINT, "SD_ID" BIGINT); + +CREATE TABLE "APP"."INDEX_PARAMS" ("INDEX_ID" BIGINT NOT NULL, "PARAM_KEY" VARCHAR(256) NOT NULL, "PARAM_VALUE" VARCHAR(4000)); + +CREATE TABLE "APP"."PARTITIONS" ("PART_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "LAST_ACCESS_TIME" INTEGER NOT NULL, "PART_NAME" VARCHAR(767), "SD_ID" BIGINT, "TBL_ID" BIGINT); + +CREATE TABLE "APP"."SERDES" ("SERDE_ID" BIGINT NOT NULL, "NAME" VARCHAR(128), "SLIB" VARCHAR(4000), "DESCRIPTION" VARCHAR(4000), "SERIALIZER_CLASS" VARCHAR(4000), "DESERIALIZER_CLASS" VARCHAR(4000), SERDE_TYPE INTEGER); + +CREATE TABLE "APP"."PART_PRIVS" ("PART_GRANT_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PART_ID" BIGINT, "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "PART_PRIV" VARCHAR(128), "AUTHORIZER" VARCHAR(128)); + +CREATE TABLE "APP"."ROLE_MAP" ("ROLE_GRANT_ID" BIGINT NOT NULL, "ADD_TIME" INTEGER NOT NULL, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "ROLE_ID" BIGINT); + +CREATE TABLE "APP"."TYPES" ("TYPES_ID" BIGINT NOT NULL, "TYPE_NAME" VARCHAR(128), "TYPE1" VARCHAR(767), "TYPE2" VARCHAR(767)); + +CREATE TABLE "APP"."GLOBAL_PRIVS" ("USER_GRANT_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "USER_PRIV" VARCHAR(128), "AUTHORIZER" VARCHAR(128)); + +CREATE TABLE "APP"."PARTITION_PARAMS" ("PART_ID" BIGINT NOT NULL, "PARAM_KEY" VARCHAR(256) NOT NULL, "PARAM_VALUE" VARCHAR(4000)); + +CREATE TABLE "APP"."PARTITION_EVENTS" ( + "PART_NAME_ID" BIGINT NOT NULL, + "CAT_NAME" VARCHAR(256), + "DB_NAME" VARCHAR(128), + "EVENT_TIME" BIGINT NOT NULL, + "EVENT_TYPE" INTEGER NOT NULL, + "PARTITION_NAME" VARCHAR(767), + "TBL_NAME" VARCHAR(256) +); + +CREATE TABLE "APP"."COLUMNS" ("SD_ID" BIGINT NOT NULL, "COMMENT" VARCHAR(256), "COLUMN_NAME" VARCHAR(128) NOT NULL, "TYPE_NAME" VARCHAR(4000) NOT NULL, "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."ROLES" ("ROLE_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "OWNER_NAME" VARCHAR(128), "ROLE_NAME" VARCHAR(128)); + +CREATE TABLE "APP"."TBLS" ("TBL_ID" BIGINT NOT NULL, "CREATE_TIME" INTEGER NOT NULL, "DB_ID" BIGINT, "LAST_ACCESS_TIME" INTEGER NOT NULL, "OWNER" VARCHAR(767), "OWNER_TYPE" VARCHAR(10), "RETENTION" INTEGER NOT NULL, "SD_ID" BIGINT, "TBL_NAME" VARCHAR(256), "TBL_TYPE" VARCHAR(128), "VIEW_EXPANDED_TEXT" LONG VARCHAR, "VIEW_ORIGINAL_TEXT" LONG VARCHAR, "IS_REWRITE_ENABLED" CHAR(1) NOT NULL DEFAULT 'N'); + +CREATE TABLE "APP"."PARTITION_KEYS" ("TBL_ID" BIGINT NOT NULL, "PKEY_COMMENT" VARCHAR(4000), "PKEY_NAME" VARCHAR(128) NOT NULL, "PKEY_TYPE" VARCHAR(767) NOT NULL, "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."PART_COL_PRIVS" ("PART_COLUMN_GRANT_ID" BIGINT NOT NULL, "COLUMN_NAME" VARCHAR(767), "CREATE_TIME" INTEGER NOT NULL, "GRANT_OPTION" SMALLINT NOT NULL, "GRANTOR" VARCHAR(128), "GRANTOR_TYPE" VARCHAR(128), "PART_ID" BIGINT, "PRINCIPAL_NAME" VARCHAR(128), "PRINCIPAL_TYPE" VARCHAR(128), "PART_COL_PRIV" VARCHAR(128), "AUTHORIZER" VARCHAR(128)); + +CREATE TABLE "APP"."SDS" ("SD_ID" BIGINT NOT NULL, "INPUT_FORMAT" VARCHAR(4000), "IS_COMPRESSED" CHAR(1) NOT NULL, "LOCATION" VARCHAR(4000), "NUM_BUCKETS" INTEGER NOT NULL, "OUTPUT_FORMAT" VARCHAR(4000), "SERDE_ID" BIGINT, "CD_ID" BIGINT, "IS_STOREDASSUBDIRECTORIES" CHAR(1) NOT NULL); + +CREATE TABLE "APP"."SEQUENCE_TABLE" ("SEQUENCE_NAME" VARCHAR(256) NOT NULL, "NEXT_VAL" BIGINT NOT NULL); + +CREATE TABLE "APP"."TAB_COL_STATS"( + "CAT_NAME" VARCHAR(256) NOT NULL, + "DB_NAME" VARCHAR(128) NOT NULL, + "TABLE_NAME" VARCHAR(256) NOT NULL, + "COLUMN_NAME" VARCHAR(767) NOT NULL, + "COLUMN_TYPE" VARCHAR(128) NOT NULL, + "LONG_LOW_VALUE" BIGINT, + "LONG_HIGH_VALUE" BIGINT, + "DOUBLE_LOW_VALUE" DOUBLE, + "DOUBLE_HIGH_VALUE" DOUBLE, + "BIG_DECIMAL_LOW_VALUE" VARCHAR(4000), + "BIG_DECIMAL_HIGH_VALUE" VARCHAR(4000), + "NUM_DISTINCTS" BIGINT, + "NUM_NULLS" BIGINT NOT NULL, + "AVG_COL_LEN" DOUBLE, + "MAX_COL_LEN" BIGINT, + "NUM_TRUES" BIGINT, + "NUM_FALSES" BIGINT, + "LAST_ANALYZED" BIGINT, + "CS_ID" BIGINT NOT NULL, + "TBL_ID" BIGINT NOT NULL, + "BIT_VECTOR" BLOB +); + +CREATE TABLE "APP"."TABLE_PARAMS" ("TBL_ID" BIGINT NOT NULL, "PARAM_KEY" VARCHAR(256) NOT NULL, "PARAM_VALUE" CLOB); + +CREATE TABLE "APP"."BUCKETING_COLS" ("SD_ID" BIGINT NOT NULL, "BUCKET_COL_NAME" VARCHAR(256), "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."TYPE_FIELDS" ("TYPE_NAME" BIGINT NOT NULL, "COMMENT" VARCHAR(256), "FIELD_NAME" VARCHAR(128) NOT NULL, "FIELD_TYPE" VARCHAR(767) NOT NULL, "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."NUCLEUS_TABLES" ("CLASS_NAME" VARCHAR(128) NOT NULL, "TABLE_NAME" VARCHAR(128) NOT NULL, "TYPE" VARCHAR(4) NOT NULL, "OWNER" VARCHAR(2) NOT NULL, "VERSION" VARCHAR(20) NOT NULL, "INTERFACE_NAME" VARCHAR(256) DEFAULT NULL); + +CREATE TABLE "APP"."SD_PARAMS" ("SD_ID" BIGINT NOT NULL, "PARAM_KEY" VARCHAR(256) NOT NULL, "PARAM_VALUE" CLOB); + +CREATE TABLE "APP"."SKEWED_STRING_LIST" ("STRING_LIST_ID" BIGINT NOT NULL); + +CREATE TABLE "APP"."SKEWED_STRING_LIST_VALUES" ("STRING_LIST_ID" BIGINT NOT NULL, "STRING_LIST_VALUE" VARCHAR(256), "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."SKEWED_COL_NAMES" ("SD_ID" BIGINT NOT NULL, "SKEWED_COL_NAME" VARCHAR(256), "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."SKEWED_COL_VALUE_LOC_MAP" ("SD_ID" BIGINT NOT NULL, "STRING_LIST_ID_KID" BIGINT NOT NULL, "LOCATION" VARCHAR(4000)); + +CREATE TABLE "APP"."SKEWED_VALUES" ("SD_ID_OID" BIGINT NOT NULL, "STRING_LIST_ID_EID" BIGINT NOT NULL, "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."MASTER_KEYS" ("KEY_ID" INTEGER NOT NULL generated always as identity (start with 1), "MASTER_KEY" VARCHAR(767)); + +CREATE TABLE "APP"."DELEGATION_TOKENS" ( "TOKEN_IDENT" VARCHAR(767) NOT NULL, "TOKEN" VARCHAR(767)); + +CREATE TABLE "APP"."PART_COL_STATS"( + "CAT_NAME" VARCHAR(256) NOT NULL, + "DB_NAME" VARCHAR(128) NOT NULL, + "TABLE_NAME" VARCHAR(256) NOT NULL, + "PARTITION_NAME" VARCHAR(767) NOT NULL, + "COLUMN_NAME" VARCHAR(767) NOT NULL, + "COLUMN_TYPE" VARCHAR(128) NOT NULL, + "LONG_LOW_VALUE" BIGINT, + "LONG_HIGH_VALUE" BIGINT, + "DOUBLE_LOW_VALUE" DOUBLE, + "DOUBLE_HIGH_VALUE" DOUBLE, + "BIG_DECIMAL_LOW_VALUE" VARCHAR(4000), + "BIG_DECIMAL_HIGH_VALUE" VARCHAR(4000), + "NUM_DISTINCTS" BIGINT, + "BIT_VECTOR" BLOB, + "NUM_NULLS" BIGINT NOT NULL, + "AVG_COL_LEN" DOUBLE, + "MAX_COL_LEN" BIGINT, + "NUM_TRUES" BIGINT, + "NUM_FALSES" BIGINT, + "LAST_ANALYZED" BIGINT, + "CS_ID" BIGINT NOT NULL, + "PART_ID" BIGINT NOT NULL +); + +CREATE TABLE "APP"."VERSION" ("VER_ID" BIGINT NOT NULL, "SCHEMA_VERSION" VARCHAR(127) NOT NULL, "VERSION_COMMENT" VARCHAR(255)); + +CREATE TABLE "APP"."FUNCS" ("FUNC_ID" BIGINT NOT NULL, "CLASS_NAME" VARCHAR(4000), "CREATE_TIME" INTEGER NOT NULL, "DB_ID" BIGINT, "FUNC_NAME" VARCHAR(128), "FUNC_TYPE" INTEGER NOT NULL, "OWNER_NAME" VARCHAR(128), "OWNER_TYPE" VARCHAR(10)); + +CREATE TABLE "APP"."FUNC_RU" ("FUNC_ID" BIGINT NOT NULL, "RESOURCE_TYPE" INTEGER NOT NULL, "RESOURCE_URI" VARCHAR(4000), "INTEGER_IDX" INTEGER NOT NULL); + +CREATE TABLE "APP"."NOTIFICATION_LOG" ( + "NL_ID" BIGINT NOT NULL, + "CAT_NAME" VARCHAR(256), + "DB_NAME" VARCHAR(128), + "EVENT_ID" BIGINT NOT NULL, + "EVENT_TIME" INTEGER NOT NULL, + "EVENT_TYPE" VARCHAR(32) NOT NULL, + "MESSAGE" CLOB, + "TBL_NAME" VARCHAR(256), + "MESSAGE_FORMAT" VARCHAR(16) +); + +CREATE TABLE "APP"."NOTIFICATION_SEQUENCE" ("NNI_ID" BIGINT NOT NULL, "NEXT_EVENT_ID" BIGINT NOT NULL); + +CREATE TABLE "APP"."KEY_CONSTRAINTS" ("CHILD_CD_ID" BIGINT, "CHILD_INTEGER_IDX" INTEGER, "CHILD_TBL_ID" BIGINT, "PARENT_CD_ID" BIGINT , "PARENT_INTEGER_IDX" INTEGER, "PARENT_TBL_ID" BIGINT NOT NULL, "POSITION" BIGINT NOT NULL, "CONSTRAINT_NAME" VARCHAR(400) NOT NULL, "CONSTRAINT_TYPE" SMALLINT NOT NULL, "UPDATE_RULE" SMALLINT, "DELETE_RULE" SMALLINT, "ENABLE_VALIDATE_RELY" SMALLINT NOT NULL, "DEFAULT_VALUE" VARCHAR(400)); + +CREATE TABLE "APP"."METASTORE_DB_PROPERTIES" ("PROPERTY_KEY" VARCHAR(255) NOT NULL, "PROPERTY_VALUE" VARCHAR(1000) NOT NULL, "DESCRIPTION" VARCHAR(1000)); + +CREATE TABLE "APP"."WM_RESOURCEPLAN" (RP_ID BIGINT NOT NULL, NAME VARCHAR(128) NOT NULL, QUERY_PARALLELISM INTEGER, STATUS VARCHAR(20) NOT NULL, DEFAULT_POOL_ID BIGINT); + +CREATE TABLE "APP"."WM_POOL" (POOL_ID BIGINT NOT NULL, RP_ID BIGINT NOT NULL, PATH VARCHAR(1024) NOT NULL, ALLOC_FRACTION DOUBLE, QUERY_PARALLELISM INTEGER, SCHEDULING_POLICY VARCHAR(1024)); + +CREATE TABLE "APP"."WM_TRIGGER" (TRIGGER_ID BIGINT NOT NULL, RP_ID BIGINT NOT NULL, NAME VARCHAR(128) NOT NULL, TRIGGER_EXPRESSION VARCHAR(1024), ACTION_EXPRESSION VARCHAR(1024), IS_IN_UNMANAGED INTEGER NOT NULL DEFAULT 0); + +CREATE TABLE "APP"."WM_POOL_TO_TRIGGER" (POOL_ID BIGINT NOT NULL, TRIGGER_ID BIGINT NOT NULL); + +CREATE TABLE "APP"."WM_MAPPING" (MAPPING_ID BIGINT NOT NULL, RP_ID BIGINT NOT NULL, ENTITY_TYPE VARCHAR(128) NOT NULL, ENTITY_NAME VARCHAR(128) NOT NULL, POOL_ID BIGINT, ORDERING INTEGER); + +CREATE TABLE "APP"."MV_CREATION_METADATA" ( + "MV_CREATION_METADATA_ID" BIGINT NOT NULL, + "CAT_NAME" VARCHAR(256) NOT NULL, + "DB_NAME" VARCHAR(128) NOT NULL, + "TBL_NAME" VARCHAR(256) NOT NULL, + "TXN_LIST" CLOB, + "MATERIALIZATION_TIME" BIGINT NOT NULL +); + +CREATE TABLE "APP"."MV_TABLES_USED" ( + "MV_CREATION_METADATA_ID" BIGINT NOT NULL, + "TBL_ID" BIGINT NOT NULL +); + +CREATE TABLE "APP"."CTLGS" ( + "CTLG_ID" BIGINT NOT NULL, + "NAME" VARCHAR(256) UNIQUE, + "DESC" VARCHAR(4000), + "LOCATION_URI" VARCHAR(4000) NOT NULL); + +-- ---------------------------------------------- +-- DML Statements +-- ---------------------------------------------- + +INSERT INTO "APP"."NOTIFICATION_SEQUENCE" ("NNI_ID", "NEXT_EVENT_ID") SELECT * FROM (VALUES (1,1)) tmp_table WHERE NOT EXISTS ( SELECT "NEXT_EVENT_ID" FROM "APP"."NOTIFICATION_SEQUENCE"); + +INSERT INTO "APP"."SEQUENCE_TABLE" ("SEQUENCE_NAME", "NEXT_VAL") SELECT * FROM (VALUES ('org.apache.hadoop.hive.metastore.model.MNotificationLog', 1)) tmp_table WHERE NOT EXISTS ( SELECT "NEXT_VAL" FROM "APP"."SEQUENCE_TABLE" WHERE "SEQUENCE_NAME" = 'org.apache.hadoop.hive.metastore.model.MNotificationLog'); + +-- ---------------------------------------------- +-- DDL Statements for indexes +-- ---------------------------------------------- + +CREATE UNIQUE INDEX "APP"."UNIQUEINDEX" ON "APP"."IDXS" ("INDEX_NAME", "ORIG_TBL_ID"); + +CREATE INDEX "APP"."TABLECOLUMNPRIVILEGEINDEX" ON "APP"."TBL_COL_PRIVS" ("AUTHORIZER", "TBL_ID", "COLUMN_NAME", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "TBL_COL_PRIV", "GRANTOR", "GRANTOR_TYPE"); + +CREATE UNIQUE INDEX "APP"."DBPRIVILEGEINDEX" ON "APP"."DB_PRIVS" ("AUTHORIZER", "DB_ID", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "DB_PRIV", "GRANTOR", "GRANTOR_TYPE"); + +CREATE INDEX "APP"."PCS_STATS_IDX" ON "APP"."PART_COL_STATS" ("CAT_NAME", "DB_NAME","TABLE_NAME","COLUMN_NAME","PARTITION_NAME"); + +CREATE INDEX "APP"."TAB_COL_STATS_IDX" ON "APP"."TAB_COL_STATS" ("CAT_NAME", "DB_NAME", "TABLE_NAME", "COLUMN_NAME"); + +CREATE INDEX "APP"."PARTPRIVILEGEINDEX" ON "APP"."PART_PRIVS" ("AUTHORIZER", "PART_ID", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "PART_PRIV", "GRANTOR", "GRANTOR_TYPE"); + +CREATE UNIQUE INDEX "APP"."ROLEENTITYINDEX" ON "APP"."ROLES" ("ROLE_NAME"); + +CREATE INDEX "APP"."TABLEPRIVILEGEINDEX" ON "APP"."TBL_PRIVS" ("AUTHORIZER", "TBL_ID", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "TBL_PRIV", "GRANTOR", "GRANTOR_TYPE"); + +CREATE UNIQUE INDEX "APP"."UNIQUETABLE" ON "APP"."TBLS" ("TBL_NAME", "DB_ID"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_DATABASE" ON "APP"."DBS" ("NAME", "CTLG_NAME"); + +CREATE UNIQUE INDEX "APP"."USERROLEMAPINDEX" ON "APP"."ROLE_MAP" ("PRINCIPAL_NAME", "ROLE_ID", "GRANTOR", "GRANTOR_TYPE"); + +CREATE UNIQUE INDEX "APP"."GLOBALPRIVILEGEINDEX" ON "APP"."GLOBAL_PRIVS" ("AUTHORIZER", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "USER_PRIV", "GRANTOR", "GRANTOR_TYPE"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_TYPE" ON "APP"."TYPES" ("TYPE_NAME"); + +CREATE INDEX "APP"."PARTITIONCOLUMNPRIVILEGEINDEX" ON "APP"."PART_COL_PRIVS" ("AUTHORIZER", "PART_ID", "COLUMN_NAME", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "PART_COL_PRIV", "GRANTOR", "GRANTOR_TYPE"); + +CREATE UNIQUE INDEX "APP"."UNIQUEPARTITION" ON "APP"."PARTITIONS" ("PART_NAME", "TBL_ID"); + +CREATE UNIQUE INDEX "APP"."UNIQUEFUNCTION" ON "APP"."FUNCS" ("FUNC_NAME", "DB_ID"); + +CREATE INDEX "APP"."FUNCS_N49" ON "APP"."FUNCS" ("DB_ID"); + +CREATE INDEX "APP"."FUNC_RU_N49" ON "APP"."FUNC_RU" ("FUNC_ID"); + +CREATE INDEX "APP"."CONSTRAINTS_PARENT_TBL_ID_INDEX" ON "APP"."KEY_CONSTRAINTS"("PARENT_TBL_ID"); + +CREATE INDEX "APP"."CONSTRAINTS_CONSTRAINT_TYPE_INDEX" ON "APP"."KEY_CONSTRAINTS"("CONSTRAINT_TYPE"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_WM_RESOURCEPLAN" ON "APP"."WM_RESOURCEPLAN" ("NAME"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_WM_POOL" ON "APP"."WM_POOL" ("RP_ID", "PATH"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_WM_TRIGGER" ON "APP"."WM_TRIGGER" ("RP_ID", "NAME"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_WM_MAPPING" ON "APP"."WM_MAPPING" ("RP_ID", "ENTITY_TYPE", "ENTITY_NAME"); + +CREATE UNIQUE INDEX "APP"."MV_UNIQUE_TABLE" ON "APP"."MV_CREATION_METADATA" ("TBL_NAME", "DB_NAME"); + +CREATE UNIQUE INDEX "APP"."UNIQUE_CATALOG" ON "APP"."CTLGS" ("NAME"); + + +-- ---------------------------------------------- +-- DDL Statements for keys +-- ---------------------------------------------- + +-- primary/unique +ALTER TABLE "APP"."IDXS" ADD CONSTRAINT "IDXS_PK" PRIMARY KEY ("INDEX_ID"); + +ALTER TABLE "APP"."TBL_COL_PRIVS" ADD CONSTRAINT "TBL_COL_PRIVS_PK" PRIMARY KEY ("TBL_COLUMN_GRANT_ID"); + +ALTER TABLE "APP"."CDS" ADD CONSTRAINT "SQL110922153006460" PRIMARY KEY ("CD_ID"); + +ALTER TABLE "APP"."DB_PRIVS" ADD CONSTRAINT "DB_PRIVS_PK" PRIMARY KEY ("DB_GRANT_ID"); + +ALTER TABLE "APP"."INDEX_PARAMS" ADD CONSTRAINT "INDEX_PARAMS_PK" PRIMARY KEY ("INDEX_ID", "PARAM_KEY"); + +ALTER TABLE "APP"."PARTITION_KEYS" ADD CONSTRAINT "PARTITION_KEY_PK" PRIMARY KEY ("TBL_ID", "PKEY_NAME"); + +ALTER TABLE "APP"."SEQUENCE_TABLE" ADD CONSTRAINT "SEQUENCE_TABLE_PK" PRIMARY KEY ("SEQUENCE_NAME"); + +ALTER TABLE "APP"."PART_PRIVS" ADD CONSTRAINT "PART_PRIVS_PK" PRIMARY KEY ("PART_GRANT_ID"); + +ALTER TABLE "APP"."SDS" ADD CONSTRAINT "SDS_PK" PRIMARY KEY ("SD_ID"); + +ALTER TABLE "APP"."SERDES" ADD CONSTRAINT "SERDES_PK" PRIMARY KEY ("SERDE_ID"); + +ALTER TABLE "APP"."COLUMNS" ADD CONSTRAINT "COLUMNS_PK" PRIMARY KEY ("SD_ID", "COLUMN_NAME"); + +ALTER TABLE "APP"."PARTITION_EVENTS" ADD CONSTRAINT "PARTITION_EVENTS_PK" PRIMARY KEY ("PART_NAME_ID"); + +ALTER TABLE "APP"."TYPE_FIELDS" ADD CONSTRAINT "TYPE_FIELDS_PK" PRIMARY KEY ("TYPE_NAME", "FIELD_NAME"); + +ALTER TABLE "APP"."ROLES" ADD CONSTRAINT "ROLES_PK" PRIMARY KEY ("ROLE_ID"); + +ALTER TABLE "APP"."TBL_PRIVS" ADD CONSTRAINT "TBL_PRIVS_PK" PRIMARY KEY ("TBL_GRANT_ID"); + +ALTER TABLE "APP"."SERDE_PARAMS" ADD CONSTRAINT "SERDE_PARAMS_PK" PRIMARY KEY ("SERDE_ID", "PARAM_KEY"); + +ALTER TABLE "APP"."NUCLEUS_TABLES" ADD CONSTRAINT "NUCLEUS_TABLES_PK" PRIMARY KEY ("CLASS_NAME"); + +ALTER TABLE "APP"."TBLS" ADD CONSTRAINT "TBLS_PK" PRIMARY KEY ("TBL_ID"); + +ALTER TABLE "APP"."SD_PARAMS" ADD CONSTRAINT "SD_PARAMS_PK" PRIMARY KEY ("SD_ID", "PARAM_KEY"); + +ALTER TABLE "APP"."DATABASE_PARAMS" ADD CONSTRAINT "DATABASE_PARAMS_PK" PRIMARY KEY ("DB_ID", "PARAM_KEY"); + +ALTER TABLE "APP"."DBS" ADD CONSTRAINT "DBS_PK" PRIMARY KEY ("DB_ID"); + +ALTER TABLE "APP"."ROLE_MAP" ADD CONSTRAINT "ROLE_MAP_PK" PRIMARY KEY ("ROLE_GRANT_ID"); + +ALTER TABLE "APP"."GLOBAL_PRIVS" ADD CONSTRAINT "GLOBAL_PRIVS_PK" PRIMARY KEY ("USER_GRANT_ID"); + +ALTER TABLE "APP"."BUCKETING_COLS" ADD CONSTRAINT "BUCKETING_COLS_PK" PRIMARY KEY ("SD_ID", "INTEGER_IDX"); + +ALTER TABLE "APP"."SORT_COLS" ADD CONSTRAINT "SORT_COLS_PK" PRIMARY KEY ("SD_ID", "INTEGER_IDX"); + +ALTER TABLE "APP"."PARTITION_KEY_VALS" ADD CONSTRAINT "PARTITION_KEY_VALS_PK" PRIMARY KEY ("PART_ID", "INTEGER_IDX"); + +ALTER TABLE "APP"."TYPES" ADD CONSTRAINT "TYPES_PK" PRIMARY KEY ("TYPES_ID"); + +ALTER TABLE "APP"."COLUMNS_V2" ADD CONSTRAINT "SQL110922153006740" PRIMARY KEY ("CD_ID", "COLUMN_NAME"); + +ALTER TABLE "APP"."PART_COL_PRIVS" ADD CONSTRAINT "PART_COL_PRIVS_PK" PRIMARY KEY ("PART_COLUMN_GRANT_ID"); + +ALTER TABLE "APP"."PARTITION_PARAMS" ADD CONSTRAINT "PARTITION_PARAMS_PK" PRIMARY KEY ("PART_ID", "PARAM_KEY"); + +ALTER TABLE "APP"."PARTITIONS" ADD CONSTRAINT "PARTITIONS_PK" PRIMARY KEY ("PART_ID"); + +ALTER TABLE "APP"."TABLE_PARAMS" ADD CONSTRAINT "TABLE_PARAMS_PK" PRIMARY KEY ("TBL_ID", "PARAM_KEY"); + +ALTER TABLE "APP"."SKEWED_STRING_LIST" ADD CONSTRAINT "SKEWED_STRING_LIST_PK" PRIMARY KEY ("STRING_LIST_ID"); + +ALTER TABLE "APP"."SKEWED_STRING_LIST_VALUES" ADD CONSTRAINT "SKEWED_STRING_LIST_VALUES_PK" PRIMARY KEY ("STRING_LIST_ID", "INTEGER_IDX"); + +ALTER TABLE "APP"."SKEWED_COL_NAMES" ADD CONSTRAINT "SKEWED_COL_NAMES_PK" PRIMARY KEY ("SD_ID", "INTEGER_IDX"); + +ALTER TABLE "APP"."SKEWED_COL_VALUE_LOC_MAP" ADD CONSTRAINT "SKEWED_COL_VALUE_LOC_MAP_PK" PRIMARY KEY ("SD_ID", "STRING_LIST_ID_KID"); + +ALTER TABLE "APP"."SKEWED_VALUES" ADD CONSTRAINT "SKEWED_VALUES_PK" PRIMARY KEY ("SD_ID_OID", "INTEGER_IDX"); + +ALTER TABLE "APP"."TAB_COL_STATS" ADD CONSTRAINT "TAB_COL_STATS_PK" PRIMARY KEY ("CS_ID"); + +ALTER TABLE "APP"."PART_COL_STATS" ADD CONSTRAINT "PART_COL_STATS_PK" PRIMARY KEY ("CS_ID"); + +ALTER TABLE "APP"."FUNCS" ADD CONSTRAINT "FUNCS_PK" PRIMARY KEY ("FUNC_ID"); + +ALTER TABLE "APP"."FUNC_RU" ADD CONSTRAINT "FUNC_RU_PK" PRIMARY KEY ("FUNC_ID", "INTEGER_IDX"); + +ALTER TABLE "APP"."NOTIFICATION_LOG" ADD CONSTRAINT "NOTIFICATION_LOG_PK" PRIMARY KEY ("NL_ID"); + +ALTER TABLE "APP"."NOTIFICATION_SEQUENCE" ADD CONSTRAINT "NOTIFICATION_SEQUENCE_PK" PRIMARY KEY ("NNI_ID"); + +ALTER TABLE "APP"."KEY_CONSTRAINTS" ADD CONSTRAINT "CONSTRAINTS_PK" PRIMARY KEY ("CONSTRAINT_NAME", "POSITION"); + +ALTER TABLE "APP"."METASTORE_DB_PROPERTIES" ADD CONSTRAINT "PROPERTY_KEY_PK" PRIMARY KEY ("PROPERTY_KEY"); + +ALTER TABLE "APP"."MV_CREATION_METADATA" ADD CONSTRAINT "MV_CREATION_METADATA_PK" PRIMARY KEY ("MV_CREATION_METADATA_ID"); + +ALTER TABLE "APP"."CTLGS" ADD CONSTRAINT "CTLG_PK" PRIMARY KEY ("CTLG_ID"); + + +-- foreign +ALTER TABLE "APP"."IDXS" ADD CONSTRAINT "IDXS_FK1" FOREIGN KEY ("ORIG_TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."IDXS" ADD CONSTRAINT "IDXS_FK2" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."IDXS" ADD CONSTRAINT "IDXS_FK3" FOREIGN KEY ("INDEX_TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TBL_COL_PRIVS" ADD CONSTRAINT "TBL_COL_PRIVS_FK1" FOREIGN KEY ("TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."DB_PRIVS" ADD CONSTRAINT "DB_PRIVS_FK1" FOREIGN KEY ("DB_ID") REFERENCES "APP"."DBS" ("DB_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."INDEX_PARAMS" ADD CONSTRAINT "INDEX_PARAMS_FK1" FOREIGN KEY ("INDEX_ID") REFERENCES "APP"."IDXS" ("INDEX_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PARTITION_KEYS" ADD CONSTRAINT "PARTITION_KEYS_FK1" FOREIGN KEY ("TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PART_PRIVS" ADD CONSTRAINT "PART_PRIVS_FK1" FOREIGN KEY ("PART_ID") REFERENCES "APP"."PARTITIONS" ("PART_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SDS" ADD CONSTRAINT "SDS_FK1" FOREIGN KEY ("SERDE_ID") REFERENCES "APP"."SERDES" ("SERDE_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SDS" ADD CONSTRAINT "SDS_FK2" FOREIGN KEY ("CD_ID") REFERENCES "APP"."CDS" ("CD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."COLUMNS" ADD CONSTRAINT "COLUMNS_FK1" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TYPE_FIELDS" ADD CONSTRAINT "TYPE_FIELDS_FK1" FOREIGN KEY ("TYPE_NAME") REFERENCES "APP"."TYPES" ("TYPES_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TBL_PRIVS" ADD CONSTRAINT "TBL_PRIVS_FK1" FOREIGN KEY ("TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SERDE_PARAMS" ADD CONSTRAINT "SERDE_PARAMS_FK1" FOREIGN KEY ("SERDE_ID") REFERENCES "APP"."SERDES" ("SERDE_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TBLS" ADD CONSTRAINT "TBLS_FK2" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TBLS" ADD CONSTRAINT "TBLS_FK1" FOREIGN KEY ("DB_ID") REFERENCES "APP"."DBS" ("DB_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."DBS" ADD CONSTRAINT "DBS_FK1" FOREIGN KEY ("CTLG_NAME") REFERENCES "APP"."CTLGS" ("NAME") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SD_PARAMS" ADD CONSTRAINT "SD_PARAMS_FK1" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."DATABASE_PARAMS" ADD CONSTRAINT "DATABASE_PARAMS_FK1" FOREIGN KEY ("DB_ID") REFERENCES "APP"."DBS" ("DB_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."ROLE_MAP" ADD CONSTRAINT "ROLE_MAP_FK1" FOREIGN KEY ("ROLE_ID") REFERENCES "APP"."ROLES" ("ROLE_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."BUCKETING_COLS" ADD CONSTRAINT "BUCKETING_COLS_FK1" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SORT_COLS" ADD CONSTRAINT "SORT_COLS_FK1" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PARTITION_KEY_VALS" ADD CONSTRAINT "PARTITION_KEY_VALS_FK1" FOREIGN KEY ("PART_ID") REFERENCES "APP"."PARTITIONS" ("PART_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."COLUMNS_V2" ADD CONSTRAINT "COLUMNS_V2_FK1" FOREIGN KEY ("CD_ID") REFERENCES "APP"."CDS" ("CD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PART_COL_PRIVS" ADD CONSTRAINT "PART_COL_PRIVS_FK1" FOREIGN KEY ("PART_ID") REFERENCES "APP"."PARTITIONS" ("PART_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PARTITION_PARAMS" ADD CONSTRAINT "PARTITION_PARAMS_FK1" FOREIGN KEY ("PART_ID") REFERENCES "APP"."PARTITIONS" ("PART_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PARTITIONS" ADD CONSTRAINT "PARTITIONS_FK1" FOREIGN KEY ("TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PARTITIONS" ADD CONSTRAINT "PARTITIONS_FK2" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TABLE_PARAMS" ADD CONSTRAINT "TABLE_PARAMS_FK1" FOREIGN KEY ("TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SKEWED_STRING_LIST_VALUES" ADD CONSTRAINT "SKEWED_STRING_LIST_VALUES_FK1" FOREIGN KEY ("STRING_LIST_ID") REFERENCES "APP"."SKEWED_STRING_LIST" ("STRING_LIST_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SKEWED_COL_NAMES" ADD CONSTRAINT "SKEWED_COL_NAMES_FK1" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SKEWED_COL_VALUE_LOC_MAP" ADD CONSTRAINT "SKEWED_COL_VALUE_LOC_MAP_FK1" FOREIGN KEY ("SD_ID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SKEWED_COL_VALUE_LOC_MAP" ADD CONSTRAINT "SKEWED_COL_VALUE_LOC_MAP_FK2" FOREIGN KEY ("STRING_LIST_ID_KID") REFERENCES "APP"."SKEWED_STRING_LIST" ("STRING_LIST_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SKEWED_VALUES" ADD CONSTRAINT "SKEWED_VALUES_FK1" FOREIGN KEY ("SD_ID_OID") REFERENCES "APP"."SDS" ("SD_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."SKEWED_VALUES" ADD CONSTRAINT "SKEWED_VALUES_FK2" FOREIGN KEY ("STRING_LIST_ID_EID") REFERENCES "APP"."SKEWED_STRING_LIST" ("STRING_LIST_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."TAB_COL_STATS" ADD CONSTRAINT "TAB_COL_STATS_FK" FOREIGN KEY ("TBL_ID") REFERENCES TBLS("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."PART_COL_STATS" ADD CONSTRAINT "PART_COL_STATS_FK" FOREIGN KEY ("PART_ID") REFERENCES PARTITIONS("PART_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."VERSION" ADD CONSTRAINT "VERSION_PK" PRIMARY KEY ("VER_ID"); + +ALTER TABLE "APP"."FUNCS" ADD CONSTRAINT "FUNCS_FK1" FOREIGN KEY ("DB_ID") REFERENCES "APP"."DBS" ("DB_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."FUNC_RU" ADD CONSTRAINT "FUNC_RU_FK1" FOREIGN KEY ("FUNC_ID") REFERENCES "APP"."FUNCS" ("FUNC_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_RESOURCEPLAN" ADD CONSTRAINT "WM_RESOURCEPLAN_PK" PRIMARY KEY ("RP_ID"); + +ALTER TABLE "APP"."WM_POOL" ADD CONSTRAINT "WM_POOL_PK" PRIMARY KEY ("POOL_ID"); + +ALTER TABLE "APP"."WM_POOL" ADD CONSTRAINT "WM_POOL_FK1" FOREIGN KEY ("RP_ID") REFERENCES "APP"."WM_RESOURCEPLAN" ("RP_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_RESOURCEPLAN" ADD CONSTRAINT "WM_RESOURCEPLAN_FK1" FOREIGN KEY ("DEFAULT_POOL_ID") REFERENCES "APP"."WM_POOL" ("POOL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_TRIGGER" ADD CONSTRAINT "WM_TRIGGER_PK" PRIMARY KEY ("TRIGGER_ID"); + +ALTER TABLE "APP"."WM_TRIGGER" ADD CONSTRAINT "WM_TRIGGER_FK1" FOREIGN KEY ("RP_ID") REFERENCES "APP"."WM_RESOURCEPLAN" ("RP_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_POOL_TO_TRIGGER" ADD CONSTRAINT "WM_POOL_TO_TRIGGER_FK1" FOREIGN KEY ("POOL_ID") REFERENCES "APP"."WM_POOL" ("POOL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_POOL_TO_TRIGGER" ADD CONSTRAINT "WM_POOL_TO_TRIGGER_FK2" FOREIGN KEY ("TRIGGER_ID") REFERENCES "APP"."WM_TRIGGER" ("TRIGGER_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_MAPPING" ADD CONSTRAINT "WM_MAPPING_PK" PRIMARY KEY ("MAPPING_ID"); + +ALTER TABLE "APP"."WM_MAPPING" ADD CONSTRAINT "WM_MAPPING_FK1" FOREIGN KEY ("RP_ID") REFERENCES "APP"."WM_RESOURCEPLAN" ("RP_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."WM_MAPPING" ADD CONSTRAINT "WM_MAPPING_FK2" FOREIGN KEY ("POOL_ID") REFERENCES "APP"."WM_POOL" ("POOL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."MV_TABLES_USED" ADD CONSTRAINT "MV_TABLES_USED_FK1" FOREIGN KEY ("MV_CREATION_METADATA_ID") REFERENCES "APP"."MV_CREATION_METADATA" ("MV_CREATION_METADATA_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."MV_TABLES_USED" ADD CONSTRAINT "MV_TABLES_USED_FK2" FOREIGN KEY ("TBL_ID") REFERENCES "APP"."TBLS" ("TBL_ID") ON DELETE NO ACTION ON UPDATE NO ACTION; + +ALTER TABLE "APP"."DBS" ADD CONSTRAINT "DBS_CTLG_FK" FOREIGN KEY ("CTLG_NAME") REFERENCES "APP"."CTLGS" ("NAME") ON DELETE NO ACTION ON UPDATE NO ACTION; + +-- ---------------------------------------------- +-- DDL Statements for checks +-- ---------------------------------------------- + +ALTER TABLE "APP"."IDXS" ADD CONSTRAINT "SQL110318025504980" CHECK (DEFERRED_REBUILD IN ('Y','N')); + +ALTER TABLE "APP"."SDS" ADD CONSTRAINT "SQL110318025505550" CHECK (IS_COMPRESSED IN ('Y','N')); + +-- ---------------------------- +-- Transaction and Lock Tables +-- ---------------------------- +CREATE TABLE TXNS ( + TXN_ID bigint PRIMARY KEY, + TXN_STATE char(1) NOT NULL, + TXN_STARTED bigint NOT NULL, + TXN_LAST_HEARTBEAT bigint NOT NULL, + TXN_USER varchar(128) NOT NULL, + TXN_HOST varchar(128) NOT NULL, + TXN_AGENT_INFO varchar(128), + TXN_META_INFO varchar(128), + TXN_HEARTBEAT_COUNT integer, + TXN_TYPE integer +); + +CREATE TABLE TXN_COMPONENTS ( + TC_TXNID bigint NOT NULL REFERENCES TXNS (TXN_ID), + TC_DATABASE varchar(128) NOT NULL, + TC_TABLE varchar(128), + TC_PARTITION varchar(767), + TC_OPERATION_TYPE char(1) NOT NULL, + TC_WRITEID bigint +); + +CREATE INDEX TC_TXNID_INDEX ON TXN_COMPONENTS (TC_TXNID); + +CREATE TABLE COMPLETED_TXN_COMPONENTS ( + CTC_TXNID bigint NOT NULL, + CTC_DATABASE varchar(128) NOT NULL, + CTC_TABLE varchar(256), + CTC_PARTITION varchar(767), + CTC_TIMESTAMP timestamp DEFAULT CURRENT_TIMESTAMP NOT NULL, + CTC_WRITEID bigint, + CTC_UPDATE_DELETE char(1) NOT NULL +); + +CREATE INDEX COMPLETED_TXN_COMPONENTS_IDX ON COMPLETED_TXN_COMPONENTS (CTC_DATABASE, CTC_TABLE, CTC_PARTITION); + +CREATE TABLE NEXT_TXN_ID ( + NTXN_NEXT bigint NOT NULL +); +INSERT INTO NEXT_TXN_ID VALUES(1); + +CREATE TABLE HIVE_LOCKS ( + HL_LOCK_EXT_ID bigint NOT NULL, + HL_LOCK_INT_ID bigint NOT NULL, + HL_TXNID bigint NOT NULL, + HL_DB varchar(128) NOT NULL, + HL_TABLE varchar(128), + HL_PARTITION varchar(767), + HL_LOCK_STATE char(1) NOT NULL, + HL_LOCK_TYPE char(1) NOT NULL, + HL_LAST_HEARTBEAT bigint NOT NULL, + HL_ACQUIRED_AT bigint, + HL_USER varchar(128) NOT NULL, + HL_HOST varchar(128) NOT NULL, + HL_HEARTBEAT_COUNT integer, + HL_AGENT_INFO varchar(128), + HL_BLOCKEDBY_EXT_ID bigint, + HL_BLOCKEDBY_INT_ID bigint, + PRIMARY KEY(HL_LOCK_EXT_ID, HL_LOCK_INT_ID) +); + +CREATE INDEX HL_TXNID_INDEX ON HIVE_LOCKS (HL_TXNID); + +CREATE TABLE NEXT_LOCK_ID ( + NL_NEXT bigint NOT NULL +); +INSERT INTO NEXT_LOCK_ID VALUES(1); + +CREATE TABLE COMPACTION_QUEUE ( + CQ_ID bigint PRIMARY KEY, + CQ_DATABASE varchar(128) NOT NULL, + CQ_TABLE varchar(128) NOT NULL, + CQ_PARTITION varchar(767), + CQ_STATE char(1) NOT NULL, + CQ_TYPE char(1) NOT NULL, + CQ_TBLPROPERTIES varchar(2048), + CQ_WORKER_ID varchar(128), + CQ_START bigint, + CQ_RUN_AS varchar(128), + CQ_HIGHEST_WRITE_ID bigint, + CQ_META_INFO varchar(2048) for bit data, + CQ_HADOOP_JOB_ID varchar(32) +); + +CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( + NCQ_NEXT bigint NOT NULL +); +INSERT INTO NEXT_COMPACTION_QUEUE_ID VALUES(1); + +CREATE TABLE COMPLETED_COMPACTIONS ( + CC_ID bigint PRIMARY KEY, + CC_DATABASE varchar(128) NOT NULL, + CC_TABLE varchar(128) NOT NULL, + CC_PARTITION varchar(767), + CC_STATE char(1) NOT NULL, + CC_TYPE char(1) NOT NULL, + CC_TBLPROPERTIES varchar(2048), + CC_WORKER_ID varchar(128), + CC_START bigint, + CC_END bigint, + CC_RUN_AS varchar(128), + CC_HIGHEST_WRITE_ID bigint, + CC_META_INFO varchar(2048) for bit data, + CC_HADOOP_JOB_ID varchar(32) +); + +CREATE TABLE AUX_TABLE ( + MT_KEY1 varchar(128) NOT NULL, + MT_KEY2 bigint NOT NULL, + MT_COMMENT varchar(255), + PRIMARY KEY(MT_KEY1, MT_KEY2) +); + +--1st 4 cols make up a PK but since WS_PARTITION is nullable we can't declare such PK +--This is a good candidate for Index orgainzed table +CREATE TABLE WRITE_SET ( + WS_DATABASE varchar(128) NOT NULL, + WS_TABLE varchar(128) NOT NULL, + WS_PARTITION varchar(767), + WS_TXNID bigint NOT NULL, + WS_COMMIT_ID bigint NOT NULL, + WS_OPERATION_TYPE char(1) NOT NULL +); + +CREATE TABLE TXN_TO_WRITE_ID ( + T2W_TXNID bigint NOT NULL, + T2W_DATABASE varchar(128) NOT NULL, + T2W_TABLE varchar(256) NOT NULL, + T2W_WRITEID bigint NOT NULL +); + +CREATE UNIQUE INDEX TBL_TO_TXN_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_TXNID); +CREATE UNIQUE INDEX TBL_TO_WRITE_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_WRITEID); + +CREATE TABLE NEXT_WRITE_ID ( + NWI_DATABASE varchar(128) NOT NULL, + NWI_TABLE varchar(256) NOT NULL, + NWI_NEXT bigint NOT NULL +); + +CREATE UNIQUE INDEX NEXT_WRITE_ID_IDX ON NEXT_WRITE_ID (NWI_DATABASE, NWI_TABLE); + +CREATE TABLE MIN_HISTORY_LEVEL ( + MHL_TXNID bigint NOT NULL, + MHL_MIN_OPEN_TXNID bigint NOT NULL, + PRIMARY KEY(MHL_TXNID) +); + +CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); + +CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( + MRL_TXN_ID BIGINT NOT NULL, + MRL_DB_NAME VARCHAR(128) NOT NULL, + MRL_TBL_NAME VARCHAR(256) NOT NULL, + MRL_LAST_HEARTBEAT BIGINT NOT NULL, + PRIMARY KEY(MRL_TXN_ID) +); + +CREATE TABLE "APP"."I_SCHEMA" ( + "SCHEMA_ID" bigint primary key, + "SCHEMA_TYPE" integer not null, + "NAME" varchar(256) unique, + "DB_ID" bigint references "APP"."DBS" ("DB_ID"), + "COMPATIBILITY" integer not null, + "VALIDATION_LEVEL" integer not null, + "CAN_EVOLVE" char(1) not null, + "SCHEMA_GROUP" varchar(256), + "DESCRIPTION" varchar(4000) +); + +CREATE TABLE "APP"."SCHEMA_VERSION" ( + "SCHEMA_VERSION_ID" bigint primary key, + "SCHEMA_ID" bigint references "APP"."I_SCHEMA" ("SCHEMA_ID"), + "VERSION" integer not null, + "CREATED_AT" bigint not null, + "CD_ID" bigint references "APP"."CDS" ("CD_ID"), + "STATE" integer not null, + "DESCRIPTION" varchar(4000), + "SCHEMA_TEXT" clob, + "FINGERPRINT" varchar(256), + "SCHEMA_VERSION_NAME" varchar(256), + "SERDE_ID" bigint references "APP"."SERDES" ("SERDE_ID") +); + +CREATE UNIQUE INDEX "APP"."UNIQUE_SCHEMA_VERSION" ON "APP"."SCHEMA_VERSION" ("SCHEMA_ID", "VERSION"); + +CREATE TABLE REPL_TXN_MAP ( + RTM_REPL_POLICY varchar(256) NOT NULL, + RTM_SRC_TXN_ID bigint NOT NULL, + RTM_TARGET_TXN_ID bigint NOT NULL, + PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID) +); + +CREATE TABLE "APP"."RUNTIME_STATS" ( + "RS_ID" bigint primary key, + "CREATE_TIME" integer not null, + "WEIGHT" integer not null, + "PAYLOAD" BLOB +); + +CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); + +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID bigint NOT NULL, + WNL_TXNID bigint NOT NULL, + WNL_WRITEID bigint NOT NULL, + WNL_DATABASE varchar(128) NOT NULL, + WNL_TABLE varchar(128) NOT NULL, + WNL_PARTITION varchar(767) NOT NULL, + WNL_TABLE_OBJ clob NOT NULL, + WNL_PARTITION_OBJ clob, + WNL_FILES clob, + WNL_EVENT_TIME integer NOT NULL, + PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION) +); +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- ----------------------------------------------------------------- +-- Record schema version. Should be the last step in the init script +-- ----------------------------------------------------------------- +INSERT INTO "APP"."VERSION" (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES (1, '3.2.0', 'Hive release version 3.2.0'); diff --git a/standalone-metastore/src/main/sql/derby/upgrade-2.3.0-to-3.0.0.derby.sql b/standalone-metastore/src/main/sql/derby/upgrade-2.3.0-to-3.0.0.derby.sql index 7b7a8a24d9..10f1373c13 100644 --- a/standalone-metastore/src/main/sql/derby/upgrade-2.3.0-to-3.0.0.derby.sql +++ b/standalone-metastore/src/main/sql/derby/upgrade-2.3.0-to-3.0.0.derby.sql @@ -244,7 +244,6 @@ CREATE TABLE MIN_HISTORY_LEVEL ( CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); - CREATE TABLE "APP"."RUNTIME_STATS" ( "RS_ID" bigint primary key, "CREATE_TIME" integer not null, diff --git a/standalone-metastore/src/main/sql/derby/upgrade-3.1.0-to-3.2.0.derby.sql b/standalone-metastore/src/main/sql/derby/upgrade-3.1.0-to-3.2.0.derby.sql new file mode 100644 index 0000000000..5fcefd66ae --- /dev/null +++ b/standalone-metastore/src/main/sql/derby/upgrade-3.1.0-to-3.2.0.derby.sql @@ -0,0 +1,19 @@ +-- Upgrade MetaStore schema from 3.1.0 to 3.2.0 + +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID bigint NOT NULL, + WNL_TXNID bigint NOT NULL, + WNL_WRITEID bigint NOT NULL, + WNL_DATABASE varchar(128) NOT NULL, + WNL_TABLE varchar(128) NOT NULL, + WNL_PARTITION varchar(767) NOT NULL, + WNL_TABLE_OBJ clob NOT NULL, + WNL_PARTITION_OBJ clob, + WNL_FILES clob, + WNL_EVENT_TIME integer NOT NULL, + PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION) +); +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- This needs to be the last thing done. Insert any changes above this line. +UPDATE "APP".VERSION SET SCHEMA_VERSION='3.2.0', VERSION_COMMENT='Hive release version 3.2.0' where VER_ID=1; diff --git a/standalone-metastore/src/main/sql/derby/upgrade.order.derby b/standalone-metastore/src/main/sql/derby/upgrade.order.derby index f43da9af19..e6eb71ae56 100644 --- a/standalone-metastore/src/main/sql/derby/upgrade.order.derby +++ b/standalone-metastore/src/main/sql/derby/upgrade.order.derby @@ -15,3 +15,4 @@ 2.2.0-to-2.3.0 2.3.0-to-3.0.0 3.0.0-to-3.1.0 +3.1.0-to-3.2.0 diff --git a/standalone-metastore/src/main/sql/mssql/hive-schema-3.2.0.mssql.sql b/standalone-metastore/src/main/sql/mssql/hive-schema-3.2.0.mssql.sql new file mode 100644 index 0000000000..9e586e8ac3 --- /dev/null +++ b/standalone-metastore/src/main/sql/mssql/hive-schema-3.2.0.mssql.sql @@ -0,0 +1,1284 @@ +-- Licensed to the Apache Software Foundation (ASF) under one or more +-- contributor license agreements. See the NOTICE file distributed with +-- this work for additional information regarding copyright ownership. +-- The ASF licenses this file to You under the Apache License, Version 2.0 +-- (the "License"); you may not use this file except in compliance with +-- the License. You may obtain a copy of the License at +-- +-- http://www.apache.org/licenses/LICENSE-2.0 +-- +-- Unless required by applicable law or agreed to in writing, software +-- distributed under the License is distributed on an "AS IS" BASIS, +-- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +-- See the License for the specific language governing permissions and +-- limitations under the License. + +------------------------------------------------------------------ +-- DataNucleus SchemaTool (ran at 08/04/2014 15:10:15) +------------------------------------------------------------------ +-- Complete schema required for the following classes:- +-- org.apache.hadoop.hive.metastore.model.MColumnDescriptor +-- org.apache.hadoop.hive.metastore.model.MDBPrivilege +-- org.apache.hadoop.hive.metastore.model.MDatabase +-- org.apache.hadoop.hive.metastore.model.MDelegationToken +-- org.apache.hadoop.hive.metastore.model.MFieldSchema +-- org.apache.hadoop.hive.metastore.model.MFunction +-- org.apache.hadoop.hive.metastore.model.MGlobalPrivilege +-- org.apache.hadoop.hive.metastore.model.MIndex +-- org.apache.hadoop.hive.metastore.model.MMasterKey +-- org.apache.hadoop.hive.metastore.model.MOrder +-- org.apache.hadoop.hive.metastore.model.MPartition +-- org.apache.hadoop.hive.metastore.model.MPartitionColumnPrivilege +-- org.apache.hadoop.hive.metastore.model.MPartitionColumnStatistics +-- org.apache.hadoop.hive.metastore.model.MPartitionEvent +-- org.apache.hadoop.hive.metastore.model.MPartitionPrivilege +-- org.apache.hadoop.hive.metastore.model.MResourceUri +-- org.apache.hadoop.hive.metastore.model.MRole +-- org.apache.hadoop.hive.metastore.model.MRoleMap +-- org.apache.hadoop.hive.metastore.model.MSerDeInfo +-- org.apache.hadoop.hive.metastore.model.MStorageDescriptor +-- org.apache.hadoop.hive.metastore.model.MStringList +-- org.apache.hadoop.hive.metastore.model.MTable +-- org.apache.hadoop.hive.metastore.model.MTableColumnPrivilege +-- org.apache.hadoop.hive.metastore.model.MTableColumnStatistics +-- org.apache.hadoop.hive.metastore.model.MTablePrivilege +-- org.apache.hadoop.hive.metastore.model.MType +-- org.apache.hadoop.hive.metastore.model.MVersionTable +-- +-- Table MASTER_KEYS for classes [org.apache.hadoop.hive.metastore.model.MMasterKey] +CREATE TABLE MASTER_KEYS +( + KEY_ID int NOT NULL, + MASTER_KEY nvarchar(767) NULL +); + +ALTER TABLE MASTER_KEYS ADD CONSTRAINT MASTER_KEYS_PK PRIMARY KEY (KEY_ID); + +-- Table IDXS for classes [org.apache.hadoop.hive.metastore.model.MIndex] +CREATE TABLE IDXS +( + INDEX_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + DEFERRED_REBUILD bit NOT NULL, + INDEX_HANDLER_CLASS nvarchar(4000) NULL, + INDEX_NAME nvarchar(128) NULL, + INDEX_TBL_ID bigint NULL, + LAST_ACCESS_TIME int NOT NULL, + ORIG_TBL_ID bigint NULL, + SD_ID bigint NULL +); + +ALTER TABLE IDXS ADD CONSTRAINT IDXS_PK PRIMARY KEY (INDEX_ID); + +-- Table PART_COL_STATS for classes [org.apache.hadoop.hive.metastore.model.MPartitionColumnStatistics] +CREATE TABLE PART_COL_STATS +( + CS_ID bigint NOT NULL, + AVG_COL_LEN float NULL, + "COLUMN_NAME" nvarchar(767) NOT NULL, + COLUMN_TYPE nvarchar(128) NOT NULL, + DB_NAME nvarchar(128) NOT NULL, + BIG_DECIMAL_HIGH_VALUE nvarchar(255) NULL, + BIG_DECIMAL_LOW_VALUE nvarchar(255) NULL, + DOUBLE_HIGH_VALUE float NULL, + DOUBLE_LOW_VALUE float NULL, + LAST_ANALYZED bigint NOT NULL, + LONG_HIGH_VALUE bigint NULL, + LONG_LOW_VALUE bigint NULL, + MAX_COL_LEN bigint NULL, + NUM_DISTINCTS bigint NULL, + BIT_VECTOR varbinary(max) NULL, + NUM_FALSES bigint NULL, + NUM_NULLS bigint NOT NULL, + NUM_TRUES bigint NULL, + PART_ID bigint NULL, + PARTITION_NAME nvarchar(767) NOT NULL, + "TABLE_NAME" nvarchar(256) NOT NULL, + "CAT_NAME" nvarchar(256) NOT NULL +); + +ALTER TABLE PART_COL_STATS ADD CONSTRAINT PART_COL_STATS_PK PRIMARY KEY (CS_ID); + +CREATE INDEX PCS_STATS_IDX ON PART_COL_STATS (CAT_NAME, DB_NAME,TABLE_NAME,COLUMN_NAME,PARTITION_NAME); + +-- Table PART_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MPartitionPrivilege] +CREATE TABLE PART_PRIVS +( + PART_GRANT_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PART_ID bigint NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + PART_PRIV nvarchar(128) NULL, + AUTHORIZER nvarchar(128) NULL +); + +ALTER TABLE PART_PRIVS ADD CONSTRAINT PART_PRIVS_PK PRIMARY KEY (PART_GRANT_ID); + +-- Table SKEWED_STRING_LIST for classes [org.apache.hadoop.hive.metastore.model.MStringList] +CREATE TABLE SKEWED_STRING_LIST +( + STRING_LIST_ID bigint NOT NULL +); + +ALTER TABLE SKEWED_STRING_LIST ADD CONSTRAINT SKEWED_STRING_LIST_PK PRIMARY KEY (STRING_LIST_ID); + +-- Table ROLES for classes [org.apache.hadoop.hive.metastore.model.MRole] +CREATE TABLE ROLES +( + ROLE_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + OWNER_NAME nvarchar(128) NULL, + ROLE_NAME nvarchar(128) NULL +); + +ALTER TABLE ROLES ADD CONSTRAINT ROLES_PK PRIMARY KEY (ROLE_ID); + +-- Table PARTITIONS for classes [org.apache.hadoop.hive.metastore.model.MPartition] +CREATE TABLE PARTITIONS +( + PART_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + LAST_ACCESS_TIME int NOT NULL, + PART_NAME nvarchar(767) NULL, + SD_ID bigint NULL, + TBL_ID bigint NULL +); + +ALTER TABLE PARTITIONS ADD CONSTRAINT PARTITIONS_PK PRIMARY KEY (PART_ID); + +-- Table CDS for classes [org.apache.hadoop.hive.metastore.model.MColumnDescriptor] +CREATE TABLE CDS +( + CD_ID bigint NOT NULL +); + +ALTER TABLE CDS ADD CONSTRAINT CDS_PK PRIMARY KEY (CD_ID); + +-- Table VERSION for classes [org.apache.hadoop.hive.metastore.model.MVersionTable] +CREATE TABLE VERSION +( + VER_ID bigint NOT NULL, + SCHEMA_VERSION nvarchar(127) NOT NULL, + VERSION_COMMENT nvarchar(255) NOT NULL +); + +ALTER TABLE VERSION ADD CONSTRAINT VERSION_PK PRIMARY KEY (VER_ID); + +-- Table GLOBAL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MGlobalPrivilege] +CREATE TABLE GLOBAL_PRIVS +( + USER_GRANT_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + USER_PRIV nvarchar(128) NULL, + AUTHORIZER nvarchar(128) NULL +); + +ALTER TABLE GLOBAL_PRIVS ADD CONSTRAINT GLOBAL_PRIVS_PK PRIMARY KEY (USER_GRANT_ID); + +-- Table PART_COL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MPartitionColumnPrivilege] +CREATE TABLE PART_COL_PRIVS +( + PART_COLUMN_GRANT_ID bigint NOT NULL, + "COLUMN_NAME" nvarchar(767) NULL, + CREATE_TIME int NOT NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PART_ID bigint NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + PART_COL_PRIV nvarchar(128) NULL, + AUTHORIZER nvarchar(128) NULL +); + +ALTER TABLE PART_COL_PRIVS ADD CONSTRAINT PART_COL_PRIVS_PK PRIMARY KEY (PART_COLUMN_GRANT_ID); + +-- Table DB_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MDBPrivilege] +CREATE TABLE DB_PRIVS +( + DB_GRANT_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + DB_ID bigint NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + DB_PRIV nvarchar(128) NULL, + AUTHORIZER nvarchar(128) NULL +); + +ALTER TABLE DB_PRIVS ADD CONSTRAINT DB_PRIVS_PK PRIMARY KEY (DB_GRANT_ID); + +-- Table TAB_COL_STATS for classes [org.apache.hadoop.hive.metastore.model.MTableColumnStatistics] +CREATE TABLE TAB_COL_STATS +( + CS_ID bigint NOT NULL, + AVG_COL_LEN float NULL, + "COLUMN_NAME" nvarchar(767) NOT NULL, + COLUMN_TYPE nvarchar(128) NOT NULL, + DB_NAME nvarchar(128) NOT NULL, + BIG_DECIMAL_HIGH_VALUE nvarchar(255) NULL, + BIG_DECIMAL_LOW_VALUE nvarchar(255) NULL, + DOUBLE_HIGH_VALUE float NULL, + DOUBLE_LOW_VALUE float NULL, + LAST_ANALYZED bigint NOT NULL, + LONG_HIGH_VALUE bigint NULL, + LONG_LOW_VALUE bigint NULL, + MAX_COL_LEN bigint NULL, + NUM_DISTINCTS bigint NULL, + BIT_VECTOR varbinary(max) NULL, + NUM_FALSES bigint NULL, + NUM_NULLS bigint NOT NULL, + NUM_TRUES bigint NULL, + TBL_ID bigint NULL, + "TABLE_NAME" nvarchar(256) NOT NULL, + "CAT_NAME" nvarchar(256) NOT NULL +); + +ALTER TABLE TAB_COL_STATS ADD CONSTRAINT TAB_COL_STATS_PK PRIMARY KEY (CS_ID); +CREATE INDEX TAB_COL_STATS_IDX ON TAB_COL_STATS (CAT_NAME, DB_NAME, TABLE_NAME, COLUMN_NAME); + +-- Table TYPES for classes [org.apache.hadoop.hive.metastore.model.MType] +CREATE TABLE TYPES +( + TYPES_ID bigint NOT NULL, + TYPE_NAME nvarchar(128) NULL, + TYPE1 nvarchar(767) NULL, + TYPE2 nvarchar(767) NULL +); + +ALTER TABLE TYPES ADD CONSTRAINT TYPES_PK PRIMARY KEY (TYPES_ID); + +-- Table TBL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MTablePrivilege] +CREATE TABLE TBL_PRIVS +( + TBL_GRANT_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + TBL_PRIV nvarchar(128) NULL, + TBL_ID bigint NULL, + AUTHORIZER nvarchar(128) NULL +); + +ALTER TABLE TBL_PRIVS ADD CONSTRAINT TBL_PRIVS_PK PRIMARY KEY (TBL_GRANT_ID); + +-- Table DBS for classes [org.apache.hadoop.hive.metastore.model.MDatabase] +CREATE TABLE DBS +( + DB_ID bigint NOT NULL, + "DESC" nvarchar(4000) NULL, + DB_LOCATION_URI nvarchar(4000) NOT NULL, + "NAME" nvarchar(128) NULL, + OWNER_NAME nvarchar(128) NULL, + OWNER_TYPE nvarchar(10) NULL, + CTLG_NAME nvarchar(256) +); + +ALTER TABLE DBS ADD CONSTRAINT DBS_PK PRIMARY KEY (DB_ID); + +-- Table TBL_COL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MTableColumnPrivilege] +CREATE TABLE TBL_COL_PRIVS +( + TBL_COLUMN_GRANT_ID bigint NOT NULL, + "COLUMN_NAME" nvarchar(767) NULL, + CREATE_TIME int NOT NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + TBL_COL_PRIV nvarchar(128) NULL, + TBL_ID bigint NULL, + AUTHORIZER nvarchar(128) NULL +); + +ALTER TABLE TBL_COL_PRIVS ADD CONSTRAINT TBL_COL_PRIVS_PK PRIMARY KEY (TBL_COLUMN_GRANT_ID); + +-- Table DELEGATION_TOKENS for classes [org.apache.hadoop.hive.metastore.model.MDelegationToken] +CREATE TABLE DELEGATION_TOKENS +( + TOKEN_IDENT nvarchar(767) NOT NULL, + TOKEN nvarchar(767) NULL +); + +ALTER TABLE DELEGATION_TOKENS ADD CONSTRAINT DELEGATION_TOKENS_PK PRIMARY KEY (TOKEN_IDENT); + +-- Table SERDES for classes [org.apache.hadoop.hive.metastore.model.MSerDeInfo] +CREATE TABLE SERDES +( + SERDE_ID bigint NOT NULL, + "NAME" nvarchar(128) NULL, + SLIB nvarchar(4000) NULL, + "DESCRIPTION" nvarchar(4000), + "SERIALIZER_CLASS" nvarchar(4000), + "DESERIALIZER_CLASS" nvarchar(4000), + "SERDE_TYPE" int +); + +ALTER TABLE SERDES ADD CONSTRAINT SERDES_PK PRIMARY KEY (SERDE_ID); + +-- Table FUNCS for classes [org.apache.hadoop.hive.metastore.model.MFunction] +CREATE TABLE FUNCS +( + FUNC_ID bigint NOT NULL, + CLASS_NAME nvarchar(4000) NULL, + CREATE_TIME int NOT NULL, + DB_ID bigint NULL, + FUNC_NAME nvarchar(128) NULL, + FUNC_TYPE int NOT NULL, + OWNER_NAME nvarchar(128) NULL, + OWNER_TYPE nvarchar(10) NULL +); + +ALTER TABLE FUNCS ADD CONSTRAINT FUNCS_PK PRIMARY KEY (FUNC_ID); + +-- Table ROLE_MAP for classes [org.apache.hadoop.hive.metastore.model.MRoleMap] +CREATE TABLE ROLE_MAP +( + ROLE_GRANT_ID bigint NOT NULL, + ADD_TIME int NOT NULL, + GRANT_OPTION smallint NOT NULL CHECK (GRANT_OPTION IN (0,1)), + GRANTOR nvarchar(128) NULL, + GRANTOR_TYPE nvarchar(128) NULL, + PRINCIPAL_NAME nvarchar(128) NULL, + PRINCIPAL_TYPE nvarchar(128) NULL, + ROLE_ID bigint NULL +); + +ALTER TABLE ROLE_MAP ADD CONSTRAINT ROLE_MAP_PK PRIMARY KEY (ROLE_GRANT_ID); + +-- Table TBLS for classes [org.apache.hadoop.hive.metastore.model.MTable] +CREATE TABLE TBLS +( + TBL_ID bigint NOT NULL, + CREATE_TIME int NOT NULL, + DB_ID bigint NULL, + LAST_ACCESS_TIME int NOT NULL, + OWNER nvarchar(767) NULL, + OWNER_TYPE nvarchar(10) NULL, + RETENTION int NOT NULL, + SD_ID bigint NULL, + TBL_NAME nvarchar(256) NULL, + TBL_TYPE nvarchar(128) NULL, + VIEW_EXPANDED_TEXT text NULL, + VIEW_ORIGINAL_TEXT text NULL, + IS_REWRITE_ENABLED bit NOT NULL DEFAULT 0 +); + +ALTER TABLE TBLS ADD CONSTRAINT TBLS_PK PRIMARY KEY (TBL_ID); + +-- Table MV_CREATION_METADATA for classes [org.apache.hadoop.hive.metastore.model.MCreationMetadata] +CREATE TABLE MV_CREATION_METADATA +( + MV_CREATION_METADATA_ID bigint NOT NULL, + CAT_NAME nvarchar(256) NOT NULL, + DB_NAME nvarchar(128) NOT NULL, + TBL_NAME nvarchar(256) NOT NULL, + TXN_LIST text NULL, + MATERIALIZATION_TIME bigint NOT NULL +); + +ALTER TABLE MV_CREATION_METADATA ADD CONSTRAINT MV_CREATION_METADATA_PK PRIMARY KEY (MV_CREATION_METADATA_ID); +CREATE INDEX MV_UNIQUE_TABLE ON MV_CREATION_METADATA (TBL_NAME,DB_NAME); + + +CREATE TABLE MV_TABLES_USED +( + MV_CREATION_METADATA_ID bigint NOT NULL, + TBL_ID bigint NOT NULL +); + +ALTER TABLE MV_TABLES_USED WITH CHECK ADD FOREIGN KEY(MV_CREATION_METADATA_ID) REFERENCES MV_CREATION_METADATA (MV_CREATION_METADATA_ID); +ALTER TABLE MV_TABLES_USED WITH CHECK ADD FOREIGN KEY(TBL_ID) REFERENCES TBLS (TBL_ID); + +-- Table SDS for classes [org.apache.hadoop.hive.metastore.model.MStorageDescriptor] +CREATE TABLE SDS +( + SD_ID bigint NOT NULL, + CD_ID bigint NULL, + INPUT_FORMAT nvarchar(4000) NULL, + IS_COMPRESSED bit NOT NULL, + IS_STOREDASSUBDIRECTORIES bit NOT NULL, + LOCATION nvarchar(4000) NULL, + NUM_BUCKETS int NOT NULL, + OUTPUT_FORMAT nvarchar(4000) NULL, + SERDE_ID bigint NULL +); + +ALTER TABLE SDS ADD CONSTRAINT SDS_PK PRIMARY KEY (SD_ID); + +-- Table PARTITION_EVENTS for classes [org.apache.hadoop.hive.metastore.model.MPartitionEvent] +CREATE TABLE PARTITION_EVENTS +( + PART_NAME_ID bigint NOT NULL, + CAT_NAME nvarchar(256) NULL, + DB_NAME nvarchar(128) NULL, + EVENT_TIME bigint NOT NULL, + EVENT_TYPE int NOT NULL, + PARTITION_NAME nvarchar(767) NULL, + TBL_NAME nvarchar(256) NULL +); + +ALTER TABLE PARTITION_EVENTS ADD CONSTRAINT PARTITION_EVENTS_PK PRIMARY KEY (PART_NAME_ID); + +-- Table SORT_COLS for join relationship +CREATE TABLE SORT_COLS +( + SD_ID bigint NOT NULL, + "COLUMN_NAME" nvarchar(767) NULL, + "ORDER" int NOT NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE SORT_COLS ADD CONSTRAINT SORT_COLS_PK PRIMARY KEY (SD_ID,INTEGER_IDX); + +-- Table SKEWED_COL_NAMES for join relationship +CREATE TABLE SKEWED_COL_NAMES +( + SD_ID bigint NOT NULL, + SKEWED_COL_NAME nvarchar(255) NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE SKEWED_COL_NAMES ADD CONSTRAINT SKEWED_COL_NAMES_PK PRIMARY KEY (SD_ID,INTEGER_IDX); + +-- Table SKEWED_COL_VALUE_LOC_MAP for join relationship +CREATE TABLE SKEWED_COL_VALUE_LOC_MAP +( + SD_ID bigint NOT NULL, + STRING_LIST_ID_KID bigint NOT NULL, + LOCATION nvarchar(4000) NULL +); + +ALTER TABLE SKEWED_COL_VALUE_LOC_MAP ADD CONSTRAINT SKEWED_COL_VALUE_LOC_MAP_PK PRIMARY KEY (SD_ID,STRING_LIST_ID_KID); + +-- Table SKEWED_STRING_LIST_VALUES for join relationship +CREATE TABLE SKEWED_STRING_LIST_VALUES +( + STRING_LIST_ID bigint NOT NULL, + STRING_LIST_VALUE nvarchar(255) NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE SKEWED_STRING_LIST_VALUES ADD CONSTRAINT SKEWED_STRING_LIST_VALUES_PK PRIMARY KEY (STRING_LIST_ID,INTEGER_IDX); + +-- Table PARTITION_KEY_VALS for join relationship +CREATE TABLE PARTITION_KEY_VALS +( + PART_ID bigint NOT NULL, + PART_KEY_VAL nvarchar(255) NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE PARTITION_KEY_VALS ADD CONSTRAINT PARTITION_KEY_VALS_PK PRIMARY KEY (PART_ID,INTEGER_IDX); + +-- Table PARTITION_KEYS for join relationship +CREATE TABLE PARTITION_KEYS +( + TBL_ID bigint NOT NULL, + PKEY_COMMENT nvarchar(4000) NULL, + PKEY_NAME nvarchar(128) NOT NULL, + PKEY_TYPE nvarchar(767) NOT NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE PARTITION_KEYS ADD CONSTRAINT PARTITION_KEY_PK PRIMARY KEY (TBL_ID,PKEY_NAME); + +-- Table SKEWED_VALUES for join relationship +CREATE TABLE SKEWED_VALUES +( + SD_ID_OID bigint NOT NULL, + STRING_LIST_ID_EID bigint NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE SKEWED_VALUES ADD CONSTRAINT SKEWED_VALUES_PK PRIMARY KEY (SD_ID_OID,INTEGER_IDX); + +-- Table SD_PARAMS for join relationship +CREATE TABLE SD_PARAMS +( + SD_ID bigint NOT NULL, + PARAM_KEY nvarchar(256) NOT NULL, + PARAM_VALUE varchar(max) NULL +); + +ALTER TABLE SD_PARAMS ADD CONSTRAINT SD_PARAMS_PK PRIMARY KEY (SD_ID,PARAM_KEY); + +-- Table FUNC_RU for join relationship +CREATE TABLE FUNC_RU +( + FUNC_ID bigint NOT NULL, + RESOURCE_TYPE int NOT NULL, + RESOURCE_URI nvarchar(4000) NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE FUNC_RU ADD CONSTRAINT FUNC_RU_PK PRIMARY KEY (FUNC_ID,INTEGER_IDX); + +-- Table TYPE_FIELDS for join relationship +CREATE TABLE TYPE_FIELDS +( + TYPE_NAME bigint NOT NULL, + COMMENT nvarchar(256) NULL, + FIELD_NAME nvarchar(128) NOT NULL, + FIELD_TYPE nvarchar(767) NOT NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE TYPE_FIELDS ADD CONSTRAINT TYPE_FIELDS_PK PRIMARY KEY (TYPE_NAME,FIELD_NAME); + +-- Table BUCKETING_COLS for join relationship +CREATE TABLE BUCKETING_COLS +( + SD_ID bigint NOT NULL, + BUCKET_COL_NAME nvarchar(255) NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE BUCKETING_COLS ADD CONSTRAINT BUCKETING_COLS_PK PRIMARY KEY (SD_ID,INTEGER_IDX); + +-- Table DATABASE_PARAMS for join relationship +CREATE TABLE DATABASE_PARAMS +( + DB_ID bigint NOT NULL, + PARAM_KEY nvarchar(180) NOT NULL, + PARAM_VALUE nvarchar(4000) NULL +); + +ALTER TABLE DATABASE_PARAMS ADD CONSTRAINT DATABASE_PARAMS_PK PRIMARY KEY (DB_ID,PARAM_KEY); + +-- Table INDEX_PARAMS for join relationship +CREATE TABLE INDEX_PARAMS +( + INDEX_ID bigint NOT NULL, + PARAM_KEY nvarchar(256) NOT NULL, + PARAM_VALUE nvarchar(4000) NULL +); + +ALTER TABLE INDEX_PARAMS ADD CONSTRAINT INDEX_PARAMS_PK PRIMARY KEY (INDEX_ID,PARAM_KEY); + +-- Table COLUMNS_V2 for join relationship +CREATE TABLE COLUMNS_V2 +( + CD_ID bigint NOT NULL, + COMMENT nvarchar(256) NULL, + "COLUMN_NAME" nvarchar(767) NOT NULL, + TYPE_NAME varchar(max) NOT NULL, + INTEGER_IDX int NOT NULL +); + +ALTER TABLE COLUMNS_V2 ADD CONSTRAINT COLUMNS_PK PRIMARY KEY (CD_ID,"COLUMN_NAME"); + +-- Table SERDE_PARAMS for join relationship +CREATE TABLE SERDE_PARAMS +( + SERDE_ID bigint NOT NULL, + PARAM_KEY nvarchar(256) NOT NULL, + PARAM_VALUE varchar(max) NULL +); + +ALTER TABLE SERDE_PARAMS ADD CONSTRAINT SERDE_PARAMS_PK PRIMARY KEY (SERDE_ID,PARAM_KEY); + +-- Table PARTITION_PARAMS for join relationship +CREATE TABLE PARTITION_PARAMS +( + PART_ID bigint NOT NULL, + PARAM_KEY nvarchar(256) NOT NULL, + PARAM_VALUE nvarchar(4000) NULL +); + +ALTER TABLE PARTITION_PARAMS ADD CONSTRAINT PARTITION_PARAMS_PK PRIMARY KEY (PART_ID,PARAM_KEY); + +-- Table TABLE_PARAMS for join relationship +CREATE TABLE TABLE_PARAMS +( + TBL_ID bigint NOT NULL, + PARAM_KEY nvarchar(256) NOT NULL, + PARAM_VALUE varchar(max) NULL +); + +ALTER TABLE TABLE_PARAMS ADD CONSTRAINT TABLE_PARAMS_PK PRIMARY KEY (TBL_ID,PARAM_KEY); + +CREATE TABLE NOTIFICATION_LOG +( + NL_ID bigint NOT NULL, + EVENT_ID bigint NOT NULL, + EVENT_TIME int NOT NULL, + EVENT_TYPE nvarchar(32) NOT NULL, + CAT_NAME nvarchar(128) NULL, + DB_NAME nvarchar(128) NULL, + TBL_NAME nvarchar(256) NULL, + MESSAGE_FORMAT nvarchar(16), + MESSAGE text NULL +); + +ALTER TABLE NOTIFICATION_LOG ADD CONSTRAINT NOTIFICATION_LOG_PK PRIMARY KEY (NL_ID); + +CREATE TABLE NOTIFICATION_SEQUENCE +( + NNI_ID bigint NOT NULL, + NEXT_EVENT_ID bigint NOT NULL +); + +ALTER TABLE NOTIFICATION_SEQUENCE ADD CONSTRAINT NOTIFICATION_SEQUENCE_PK PRIMARY KEY (NNI_ID); + +-- Tables to manage resource plans. + +CREATE TABLE WM_RESOURCEPLAN +( + RP_ID bigint NOT NULL, + "NAME" nvarchar(128) NOT NULL, + QUERY_PARALLELISM int, + STATUS nvarchar(20) NOT NULL, + DEFAULT_POOL_ID bigint +); + +ALTER TABLE WM_RESOURCEPLAN ADD CONSTRAINT WM_RESOURCEPLAN_PK PRIMARY KEY (RP_ID); + +CREATE TABLE WM_POOL +( + POOL_ID bigint NOT NULL, + RP_ID bigint NOT NULL, + PATH nvarchar(1024) NOT NULL, + ALLOC_FRACTION float, + QUERY_PARALLELISM int, + SCHEDULING_POLICY nvarchar(1024) +); + +ALTER TABLE WM_POOL ADD CONSTRAINT WM_POOL_PK PRIMARY KEY (POOL_ID); + +CREATE TABLE WM_TRIGGER +( + TRIGGER_ID bigint NOT NULL, + RP_ID bigint NOT NULL, + "NAME" nvarchar(128) NOT NULL, + TRIGGER_EXPRESSION nvarchar(1024), + ACTION_EXPRESSION nvarchar(1024), + IS_IN_UNMANAGED bit NOT NULL DEFAULT 0 +); + +ALTER TABLE WM_TRIGGER ADD CONSTRAINT WM_TRIGGER_PK PRIMARY KEY (TRIGGER_ID); + +CREATE TABLE WM_POOL_TO_TRIGGER +( + POOL_ID bigint NOT NULL, + TRIGGER_ID bigint NOT NULL +); + +ALTER TABLE WM_POOL_TO_TRIGGER ADD CONSTRAINT WM_POOL_TO_TRIGGER_PK PRIMARY KEY (POOL_ID, TRIGGER_ID); + +CREATE TABLE WM_MAPPING +( + MAPPING_ID bigint NOT NULL, + RP_ID bigint NOT NULL, + ENTITY_TYPE nvarchar(128) NOT NULL, + ENTITY_NAME nvarchar(128) NOT NULL, + POOL_ID bigint, + ORDERING int +); + +ALTER TABLE WM_MAPPING ADD CONSTRAINT WM_MAPPING_PK PRIMARY KEY (MAPPING_ID); + +CREATE TABLE CTLGS ( + CTLG_ID bigint primary key, + "NAME" nvarchar(256), + "DESC" nvarchar(4000), + LOCATION_URI nvarchar(4000) not null +); + +CREATE UNIQUE INDEX UNIQUE_CTLG ON CTLGS ("NAME"); + +-- Constraints for table MASTER_KEYS for class(es) [org.apache.hadoop.hive.metastore.model.MMasterKey] + +-- Constraints for table IDXS for class(es) [org.apache.hadoop.hive.metastore.model.MIndex] +ALTER TABLE IDXS ADD CONSTRAINT IDXS_FK1 FOREIGN KEY (INDEX_TBL_ID) REFERENCES TBLS (TBL_ID) ; + +ALTER TABLE IDXS ADD CONSTRAINT IDXS_FK2 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +ALTER TABLE IDXS ADD CONSTRAINT IDXS_FK3 FOREIGN KEY (ORIG_TBL_ID) REFERENCES TBLS (TBL_ID) ; + +CREATE UNIQUE INDEX UNIQUEINDEX ON IDXS (INDEX_NAME,ORIG_TBL_ID); + +CREATE INDEX IDXS_N51 ON IDXS (SD_ID); + +CREATE INDEX IDXS_N50 ON IDXS (ORIG_TBL_ID); + +CREATE INDEX IDXS_N49 ON IDXS (INDEX_TBL_ID); + + +-- Constraints for table PART_COL_STATS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionColumnStatistics] +ALTER TABLE PART_COL_STATS ADD CONSTRAINT PART_COL_STATS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) ; + +CREATE INDEX PART_COL_STATS_N49 ON PART_COL_STATS (PART_ID); + + +-- Constraints for table PART_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionPrivilege] +ALTER TABLE PART_PRIVS ADD CONSTRAINT PART_PRIVS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) ; + +CREATE INDEX PARTPRIVILEGEINDEX ON PART_PRIVS (AUTHORIZER,PART_ID,PRINCIPAL_NAME,PRINCIPAL_TYPE,PART_PRIV,GRANTOR,GRANTOR_TYPE); + +CREATE INDEX PART_PRIVS_N49 ON PART_PRIVS (PART_ID); + + +-- Constraints for table SKEWED_STRING_LIST for class(es) [org.apache.hadoop.hive.metastore.model.MStringList] + +-- Constraints for table ROLES for class(es) [org.apache.hadoop.hive.metastore.model.MRole] +CREATE UNIQUE INDEX ROLEENTITYINDEX ON ROLES (ROLE_NAME); + + +-- Constraints for table PARTITIONS for class(es) [org.apache.hadoop.hive.metastore.model.MPartition] +ALTER TABLE PARTITIONS ADD CONSTRAINT PARTITIONS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) ; + +ALTER TABLE PARTITIONS ADD CONSTRAINT PARTITIONS_FK2 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +CREATE INDEX PARTITIONS_N49 ON PARTITIONS (SD_ID); + +CREATE INDEX PARTITIONS_N50 ON PARTITIONS (TBL_ID); + +CREATE UNIQUE INDEX UNIQUEPARTITION ON PARTITIONS (PART_NAME,TBL_ID); + + +-- Constraints for table CDS for class(es) [org.apache.hadoop.hive.metastore.model.MColumnDescriptor] + +-- Constraints for table VERSION for class(es) [org.apache.hadoop.hive.metastore.model.MVersionTable] + +-- Constraints for table GLOBAL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MGlobalPrivilege] +CREATE UNIQUE INDEX GLOBALPRIVILEGEINDEX ON GLOBAL_PRIVS (AUTHORIZER,PRINCIPAL_NAME,PRINCIPAL_TYPE,USER_PRIV,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table PART_COL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionColumnPrivilege] +ALTER TABLE PART_COL_PRIVS ADD CONSTRAINT PART_COL_PRIVS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) ; + +CREATE INDEX PART_COL_PRIVS_N49 ON PART_COL_PRIVS (PART_ID); + +CREATE INDEX PARTITIONCOLUMNPRIVILEGEINDEX ON PART_COL_PRIVS (AUTHORIZER,PART_ID,"COLUMN_NAME",PRINCIPAL_NAME,PRINCIPAL_TYPE,PART_COL_PRIV,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table DB_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MDBPrivilege] +ALTER TABLE DB_PRIVS ADD CONSTRAINT DB_PRIVS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) ; + +CREATE UNIQUE INDEX DBPRIVILEGEINDEX ON DB_PRIVS (AUTHORIZER,DB_ID,PRINCIPAL_NAME,PRINCIPAL_TYPE,DB_PRIV,GRANTOR,GRANTOR_TYPE); + +CREATE INDEX DB_PRIVS_N49 ON DB_PRIVS (DB_ID); + + +-- Constraints for table TAB_COL_STATS for class(es) [org.apache.hadoop.hive.metastore.model.MTableColumnStatistics] +ALTER TABLE TAB_COL_STATS ADD CONSTRAINT TAB_COL_STATS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) ; + +CREATE INDEX TAB_COL_STATS_N49 ON TAB_COL_STATS (TBL_ID); + + +-- Constraints for table TYPES for class(es) [org.apache.hadoop.hive.metastore.model.MType] +CREATE UNIQUE INDEX UNIQUETYPE ON TYPES (TYPE_NAME); + + +-- Constraints for table TBL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MTablePrivilege] +ALTER TABLE TBL_PRIVS ADD CONSTRAINT TBL_PRIVS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) ; + +CREATE INDEX TBL_PRIVS_N49 ON TBL_PRIVS (TBL_ID); + +CREATE INDEX TABLEPRIVILEGEINDEX ON TBL_PRIVS (AUTHORIZER,TBL_ID,PRINCIPAL_NAME,PRINCIPAL_TYPE,TBL_PRIV,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table DBS for class(es) [org.apache.hadoop.hive.metastore.model.MDatabase] +CREATE UNIQUE INDEX UNIQUEDATABASE ON DBS ("NAME", "CTLG_NAME"); + + +-- Constraints for table TBL_COL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MTableColumnPrivilege] +ALTER TABLE TBL_COL_PRIVS ADD CONSTRAINT TBL_COL_PRIVS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) ; + +CREATE INDEX TABLECOLUMNPRIVILEGEINDEX ON TBL_COL_PRIVS (AUTHORIZER,TBL_ID,"COLUMN_NAME",PRINCIPAL_NAME,PRINCIPAL_TYPE,TBL_COL_PRIV,GRANTOR,GRANTOR_TYPE); + +CREATE INDEX TBL_COL_PRIVS_N49 ON TBL_COL_PRIVS (TBL_ID); + + +-- Constraints for table DELEGATION_TOKENS for class(es) [org.apache.hadoop.hive.metastore.model.MDelegationToken] + +-- Constraints for table SERDES for class(es) [org.apache.hadoop.hive.metastore.model.MSerDeInfo] + +-- Constraints for table FUNCS for class(es) [org.apache.hadoop.hive.metastore.model.MFunction] +ALTER TABLE FUNCS ADD CONSTRAINT FUNCS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) ; + +CREATE UNIQUE INDEX UNIQUEFUNCTION ON FUNCS (FUNC_NAME,DB_ID); + +CREATE INDEX FUNCS_N49 ON FUNCS (DB_ID); + + +-- Constraints for table ROLE_MAP for class(es) [org.apache.hadoop.hive.metastore.model.MRoleMap] +ALTER TABLE ROLE_MAP ADD CONSTRAINT ROLE_MAP_FK1 FOREIGN KEY (ROLE_ID) REFERENCES ROLES (ROLE_ID) ; + +CREATE INDEX ROLE_MAP_N49 ON ROLE_MAP (ROLE_ID); + +CREATE UNIQUE INDEX USERROLEMAPINDEX ON ROLE_MAP (PRINCIPAL_NAME,ROLE_ID,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table TBLS for class(es) [org.apache.hadoop.hive.metastore.model.MTable] +ALTER TABLE TBLS ADD CONSTRAINT TBLS_FK2 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +ALTER TABLE TBLS ADD CONSTRAINT TBLS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) ; + +CREATE INDEX TBLS_N50 ON TBLS (SD_ID); + +CREATE UNIQUE INDEX UNIQUETABLE ON TBLS (TBL_NAME,DB_ID); + +CREATE INDEX TBLS_N49 ON TBLS (DB_ID); + + +-- Constraints for table SDS for class(es) [org.apache.hadoop.hive.metastore.model.MStorageDescriptor] +ALTER TABLE SDS ADD CONSTRAINT SDS_FK1 FOREIGN KEY (SERDE_ID) REFERENCES SERDES (SERDE_ID) ; + +ALTER TABLE SDS ADD CONSTRAINT SDS_FK2 FOREIGN KEY (CD_ID) REFERENCES CDS (CD_ID) ; + +CREATE INDEX SDS_N50 ON SDS (CD_ID); + +CREATE INDEX SDS_N49 ON SDS (SERDE_ID); + + +-- Constraints for table PARTITION_EVENTS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionEvent] +CREATE INDEX PARTITIONEVENTINDEX ON PARTITION_EVENTS (PARTITION_NAME); + + +-- Constraints for table SORT_COLS +ALTER TABLE SORT_COLS ADD CONSTRAINT SORT_COLS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +CREATE INDEX SORT_COLS_N49 ON SORT_COLS (SD_ID); + + +-- Constraints for table SKEWED_COL_NAMES +ALTER TABLE SKEWED_COL_NAMES ADD CONSTRAINT SKEWED_COL_NAMES_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +CREATE INDEX SKEWED_COL_NAMES_N49 ON SKEWED_COL_NAMES (SD_ID); + + +-- Constraints for table SKEWED_COL_VALUE_LOC_MAP +ALTER TABLE SKEWED_COL_VALUE_LOC_MAP ADD CONSTRAINT SKEWED_COL_VALUE_LOC_MAP_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +ALTER TABLE SKEWED_COL_VALUE_LOC_MAP ADD CONSTRAINT SKEWED_COL_VALUE_LOC_MAP_FK2 FOREIGN KEY (STRING_LIST_ID_KID) REFERENCES SKEWED_STRING_LIST (STRING_LIST_ID) ; + +CREATE INDEX SKEWED_COL_VALUE_LOC_MAP_N50 ON SKEWED_COL_VALUE_LOC_MAP (STRING_LIST_ID_KID); + +CREATE INDEX SKEWED_COL_VALUE_LOC_MAP_N49 ON SKEWED_COL_VALUE_LOC_MAP (SD_ID); + + +-- Constraints for table SKEWED_STRING_LIST_VALUES +ALTER TABLE SKEWED_STRING_LIST_VALUES ADD CONSTRAINT SKEWED_STRING_LIST_VALUES_FK1 FOREIGN KEY (STRING_LIST_ID) REFERENCES SKEWED_STRING_LIST (STRING_LIST_ID) ; + +CREATE INDEX SKEWED_STRING_LIST_VALUES_N49 ON SKEWED_STRING_LIST_VALUES (STRING_LIST_ID); + + +-- Constraints for table PARTITION_KEY_VALS +ALTER TABLE PARTITION_KEY_VALS ADD CONSTRAINT PARTITION_KEY_VALS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) ; + +CREATE INDEX PARTITION_KEY_VALS_N49 ON PARTITION_KEY_VALS (PART_ID); + + +-- Constraints for table PARTITION_KEYS +ALTER TABLE PARTITION_KEYS ADD CONSTRAINT PARTITION_KEYS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) ; + +CREATE INDEX PARTITION_KEYS_N49 ON PARTITION_KEYS (TBL_ID); + + +-- Constraints for table SKEWED_VALUES +ALTER TABLE SKEWED_VALUES ADD CONSTRAINT SKEWED_VALUES_FK1 FOREIGN KEY (SD_ID_OID) REFERENCES SDS (SD_ID) ; + +ALTER TABLE SKEWED_VALUES ADD CONSTRAINT SKEWED_VALUES_FK2 FOREIGN KEY (STRING_LIST_ID_EID) REFERENCES SKEWED_STRING_LIST (STRING_LIST_ID) ; + +CREATE INDEX SKEWED_VALUES_N50 ON SKEWED_VALUES (SD_ID_OID); + +CREATE INDEX SKEWED_VALUES_N49 ON SKEWED_VALUES (STRING_LIST_ID_EID); + + +-- Constraints for table SD_PARAMS +ALTER TABLE SD_PARAMS ADD CONSTRAINT SD_PARAMS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +CREATE INDEX SD_PARAMS_N49 ON SD_PARAMS (SD_ID); + + +-- Constraints for table FUNC_RU +ALTER TABLE FUNC_RU ADD CONSTRAINT FUNC_RU_FK1 FOREIGN KEY (FUNC_ID) REFERENCES FUNCS (FUNC_ID) ; + +CREATE INDEX FUNC_RU_N49 ON FUNC_RU (FUNC_ID); + + +-- Constraints for table TYPE_FIELDS +ALTER TABLE TYPE_FIELDS ADD CONSTRAINT TYPE_FIELDS_FK1 FOREIGN KEY (TYPE_NAME) REFERENCES TYPES (TYPES_ID) ; + +CREATE INDEX TYPE_FIELDS_N49 ON TYPE_FIELDS (TYPE_NAME); + + +-- Constraints for table BUCKETING_COLS +ALTER TABLE BUCKETING_COLS ADD CONSTRAINT BUCKETING_COLS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) ; + +CREATE INDEX BUCKETING_COLS_N49 ON BUCKETING_COLS (SD_ID); + + +-- Constraints for table DATABASE_PARAMS +ALTER TABLE DATABASE_PARAMS ADD CONSTRAINT DATABASE_PARAMS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) ; + +CREATE INDEX DATABASE_PARAMS_N49 ON DATABASE_PARAMS (DB_ID); + + +-- Constraints for table INDEX_PARAMS +ALTER TABLE INDEX_PARAMS ADD CONSTRAINT INDEX_PARAMS_FK1 FOREIGN KEY (INDEX_ID) REFERENCES IDXS (INDEX_ID) ; + +CREATE INDEX INDEX_PARAMS_N49 ON INDEX_PARAMS (INDEX_ID); + + +-- Constraints for table COLUMNS_V2 +ALTER TABLE COLUMNS_V2 ADD CONSTRAINT COLUMNS_V2_FK1 FOREIGN KEY (CD_ID) REFERENCES CDS (CD_ID) ; + +CREATE INDEX COLUMNS_V2_N49 ON COLUMNS_V2 (CD_ID); + + +-- Constraints for table SERDE_PARAMS +ALTER TABLE SERDE_PARAMS ADD CONSTRAINT SERDE_PARAMS_FK1 FOREIGN KEY (SERDE_ID) REFERENCES SERDES (SERDE_ID) ; + +CREATE INDEX SERDE_PARAMS_N49 ON SERDE_PARAMS (SERDE_ID); + + +-- Constraints for table PARTITION_PARAMS +ALTER TABLE PARTITION_PARAMS ADD CONSTRAINT PARTITION_PARAMS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) ; + +CREATE INDEX PARTITION_PARAMS_N49 ON PARTITION_PARAMS (PART_ID); + + +-- Constraints for table TABLE_PARAMS +ALTER TABLE TABLE_PARAMS ADD CONSTRAINT TABLE_PARAMS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) ; + +CREATE INDEX TABLE_PARAMS_N49 ON TABLE_PARAMS (TBL_ID); + +-- Constraints for resource plan tables. + +CREATE UNIQUE INDEX UNIQUE_WM_RESOURCEPLAN ON WM_RESOURCEPLAN ("NAME"); + +CREATE UNIQUE INDEX UNIQUE_WM_POOL ON WM_POOL (RP_ID, PATH); + +ALTER TABLE WM_RESOURCEPLAN ADD CONSTRAINT WM_RESOURCEPLAN_FK1 FOREIGN KEY (DEFAULT_POOL_ID) REFERENCES WM_POOL (POOL_ID); + +ALTER TABLE WM_POOL ADD CONSTRAINT WM_POOL_FK1 FOREIGN KEY (RP_ID) REFERENCES WM_RESOURCEPLAN (RP_ID); + +CREATE UNIQUE INDEX UNIQUE_WM_TRIGGER ON WM_TRIGGER (RP_ID, "NAME"); + +ALTER TABLE WM_TRIGGER ADD CONSTRAINT WM_TRIGGER_FK1 FOREIGN KEY (RP_ID) REFERENCES WM_RESOURCEPLAN (RP_ID); + +ALTER TABLE WM_POOL_TO_TRIGGER ADD CONSTRAINT WM_POOL_TO_TRIGGER_FK1 FOREIGN KEY (POOL_ID) REFERENCES WM_POOL (POOL_ID); + +ALTER TABLE WM_POOL_TO_TRIGGER ADD CONSTRAINT WM_POOL_TO_TRIGGER_FK2 FOREIGN KEY (TRIGGER_ID) REFERENCES WM_TRIGGER (TRIGGER_ID); + +CREATE UNIQUE INDEX UNIQUE_WM_MAPPING ON WM_MAPPING (RP_ID, ENTITY_TYPE, ENTITY_NAME); + +ALTER TABLE WM_MAPPING ADD CONSTRAINT WM_MAPPING_FK1 FOREIGN KEY (RP_ID) REFERENCES WM_RESOURCEPLAN (RP_ID); + +ALTER TABLE DBS ADD CONSTRAINT "DBS_FK1" FOREIGN KEY ("CTLG_NAME") REFERENCES CTLGS ("NAME"); +-- ----------------------------------------------------------------------------------------------------------------------------------------------- +-- Transaction and Lock Tables +-- These are not part of package jdo, so if you are going to regenerate this file you need to manually add the following section back to the file. +-- ----------------------------------------------------------------------------------------------------------------------------------------------- +CREATE TABLE COMPACTION_QUEUE( + CQ_ID bigint NOT NULL, + CQ_DATABASE nvarchar(128) NOT NULL, + CQ_TABLE nvarchar(128) NOT NULL, + CQ_PARTITION nvarchar(767) NULL, + CQ_STATE char(1) NOT NULL, + CQ_TYPE char(1) NOT NULL, + CQ_TBLPROPERTIES nvarchar(2048) NULL, + CQ_WORKER_ID nvarchar(128) NULL, + CQ_START bigint NULL, + CQ_RUN_AS nvarchar(128) NULL, + CQ_HIGHEST_WRITE_ID bigint NULL, + CQ_META_INFO varbinary(2048) NULL, + CQ_HADOOP_JOB_ID nvarchar(128) NULL, +PRIMARY KEY CLUSTERED +( + CQ_ID ASC +) +); + +CREATE TABLE COMPLETED_COMPACTIONS ( + CC_ID bigint NOT NULL, + CC_DATABASE nvarchar(128) NOT NULL, + CC_TABLE nvarchar(128) NOT NULL, + CC_PARTITION nvarchar(767) NULL, + CC_STATE char(1) NOT NULL, + CC_TYPE char(1) NOT NULL, + CC_TBLPROPERTIES nvarchar(2048) NULL, + CC_WORKER_ID nvarchar(128) NULL, + CC_START bigint NULL, + CC_END bigint NULL, + CC_RUN_AS nvarchar(128) NULL, + CC_HIGHEST_WRITE_ID bigint NULL, + CC_META_INFO varbinary(2048) NULL, + CC_HADOOP_JOB_ID nvarchar(128) NULL, +PRIMARY KEY CLUSTERED +( + CC_ID ASC +) +); + +CREATE TABLE COMPLETED_TXN_COMPONENTS( + CTC_TXNID bigint NOT NULL, + CTC_DATABASE nvarchar(128) NOT NULL, + CTC_TABLE nvarchar(128) NULL, + CTC_PARTITION nvarchar(767) NULL, + CTC_TIMESTAMP datetime2 DEFAULT CURRENT_TIMESTAMP NOT NULL, + CTC_WRITEID bigint, + CTC_UPDATE_DELETE char(1) NOT NULL +); + +CREATE INDEX COMPLETED_TXN_COMPONENTS_IDX ON COMPLETED_TXN_COMPONENTS (CTC_DATABASE, CTC_TABLE, CTC_PARTITION); + +CREATE TABLE HIVE_LOCKS( + HL_LOCK_EXT_ID bigint NOT NULL, + HL_LOCK_INT_ID bigint NOT NULL, + HL_TXNID bigint NOT NULL, + HL_DB nvarchar(128) NOT NULL, + HL_TABLE nvarchar(128) NULL, + HL_PARTITION nvarchar(767) NULL, + HL_LOCK_STATE char(1) NOT NULL, + HL_LOCK_TYPE char(1) NOT NULL, + HL_LAST_HEARTBEAT bigint NOT NULL, + HL_ACQUIRED_AT bigint NULL, + HL_USER nvarchar(128) NOT NULL, + HL_HOST nvarchar(128) NOT NULL, + HL_HEARTBEAT_COUNT int NULL, + HL_AGENT_INFO nvarchar(128) NULL, + HL_BLOCKEDBY_EXT_ID bigint NULL, + HL_BLOCKEDBY_INT_ID bigint NULL, +PRIMARY KEY CLUSTERED +( + HL_LOCK_EXT_ID ASC, + HL_LOCK_INT_ID ASC +) +); + +CREATE TABLE NEXT_COMPACTION_QUEUE_ID( + NCQ_NEXT bigint NOT NULL +); + +INSERT INTO NEXT_COMPACTION_QUEUE_ID VALUES(1); + +CREATE TABLE NEXT_LOCK_ID( + NL_NEXT bigint NOT NULL +); + +INSERT INTO NEXT_LOCK_ID VALUES(1); + +CREATE TABLE NEXT_TXN_ID( + NTXN_NEXT bigint NOT NULL +); + +INSERT INTO NEXT_TXN_ID VALUES(1); + +CREATE TABLE TXNS( + TXN_ID bigint NOT NULL, + TXN_STATE char(1) NOT NULL, + TXN_STARTED bigint NOT NULL, + TXN_LAST_HEARTBEAT bigint NOT NULL, + TXN_USER nvarchar(128) NOT NULL, + TXN_HOST nvarchar(128) NOT NULL, + TXN_AGENT_INFO nvarchar(128) NULL, + TXN_META_INFO nvarchar(128) NULL, + TXN_HEARTBEAT_COUNT int NULL, + TXN_TYPE int NULL, +PRIMARY KEY CLUSTERED +( + TXN_ID ASC +) +); + +CREATE TABLE TXN_COMPONENTS( + TC_TXNID bigint NOT NULL, + TC_DATABASE nvarchar(128) NOT NULL, + TC_TABLE nvarchar(128) NULL, + TC_PARTITION nvarchar(767) NULL, + TC_OPERATION_TYPE char(1) NOT NULL, + TC_WRITEID bigint +); + +ALTER TABLE TXN_COMPONENTS WITH CHECK ADD FOREIGN KEY(TC_TXNID) REFERENCES TXNS (TXN_ID); + +CREATE INDEX TC_TXNID_INDEX ON TXN_COMPONENTS (TC_TXNID); + +CREATE TABLE AUX_TABLE ( + MT_KEY1 nvarchar(128) NOT NULL, + MT_KEY2 bigint NOT NULL, + MT_COMMENT nvarchar(255) NULL, + PRIMARY KEY CLUSTERED +( + MT_KEY1 ASC, + MT_KEY2 ASC +) +); + +CREATE TABLE KEY_CONSTRAINTS +( + CHILD_CD_ID BIGINT, + CHILD_INTEGER_IDX INT, + CHILD_TBL_ID BIGINT, + PARENT_CD_ID BIGINT, + PARENT_INTEGER_IDX INT NOT NULL, + PARENT_TBL_ID BIGINT NOT NULL, + POSITION INT NOT NULL, + CONSTRAINT_NAME VARCHAR(400) NOT NULL, + CONSTRAINT_TYPE SMALLINT NOT NULL, + UPDATE_RULE SMALLINT, + DELETE_RULE SMALLINT, + ENABLE_VALIDATE_RELY SMALLINT NOT NULL, + DEFAULT_VALUE VARCHAR(400) +) ; + +ALTER TABLE KEY_CONSTRAINTS ADD CONSTRAINT CONSTRAINTS_PK PRIMARY KEY (CONSTRAINT_NAME, POSITION); + +CREATE INDEX CONSTRAINTS_PARENT_TBL_ID__INDEX ON KEY_CONSTRAINTS(PARENT_TBL_ID); + +CREATE INDEX CONSTRAINTS_CONSTRAINT_TYPE_INDEX ON KEY_CONSTRAINTS(CONSTRAINT_TYPE); + +CREATE TABLE WRITE_SET ( + WS_DATABASE nvarchar(128) NOT NULL, + WS_TABLE nvarchar(128) NOT NULL, + WS_PARTITION nvarchar(767), + WS_TXNID bigint NOT NULL, + WS_COMMIT_ID bigint NOT NULL, + WS_OPERATION_TYPE char(1) NOT NULL +); + +CREATE TABLE METASTORE_DB_PROPERTIES ( + PROPERTY_KEY VARCHAR(255) NOT NULL, + PROPERTY_VALUE VARCHAR(1000) NOT NULL, + DESCRIPTION VARCHAR(1000) +); + +ALTER TABLE METASTORE_DB_PROPERTIES ADD CONSTRAINT PROPERTY_KEY_PK PRIMARY KEY (PROPERTY_KEY); + +CREATE TABLE TXN_TO_WRITE_ID ( + T2W_TXNID bigint NOT NULL, + T2W_DATABASE nvarchar(128) NOT NULL, + T2W_TABLE nvarchar(256) NOT NULL, + T2W_WRITEID bigint NOT NULL +); + +CREATE UNIQUE INDEX TBL_TO_TXN_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_TXNID); +CREATE UNIQUE INDEX TBL_TO_WRITE_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_WRITEID); + +CREATE TABLE NEXT_WRITE_ID ( + NWI_DATABASE nvarchar(128) NOT NULL, + NWI_TABLE nvarchar(256) NOT NULL, + NWI_NEXT bigint NOT NULL +); + +CREATE UNIQUE INDEX NEXT_WRITE_ID_IDX ON NEXT_WRITE_ID (NWI_DATABASE, NWI_TABLE); + +CREATE TABLE MIN_HISTORY_LEVEL ( + MHL_TXNID bigint NOT NULL, + MHL_MIN_OPEN_TXNID bigint NOT NULL, +PRIMARY KEY CLUSTERED +( + MHL_TXNID ASC +) +); + +CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); + +CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( + MRL_TXN_ID bigint NOT NULL, + MRL_DB_NAME nvarchar(128) NOT NULL, + MRL_TBL_NAME nvarchar(256) NOT NULL, + MRL_LAST_HEARTBEAT bigint NOT NULL, +PRIMARY KEY CLUSTERED +( + MRL_TXN_ID ASC +) +); + +CREATE TABLE "I_SCHEMA" ( + "SCHEMA_ID" bigint primary key, + "SCHEMA_TYPE" int not null, + "NAME" nvarchar(256) unique, + "DB_ID" bigint references "DBS" ("DB_ID"), + "COMPATIBILITY" int not null, + "VALIDATION_LEVEL" int not null, + "CAN_EVOLVE" bit not null, + "SCHEMA_GROUP" nvarchar(256), + "DESCRIPTION" nvarchar(4000), +); + +CREATE TABLE "SCHEMA_VERSION" ( + "SCHEMA_VERSION_ID" bigint primary key, + "SCHEMA_ID" bigint references "I_SCHEMA" ("SCHEMA_ID"), + "VERSION" int not null, + "CREATED_AT" bigint not null, + "CD_ID" bigint references "CDS" ("CD_ID"), + "STATE" int not null, + "DESCRIPTION" nvarchar(4000), + "SCHEMA_TEXT" varchar(max), + "FINGERPRINT" nvarchar(256), + "SCHEMA_VERSION_NAME" nvarchar(256), + "SERDE_ID" bigint references "SERDES" ("SERDE_ID"), + unique ("SCHEMA_ID", "VERSION") +); + +CREATE TABLE REPL_TXN_MAP ( + RTM_REPL_POLICY nvarchar(256) NOT NULL, + RTM_SRC_TXN_ID bigint NOT NULL, + RTM_TARGET_TXN_ID bigint NOT NULL +); + +ALTER TABLE REPL_TXN_MAP ADD CONSTRAINT REPL_TXN_MAP_PK PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID); + +-- Table SEQUENCE_TABLE is an internal table required by DataNucleus. +-- NOTE: Some versions of SchemaTool do not automatically generate this table. +-- See http://www.datanucleus.org/servlet/jira/browse/NUCRDBMS-416 +CREATE TABLE SEQUENCE_TABLE +( + SEQUENCE_NAME nvarchar(256) NOT NULL, + NEXT_VAL bigint NOT NULL +); + +CREATE UNIQUE INDEX PART_TABLE_PK ON SEQUENCE_TABLE (SEQUENCE_NAME); + +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MNotificationLog', 1); + +CREATE TABLE RUNTIME_STATS ( + RS_ID bigint primary key, + CREATE_TIME bigint NOT NULL, + WEIGHT bigint NOT NULL, + PAYLOAD varbinary(max) +); + +CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); + +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID bigint NOT NULL, + WNL_TXNID bigint NOT NULL, + WNL_WRITEID bigint NOT NULL, + WNL_DATABASE nvarchar(128) NOT NULL, + WNL_TABLE nvarchar(128) NOT NULL, + WNL_PARTITION nvarchar(767) NOT NULL, + WNL_TABLE_OBJ text NOT NULL, + WNL_PARTITION_OBJ text, + WNL_FILES text, + WNL_EVENT_TIME int NOT NULL +); + +ALTER TABLE TXN_WRITE_NOTIFICATION_LOG ADD CONSTRAINT TXN_WRITE_NOTIFICATION_LOG_PK PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION); + +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- ----------------------------------------------------------------- +-- Record schema version. Should be the last step in the init script +-- ----------------------------------------------------------------- +INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES (1, '3.2.0', 'Hive release version 3.2.0'); diff --git a/standalone-metastore/src/main/sql/mssql/upgrade-3.1.0-to-3.2.0.mssql.sql b/standalone-metastore/src/main/sql/mssql/upgrade-3.1.0-to-3.2.0.mssql.sql new file mode 100644 index 0000000000..86bd3cc2e4 --- /dev/null +++ b/standalone-metastore/src/main/sql/mssql/upgrade-3.1.0-to-3.2.0.mssql.sql @@ -0,0 +1,21 @@ +SELECT 'Upgrading MetaStore schema from 3.1.0 to 3.2.0' AS MESSAGE; + +-- HIVE-19267 +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID bigint NOT NULL, + WNL_TXNID bigint NOT NULL, + WNL_WRITEID bigint NOT NULL, + WNL_DATABASE nvarchar(128) NOT NULL, + WNL_TABLE nvarchar(128) NOT NULL, + WNL_PARTITION nvarchar(767) NOT NULL, + WNL_TABLE_OBJ text NOT NULL, + WNL_PARTITION_OBJ text, + WNL_FILES text, + WNL_EVENT_TIME int NOT NULL +); +ALTER TABLE TXN_WRITE_NOTIFICATION_LOG ADD CONSTRAINT TXN_WRITE_NOTIFICATION_LOG_PK PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION); +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- These lines need to be last. Insert any changes above. +UPDATE VERSION SET SCHEMA_VERSION='3.2.0', VERSION_COMMENT='Hive release version 3.2.0' where VER_ID=1; +SELECT 'Finished upgrading MetaStore schema from 3.1.0 to 3.2.0' AS MESSAGE; diff --git a/standalone-metastore/src/main/sql/mssql/upgrade.order.mssql b/standalone-metastore/src/main/sql/mssql/upgrade.order.mssql index 5572c26fe1..f9f2905842 100644 --- a/standalone-metastore/src/main/sql/mssql/upgrade.order.mssql +++ b/standalone-metastore/src/main/sql/mssql/upgrade.order.mssql @@ -9,3 +9,4 @@ 2.2.0-to-2.3.0 2.3.0-to-3.0.0 3.0.0-to-3.1.0 +3.1.0-to-3.2.0 diff --git a/standalone-metastore/src/main/sql/mysql/hive-schema-3.0.0.mysql.sql b/standalone-metastore/src/main/sql/mysql/hive-schema-3.0.0.mysql.sql index c54df55ed9..c65af1e80c 100644 --- a/standalone-metastore/src/main/sql/mysql/hive-schema-3.0.0.mysql.sql +++ b/standalone-metastore/src/main/sql/mysql/hive-schema-3.0.0.mysql.sql @@ -1155,7 +1155,6 @@ CREATE TABLE REPL_TXN_MAP ( PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID) ) ENGINE=InnoDB DEFAULT CHARSET=latin1; - CREATE TABLE RUNTIME_STATS ( RS_ID bigint primary key, CREATE_TIME bigint NOT NULL, diff --git a/standalone-metastore/src/main/sql/mysql/hive-schema-3.2.0.mysql.sql b/standalone-metastore/src/main/sql/mysql/hive-schema-3.2.0.mysql.sql new file mode 100644 index 0000000000..bb2668cb41 --- /dev/null +++ b/standalone-metastore/src/main/sql/mysql/hive-schema-3.2.0.mysql.sql @@ -0,0 +1,1218 @@ +-- MySQL dump 10.13 Distrib 5.5.25, for osx10.6 (i386) +-- +-- Host: localhost Database: test +-- ------------------------------------------------------ +-- Server version 5.5.25 + +/*!40101 SET @OLD_CHARACTER_SET_CLIENT=@@CHARACTER_SET_CLIENT */; +/*!40101 SET @OLD_CHARACTER_SET_RESULTS=@@CHARACTER_SET_RESULTS */; +/*!40101 SET @OLD_COLLATION_CONNECTION=@@COLLATION_CONNECTION */; +/*!40101 SET NAMES utf8 */; +/*!40103 SET @OLD_TIME_ZONE=@@TIME_ZONE */; +/*!40103 SET TIME_ZONE='+00:00' */; +/*!40014 SET @OLD_UNIQUE_CHECKS=@@UNIQUE_CHECKS, UNIQUE_CHECKS=0 */; +/*!40014 SET @OLD_FOREIGN_KEY_CHECKS=@@FOREIGN_KEY_CHECKS, FOREIGN_KEY_CHECKS=0 */; +/*!40101 SET @OLD_SQL_MODE=@@SQL_MODE, SQL_MODE='NO_AUTO_VALUE_ON_ZERO' */; +/*!40111 SET @OLD_SQL_NOTES=@@SQL_NOTES, SQL_NOTES=0 */; + +-- +-- Table structure for table `BUCKETING_COLS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `BUCKETING_COLS` ( + `SD_ID` bigint(20) NOT NULL, + `BUCKET_COL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`SD_ID`,`INTEGER_IDX`), + KEY `BUCKETING_COLS_N49` (`SD_ID`), + CONSTRAINT `BUCKETING_COLS_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `CDS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `CDS` ( + `CD_ID` bigint(20) NOT NULL, + PRIMARY KEY (`CD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `COLUMNS_V2` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `COLUMNS_V2` ( + `CD_ID` bigint(20) NOT NULL, + `COMMENT` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TYPE_NAME` MEDIUMTEXT DEFAULT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`CD_ID`,`COLUMN_NAME`), + KEY `COLUMNS_V2_N49` (`CD_ID`), + CONSTRAINT `COLUMNS_V2_FK1` FOREIGN KEY (`CD_ID`) REFERENCES `CDS` (`CD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `DATABASE_PARAMS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `DATABASE_PARAMS` ( + `DB_ID` bigint(20) NOT NULL, + `PARAM_KEY` varchar(180) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARAM_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`DB_ID`,`PARAM_KEY`), + KEY `DATABASE_PARAMS_N49` (`DB_ID`), + CONSTRAINT `DATABASE_PARAMS_FK1` FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +CREATE TABLE `CTLGS` ( + `CTLG_ID` BIGINT PRIMARY KEY, + `NAME` VARCHAR(256), + `DESC` VARCHAR(4000), + `LOCATION_URI` VARCHAR(4000) NOT NULL, + UNIQUE KEY `UNIQUE_CATALOG` (`NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + + +-- +-- Table structure for table `DBS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `DBS` ( + `DB_ID` bigint(20) NOT NULL, + `DESC` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `DB_LOCATION_URI` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `OWNER_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `OWNER_TYPE` varchar(10) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `CTLG_NAME` varchar(256) NOT NULL, + PRIMARY KEY (`DB_ID`), + UNIQUE KEY `UNIQUE_DATABASE` (`NAME`, `CTLG_NAME`), + CONSTRAINT `CTLG_FK1` FOREIGN KEY (`CTLG_NAME`) REFERENCES `CTLGS` (`NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `DB_PRIVS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `DB_PRIVS` ( + `DB_GRANT_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `DB_ID` bigint(20) DEFAULT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `DB_PRIV` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `AUTHORIZER` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`DB_GRANT_ID`), + UNIQUE KEY `DBPRIVILEGEINDEX` (`AUTHORIZER`,`DB_ID`,`PRINCIPAL_NAME`,`PRINCIPAL_TYPE`,`DB_PRIV`,`GRANTOR`,`GRANTOR_TYPE`), + KEY `DB_PRIVS_N49` (`DB_ID`), + CONSTRAINT `DB_PRIVS_FK1` FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `GLOBAL_PRIVS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `GLOBAL_PRIVS` ( + `USER_GRANT_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `USER_PRIV` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `AUTHORIZER` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`USER_GRANT_ID`), + UNIQUE KEY `GLOBALPRIVILEGEINDEX` (`AUTHORIZER`,`PRINCIPAL_NAME`,`PRINCIPAL_TYPE`,`USER_PRIV`,`GRANTOR`,`GRANTOR_TYPE`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `IDXS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `IDXS` ( + `INDEX_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `DEFERRED_REBUILD` bit(1) NOT NULL, + `INDEX_HANDLER_CLASS` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `INDEX_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `INDEX_TBL_ID` bigint(20) DEFAULT NULL, + `LAST_ACCESS_TIME` int(11) NOT NULL, + `ORIG_TBL_ID` bigint(20) DEFAULT NULL, + `SD_ID` bigint(20) DEFAULT NULL, + PRIMARY KEY (`INDEX_ID`), + UNIQUE KEY `UNIQUEINDEX` (`INDEX_NAME`,`ORIG_TBL_ID`), + KEY `IDXS_N51` (`SD_ID`), + KEY `IDXS_N50` (`INDEX_TBL_ID`), + KEY `IDXS_N49` (`ORIG_TBL_ID`), + CONSTRAINT `IDXS_FK1` FOREIGN KEY (`ORIG_TBL_ID`) REFERENCES `TBLS` (`TBL_ID`), + CONSTRAINT `IDXS_FK2` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`), + CONSTRAINT `IDXS_FK3` FOREIGN KEY (`INDEX_TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `INDEX_PARAMS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `INDEX_PARAMS` ( + `INDEX_ID` bigint(20) NOT NULL, + `PARAM_KEY` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARAM_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`INDEX_ID`,`PARAM_KEY`), + KEY `INDEX_PARAMS_N49` (`INDEX_ID`), + CONSTRAINT `INDEX_PARAMS_FK1` FOREIGN KEY (`INDEX_ID`) REFERENCES `IDXS` (`INDEX_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `NUCLEUS_TABLES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `NUCLEUS_TABLES` ( + `CLASS_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TABLE_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TYPE` varchar(4) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `OWNER` varchar(2) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `VERSION` varchar(20) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `INTERFACE_NAME` varchar(255) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`CLASS_NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PARTITIONS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PARTITIONS` ( + `PART_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `LAST_ACCESS_TIME` int(11) NOT NULL, + `PART_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `SD_ID` bigint(20) DEFAULT NULL, + `TBL_ID` bigint(20) DEFAULT NULL, + PRIMARY KEY (`PART_ID`), + UNIQUE KEY `UNIQUEPARTITION` (`PART_NAME`,`TBL_ID`), + KEY `PARTITIONS_N49` (`TBL_ID`), + KEY `PARTITIONS_N50` (`SD_ID`), + CONSTRAINT `PARTITIONS_FK1` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`), + CONSTRAINT `PARTITIONS_FK2` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PARTITION_EVENTS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PARTITION_EVENTS` ( + `PART_NAME_ID` bigint(20) NOT NULL, + `CAT_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `DB_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `EVENT_TIME` bigint(20) NOT NULL, + `EVENT_TYPE` int(11) NOT NULL, + `PARTITION_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`PART_NAME_ID`), + KEY `PARTITIONEVENTINDEX` (`PARTITION_NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PARTITION_KEYS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PARTITION_KEYS` ( + `TBL_ID` bigint(20) NOT NULL, + `PKEY_COMMENT` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PKEY_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PKEY_TYPE` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`TBL_ID`,`PKEY_NAME`), + KEY `PARTITION_KEYS_N49` (`TBL_ID`), + CONSTRAINT `PARTITION_KEYS_FK1` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PARTITION_KEY_VALS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PARTITION_KEY_VALS` ( + `PART_ID` bigint(20) NOT NULL, + `PART_KEY_VAL` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`PART_ID`,`INTEGER_IDX`), + KEY `PARTITION_KEY_VALS_N49` (`PART_ID`), + CONSTRAINT `PARTITION_KEY_VALS_FK1` FOREIGN KEY (`PART_ID`) REFERENCES `PARTITIONS` (`PART_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PARTITION_PARAMS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PARTITION_PARAMS` ( + `PART_ID` bigint(20) NOT NULL, + `PARAM_KEY` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARAM_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`PART_ID`,`PARAM_KEY`), + KEY `PARTITION_PARAMS_N49` (`PART_ID`), + CONSTRAINT `PARTITION_PARAMS_FK1` FOREIGN KEY (`PART_ID`) REFERENCES `PARTITIONS` (`PART_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PART_COL_PRIVS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PART_COL_PRIVS` ( + `PART_COLUMN_GRANT_ID` bigint(20) NOT NULL, + `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `CREATE_TIME` int(11) NOT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PART_ID` bigint(20) DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PART_COL_PRIV` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `AUTHORIZER` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`PART_COLUMN_GRANT_ID`), + KEY `PART_COL_PRIVS_N49` (`PART_ID`), + KEY `PARTITIONCOLUMNPRIVILEGEINDEX` (`AUTHORIZER`,`PART_ID`,`COLUMN_NAME`,`PRINCIPAL_NAME`,`PRINCIPAL_TYPE`,`PART_COL_PRIV`,`GRANTOR`,`GRANTOR_TYPE`), + CONSTRAINT `PART_COL_PRIVS_FK1` FOREIGN KEY (`PART_ID`) REFERENCES `PARTITIONS` (`PART_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `PART_PRIVS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `PART_PRIVS` ( + `PART_GRANT_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PART_ID` bigint(20) DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PART_PRIV` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `AUTHORIZER` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`PART_GRANT_ID`), + KEY `PARTPRIVILEGEINDEX` (`AUTHORIZER`,`PART_ID`,`PRINCIPAL_NAME`,`PRINCIPAL_TYPE`,`PART_PRIV`,`GRANTOR`,`GRANTOR_TYPE`), + KEY `PART_PRIVS_N49` (`PART_ID`), + CONSTRAINT `PART_PRIVS_FK1` FOREIGN KEY (`PART_ID`) REFERENCES `PARTITIONS` (`PART_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `ROLES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `ROLES` ( + `ROLE_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `OWNER_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `ROLE_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`ROLE_ID`), + UNIQUE KEY `ROLEENTITYINDEX` (`ROLE_NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `ROLE_MAP` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `ROLE_MAP` ( + `ROLE_GRANT_ID` bigint(20) NOT NULL, + `ADD_TIME` int(11) NOT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `ROLE_ID` bigint(20) DEFAULT NULL, + PRIMARY KEY (`ROLE_GRANT_ID`), + UNIQUE KEY `USERROLEMAPINDEX` (`PRINCIPAL_NAME`,`ROLE_ID`,`GRANTOR`,`GRANTOR_TYPE`), + KEY `ROLE_MAP_N49` (`ROLE_ID`), + CONSTRAINT `ROLE_MAP_FK1` FOREIGN KEY (`ROLE_ID`) REFERENCES `ROLES` (`ROLE_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SDS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SDS` ( + `SD_ID` bigint(20) NOT NULL, + `CD_ID` bigint(20) DEFAULT NULL, + `INPUT_FORMAT` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `IS_COMPRESSED` bit(1) NOT NULL, + `IS_STOREDASSUBDIRECTORIES` bit(1) NOT NULL, + `LOCATION` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `NUM_BUCKETS` int(11) NOT NULL, + `OUTPUT_FORMAT` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `SERDE_ID` bigint(20) DEFAULT NULL, + PRIMARY KEY (`SD_ID`), + KEY `SDS_N49` (`SERDE_ID`), + KEY `SDS_N50` (`CD_ID`), + CONSTRAINT `SDS_FK1` FOREIGN KEY (`SERDE_ID`) REFERENCES `SERDES` (`SERDE_ID`), + CONSTRAINT `SDS_FK2` FOREIGN KEY (`CD_ID`) REFERENCES `CDS` (`CD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SD_PARAMS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SD_PARAMS` ( + `SD_ID` bigint(20) NOT NULL, + `PARAM_KEY` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARAM_VALUE` MEDIUMTEXT CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`SD_ID`,`PARAM_KEY`), + KEY `SD_PARAMS_N49` (`SD_ID`), + CONSTRAINT `SD_PARAMS_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SEQUENCE_TABLE` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SEQUENCE_TABLE` ( + `SEQUENCE_NAME` varchar(255) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `NEXT_VAL` bigint(20) NOT NULL, + PRIMARY KEY (`SEQUENCE_NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +INSERT INTO `SEQUENCE_TABLE` (`SEQUENCE_NAME`, `NEXT_VAL`) VALUES ('org.apache.hadoop.hive.metastore.model.MNotificationLog', 1); + +-- +-- Table structure for table `SERDES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SERDES` ( + `SERDE_ID` bigint(20) NOT NULL, + `NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `SLIB` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `DESCRIPTION` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `SERIALIZER_CLASS` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `DESERIALIZER_CLASS` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `SERDE_TYPE` integer, + PRIMARY KEY (`SERDE_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SERDE_PARAMS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SERDE_PARAMS` ( + `SERDE_ID` bigint(20) NOT NULL, + `PARAM_KEY` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARAM_VALUE` MEDIUMTEXT CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`SERDE_ID`,`PARAM_KEY`), + KEY `SERDE_PARAMS_N49` (`SERDE_ID`), + CONSTRAINT `SERDE_PARAMS_FK1` FOREIGN KEY (`SERDE_ID`) REFERENCES `SERDES` (`SERDE_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SKEWED_COL_NAMES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SKEWED_COL_NAMES` ( + `SD_ID` bigint(20) NOT NULL, + `SKEWED_COL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`SD_ID`,`INTEGER_IDX`), + KEY `SKEWED_COL_NAMES_N49` (`SD_ID`), + CONSTRAINT `SKEWED_COL_NAMES_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SKEWED_COL_VALUE_LOC_MAP` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SKEWED_COL_VALUE_LOC_MAP` ( + `SD_ID` bigint(20) NOT NULL, + `STRING_LIST_ID_KID` bigint(20) NOT NULL, + `LOCATION` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`SD_ID`,`STRING_LIST_ID_KID`), + KEY `SKEWED_COL_VALUE_LOC_MAP_N49` (`STRING_LIST_ID_KID`), + KEY `SKEWED_COL_VALUE_LOC_MAP_N50` (`SD_ID`), + CONSTRAINT `SKEWED_COL_VALUE_LOC_MAP_FK2` FOREIGN KEY (`STRING_LIST_ID_KID`) REFERENCES `SKEWED_STRING_LIST` (`STRING_LIST_ID`), + CONSTRAINT `SKEWED_COL_VALUE_LOC_MAP_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SKEWED_STRING_LIST` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SKEWED_STRING_LIST` ( + `STRING_LIST_ID` bigint(20) NOT NULL, + PRIMARY KEY (`STRING_LIST_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SKEWED_STRING_LIST_VALUES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SKEWED_STRING_LIST_VALUES` ( + `STRING_LIST_ID` bigint(20) NOT NULL, + `STRING_LIST_VALUE` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`STRING_LIST_ID`,`INTEGER_IDX`), + KEY `SKEWED_STRING_LIST_VALUES_N49` (`STRING_LIST_ID`), + CONSTRAINT `SKEWED_STRING_LIST_VALUES_FK1` FOREIGN KEY (`STRING_LIST_ID`) REFERENCES `SKEWED_STRING_LIST` (`STRING_LIST_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SKEWED_VALUES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SKEWED_VALUES` ( + `SD_ID_OID` bigint(20) NOT NULL, + `STRING_LIST_ID_EID` bigint(20) NOT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`SD_ID_OID`,`INTEGER_IDX`), + KEY `SKEWED_VALUES_N50` (`SD_ID_OID`), + KEY `SKEWED_VALUES_N49` (`STRING_LIST_ID_EID`), + CONSTRAINT `SKEWED_VALUES_FK2` FOREIGN KEY (`STRING_LIST_ID_EID`) REFERENCES `SKEWED_STRING_LIST` (`STRING_LIST_ID`), + CONSTRAINT `SKEWED_VALUES_FK1` FOREIGN KEY (`SD_ID_OID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `SORT_COLS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `SORT_COLS` ( + `SD_ID` bigint(20) NOT NULL, + `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `ORDER` int(11) NOT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`SD_ID`,`INTEGER_IDX`), + KEY `SORT_COLS_N49` (`SD_ID`), + CONSTRAINT `SORT_COLS_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `TABLE_PARAMS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `TABLE_PARAMS` ( + `TBL_ID` bigint(20) NOT NULL, + `PARAM_KEY` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARAM_VALUE` MEDIUMTEXT CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`TBL_ID`,`PARAM_KEY`), + KEY `TABLE_PARAMS_N49` (`TBL_ID`), + CONSTRAINT `TABLE_PARAMS_FK1` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `MV_CREATION_METADATA` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `MV_CREATION_METADATA` ( + `MV_CREATION_METADATA_ID` bigint(20) NOT NULL, + `CAT_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `DB_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TXN_LIST` TEXT DEFAULT NULL, + `MATERIALIZATION_TIME` bigint(20) NOT NULL, + PRIMARY KEY (`MV_CREATION_METADATA_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +CREATE INDEX MV_UNIQUE_TABLE ON MV_CREATION_METADATA (TBL_NAME, DB_NAME) USING BTREE; + +-- +-- Table structure for table `TBLS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `TBLS` ( + `TBL_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `DB_ID` bigint(20) DEFAULT NULL, + `LAST_ACCESS_TIME` int(11) NOT NULL, + `OWNER` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `OWNER_TYPE` varchar(10) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `RETENTION` int(11) NOT NULL, + `SD_ID` bigint(20) DEFAULT NULL, + `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TBL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `VIEW_EXPANDED_TEXT` mediumtext, + `VIEW_ORIGINAL_TEXT` mediumtext, + `IS_REWRITE_ENABLED` bit(1) NOT NULL DEFAULT 0, + PRIMARY KEY (`TBL_ID`), + UNIQUE KEY `UNIQUETABLE` (`TBL_NAME`,`DB_ID`), + KEY `TBLS_N50` (`SD_ID`), + KEY `TBLS_N49` (`DB_ID`), + CONSTRAINT `TBLS_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`), + CONSTRAINT `TBLS_FK2` FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `MV_TABLES_USED` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `MV_TABLES_USED` ( + `MV_CREATION_METADATA_ID` bigint(20) NOT NULL, + `TBL_ID` bigint(20) NOT NULL, + CONSTRAINT `MV_TABLES_USED_FK1` FOREIGN KEY (`MV_CREATION_METADATA_ID`) REFERENCES `MV_CREATION_METADATA` (`MV_CREATION_METADATA_ID`), + CONSTRAINT `MV_TABLES_USED_FK2` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `TBL_COL_PRIVS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `TBL_COL_PRIVS` ( + `TBL_COLUMN_GRANT_ID` bigint(20) NOT NULL, + `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `CREATE_TIME` int(11) NOT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TBL_COL_PRIV` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TBL_ID` bigint(20) DEFAULT NULL, + `AUTHORIZER` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`TBL_COLUMN_GRANT_ID`), + KEY `TABLECOLUMNPRIVILEGEINDEX` (`AUTHORIZER`,`TBL_ID`,`COLUMN_NAME`,`PRINCIPAL_NAME`,`PRINCIPAL_TYPE`,`TBL_COL_PRIV`,`GRANTOR`,`GRANTOR_TYPE`), + KEY `TBL_COL_PRIVS_N49` (`TBL_ID`), + CONSTRAINT `TBL_COL_PRIVS_FK1` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `TBL_PRIVS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `TBL_PRIVS` ( + `TBL_GRANT_ID` bigint(20) NOT NULL, + `CREATE_TIME` int(11) NOT NULL, + `GRANT_OPTION` smallint(6) NOT NULL, + `GRANTOR` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `GRANTOR_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `PRINCIPAL_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TBL_PRIV` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TBL_ID` bigint(20) DEFAULT NULL, + `AUTHORIZER` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`TBL_GRANT_ID`), + KEY `TBL_PRIVS_N49` (`TBL_ID`), + KEY `TABLEPRIVILEGEINDEX` (`AUTHORIZER`,`TBL_ID`,`PRINCIPAL_NAME`,`PRINCIPAL_TYPE`,`TBL_PRIV`,`GRANTOR`,`GRANTOR_TYPE`), + CONSTRAINT `TBL_PRIVS_FK1` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `TAB_COL_STATS` +-- +CREATE TABLE IF NOT EXISTS `TAB_COL_STATS` ( + `CS_ID` bigint(20) NOT NULL, + `CAT_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `DB_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TABLE_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `COLUMN_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TBL_ID` bigint(20) NOT NULL, + `LONG_LOW_VALUE` bigint(20), + `LONG_HIGH_VALUE` bigint(20), + `DOUBLE_HIGH_VALUE` double(53,4), + `DOUBLE_LOW_VALUE` double(53,4), + `BIG_DECIMAL_LOW_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin, + `BIG_DECIMAL_HIGH_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin, + `NUM_NULLS` bigint(20) NOT NULL, + `NUM_DISTINCTS` bigint(20), + `BIT_VECTOR` blob, + `AVG_COL_LEN` double(53,4), + `MAX_COL_LEN` bigint(20), + `NUM_TRUES` bigint(20), + `NUM_FALSES` bigint(20), + `LAST_ANALYZED` bigint(20) NOT NULL, + PRIMARY KEY (`CS_ID`), + CONSTRAINT `TAB_COL_STATS_FK` FOREIGN KEY (`TBL_ID`) REFERENCES `TBLS` (`TBL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX TAB_COL_STATS_IDX ON TAB_COL_STATS (CAT_NAME, DB_NAME, TABLE_NAME, COLUMN_NAME) USING BTREE; +-- +-- Table structure for table `PART_COL_STATS` +-- +CREATE TABLE IF NOT EXISTS `PART_COL_STATS` ( + `CS_ID` bigint(20) NOT NULL, + `CAT_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `DB_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `TABLE_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PARTITION_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `COLUMN_TYPE` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `PART_ID` bigint(20) NOT NULL, + `LONG_LOW_VALUE` bigint(20), + `LONG_HIGH_VALUE` bigint(20), + `DOUBLE_HIGH_VALUE` double(53,4), + `DOUBLE_LOW_VALUE` double(53,4), + `BIG_DECIMAL_LOW_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin, + `BIG_DECIMAL_HIGH_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin, + `NUM_NULLS` bigint(20) NOT NULL, + `NUM_DISTINCTS` bigint(20), + `BIT_VECTOR` blob, + `AVG_COL_LEN` double(53,4), + `MAX_COL_LEN` bigint(20), + `NUM_TRUES` bigint(20), + `NUM_FALSES` bigint(20), + `LAST_ANALYZED` bigint(20) NOT NULL, + PRIMARY KEY (`CS_ID`), + CONSTRAINT `PART_COL_STATS_FK` FOREIGN KEY (`PART_ID`) REFERENCES `PARTITIONS` (`PART_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX PCS_STATS_IDX ON PART_COL_STATS (CAT_NAME, DB_NAME,TABLE_NAME,COLUMN_NAME,PARTITION_NAME) USING BTREE; + +-- +-- Table structure for table `TYPES` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `TYPES` ( + `TYPES_ID` bigint(20) NOT NULL, + `TYPE_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TYPE1` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `TYPE2` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + PRIMARY KEY (`TYPES_ID`), + UNIQUE KEY `UNIQUE_TYPE` (`TYPE_NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +/*!40101 SET character_set_client = @saved_cs_client */; + +-- +-- Table structure for table `TYPE_FIELDS` +-- + +/*!40101 SET @saved_cs_client = @@character_set_client */; +/*!40101 SET character_set_client = utf8 */; +CREATE TABLE IF NOT EXISTS `TYPE_FIELDS` ( + `TYPE_NAME` bigint(20) NOT NULL, + `COMMENT` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, + `FIELD_NAME` varchar(128) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `FIELD_TYPE` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, + `INTEGER_IDX` int(11) NOT NULL, + PRIMARY KEY (`TYPE_NAME`,`FIELD_NAME`), + KEY `TYPE_FIELDS_N49` (`TYPE_NAME`), + CONSTRAINT `TYPE_FIELDS_FK1` FOREIGN KEY (`TYPE_NAME`) REFERENCES `TYPES` (`TYPES_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +-- Table `MASTER_KEYS` for classes [org.apache.hadoop.hive.metastore.model.MMasterKey] +CREATE TABLE IF NOT EXISTS `MASTER_KEYS` +( + `KEY_ID` INTEGER NOT NULL AUTO_INCREMENT, + `MASTER_KEY` VARCHAR(767) BINARY NULL, + PRIMARY KEY (`KEY_ID`) +) ENGINE=INNODB DEFAULT CHARSET=latin1; + +-- Table `DELEGATION_TOKENS` for classes [org.apache.hadoop.hive.metastore.model.MDelegationToken] +CREATE TABLE IF NOT EXISTS `DELEGATION_TOKENS` +( + `TOKEN_IDENT` VARCHAR(767) BINARY NOT NULL, + `TOKEN` VARCHAR(767) BINARY NULL, + PRIMARY KEY (`TOKEN_IDENT`) +) ENGINE=INNODB DEFAULT CHARSET=latin1; + +-- +-- Table structure for VERSION +-- +CREATE TABLE IF NOT EXISTS `VERSION` ( + `VER_ID` BIGINT NOT NULL, + `SCHEMA_VERSION` VARCHAR(127) NOT NULL, + `VERSION_COMMENT` VARCHAR(255), + PRIMARY KEY (`VER_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +-- +-- Table structure for table FUNCS +-- +CREATE TABLE IF NOT EXISTS `FUNCS` ( + `FUNC_ID` BIGINT(20) NOT NULL, + `CLASS_NAME` VARCHAR(4000) CHARACTER SET latin1 COLLATE latin1_bin, + `CREATE_TIME` INT(11) NOT NULL, + `DB_ID` BIGINT(20), + `FUNC_NAME` VARCHAR(128) CHARACTER SET latin1 COLLATE latin1_bin, + `FUNC_TYPE` INT(11) NOT NULL, + `OWNER_NAME` VARCHAR(128) CHARACTER SET latin1 COLLATE latin1_bin, + `OWNER_TYPE` VARCHAR(10) CHARACTER SET latin1 COLLATE latin1_bin, + PRIMARY KEY (`FUNC_ID`), + UNIQUE KEY `UNIQUEFUNCTION` (`FUNC_NAME`, `DB_ID`), + KEY `FUNCS_N49` (`DB_ID`), + CONSTRAINT `FUNCS_FK1` FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +-- +-- Table structure for table FUNC_RU +-- +CREATE TABLE IF NOT EXISTS `FUNC_RU` ( + `FUNC_ID` BIGINT(20) NOT NULL, + `RESOURCE_TYPE` INT(11) NOT NULL, + `RESOURCE_URI` VARCHAR(4000) CHARACTER SET latin1 COLLATE latin1_bin, + `INTEGER_IDX` INT(11) NOT NULL, + PRIMARY KEY (`FUNC_ID`, `INTEGER_IDX`), + CONSTRAINT `FUNC_RU_FK1` FOREIGN KEY (`FUNC_ID`) REFERENCES `FUNCS` (`FUNC_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE IF NOT EXISTS `NOTIFICATION_LOG` +( + `NL_ID` BIGINT(20) NOT NULL, + `EVENT_ID` BIGINT(20) NOT NULL, + `EVENT_TIME` INT(11) NOT NULL, + `EVENT_TYPE` varchar(32) NOT NULL, + `CAT_NAME` varchar(256), + `DB_NAME` varchar(128), + `TBL_NAME` varchar(256), + `MESSAGE` longtext, + `MESSAGE_FORMAT` varchar(16), + PRIMARY KEY (`NL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE IF NOT EXISTS `NOTIFICATION_SEQUENCE` +( + `NNI_ID` BIGINT(20) NOT NULL, + `NEXT_EVENT_ID` BIGINT(20) NOT NULL, + PRIMARY KEY (`NNI_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +INSERT INTO `NOTIFICATION_SEQUENCE` (`NNI_ID`, `NEXT_EVENT_ID`) SELECT * from (select 1 as `NNI_ID`, 1 as `NOTIFICATION_SEQUENCE`) a WHERE (SELECT COUNT(*) FROM `NOTIFICATION_SEQUENCE`) = 0; + +CREATE TABLE IF NOT EXISTS `KEY_CONSTRAINTS` +( + `CHILD_CD_ID` BIGINT, + `CHILD_INTEGER_IDX` INT(11), + `CHILD_TBL_ID` BIGINT, + `PARENT_CD_ID` BIGINT, + `PARENT_INTEGER_IDX` INT(11) NOT NULL, + `PARENT_TBL_ID` BIGINT NOT NULL, + `POSITION` BIGINT NOT NULL, + `CONSTRAINT_NAME` VARCHAR(400) NOT NULL, + `CONSTRAINT_TYPE` SMALLINT(6) NOT NULL, + `UPDATE_RULE` SMALLINT(6), + `DELETE_RULE` SMALLINT(6), + `ENABLE_VALIDATE_RELY` SMALLINT(6) NOT NULL, + `DEFAULT_VALUE` VARCHAR(400), + PRIMARY KEY (`CONSTRAINT_NAME`, `POSITION`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX `CONSTRAINTS_PARENT_TABLE_ID_INDEX` ON KEY_CONSTRAINTS (`PARENT_TBL_ID`) USING BTREE; + +CREATE INDEX `CONSTRAINTS_CONSTRAINT_TYPE_INDEX` ON KEY_CONSTRAINTS (`CONSTRAINT_TYPE`) USING BTREE; + +-- ----------------------------- +-- Metastore DB Properties table +-- ----------------------------- +CREATE TABLE IF NOT EXISTS `METASTORE_DB_PROPERTIES` ( + `PROPERTY_KEY` varchar(255) NOT NULL, + `PROPERTY_VALUE` varchar(1000) NOT NULL, + `DESCRIPTION` varchar(1000), + PRIMARY KEY(`PROPERTY_KEY`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + + +-- --------------------- +-- Resource plan tables. +-- --------------------- +CREATE TABLE IF NOT EXISTS WM_RESOURCEPLAN ( + `RP_ID` bigint(20) NOT NULL, + `NAME` varchar(128) NOT NULL, + `QUERY_PARALLELISM` int(11), + `STATUS` varchar(20) NOT NULL, + `DEFAULT_POOL_ID` bigint(20), + PRIMARY KEY (`RP_ID`), + UNIQUE KEY `UNIQUE_WM_RESOURCEPLAN` (`NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE IF NOT EXISTS WM_POOL +( + `POOL_ID` bigint(20) NOT NULL, + `RP_ID` bigint(20) NOT NULL, + `PATH` varchar(767) NOT NULL, + `ALLOC_FRACTION` DOUBLE, + `QUERY_PARALLELISM` int(11), + `SCHEDULING_POLICY` varchar(767), + PRIMARY KEY (`POOL_ID`), + UNIQUE KEY `UNIQUE_WM_POOL` (`RP_ID`, `PATH`), + CONSTRAINT `WM_POOL_FK1` FOREIGN KEY (`RP_ID`) REFERENCES `WM_RESOURCEPLAN` (`RP_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +ALTER TABLE `WM_RESOURCEPLAN` ADD CONSTRAINT `WM_RESOURCEPLAN_FK1` FOREIGN KEY (`DEFAULT_POOL_ID`) REFERENCES `WM_POOL`(`POOL_ID`); + +CREATE TABLE IF NOT EXISTS WM_TRIGGER +( + `TRIGGER_ID` bigint(20) NOT NULL, + `RP_ID` bigint(20) NOT NULL, + `NAME` varchar(128) NOT NULL, + `TRIGGER_EXPRESSION` varchar(1024), + `ACTION_EXPRESSION` varchar(1024), + `IS_IN_UNMANAGED` bit(1) NOT NULL DEFAULT 0, + PRIMARY KEY (`TRIGGER_ID`), + UNIQUE KEY `UNIQUE_WM_TRIGGER` (`RP_ID`, `NAME`), + CONSTRAINT `WM_TRIGGER_FK1` FOREIGN KEY (`RP_ID`) REFERENCES `WM_RESOURCEPLAN` (`RP_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE IF NOT EXISTS WM_POOL_TO_TRIGGER +( + `POOL_ID` bigint(20) NOT NULL, + `TRIGGER_ID` bigint(20) NOT NULL, + PRIMARY KEY (`POOL_ID`, `TRIGGER_ID`), + CONSTRAINT `WM_POOL_TO_TRIGGER_FK1` FOREIGN KEY (`POOL_ID`) REFERENCES `WM_POOL` (`POOL_ID`), + CONSTRAINT `WM_POOL_TO_TRIGGER_FK2` FOREIGN KEY (`TRIGGER_ID`) REFERENCES `WM_TRIGGER` (`TRIGGER_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE IF NOT EXISTS WM_MAPPING +( + `MAPPING_ID` bigint(20) NOT NULL, + `RP_ID` bigint(20) NOT NULL, + `ENTITY_TYPE` varchar(128) NOT NULL, + `ENTITY_NAME` varchar(128) NOT NULL, + `POOL_ID` bigint(20), + `ORDERING` int, + PRIMARY KEY (`MAPPING_ID`), + UNIQUE KEY `UNIQUE_WM_MAPPING` (`RP_ID`, `ENTITY_TYPE`, `ENTITY_NAME`), + CONSTRAINT `WM_MAPPING_FK1` FOREIGN KEY (`RP_ID`) REFERENCES `WM_RESOURCEPLAN` (`RP_ID`), + CONSTRAINT `WM_MAPPING_FK2` FOREIGN KEY (`POOL_ID`) REFERENCES `WM_POOL` (`POOL_ID`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +-- ---------------------------- +-- Transaction and Lock Tables +-- ---------------------------- +CREATE TABLE TXNS ( + TXN_ID bigint PRIMARY KEY, + TXN_STATE char(1) NOT NULL, + TXN_STARTED bigint NOT NULL, + TXN_LAST_HEARTBEAT bigint NOT NULL, + TXN_USER varchar(128) NOT NULL, + TXN_HOST varchar(128) NOT NULL, + TXN_AGENT_INFO varchar(128), + TXN_META_INFO varchar(128), + TXN_HEARTBEAT_COUNT int, + TXN_TYPE int +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE TXN_COMPONENTS ( + TC_TXNID bigint NOT NULL, + TC_DATABASE varchar(128) NOT NULL, + TC_TABLE varchar(128), + TC_PARTITION varchar(767), + TC_OPERATION_TYPE char(1) NOT NULL, + TC_WRITEID bigint, + FOREIGN KEY (TC_TXNID) REFERENCES TXNS (TXN_ID) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX TC_TXNID_INDEX ON TXN_COMPONENTS (TC_TXNID); + +CREATE TABLE COMPLETED_TXN_COMPONENTS ( + CTC_TXNID bigint NOT NULL, + CTC_DATABASE varchar(128) NOT NULL, + CTC_TABLE varchar(256), + CTC_PARTITION varchar(767), + CTC_TIMESTAMP timestamp DEFAULT CURRENT_TIMESTAMP NOT NULL, + CTC_WRITEID bigint, + CTC_UPDATE_DELETE char(1) NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX COMPLETED_TXN_COMPONENTS_IDX ON COMPLETED_TXN_COMPONENTS (CTC_DATABASE, CTC_TABLE, CTC_PARTITION) USING BTREE; + +CREATE TABLE NEXT_TXN_ID ( + NTXN_NEXT bigint NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +INSERT INTO NEXT_TXN_ID VALUES(1); + +CREATE TABLE HIVE_LOCKS ( + HL_LOCK_EXT_ID bigint NOT NULL, + HL_LOCK_INT_ID bigint NOT NULL, + HL_TXNID bigint NOT NULL, + HL_DB varchar(128) NOT NULL, + HL_TABLE varchar(128), + HL_PARTITION varchar(767), + HL_LOCK_STATE char(1) not null, + HL_LOCK_TYPE char(1) not null, + HL_LAST_HEARTBEAT bigint NOT NULL, + HL_ACQUIRED_AT bigint, + HL_USER varchar(128) NOT NULL, + HL_HOST varchar(128) NOT NULL, + HL_HEARTBEAT_COUNT int, + HL_AGENT_INFO varchar(128), + HL_BLOCKEDBY_EXT_ID bigint, + HL_BLOCKEDBY_INT_ID bigint, + PRIMARY KEY(HL_LOCK_EXT_ID, HL_LOCK_INT_ID), + KEY HIVE_LOCK_TXNID_INDEX (HL_TXNID) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX HL_TXNID_IDX ON HIVE_LOCKS (HL_TXNID); + +CREATE TABLE NEXT_LOCK_ID ( + NL_NEXT bigint NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +INSERT INTO NEXT_LOCK_ID VALUES(1); + +CREATE TABLE COMPACTION_QUEUE ( + CQ_ID bigint PRIMARY KEY, + CQ_DATABASE varchar(128) NOT NULL, + CQ_TABLE varchar(128) NOT NULL, + CQ_PARTITION varchar(767), + CQ_STATE char(1) NOT NULL, + CQ_TYPE char(1) NOT NULL, + CQ_TBLPROPERTIES varchar(2048), + CQ_WORKER_ID varchar(128), + CQ_START bigint, + CQ_RUN_AS varchar(128), + CQ_HIGHEST_WRITE_ID bigint, + CQ_META_INFO varbinary(2048), + CQ_HADOOP_JOB_ID varchar(32) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE COMPLETED_COMPACTIONS ( + CC_ID bigint PRIMARY KEY, + CC_DATABASE varchar(128) NOT NULL, + CC_TABLE varchar(128) NOT NULL, + CC_PARTITION varchar(767), + CC_STATE char(1) NOT NULL, + CC_TYPE char(1) NOT NULL, + CC_TBLPROPERTIES varchar(2048), + CC_WORKER_ID varchar(128), + CC_START bigint, + CC_END bigint, + CC_RUN_AS varchar(128), + CC_HIGHEST_WRITE_ID bigint, + CC_META_INFO varbinary(2048), + CC_HADOOP_JOB_ID varchar(32) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( + NCQ_NEXT bigint NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +INSERT INTO NEXT_COMPACTION_QUEUE_ID VALUES(1); + +CREATE TABLE AUX_TABLE ( + MT_KEY1 varchar(128) NOT NULL, + MT_KEY2 bigint NOT NULL, + MT_COMMENT varchar(255), + PRIMARY KEY(MT_KEY1, MT_KEY2) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE WRITE_SET ( + WS_DATABASE varchar(128) NOT NULL, + WS_TABLE varchar(128) NOT NULL, + WS_PARTITION varchar(767), + WS_TXNID bigint NOT NULL, + WS_COMMIT_ID bigint NOT NULL, + WS_OPERATION_TYPE char(1) NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE TXN_TO_WRITE_ID ( + T2W_TXNID bigint NOT NULL, + T2W_DATABASE varchar(128) NOT NULL, + T2W_TABLE varchar(256) NOT NULL, + T2W_WRITEID bigint NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE UNIQUE INDEX TBL_TO_TXN_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_TXNID); +CREATE UNIQUE INDEX TBL_TO_WRITE_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_WRITEID); + +CREATE TABLE NEXT_WRITE_ID ( + NWI_DATABASE varchar(128) NOT NULL, + NWI_TABLE varchar(256) NOT NULL, + NWI_NEXT bigint NOT NULL +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE UNIQUE INDEX NEXT_WRITE_ID_IDX ON NEXT_WRITE_ID (NWI_DATABASE, NWI_TABLE); + +CREATE TABLE MIN_HISTORY_LEVEL ( + MHL_TXNID bigint NOT NULL, + MHL_MIN_OPEN_TXNID bigint NOT NULL, + PRIMARY KEY(MHL_TXNID) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); + +CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( + MRL_TXN_ID bigint NOT NULL, + MRL_DB_NAME VARCHAR(128) NOT NULL, + MRL_TBL_NAME VARCHAR(256) NOT NULL, + MRL_LAST_HEARTBEAT bigint NOT NULL, + PRIMARY KEY(MRL_TXN_ID) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE `I_SCHEMA` ( + `SCHEMA_ID` BIGINT PRIMARY KEY, + `SCHEMA_TYPE` INTEGER NOT NULL, + `NAME` VARCHAR(256), + `DB_ID` BIGINT, + `COMPATIBILITY` INTEGER NOT NULL, + `VALIDATION_LEVEL` INTEGER NOT NULL, + `CAN_EVOLVE` bit(1) NOT NULL, + `SCHEMA_GROUP` VARCHAR(256), + `DESCRIPTION` VARCHAR(4000), + FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`), + KEY `UNIQUE_NAME` (`NAME`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE `SCHEMA_VERSION` ( + `SCHEMA_VERSION_ID` bigint primary key, + `SCHEMA_ID` BIGINT, + `VERSION` INTEGER NOT NULL, + `CREATED_AT` BIGINT NOT NULL, + `CD_ID` BIGINT, + `STATE` INTEGER NOT NULL, + `DESCRIPTION` VARCHAR(4000), + `SCHEMA_TEXT` mediumtext, + `FINGERPRINT` VARCHAR(256), + `SCHEMA_VERSION_NAME` VARCHAR(256), + `SERDE_ID` bigint, + FOREIGN KEY (`SCHEMA_ID`) REFERENCES `I_SCHEMA` (`SCHEMA_ID`), + FOREIGN KEY (`CD_ID`) REFERENCES `CDS` (`CD_ID`), + FOREIGN KEY (`SERDE_ID`) REFERENCES `SERDES` (`SERDE_ID`), + KEY `UNIQUE_VERSION` (`SCHEMA_ID`, `VERSION`) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE TABLE REPL_TXN_MAP ( + RTM_REPL_POLICY varchar(256) NOT NULL, + RTM_SRC_TXN_ID bigint NOT NULL, + RTM_TARGET_TXN_ID bigint NOT NULL, + PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + + +CREATE TABLE RUNTIME_STATS ( + RS_ID bigint primary key, + CREATE_TIME bigint NOT NULL, + WEIGHT bigint NOT NULL, + PAYLOAD blob +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); + +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID bigint NOT NULL, + WNL_TXNID bigint NOT NULL, + WNL_WRITEID bigint NOT NULL, + WNL_DATABASE varchar(128) NOT NULL, + WNL_TABLE varchar(128) NOT NULL, + WNL_PARTITION varchar(767) NOT NULL, + WNL_TABLE_OBJ longtext NOT NULL, + WNL_PARTITION_OBJ longtext, + WNL_FILES longtext, + WNL_EVENT_TIME INT(11) NOT NULL, + PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; + +INSERT INTO `SEQUENCE_TABLE` (`SEQUENCE_NAME`, `NEXT_VAL`) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- ----------------------------------------------------------------- +-- Record schema version. Should be the last step in the init script +-- ----------------------------------------------------------------- +INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES (1, '3.2.0', 'Hive release version 3.2.0'); + +/*!40101 SET character_set_client = @saved_cs_client */; +/*!40103 SET TIME_ZONE=@OLD_TIME_ZONE */; + +/*!40101 SET SQL_MODE=@OLD_SQL_MODE */; +/*!40014 SET FOREIGN_KEY_CHECKS=@OLD_FOREIGN_KEY_CHECKS */; +/*!40014 SET UNIQUE_CHECKS=@OLD_UNIQUE_CHECKS */; +/*!40101 SET CHARACTER_SET_CLIENT=@OLD_CHARACTER_SET_CLIENT */; +/*!40101 SET CHARACTER_SET_RESULTS=@OLD_CHARACTER_SET_RESULTS */; +/*!40101 SET COLLATION_CONNECTION=@OLD_COLLATION_CONNECTION */; +/*!40111 SET SQL_NOTES=@OLD_SQL_NOTES */; + +-- Dump completed on 2012-08-23 0:56:31 diff --git a/standalone-metastore/src/main/sql/mysql/upgrade-2.3.0-to-3.0.0.mysql.sql b/standalone-metastore/src/main/sql/mysql/upgrade-2.3.0-to-3.0.0.mysql.sql index 9b87563b8a..786e38ac6e 100644 --- a/standalone-metastore/src/main/sql/mysql/upgrade-2.3.0-to-3.0.0.mysql.sql +++ b/standalone-metastore/src/main/sql/mysql/upgrade-2.3.0-to-3.0.0.mysql.sql @@ -319,8 +319,8 @@ UPDATE COMPLETED_TXN_COMPONENTS SET CTC_WRITEID = CTC_TXNID; ALTER TABLE TXN_COMPONENTS MODIFY COLUMN TC_TABLE varchar(128) NULL; +ALTER TABLE `TBLS` ADD COLUMN `OWNER_TYPE` VARCHAR(10) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL; + -- These lines need to be last. Insert any changes above. UPDATE VERSION SET SCHEMA_VERSION='3.0.0', VERSION_COMMENT='Hive release version 3.0.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 2.3.0 to 3.0.0' AS ' '; - -ALTER TABLE `TBLS` ADD COLUMN `OWNER_TYPE` VARCHAR(10) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL; \ No newline at end of file diff --git a/standalone-metastore/src/main/sql/mysql/upgrade-3.1.0-to-3.2.0.mysql.sql b/standalone-metastore/src/main/sql/mysql/upgrade-3.1.0-to-3.2.0.mysql.sql new file mode 100644 index 0000000000..b253fce65f --- /dev/null +++ b/standalone-metastore/src/main/sql/mysql/upgrade-3.1.0-to-3.2.0.mysql.sql @@ -0,0 +1,21 @@ +SELECT 'Upgrading MetaStore schema from 3.1.0 to 3.2.0' AS ' '; + +-- HIVE-19267 +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID bigint NOT NULL, + WNL_TXNID bigint NOT NULL, + WNL_WRITEID bigint NOT NULL, + WNL_DATABASE varchar(128) NOT NULL, + WNL_TABLE varchar(128) NOT NULL, + WNL_PARTITION varchar(767) NOT NULL, + WNL_TABLE_OBJ longtext NOT NULL, + WNL_PARTITION_OBJ longtext, + WNL_FILES longtext, + WNL_EVENT_TIME INT(11) NOT NULL, + PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION) +) ENGINE=InnoDB DEFAULT CHARSET=latin1; +INSERT INTO `SEQUENCE_TABLE` (`SEQUENCE_NAME`, `NEXT_VAL`) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- These lines need to be last. Insert any changes above. +UPDATE VERSION SET SCHEMA_VERSION='3.2.0', VERSION_COMMENT='Hive release version 3.2.0' where VER_ID=1; +SELECT 'Finished upgrading MetaStore schema from 3.1.0 to 3.2.0' AS ' '; diff --git a/standalone-metastore/src/main/sql/mysql/upgrade.order.mysql b/standalone-metastore/src/main/sql/mysql/upgrade.order.mysql index f43da9af19..e6eb71ae56 100644 --- a/standalone-metastore/src/main/sql/mysql/upgrade.order.mysql +++ b/standalone-metastore/src/main/sql/mysql/upgrade.order.mysql @@ -15,3 +15,4 @@ 2.2.0-to-2.3.0 2.3.0-to-3.0.0 3.0.0-to-3.1.0 +3.1.0-to-3.2.0 diff --git a/standalone-metastore/src/main/sql/oracle/hive-schema-3.0.0.oracle.sql b/standalone-metastore/src/main/sql/oracle/hive-schema-3.0.0.oracle.sql index 63cc1f75e2..3e2e282f61 100644 --- a/standalone-metastore/src/main/sql/oracle/hive-schema-3.0.0.oracle.sql +++ b/standalone-metastore/src/main/sql/oracle/hive-schema-3.0.0.oracle.sql @@ -1134,7 +1134,6 @@ CREATE TABLE RUNTIME_STATS ( CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); - -- ----------------------------------------------------------------- -- Record schema version. Should be the last step in the init script -- ----------------------------------------------------------------- diff --git a/standalone-metastore/src/main/sql/oracle/hive-schema-3.1.0.oracle.sql b/standalone-metastore/src/main/sql/oracle/hive-schema-3.1.0.oracle.sql index 90388e70e4..b738ebed73 100644 --- a/standalone-metastore/src/main/sql/oracle/hive-schema-3.1.0.oracle.sql +++ b/standalone-metastore/src/main/sql/oracle/hive-schema-3.1.0.oracle.sql @@ -1153,7 +1153,6 @@ CREATE TABLE RUNTIME_STATS ( CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); - -- ----------------------------------------------------------------- -- Record schema version. Should be the last step in the init script -- ----------------------------------------------------------------- diff --git a/standalone-metastore/src/main/sql/oracle/hive-schema-3.2.0.oracle.sql b/standalone-metastore/src/main/sql/oracle/hive-schema-3.2.0.oracle.sql new file mode 100644 index 0000000000..a29752df07 --- /dev/null +++ b/standalone-metastore/src/main/sql/oracle/hive-schema-3.2.0.oracle.sql @@ -0,0 +1,1175 @@ +-- Table SEQUENCE_TABLE is an internal table required by DataNucleus. +-- NOTE: Some versions of SchemaTool do not automatically generate this table. +-- See http://www.datanucleus.org/servlet/jira/browse/NUCRDBMS-416 +CREATE TABLE SEQUENCE_TABLE +( + SEQUENCE_NAME VARCHAR2(255) NOT NULL, + NEXT_VAL NUMBER NOT NULL +); + +ALTER TABLE SEQUENCE_TABLE ADD CONSTRAINT PART_TABLE_PK PRIMARY KEY (SEQUENCE_NAME); + +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MNotificationLog', 1); + +-- Table NUCLEUS_TABLES is an internal table required by DataNucleus. +-- This table is required if datanucleus.autoStartMechanism=SchemaTable +-- NOTE: Some versions of SchemaTool do not automatically generate this table. +-- See http://www.datanucleus.org/servlet/jira/browse/NUCRDBMS-416 +CREATE TABLE NUCLEUS_TABLES +( + CLASS_NAME VARCHAR2(128) NOT NULL, + TABLE_NAME VARCHAR2(128) NOT NULL, + TYPE VARCHAR2(4) NOT NULL, + OWNER VARCHAR2(2) NOT NULL, + VERSION VARCHAR2(20) NOT NULL, + INTERFACE_NAME VARCHAR2(255) NULL +); + +ALTER TABLE NUCLEUS_TABLES ADD CONSTRAINT NUCLEUS_TABLES_PK PRIMARY KEY (CLASS_NAME); + +-- Table PART_COL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MPartitionColumnPrivilege] +CREATE TABLE PART_COL_PRIVS +( + PART_COLUMN_GRANT_ID NUMBER NOT NULL, + "COLUMN_NAME" VARCHAR2(767) NULL, + CREATE_TIME NUMBER (10) NOT NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PART_ID NUMBER NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + PART_COL_PRIV VARCHAR2(128) NULL, + AUTHORIZER VARCHAR2(128) NULL +); + +ALTER TABLE PART_COL_PRIVS ADD CONSTRAINT PART_COL_PRIVS_PK PRIMARY KEY (PART_COLUMN_GRANT_ID); + +-- Table CDS. +CREATE TABLE CDS +( + CD_ID NUMBER NOT NULL +); + +ALTER TABLE CDS ADD CONSTRAINT CDS_PK PRIMARY KEY (CD_ID); + +-- Table COLUMNS_V2 for join relationship +CREATE TABLE COLUMNS_V2 +( + CD_ID NUMBER NOT NULL, + "COMMENT" VARCHAR2(256) NULL, + "COLUMN_NAME" VARCHAR2(767) NOT NULL, + TYPE_NAME CLOB NOT NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE COLUMNS_V2 ADD CONSTRAINT COLUMNS_V2_PK PRIMARY KEY (CD_ID,"COLUMN_NAME"); + +-- Table PARTITION_KEY_VALS for join relationship +CREATE TABLE PARTITION_KEY_VALS +( + PART_ID NUMBER NOT NULL, + PART_KEY_VAL VARCHAR2(256) NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE PARTITION_KEY_VALS ADD CONSTRAINT PARTITION_KEY_VALS_PK PRIMARY KEY (PART_ID,INTEGER_IDX); + +CREATE TABLE CTLGS ( + CTLG_ID NUMBER PRIMARY KEY, + "NAME" VARCHAR2(256), + "DESC" VARCHAR2(4000), + LOCATION_URI VARCHAR2(4000) NOT NULL, + UNIQUE ("NAME") +); + +-- Table DBS for classes [org.apache.hadoop.hive.metastore.model.MDatabase] +CREATE TABLE DBS +( + DB_ID NUMBER NOT NULL, + "DESC" VARCHAR2(4000) NULL, + DB_LOCATION_URI VARCHAR2(4000) NOT NULL, + "NAME" VARCHAR2(128) NULL, + OWNER_NAME VARCHAR2(128) NULL, + OWNER_TYPE VARCHAR2(10) NULL, + CTLG_NAME VARCHAR2(256) +); + +ALTER TABLE DBS ADD CONSTRAINT DBS_PK PRIMARY KEY (DB_ID); + +-- Table PARTITION_PARAMS for join relationship +CREATE TABLE PARTITION_PARAMS +( + PART_ID NUMBER NOT NULL, + PARAM_KEY VARCHAR2(256) NOT NULL, + PARAM_VALUE VARCHAR2(4000) NULL +); + +ALTER TABLE PARTITION_PARAMS ADD CONSTRAINT PARTITION_PARAMS_PK PRIMARY KEY (PART_ID,PARAM_KEY); + +-- Table SERDES for classes [org.apache.hadoop.hive.metastore.model.MSerDeInfo] +CREATE TABLE SERDES +( + SERDE_ID NUMBER NOT NULL, + "NAME" VARCHAR2(128) NULL, + SLIB VARCHAR2(4000) NULL, + "DESCRIPTION" VARCHAR2(4000), + "SERIALIZER_CLASS" VARCHAR2(4000), + "DESERIALIZER_CLASS" VARCHAR2(4000), + "SERDE_TYPE" NUMBER +); + +ALTER TABLE SERDES ADD CONSTRAINT SERDES_PK PRIMARY KEY (SERDE_ID); + +-- Table TYPES for classes [org.apache.hadoop.hive.metastore.model.MType] +CREATE TABLE TYPES +( + TYPES_ID NUMBER NOT NULL, + TYPE_NAME VARCHAR2(128) NULL, + TYPE1 VARCHAR2(767) NULL, + TYPE2 VARCHAR2(767) NULL +); + +ALTER TABLE TYPES ADD CONSTRAINT TYPES_PK PRIMARY KEY (TYPES_ID); + +-- Table PARTITION_KEYS for join relationship +CREATE TABLE PARTITION_KEYS +( + TBL_ID NUMBER NOT NULL, + PKEY_COMMENT VARCHAR2(4000) NULL, + PKEY_NAME VARCHAR2(128) NOT NULL, + PKEY_TYPE VARCHAR2(767) NOT NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE PARTITION_KEYS ADD CONSTRAINT PARTITION_KEY_PK PRIMARY KEY (TBL_ID,PKEY_NAME); + +-- Table ROLES for classes [org.apache.hadoop.hive.metastore.model.MRole] +CREATE TABLE ROLES +( + ROLE_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + OWNER_NAME VARCHAR2(128) NULL, + ROLE_NAME VARCHAR2(128) NULL +); + +ALTER TABLE ROLES ADD CONSTRAINT ROLES_PK PRIMARY KEY (ROLE_ID); + +-- Table PARTITIONS for classes [org.apache.hadoop.hive.metastore.model.MPartition] +CREATE TABLE PARTITIONS +( + PART_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + LAST_ACCESS_TIME NUMBER (10) NOT NULL, + PART_NAME VARCHAR2(767) NULL, + SD_ID NUMBER NULL, + TBL_ID NUMBER NULL +); + +ALTER TABLE PARTITIONS ADD CONSTRAINT PARTITIONS_PK PRIMARY KEY (PART_ID); + +-- Table INDEX_PARAMS for join relationship +CREATE TABLE INDEX_PARAMS +( + INDEX_ID NUMBER NOT NULL, + PARAM_KEY VARCHAR2(256) NOT NULL, + PARAM_VALUE VARCHAR2(4000) NULL +); + +ALTER TABLE INDEX_PARAMS ADD CONSTRAINT INDEX_PARAMS_PK PRIMARY KEY (INDEX_ID,PARAM_KEY); + +-- Table TBL_COL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MTableColumnPrivilege] +CREATE TABLE TBL_COL_PRIVS +( + TBL_COLUMN_GRANT_ID NUMBER NOT NULL, + "COLUMN_NAME" VARCHAR2(767) NULL, + CREATE_TIME NUMBER (10) NOT NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + TBL_COL_PRIV VARCHAR2(128) NULL, + TBL_ID NUMBER NULL, + AUTHORIZER VARCHAR2(128) NULL +); + +ALTER TABLE TBL_COL_PRIVS ADD CONSTRAINT TBL_COL_PRIVS_PK PRIMARY KEY (TBL_COLUMN_GRANT_ID); + +-- Table IDXS for classes [org.apache.hadoop.hive.metastore.model.MIndex] +CREATE TABLE IDXS +( + INDEX_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + DEFERRED_REBUILD NUMBER(1) NOT NULL CHECK (DEFERRED_REBUILD IN (1,0)), + INDEX_HANDLER_CLASS VARCHAR2(4000) NULL, + INDEX_NAME VARCHAR2(128) NULL, + INDEX_TBL_ID NUMBER NULL, + LAST_ACCESS_TIME NUMBER (10) NOT NULL, + ORIG_TBL_ID NUMBER NULL, + SD_ID NUMBER NULL +); + +ALTER TABLE IDXS ADD CONSTRAINT IDXS_PK PRIMARY KEY (INDEX_ID); + +-- Table BUCKETING_COLS for join relationship +CREATE TABLE BUCKETING_COLS +( + SD_ID NUMBER NOT NULL, + BUCKET_COL_NAME VARCHAR2(256) NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE BUCKETING_COLS ADD CONSTRAINT BUCKETING_COLS_PK PRIMARY KEY (SD_ID,INTEGER_IDX); + +-- Table TYPE_FIELDS for join relationship +CREATE TABLE TYPE_FIELDS +( + TYPE_NAME NUMBER NOT NULL, + "COMMENT" VARCHAR2(256) NULL, + FIELD_NAME VARCHAR2(128) NOT NULL, + FIELD_TYPE VARCHAR2(767) NOT NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE TYPE_FIELDS ADD CONSTRAINT TYPE_FIELDS_PK PRIMARY KEY (TYPE_NAME,FIELD_NAME); + +-- Table SD_PARAMS for join relationship +CREATE TABLE SD_PARAMS +( + SD_ID NUMBER NOT NULL, + PARAM_KEY VARCHAR2(256) NOT NULL, + PARAM_VALUE CLOB NULL +); + +ALTER TABLE SD_PARAMS ADD CONSTRAINT SD_PARAMS_PK PRIMARY KEY (SD_ID,PARAM_KEY); + +-- Table GLOBAL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MGlobalPrivilege] +CREATE TABLE GLOBAL_PRIVS +( + USER_GRANT_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + USER_PRIV VARCHAR2(128) NULL, + AUTHORIZER VARCHAR2(128) NULL +); + +ALTER TABLE GLOBAL_PRIVS ADD CONSTRAINT GLOBAL_PRIVS_PK PRIMARY KEY (USER_GRANT_ID); + +-- Table SDS for classes [org.apache.hadoop.hive.metastore.model.MStorageDescriptor] +CREATE TABLE SDS +( + SD_ID NUMBER NOT NULL, + CD_ID NUMBER NULL, + INPUT_FORMAT VARCHAR2(4000) NULL, + IS_COMPRESSED NUMBER(1) NOT NULL CHECK (IS_COMPRESSED IN (1,0)), + LOCATION VARCHAR2(4000) NULL, + NUM_BUCKETS NUMBER (10) NOT NULL, + OUTPUT_FORMAT VARCHAR2(4000) NULL, + SERDE_ID NUMBER NULL, + IS_STOREDASSUBDIRECTORIES NUMBER(1) NOT NULL CHECK (IS_STOREDASSUBDIRECTORIES IN (1,0)) +); + +ALTER TABLE SDS ADD CONSTRAINT SDS_PK PRIMARY KEY (SD_ID); + +-- Table TABLE_PARAMS for join relationship +CREATE TABLE TABLE_PARAMS +( + TBL_ID NUMBER NOT NULL, + PARAM_KEY VARCHAR2(256) NOT NULL, + PARAM_VALUE CLOB NULL +); + +ALTER TABLE TABLE_PARAMS ADD CONSTRAINT TABLE_PARAMS_PK PRIMARY KEY (TBL_ID,PARAM_KEY); + +-- Table SORT_COLS for join relationship +CREATE TABLE SORT_COLS +( + SD_ID NUMBER NOT NULL, + "COLUMN_NAME" VARCHAR2(767) NULL, + "ORDER" NUMBER (10) NOT NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE SORT_COLS ADD CONSTRAINT SORT_COLS_PK PRIMARY KEY (SD_ID,INTEGER_IDX); + +-- Table TBL_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MTablePrivilege] +CREATE TABLE TBL_PRIVS +( + TBL_GRANT_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + TBL_PRIV VARCHAR2(128) NULL, + TBL_ID NUMBER NULL, + AUTHORIZER VARCHAR2(128) NULL +); + +ALTER TABLE TBL_PRIVS ADD CONSTRAINT TBL_PRIVS_PK PRIMARY KEY (TBL_GRANT_ID); + +-- Table DATABASE_PARAMS for join relationship +CREATE TABLE DATABASE_PARAMS +( + DB_ID NUMBER NOT NULL, + PARAM_KEY VARCHAR2(180) NOT NULL, + PARAM_VALUE VARCHAR2(4000) NULL +); + +ALTER TABLE DATABASE_PARAMS ADD CONSTRAINT DATABASE_PARAMS_PK PRIMARY KEY (DB_ID,PARAM_KEY); + +-- Table ROLE_MAP for classes [org.apache.hadoop.hive.metastore.model.MRoleMap] +CREATE TABLE ROLE_MAP +( + ROLE_GRANT_ID NUMBER NOT NULL, + ADD_TIME NUMBER (10) NOT NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + ROLE_ID NUMBER NULL +); + +ALTER TABLE ROLE_MAP ADD CONSTRAINT ROLE_MAP_PK PRIMARY KEY (ROLE_GRANT_ID); + +-- Table SERDE_PARAMS for join relationship +CREATE TABLE SERDE_PARAMS +( + SERDE_ID NUMBER NOT NULL, + PARAM_KEY VARCHAR2(256) NOT NULL, + PARAM_VALUE CLOB NULL +); + +ALTER TABLE SERDE_PARAMS ADD CONSTRAINT SERDE_PARAMS_PK PRIMARY KEY (SERDE_ID,PARAM_KEY); + +-- Table PART_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MPartitionPrivilege] +CREATE TABLE PART_PRIVS +( + PART_GRANT_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PART_ID NUMBER NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + PART_PRIV VARCHAR2(128) NULL, + AUTHORIZER VARCHAR2(128) NULL +); + +ALTER TABLE PART_PRIVS ADD CONSTRAINT PART_PRIVS_PK PRIMARY KEY (PART_GRANT_ID); + +-- Table DB_PRIVS for classes [org.apache.hadoop.hive.metastore.model.MDBPrivilege] +CREATE TABLE DB_PRIVS +( + DB_GRANT_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + DB_ID NUMBER NULL, + GRANT_OPTION NUMBER (5) NOT NULL, + GRANTOR VARCHAR2(128) NULL, + GRANTOR_TYPE VARCHAR2(128) NULL, + PRINCIPAL_NAME VARCHAR2(128) NULL, + PRINCIPAL_TYPE VARCHAR2(128) NULL, + DB_PRIV VARCHAR2(128) NULL, + AUTHORIZER VARCHAR2(128) NULL +); + +ALTER TABLE DB_PRIVS ADD CONSTRAINT DB_PRIVS_PK PRIMARY KEY (DB_GRANT_ID); + +-- Table TBLS for classes [org.apache.hadoop.hive.metastore.model.MTable] +CREATE TABLE TBLS +( + TBL_ID NUMBER NOT NULL, + CREATE_TIME NUMBER (10) NOT NULL, + DB_ID NUMBER NULL, + LAST_ACCESS_TIME NUMBER (10) NOT NULL, + OWNER VARCHAR2(767) NULL, + OWNER_TYPE VARCHAR2(10) NULL, + RETENTION NUMBER (10) NOT NULL, + SD_ID NUMBER NULL, + TBL_NAME VARCHAR2(256) NULL, + TBL_TYPE VARCHAR2(128) NULL, + VIEW_EXPANDED_TEXT CLOB NULL, + VIEW_ORIGINAL_TEXT CLOB NULL, + IS_REWRITE_ENABLED NUMBER(1) DEFAULT 0 NOT NULL CHECK (IS_REWRITE_ENABLED IN (1,0)) +); + +ALTER TABLE TBLS ADD CONSTRAINT TBLS_PK PRIMARY KEY (TBL_ID); + +-- Table MV_CREATION_METADATA for classes [org.apache.hadoop.hive.metastore.model.MCreationMetadata] +CREATE TABLE MV_CREATION_METADATA +( + MV_CREATION_METADATA_ID NUMBER NOT NULL, + CAT_NAME VARCHAR2(256) NOT NULL, + DB_NAME VARCHAR2(128) NOT NULL, + TBL_NAME VARCHAR2(256) NOT NULL, + TXN_LIST CLOB NULL, + MATERIALIZATION_TIME NUMBER NOT NULL +); + +ALTER TABLE MV_CREATION_METADATA ADD CONSTRAINT MV_CREATION_METADATA_PK PRIMARY KEY (MV_CREATION_METADATA_ID); + +CREATE UNIQUE INDEX UNIQUE_TABLE ON MV_CREATION_METADATA ("DB_NAME", "TBL_NAME"); + +-- Table MV_CREATION_METADATA for classes [org.apache.hadoop.hive.metastore.model.MCreationMetadata] +CREATE TABLE MV_TABLES_USED +( + MV_CREATION_METADATA_ID NUMBER NOT NULL, + TBL_ID NUMBER NOT NULL +); + +-- Table PARTITION_EVENTS for classes [org.apache.hadoop.hive.metastore.model.MPartitionEvent] +CREATE TABLE PARTITION_EVENTS +( + PART_NAME_ID NUMBER NOT NULL, + CAT_NAME VARCHAR2(256) NULL, + DB_NAME VARCHAR2(128) NULL, + EVENT_TIME NUMBER NOT NULL, + EVENT_TYPE NUMBER (10) NOT NULL, + PARTITION_NAME VARCHAR2(767) NULL, + TBL_NAME VARCHAR2(256) NULL +); + +ALTER TABLE PARTITION_EVENTS ADD CONSTRAINT PARTITION_EVENTS_PK PRIMARY KEY (PART_NAME_ID); + +-- Table SKEWED_STRING_LIST for classes [org.apache.hadoop.hive.metastore.model.MStringList] +CREATE TABLE SKEWED_STRING_LIST +( + STRING_LIST_ID NUMBER NOT NULL +); + +ALTER TABLE SKEWED_STRING_LIST ADD CONSTRAINT SKEWED_STRING_LIST_PK PRIMARY KEY (STRING_LIST_ID); + +CREATE TABLE SKEWED_STRING_LIST_VALUES +( + STRING_LIST_ID NUMBER NOT NULL, + "STRING_LIST_VALUE" VARCHAR2(256) NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE SKEWED_STRING_LIST_VALUES ADD CONSTRAINT SKEWED_STRING_LIST_VALUES_PK PRIMARY KEY (STRING_LIST_ID,INTEGER_IDX); + +ALTER TABLE SKEWED_STRING_LIST_VALUES ADD CONSTRAINT SKEWED_STRING_LIST_VALUES_FK1 FOREIGN KEY (STRING_LIST_ID) REFERENCES SKEWED_STRING_LIST (STRING_LIST_ID) INITIALLY DEFERRED ; + +CREATE TABLE SKEWED_COL_NAMES +( + SD_ID NUMBER NOT NULL, + "SKEWED_COL_NAME" VARCHAR2(256) NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE SKEWED_COL_NAMES ADD CONSTRAINT SKEWED_COL_NAMES_PK PRIMARY KEY (SD_ID,INTEGER_IDX); + +ALTER TABLE SKEWED_COL_NAMES ADD CONSTRAINT SKEWED_COL_NAMES_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE TABLE SKEWED_COL_VALUE_LOC_MAP +( + SD_ID NUMBER NOT NULL, + STRING_LIST_ID_KID NUMBER NOT NULL, + "LOCATION" VARCHAR2(4000) NULL +); + +CREATE TABLE MASTER_KEYS +( + KEY_ID NUMBER (10) NOT NULL, + MASTER_KEY VARCHAR2(767) NULL +); + +CREATE TABLE DELEGATION_TOKENS +( + TOKEN_IDENT VARCHAR2(767) NOT NULL, + TOKEN VARCHAR2(767) NULL +); + +ALTER TABLE SKEWED_COL_VALUE_LOC_MAP ADD CONSTRAINT SKEWED_COL_VALUE_LOC_MAP_PK PRIMARY KEY (SD_ID,STRING_LIST_ID_KID); + +ALTER TABLE SKEWED_COL_VALUE_LOC_MAP ADD CONSTRAINT SKEWED_COL_VALUE_LOC_MAP_FK1 FOREIGN KEY (STRING_LIST_ID_KID) REFERENCES SKEWED_STRING_LIST (STRING_LIST_ID) INITIALLY DEFERRED ; + +ALTER TABLE SKEWED_COL_VALUE_LOC_MAP ADD CONSTRAINT SKEWED_COL_VALUE_LOC_MAP_FK2 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE TABLE SKEWED_VALUES +( + SD_ID_OID NUMBER NOT NULL, + STRING_LIST_ID_EID NUMBER NOT NULL, + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE SKEWED_VALUES ADD CONSTRAINT SKEWED_VALUES_PK PRIMARY KEY (SD_ID_OID,INTEGER_IDX); + +ALTER TABLE SKEWED_VALUES ADD CONSTRAINT SKEWED_VALUES_FK1 FOREIGN KEY (STRING_LIST_ID_EID) REFERENCES SKEWED_STRING_LIST (STRING_LIST_ID) INITIALLY DEFERRED ; + +ALTER TABLE SKEWED_VALUES ADD CONSTRAINT SKEWED_VALUES_FK2 FOREIGN KEY (SD_ID_OID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +ALTER TABLE DBS ADD CONSTRAINT CTLGS_FK FOREIGN KEY (CTLG_NAME) REFERENCES CTLGS ("NAME") INITIALLY DEFERRED; + +-- column statistics + +CREATE TABLE TAB_COL_STATS ( + CS_ID NUMBER NOT NULL, + CAT_NAME VARCHAR2(256) NOT NULL, + DB_NAME VARCHAR2(128) NOT NULL, + TABLE_NAME VARCHAR2(256) NOT NULL, + COLUMN_NAME VARCHAR2(767) NOT NULL, + COLUMN_TYPE VARCHAR2(128) NOT NULL, + TBL_ID NUMBER NOT NULL, + LONG_LOW_VALUE NUMBER, + LONG_HIGH_VALUE NUMBER, + DOUBLE_LOW_VALUE NUMBER, + DOUBLE_HIGH_VALUE NUMBER, + BIG_DECIMAL_LOW_VALUE VARCHAR2(4000), + BIG_DECIMAL_HIGH_VALUE VARCHAR2(4000), + NUM_NULLS NUMBER NOT NULL, + NUM_DISTINCTS NUMBER, + BIT_VECTOR BLOB, + AVG_COL_LEN NUMBER, + MAX_COL_LEN NUMBER, + NUM_TRUES NUMBER, + NUM_FALSES NUMBER, + LAST_ANALYZED NUMBER NOT NULL +); + +ALTER TABLE TAB_COL_STATS ADD CONSTRAINT TAB_COL_STATS_PKEY PRIMARY KEY (CS_ID); + +ALTER TABLE TAB_COL_STATS ADD CONSTRAINT TAB_COL_STATS_FK FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +CREATE INDEX TAB_COL_STATS_N49 ON TAB_COL_STATS(TBL_ID); + +CREATE INDEX TAB_COL_STATS_IDX ON TAB_COL_STATS (CAT_NAME, DB_NAME, TABLE_NAME, COLUMN_NAME); + +CREATE TABLE VERSION ( + VER_ID NUMBER NOT NULL, + SCHEMA_VERSION VARCHAR(127) NOT NULL, + VERSION_COMMENT VARCHAR(255) +); +ALTER TABLE VERSION ADD CONSTRAINT VERSION_PK PRIMARY KEY (VER_ID); + +CREATE TABLE PART_COL_STATS ( + CS_ID NUMBER NOT NULL, + CAT_NAME VARCHAR2(256) NOT NULL, + DB_NAME VARCHAR2(128) NOT NULL, + TABLE_NAME VARCHAR2(256) NOT NULL, + PARTITION_NAME VARCHAR2(767) NOT NULL, + COLUMN_NAME VARCHAR2(767) NOT NULL, + COLUMN_TYPE VARCHAR2(128) NOT NULL, + PART_ID NUMBER NOT NULL, + LONG_LOW_VALUE NUMBER, + LONG_HIGH_VALUE NUMBER, + DOUBLE_LOW_VALUE NUMBER, + DOUBLE_HIGH_VALUE NUMBER, + BIG_DECIMAL_LOW_VALUE VARCHAR2(4000), + BIG_DECIMAL_HIGH_VALUE VARCHAR2(4000), + NUM_NULLS NUMBER NOT NULL, + NUM_DISTINCTS NUMBER, + BIT_VECTOR BLOB, + AVG_COL_LEN NUMBER, + MAX_COL_LEN NUMBER, + NUM_TRUES NUMBER, + NUM_FALSES NUMBER, + LAST_ANALYZED NUMBER NOT NULL +); + +ALTER TABLE PART_COL_STATS ADD CONSTRAINT PART_COL_STATS_PKEY PRIMARY KEY (CS_ID); + +ALTER TABLE PART_COL_STATS ADD CONSTRAINT PART_COL_STATS_FK FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) INITIALLY DEFERRED; + +CREATE INDEX PART_COL_STATS_N49 ON PART_COL_STATS (PART_ID); + +CREATE INDEX PCS_STATS_IDX ON PART_COL_STATS (CAT_NAME, DB_NAME,TABLE_NAME,COLUMN_NAME,PARTITION_NAME); + +CREATE TABLE FUNCS ( + FUNC_ID NUMBER NOT NULL, + CLASS_NAME VARCHAR2(4000), + CREATE_TIME NUMBER(10) NOT NULL, + DB_ID NUMBER, + FUNC_NAME VARCHAR2(128), + FUNC_TYPE NUMBER(10) NOT NULL, + OWNER_NAME VARCHAR2(128), + OWNER_TYPE VARCHAR2(10) +); + +ALTER TABLE FUNCS ADD CONSTRAINT FUNCS_PK PRIMARY KEY (FUNC_ID); + +CREATE TABLE FUNC_RU ( + FUNC_ID NUMBER NOT NULL, + RESOURCE_TYPE NUMBER(10) NOT NULL, + RESOURCE_URI VARCHAR2(4000), + INTEGER_IDX NUMBER(10) NOT NULL +); + +ALTER TABLE FUNC_RU ADD CONSTRAINT FUNC_RU_PK PRIMARY KEY (FUNC_ID, INTEGER_IDX); + +CREATE TABLE NOTIFICATION_LOG +( + NL_ID NUMBER NOT NULL, + EVENT_ID NUMBER NOT NULL, + EVENT_TIME NUMBER(10) NOT NULL, + EVENT_TYPE VARCHAR2(32) NOT NULL, + CAT_NAME VARCHAR2(256), + DB_NAME VARCHAR2(128), + TBL_NAME VARCHAR2(256), + MESSAGE CLOB NULL, + MESSAGE_FORMAT VARCHAR(16) NULL +); + +ALTER TABLE NOTIFICATION_LOG ADD CONSTRAINT NOTIFICATION_LOG_PK PRIMARY KEY (NL_ID); + +CREATE TABLE NOTIFICATION_SEQUENCE +( + NNI_ID NUMBER NOT NULL, + NEXT_EVENT_ID NUMBER NOT NULL +); + +ALTER TABLE NOTIFICATION_SEQUENCE ADD CONSTRAINT NOTIFICATION_SEQUENCE_PK PRIMARY KEY (NNI_ID); + +INSERT INTO NOTIFICATION_SEQUENCE (NNI_ID, NEXT_EVENT_ID) SELECT 1,1 FROM DUAL WHERE NOT EXISTS ( SELECT NEXT_EVENT_ID FROM NOTIFICATION_SEQUENCE); + +-- Tables to manage resource plans. + +CREATE TABLE WM_RESOURCEPLAN +( + RP_ID NUMBER NOT NULL, + "NAME" VARCHAR2(128) NOT NULL, + QUERY_PARALLELISM NUMBER(10), + STATUS VARCHAR2(20) NOT NULL, + DEFAULT_POOL_ID NUMBER +); + +ALTER TABLE WM_RESOURCEPLAN ADD CONSTRAINT WM_RESOURCEPLAN_PK PRIMARY KEY (RP_ID); + +CREATE TABLE WM_POOL +( + POOL_ID NUMBER NOT NULL, + RP_ID NUMBER NOT NULL, + PATH VARCHAR2(1024) NOT NULL, + ALLOC_FRACTION NUMBER, + QUERY_PARALLELISM NUMBER(10), + SCHEDULING_POLICY VARCHAR2(1024) +); + +ALTER TABLE WM_POOL ADD CONSTRAINT WM_POOL_PK PRIMARY KEY (POOL_ID); + +CREATE TABLE WM_TRIGGER +( + TRIGGER_ID NUMBER NOT NULL, + RP_ID NUMBER NOT NULL, + "NAME" VARCHAR2(128) NOT NULL, + TRIGGER_EXPRESSION VARCHAR2(1024), + ACTION_EXPRESSION VARCHAR2(1024), + IS_IN_UNMANAGED NUMBER(1) DEFAULT 0 NOT NULL CHECK (IS_IN_UNMANAGED IN (1,0)) +); + +ALTER TABLE WM_TRIGGER ADD CONSTRAINT WM_TRIGGER_PK PRIMARY KEY (TRIGGER_ID); + +CREATE TABLE WM_POOL_TO_TRIGGER +( + POOL_ID NUMBER NOT NULL, + TRIGGER_ID NUMBER NOT NULL +); + +ALTER TABLE WM_POOL_TO_TRIGGER ADD CONSTRAINT WM_POOL_TO_TRIGGER_PK PRIMARY KEY (POOL_ID, TRIGGER_ID); + +CREATE TABLE WM_MAPPING +( + MAPPING_ID NUMBER NOT NULL, + RP_ID NUMBER NOT NULL, + ENTITY_TYPE VARCHAR2(128) NOT NULL, + ENTITY_NAME VARCHAR2(128) NOT NULL, + POOL_ID NUMBER NOT NULL, + ORDERING NUMBER(10) +); + +ALTER TABLE WM_MAPPING ADD CONSTRAINT WM_MAPPING_PK PRIMARY KEY (MAPPING_ID); + +-- Constraints for table PART_COL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionColumnPrivilege] +ALTER TABLE PART_COL_PRIVS ADD CONSTRAINT PART_COL_PRIVS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) INITIALLY DEFERRED ; + +CREATE INDEX PART_COL_PRIVS_N49 ON PART_COL_PRIVS (PART_ID); + +CREATE INDEX PARTITIONCOLUMNPRIVILEGEINDEX ON PART_COL_PRIVS (AUTHORIZER,PART_ID,"COLUMN_NAME",PRINCIPAL_NAME,PRINCIPAL_TYPE,PART_COL_PRIV,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table COLUMNS_V2 +ALTER TABLE COLUMNS_V2 ADD CONSTRAINT COLUMNS_V2_FK1 FOREIGN KEY (CD_ID) REFERENCES CDS (CD_ID) INITIALLY DEFERRED ; + +CREATE INDEX COLUMNS_V2_N49 ON COLUMNS_V2 (CD_ID); + + +-- Constraints for table PARTITION_KEY_VALS +ALTER TABLE PARTITION_KEY_VALS ADD CONSTRAINT PARTITION_KEY_VALS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) INITIALLY DEFERRED ; + +CREATE INDEX PARTITION_KEY_VALS_N49 ON PARTITION_KEY_VALS (PART_ID); + + +-- Constraints for table DBS for class(es) [org.apache.hadoop.hive.metastore.model.MDatabase] +CREATE UNIQUE INDEX UNIQUE_DATABASE ON DBS ("NAME", CTLG_NAME); + + +-- Constraints for table PARTITION_PARAMS +ALTER TABLE PARTITION_PARAMS ADD CONSTRAINT PARTITION_PARAMS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) INITIALLY DEFERRED ; + +CREATE INDEX PARTITION_PARAMS_N49 ON PARTITION_PARAMS (PART_ID); + + +-- Constraints for table SERDES for class(es) [org.apache.hadoop.hive.metastore.model.MSerDeInfo] + +-- Constraints for table TYPES for class(es) [org.apache.hadoop.hive.metastore.model.MType] +CREATE UNIQUE INDEX UNIQUE_TYPE ON TYPES (TYPE_NAME); + + +-- Constraints for table PARTITION_KEYS +ALTER TABLE PARTITION_KEYS ADD CONSTRAINT PARTITION_KEYS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +CREATE INDEX PARTITION_KEYS_N49 ON PARTITION_KEYS (TBL_ID); + + +-- Constraints for table ROLES for class(es) [org.apache.hadoop.hive.metastore.model.MRole] +CREATE UNIQUE INDEX ROLEENTITYINDEX ON ROLES (ROLE_NAME); + + +-- Constraints for table PARTITIONS for class(es) [org.apache.hadoop.hive.metastore.model.MPartition] +ALTER TABLE PARTITIONS ADD CONSTRAINT PARTITIONS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +ALTER TABLE PARTITIONS ADD CONSTRAINT PARTITIONS_FK2 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE INDEX PARTITIONS_N49 ON PARTITIONS (SD_ID); + +CREATE INDEX PARTITIONS_N50 ON PARTITIONS (TBL_ID); + +CREATE UNIQUE INDEX UNIQUEPARTITION ON PARTITIONS (PART_NAME,TBL_ID); + + +-- Constraints for table INDEX_PARAMS +ALTER TABLE INDEX_PARAMS ADD CONSTRAINT INDEX_PARAMS_FK1 FOREIGN KEY (INDEX_ID) REFERENCES IDXS (INDEX_ID) INITIALLY DEFERRED ; + +CREATE INDEX INDEX_PARAMS_N49 ON INDEX_PARAMS (INDEX_ID); + + +-- Constraints for table TBL_COL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MTableColumnPrivilege] +ALTER TABLE TBL_COL_PRIVS ADD CONSTRAINT TBL_COL_PRIVS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +CREATE INDEX TABLECOLUMNPRIVILEGEINDEX ON TBL_COL_PRIVS (AUTHORIZER,TBL_ID,"COLUMN_NAME",PRINCIPAL_NAME,PRINCIPAL_TYPE,TBL_COL_PRIV,GRANTOR,GRANTOR_TYPE); + +CREATE INDEX TBL_COL_PRIVS_N49 ON TBL_COL_PRIVS (TBL_ID); + + +-- Constraints for table IDXS for class(es) [org.apache.hadoop.hive.metastore.model.MIndex] +ALTER TABLE IDXS ADD CONSTRAINT IDXS_FK2 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +ALTER TABLE IDXS ADD CONSTRAINT IDXS_FK1 FOREIGN KEY (ORIG_TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +ALTER TABLE IDXS ADD CONSTRAINT IDXS_FK3 FOREIGN KEY (INDEX_TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +CREATE UNIQUE INDEX UNIQUEINDEX ON IDXS (INDEX_NAME,ORIG_TBL_ID); + +CREATE INDEX IDXS_N50 ON IDXS (INDEX_TBL_ID); + +CREATE INDEX IDXS_N51 ON IDXS (SD_ID); + +CREATE INDEX IDXS_N49 ON IDXS (ORIG_TBL_ID); + + +-- Constraints for table BUCKETING_COLS +ALTER TABLE BUCKETING_COLS ADD CONSTRAINT BUCKETING_COLS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE INDEX BUCKETING_COLS_N49 ON BUCKETING_COLS (SD_ID); + + +-- Constraints for table TYPE_FIELDS +ALTER TABLE TYPE_FIELDS ADD CONSTRAINT TYPE_FIELDS_FK1 FOREIGN KEY (TYPE_NAME) REFERENCES TYPES (TYPES_ID) INITIALLY DEFERRED ; + +CREATE INDEX TYPE_FIELDS_N49 ON TYPE_FIELDS (TYPE_NAME); + + +-- Constraints for table SD_PARAMS +ALTER TABLE SD_PARAMS ADD CONSTRAINT SD_PARAMS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE INDEX SD_PARAMS_N49 ON SD_PARAMS (SD_ID); + + +-- Constraints for table GLOBAL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MGlobalPrivilege] +CREATE UNIQUE INDEX GLOBALPRIVILEGEINDEX ON GLOBAL_PRIVS (AUTHORIZER,PRINCIPAL_NAME,PRINCIPAL_TYPE,USER_PRIV,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table SDS for class(es) [org.apache.hadoop.hive.metastore.model.MStorageDescriptor] +ALTER TABLE SDS ADD CONSTRAINT SDS_FK1 FOREIGN KEY (SERDE_ID) REFERENCES SERDES (SERDE_ID) INITIALLY DEFERRED ; +ALTER TABLE SDS ADD CONSTRAINT SDS_FK2 FOREIGN KEY (CD_ID) REFERENCES CDS (CD_ID) INITIALLY DEFERRED ; + +CREATE INDEX SDS_N49 ON SDS (SERDE_ID); +CREATE INDEX SDS_N50 ON SDS (CD_ID); + + +-- Constraints for table TABLE_PARAMS +ALTER TABLE TABLE_PARAMS ADD CONSTRAINT TABLE_PARAMS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +CREATE INDEX TABLE_PARAMS_N49 ON TABLE_PARAMS (TBL_ID); + + +-- Constraints for table SORT_COLS +ALTER TABLE SORT_COLS ADD CONSTRAINT SORT_COLS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE INDEX SORT_COLS_N49 ON SORT_COLS (SD_ID); + + +-- Constraints for table TBL_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MTablePrivilege] +ALTER TABLE TBL_PRIVS ADD CONSTRAINT TBL_PRIVS_FK1 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID) INITIALLY DEFERRED ; + +CREATE INDEX TBL_PRIVS_N49 ON TBL_PRIVS (TBL_ID); + +CREATE INDEX TABLEPRIVILEGEINDEX ON TBL_PRIVS (AUTHORIZER,TBL_ID,PRINCIPAL_NAME,PRINCIPAL_TYPE,TBL_PRIV,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table DATABASE_PARAMS +ALTER TABLE DATABASE_PARAMS ADD CONSTRAINT DATABASE_PARAMS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) INITIALLY DEFERRED ; + +CREATE INDEX DATABASE_PARAMS_N49 ON DATABASE_PARAMS (DB_ID); + + +-- Constraints for table ROLE_MAP for class(es) [org.apache.hadoop.hive.metastore.model.MRoleMap] +ALTER TABLE ROLE_MAP ADD CONSTRAINT ROLE_MAP_FK1 FOREIGN KEY (ROLE_ID) REFERENCES ROLES (ROLE_ID) INITIALLY DEFERRED ; + +CREATE INDEX ROLE_MAP_N49 ON ROLE_MAP (ROLE_ID); + +CREATE UNIQUE INDEX USERROLEMAPINDEX ON ROLE_MAP (PRINCIPAL_NAME,ROLE_ID,GRANTOR,GRANTOR_TYPE); + + +-- Constraints for table SERDE_PARAMS +ALTER TABLE SERDE_PARAMS ADD CONSTRAINT SERDE_PARAMS_FK1 FOREIGN KEY (SERDE_ID) REFERENCES SERDES (SERDE_ID) INITIALLY DEFERRED ; + +CREATE INDEX SERDE_PARAMS_N49 ON SERDE_PARAMS (SERDE_ID); + + +-- Constraints for table PART_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionPrivilege] +ALTER TABLE PART_PRIVS ADD CONSTRAINT PART_PRIVS_FK1 FOREIGN KEY (PART_ID) REFERENCES PARTITIONS (PART_ID) INITIALLY DEFERRED ; + +CREATE INDEX PARTPRIVILEGEINDEX ON PART_PRIVS (AUTHORIZER,PART_ID,PRINCIPAL_NAME,PRINCIPAL_TYPE,PART_PRIV,GRANTOR,GRANTOR_TYPE); + +CREATE INDEX PART_PRIVS_N49 ON PART_PRIVS (PART_ID); + + +-- Constraints for table DB_PRIVS for class(es) [org.apache.hadoop.hive.metastore.model.MDBPrivilege] +ALTER TABLE DB_PRIVS ADD CONSTRAINT DB_PRIVS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) INITIALLY DEFERRED ; + +CREATE UNIQUE INDEX DBPRIVILEGEINDEX ON DB_PRIVS (AUTHORIZER,DB_ID,PRINCIPAL_NAME,PRINCIPAL_TYPE,DB_PRIV,GRANTOR,GRANTOR_TYPE); + +CREATE INDEX DB_PRIVS_N49 ON DB_PRIVS (DB_ID); + + +-- Constraints for table TBLS for class(es) [org.apache.hadoop.hive.metastore.model.MTable] +ALTER TABLE TBLS ADD CONSTRAINT TBLS_FK2 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) INITIALLY DEFERRED ; + +ALTER TABLE TBLS ADD CONSTRAINT TBLS_FK1 FOREIGN KEY (SD_ID) REFERENCES SDS (SD_ID) INITIALLY DEFERRED ; + +CREATE INDEX TBLS_N49 ON TBLS (DB_ID); + +CREATE UNIQUE INDEX UNIQUETABLE ON TBLS (TBL_NAME,DB_ID); + +CREATE INDEX TBLS_N50 ON TBLS (SD_ID); + + +-- Constraints for table PARTITION_EVENTS for class(es) [org.apache.hadoop.hive.metastore.model.MPartitionEvent] +CREATE INDEX PARTITIONEVENTINDEX ON PARTITION_EVENTS (PARTITION_NAME); + + +-- Constraints for table FUNCS for class(es) [org.apache.hadoop.hive.metastore.model.MFunctions] +ALTER TABLE FUNCS ADD CONSTRAINT FUNCS_FK1 FOREIGN KEY (DB_ID) REFERENCES DBS (DB_ID) INITIALLY DEFERRED; + +CREATE UNIQUE INDEX UNIQUEFUNCTION ON FUNCS (FUNC_NAME, DB_ID); + +CREATE INDEX FUNCS_N49 ON FUNCS (DB_ID); + + +-- Constraints for table FUNC_RU for class(es) [org.apache.hadoop.hive.metastore.model.MFunctions] +ALTER TABLE FUNC_RU ADD CONSTRAINT FUNC_RU_FK1 FOREIGN KEY (FUNC_ID) REFERENCES FUNCS (FUNC_ID) INITIALLY DEFERRED; + +CREATE INDEX FUNC_RU_N49 ON FUNC_RU (FUNC_ID); + +CREATE TABLE KEY_CONSTRAINTS +( + CHILD_CD_ID NUMBER, + CHILD_INTEGER_IDX NUMBER, + CHILD_TBL_ID NUMBER, + PARENT_CD_ID NUMBER, + PARENT_INTEGER_IDX NUMBER NOT NULL, + PARENT_TBL_ID NUMBER NOT NULL, + POSITION NUMBER NOT NULL, + CONSTRAINT_NAME VARCHAR(400) NOT NULL, + CONSTRAINT_TYPE NUMBER NOT NULL, + UPDATE_RULE NUMBER, + DELETE_RULE NUMBER, + ENABLE_VALIDATE_RELY NUMBER NOT NULL, + DEFAULT_VALUE VARCHAR(400) +) ; + +ALTER TABLE KEY_CONSTRAINTS ADD CONSTRAINT CONSTRAINTS_PK PRIMARY KEY (CONSTRAINT_NAME, POSITION); + +CREATE INDEX CONSTRAINTS_PT_INDEX ON KEY_CONSTRAINTS(PARENT_TBL_ID); + +CREATE INDEX CONSTRAINTS_CT_INDEX ON KEY_CONSTRAINTS(CONSTRAINT_TYPE); + +-- Table for METASTORE_DB_PROPERTIES and its constraints +CREATE TABLE METASTORE_DB_PROPERTIES +( + PROPERTY_KEY VARCHAR(255) NOT NULL, + PROPERTY_VALUE VARCHAR(1000) NOT NULL, + DESCRIPTION VARCHAR(1000) +); + +ALTER TABLE METASTORE_DB_PROPERTIES ADD CONSTRAINT PROPERTY_KEY_PK PRIMARY KEY (PROPERTY_KEY); + +-- Constraints for resource plan tables. + +CREATE UNIQUE INDEX UNIQUE_WM_RESOURCEPLAN ON WM_RESOURCEPLAN ("NAME"); + +CREATE UNIQUE INDEX UNIQUE_WM_POOL ON WM_POOL (RP_ID, PATH); + +ALTER TABLE WM_RESOURCEPLAN ADD CONSTRAINT WM_RESOURCEPLAN_FK1 FOREIGN KEY (DEFAULT_POOL_ID) REFERENCES WM_POOL (POOL_ID); + +ALTER TABLE WM_POOL ADD CONSTRAINT WM_POOL_FK1 FOREIGN KEY (RP_ID) REFERENCES WM_RESOURCEPLAN (RP_ID); + +CREATE UNIQUE INDEX UNIQUE_WM_TRIGGER ON WM_TRIGGER (RP_ID, "NAME"); + +ALTER TABLE WM_TRIGGER ADD CONSTRAINT WM_TRIGGER_FK1 FOREIGN KEY (RP_ID) REFERENCES WM_RESOURCEPLAN (RP_ID); + +ALTER TABLE WM_POOL_TO_TRIGGER ADD CONSTRAINT WM_POOL_TO_TRIGGER_FK1 FOREIGN KEY (POOL_ID) REFERENCES WM_POOL (POOL_ID); + +ALTER TABLE WM_POOL_TO_TRIGGER ADD CONSTRAINT WM_POOL_TO_TRIGGER_FK2 FOREIGN KEY (TRIGGER_ID) REFERENCES WM_TRIGGER (TRIGGER_ID); + +CREATE UNIQUE INDEX UNIQUE_WM_MAPPING ON WM_MAPPING (RP_ID, ENTITY_TYPE, ENTITY_NAME); + +ALTER TABLE WM_MAPPING ADD CONSTRAINT WM_MAPPING_FK1 FOREIGN KEY (RP_ID) REFERENCES WM_RESOURCEPLAN (RP_ID); + +ALTER TABLE WM_MAPPING ADD CONSTRAINT WM_MAPPING_FK2 FOREIGN KEY (POOL_ID) REFERENCES WM_POOL (POOL_ID); + +ALTER TABLE MV_TABLES_USED ADD CONSTRAINT MV_TABLES_USED_FK1 FOREIGN KEY (MV_CREATION_METADATA_ID) REFERENCES MV_CREATION_METADATA (MV_CREATION_METADATA_ID); + +ALTER TABLE MV_TABLES_USED ADD CONSTRAINT MV_TABLES_USED_FK2 FOREIGN KEY (TBL_ID) REFERENCES TBLS (TBL_ID); + +------------------------------ +-- Transaction and lock tables +------------------------------ +CREATE TABLE TXNS ( + TXN_ID NUMBER(19) PRIMARY KEY, + TXN_STATE char(1) NOT NULL, + TXN_STARTED NUMBER(19) NOT NULL, + TXN_LAST_HEARTBEAT NUMBER(19) NOT NULL, + TXN_USER varchar(128) NOT NULL, + TXN_HOST varchar(128) NOT NULL, + TXN_AGENT_INFO varchar2(128), + TXN_META_INFO varchar2(128), + TXN_HEARTBEAT_COUNT number(10), + TXN_TYPE number(10) +) ROWDEPENDENCIES; + +CREATE TABLE TXN_COMPONENTS ( + TC_TXNID NUMBER(19) NOT NULL REFERENCES TXNS (TXN_ID), + TC_DATABASE VARCHAR2(128) NOT NULL, + TC_TABLE VARCHAR2(128), + TC_PARTITION VARCHAR2(767) NULL, + TC_OPERATION_TYPE char(1) NOT NULL, + TC_WRITEID NUMBER(19) +) ROWDEPENDENCIES; + +CREATE INDEX TC_TXNID_INDEX ON TXN_COMPONENTS (TC_TXNID); + +CREATE TABLE COMPLETED_TXN_COMPONENTS ( + CTC_TXNID NUMBER(19) NOT NULL, + CTC_DATABASE VARCHAR2(128) NOT NULL, + CTC_TABLE VARCHAR2(256), + CTC_PARTITION VARCHAR2(767), + CTC_TIMESTAMP timestamp DEFAULT CURRENT_TIMESTAMP NOT NULL, + CTC_WRITEID NUMBER(19), + CTC_UPDATE_DELETE CHAR(1) NOT NULL +) ROWDEPENDENCIES; + +CREATE INDEX COMPLETED_TXN_COMPONENTS_INDEX ON COMPLETED_TXN_COMPONENTS (CTC_DATABASE, CTC_TABLE, CTC_PARTITION); + +CREATE TABLE NEXT_TXN_ID ( + NTXN_NEXT NUMBER(19) NOT NULL +); +INSERT INTO NEXT_TXN_ID VALUES(1); + +CREATE TABLE HIVE_LOCKS ( + HL_LOCK_EXT_ID NUMBER(19) NOT NULL, + HL_LOCK_INT_ID NUMBER(19) NOT NULL, + HL_TXNID NUMBER(19) NOT NULL, + HL_DB VARCHAR2(128) NOT NULL, + HL_TABLE VARCHAR2(128), + HL_PARTITION VARCHAR2(767), + HL_LOCK_STATE CHAR(1) NOT NULL, + HL_LOCK_TYPE CHAR(1) NOT NULL, + HL_LAST_HEARTBEAT NUMBER(19) NOT NULL, + HL_ACQUIRED_AT NUMBER(19), + HL_USER varchar(128) NOT NULL, + HL_HOST varchar(128) NOT NULL, + HL_HEARTBEAT_COUNT number(10), + HL_AGENT_INFO varchar2(128), + HL_BLOCKEDBY_EXT_ID number(19), + HL_BLOCKEDBY_INT_ID number(19), + PRIMARY KEY(HL_LOCK_EXT_ID, HL_LOCK_INT_ID) +) ROWDEPENDENCIES; + +CREATE INDEX HL_TXNID_INDEX ON HIVE_LOCKS (HL_TXNID); + +CREATE TABLE NEXT_LOCK_ID ( + NL_NEXT NUMBER(19) NOT NULL +); +INSERT INTO NEXT_LOCK_ID VALUES(1); + +CREATE TABLE COMPACTION_QUEUE ( + CQ_ID NUMBER(19) PRIMARY KEY, + CQ_DATABASE varchar(128) NOT NULL, + CQ_TABLE varchar(128) NOT NULL, + CQ_PARTITION varchar(767), + CQ_STATE char(1) NOT NULL, + CQ_TYPE char(1) NOT NULL, + CQ_TBLPROPERTIES varchar(2048), + CQ_WORKER_ID varchar(128), + CQ_START NUMBER(19), + CQ_RUN_AS varchar(128), + CQ_HIGHEST_WRITE_ID NUMBER(19), + CQ_META_INFO BLOB, + CQ_HADOOP_JOB_ID varchar2(32) +) ROWDEPENDENCIES; + +CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( + NCQ_NEXT NUMBER(19) NOT NULL +); +INSERT INTO NEXT_COMPACTION_QUEUE_ID VALUES(1); + +CREATE TABLE COMPLETED_COMPACTIONS ( + CC_ID NUMBER(19) PRIMARY KEY, + CC_DATABASE varchar(128) NOT NULL, + CC_TABLE varchar(128) NOT NULL, + CC_PARTITION varchar(767), + CC_STATE char(1) NOT NULL, + CC_TYPE char(1) NOT NULL, + CC_TBLPROPERTIES varchar(2048), + CC_WORKER_ID varchar(128), + CC_START NUMBER(19), + CC_END NUMBER(19), + CC_RUN_AS varchar(128), + CC_HIGHEST_WRITE_ID NUMBER(19), + CC_META_INFO BLOB, + CC_HADOOP_JOB_ID varchar2(32) +) ROWDEPENDENCIES; + +CREATE TABLE AUX_TABLE ( + MT_KEY1 varchar2(128) NOT NULL, + MT_KEY2 number(19) NOT NULL, + MT_COMMENT varchar2(255), + PRIMARY KEY(MT_KEY1, MT_KEY2) +); + +CREATE TABLE WRITE_SET ( + WS_DATABASE varchar2(128) NOT NULL, + WS_TABLE varchar2(128) NOT NULL, + WS_PARTITION varchar2(767), + WS_TXNID number(19) NOT NULL, + WS_COMMIT_ID number(19) NOT NULL, + WS_OPERATION_TYPE char(1) NOT NULL +); + +CREATE TABLE TXN_TO_WRITE_ID ( + T2W_TXNID NUMBER(19) NOT NULL, + T2W_DATABASE VARCHAR2(128) NOT NULL, + T2W_TABLE VARCHAR2(256) NOT NULL, + T2W_WRITEID NUMBER(19) NOT NULL +); + +CREATE UNIQUE INDEX TBL_TO_TXN_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_TXNID); +CREATE UNIQUE INDEX TBL_TO_WRITE_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_WRITEID); + +CREATE TABLE NEXT_WRITE_ID ( + NWI_DATABASE VARCHAR2(128) NOT NULL, + NWI_TABLE VARCHAR2(256) NOT NULL, + NWI_NEXT NUMBER(19) NOT NULL +); + +CREATE UNIQUE INDEX NEXT_WRITE_ID_IDX ON NEXT_WRITE_ID (NWI_DATABASE, NWI_TABLE); + +CREATE TABLE MIN_HISTORY_LEVEL ( + MHL_TXNID NUMBER(19) NOT NULL, + MHL_MIN_OPEN_TXNID NUMBER(19) NOT NULL, + PRIMARY KEY(MHL_TXNID) +); + +CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); + +CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( + MRL_TXN_ID NUMBER NOT NULL, + MRL_DB_NAME VARCHAR(128) NOT NULL, + MRL_TBL_NAME VARCHAR(256) NOT NULL, + MRL_LAST_HEARTBEAT NUMBER NOT NULL, + PRIMARY KEY(MRL_TXN_ID) +); + +CREATE TABLE "I_SCHEMA" ( + "SCHEMA_ID" number primary key, + "SCHEMA_TYPE" number not null, + "NAME" varchar2(256) unique, + "DB_ID" number references "DBS" ("DB_ID"), + "COMPATIBILITY" number not null, + "VALIDATION_LEVEL" number not null, + "CAN_EVOLVE" number(1) not null, + "SCHEMA_GROUP" varchar2(256), + "DESCRIPTION" varchar2(4000) +); + +CREATE TABLE "SCHEMA_VERSION" ( + "SCHEMA_VERSION_ID" number primary key, + "SCHEMA_ID" number references "I_SCHEMA" ("SCHEMA_ID"), + "VERSION" number not null, + "CREATED_AT" number not null, + "CD_ID" number references "CDS" ("CD_ID"), + "STATE" number not null, + "DESCRIPTION" varchar2(4000), + "SCHEMA_TEXT" clob, + "FINGERPRINT" varchar2(256), + "SCHEMA_VERSION_NAME" varchar2(256), + "SERDE_ID" number references "SERDES" ("SERDE_ID"), + UNIQUE ("SCHEMA_ID", "VERSION") +); + +CREATE TABLE REPL_TXN_MAP ( + RTM_REPL_POLICY varchar(256) NOT NULL, + RTM_SRC_TXN_ID number(19) NOT NULL, + RTM_TARGET_TXN_ID number(19) NOT NULL, + PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID) +); + +CREATE TABLE RUNTIME_STATS ( + RS_ID NUMBER primary key, + CREATE_TIME NUMBER(10) NOT NULL, + WEIGHT NUMBER(10) NOT NULL, + PAYLOAD BLOB +); + +CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); + +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID number(19) NOT NULL, + WNL_TXNID number(19) NOT NULL, + WNL_WRITEID number(19) NOT NULL, + WNL_DATABASE varchar(128) NOT NULL, + WNL_TABLE varchar(128) NOT NULL, + WNL_PARTITION varchar(767) NOT NULL, + WNL_TABLE_OBJ clob NOT NULL, + WNL_PARTITION_OBJ clob, + WNL_FILES clob, + WNL_EVENT_TIME number(10) NOT NULL, + PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION) +); + +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- ----------------------------------------------------------------- +-- Record schema version. Should be the last step in the init script +-- ----------------------------------------------------------------- +INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES (1, '3.2.0', 'Hive release version 3.2.0'); diff --git a/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql b/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql index ce3437f723..71f5034446 100644 --- a/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql +++ b/standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql @@ -335,8 +335,8 @@ INSERT INTO TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_TXNID, T2W_WRITEID) UPDATE TXN_COMPONENTS SET TC_WRITEID = TC_TXNID; UPDATE COMPLETED_TXN_COMPONENTS SET CTC_WRITEID = CTC_TXNID; +ALTER TABLE TBLS ADD OWNER_TYPE VARCHAR2(10) NULL; + -- These lines need to be last. Insert any changes above. UPDATE VERSION SET SCHEMA_VERSION='3.0.0', VERSION_COMMENT='Hive release version 3.0.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 2.3.0 to 3.0.0' AS Status from dual; - -ALTER TABLE TBLS ADD OWNER_TYPE VARCHAR2(10) NULL; \ No newline at end of file diff --git a/standalone-metastore/src/main/sql/oracle/upgrade-3.1.0-to-3.2.0.oracle.sql b/standalone-metastore/src/main/sql/oracle/upgrade-3.1.0-to-3.2.0.oracle.sql new file mode 100644 index 0000000000..7733f173bf --- /dev/null +++ b/standalone-metastore/src/main/sql/oracle/upgrade-3.1.0-to-3.2.0.oracle.sql @@ -0,0 +1,21 @@ +SELECT 'Upgrading MetaStore schema from 3.1.0 to 3.2.0' AS Status from dual; + +-- HIVE-19267 +CREATE TABLE TXN_WRITE_NOTIFICATION_LOG ( + WNL_ID number(19) NOT NULL, + WNL_TXNID number(19) NOT NULL, + WNL_WRITEID number(19) NOT NULL, + WNL_DATABASE varchar(128) NOT NULL, + WNL_TABLE varchar(128) NOT NULL, + WNL_PARTITION varchar(767) NOT NULL, + WNL_TABLE_OBJ clob NOT NULL, + WNL_PARTITION_OBJ clob, + WNL_FILES clob, + WNL_EVENT_TIME number(10) NOT NULL, + PRIMARY KEY (WNL_TXNID, WNL_DATABASE, WNL_TABLE, WNL_PARTITION) +); +INSERT INTO SEQUENCE_TABLE (SEQUENCE_NAME, NEXT_VAL) VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- These lines need to be last. Insert any changes above. +UPDATE VERSION SET SCHEMA_VERSION='3.2.0', VERSION_COMMENT='Hive release version 3.2.0' where VER_ID=1; +SELECT 'Finished upgrading MetaStore schema from 3.1.0 to 3.2.0' AS Status from dual; diff --git a/standalone-metastore/src/main/sql/oracle/upgrade.order.oracle b/standalone-metastore/src/main/sql/oracle/upgrade.order.oracle index 72b8303196..bfc2a86395 100644 --- a/standalone-metastore/src/main/sql/oracle/upgrade.order.oracle +++ b/standalone-metastore/src/main/sql/oracle/upgrade.order.oracle @@ -11,3 +11,4 @@ 2.2.0-to-2.3.0 2.3.0-to-3.0.0 3.0.0-to-3.1.0 +3.1.0-to-3.2.0 diff --git a/standalone-metastore/src/main/sql/postgres/hive-schema-3.0.0.postgres.sql b/standalone-metastore/src/main/sql/postgres/hive-schema-3.0.0.postgres.sql index 97697f8bf8..b89c87f41e 100644 --- a/standalone-metastore/src/main/sql/postgres/hive-schema-3.0.0.postgres.sql +++ b/standalone-metastore/src/main/sql/postgres/hive-schema-3.0.0.postgres.sql @@ -1812,7 +1812,6 @@ CREATE TABLE REPL_TXN_MAP ( PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID) ); - CREATE TABLE RUNTIME_STATS ( RS_ID bigint primary key, CREATE_TIME bigint NOT NULL, @@ -1822,7 +1821,6 @@ CREATE TABLE RUNTIME_STATS ( CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); - -- ----------------------------------------------------------------- -- Record schema version. Should be the last step in the init script -- ----------------------------------------------------------------- diff --git a/standalone-metastore/src/main/sql/postgres/hive-schema-3.1.0.postgres.sql b/standalone-metastore/src/main/sql/postgres/hive-schema-3.1.0.postgres.sql index 028d091606..3281c1f294 100644 --- a/standalone-metastore/src/main/sql/postgres/hive-schema-3.1.0.postgres.sql +++ b/standalone-metastore/src/main/sql/postgres/hive-schema-3.1.0.postgres.sql @@ -1844,7 +1844,6 @@ CREATE TABLE RUNTIME_STATS ( CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); - -- ----------------------------------------------------------------- -- Record schema version. Should be the last step in the init script -- ----------------------------------------------------------------- diff --git a/standalone-metastore/src/main/sql/postgres/hive-schema-3.2.0.postgres.sql b/standalone-metastore/src/main/sql/postgres/hive-schema-3.2.0.postgres.sql new file mode 100644 index 0000000000..828ac3c923 --- /dev/null +++ b/standalone-metastore/src/main/sql/postgres/hive-schema-3.2.0.postgres.sql @@ -0,0 +1,1866 @@ +-- +-- PostgreSQL database dump +-- + +SET statement_timeout = 0; +SET client_encoding = 'UTF8'; +SET standard_conforming_strings = off; +SET check_function_bodies = false; +SET client_min_messages = warning; +SET escape_string_warning = off; + +SET search_path = public, pg_catalog; + +SET default_tablespace = ''; + +SET default_with_oids = false; + +-- +-- Name: BUCKETING_COLS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "BUCKETING_COLS" ( + "SD_ID" bigint NOT NULL, + "BUCKET_COL_NAME" character varying(256) DEFAULT NULL::character varying, + "INTEGER_IDX" bigint NOT NULL +); + + +-- +-- Name: CDS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "CDS" ( + "CD_ID" bigint NOT NULL +); + + +-- +-- Name: COLUMNS_V2; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "COLUMNS_V2" ( + "CD_ID" bigint NOT NULL, + "COMMENT" character varying(4000), + "COLUMN_NAME" character varying(767) NOT NULL, + "TYPE_NAME" text, + "INTEGER_IDX" integer NOT NULL +); + + +-- +-- Name: DATABASE_PARAMS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "DATABASE_PARAMS" ( + "DB_ID" bigint NOT NULL, + "PARAM_KEY" character varying(180) NOT NULL, + "PARAM_VALUE" character varying(4000) DEFAULT NULL::character varying +); + + +CREATE TABLE "CTLGS" ( + "CTLG_ID" BIGINT PRIMARY KEY, + "NAME" VARCHAR(256) UNIQUE, + "DESC" VARCHAR(4000), + "LOCATION_URI" VARCHAR(4000) NOT NULL +); + +-- +-- Name: DBS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "DBS" ( + "DB_ID" bigint NOT NULL, + "DESC" character varying(4000) DEFAULT NULL::character varying, + "DB_LOCATION_URI" character varying(4000) NOT NULL, + "NAME" character varying(128) DEFAULT NULL::character varying, + "OWNER_NAME" character varying(128) DEFAULT NULL::character varying, + "OWNER_TYPE" character varying(10) DEFAULT NULL::character varying, + "CTLG_NAME" varchar(256) +); + + +-- +-- Name: DB_PRIVS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "DB_PRIVS" ( + "DB_GRANT_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "DB_ID" bigint, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "DB_PRIV" character varying(128) DEFAULT NULL::character varying, + "AUTHORIZER" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: GLOBAL_PRIVS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "GLOBAL_PRIVS" ( + "USER_GRANT_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "USER_PRIV" character varying(128) DEFAULT NULL::character varying, + "AUTHORIZER" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: IDXS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "IDXS" ( + "INDEX_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "DEFERRED_REBUILD" boolean NOT NULL, + "INDEX_HANDLER_CLASS" character varying(4000) DEFAULT NULL::character varying, + "INDEX_NAME" character varying(128) DEFAULT NULL::character varying, + "INDEX_TBL_ID" bigint, + "LAST_ACCESS_TIME" bigint NOT NULL, + "ORIG_TBL_ID" bigint, + "SD_ID" bigint +); + + +-- +-- Name: INDEX_PARAMS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "INDEX_PARAMS" ( + "INDEX_ID" bigint NOT NULL, + "PARAM_KEY" character varying(256) NOT NULL, + "PARAM_VALUE" character varying(4000) DEFAULT NULL::character varying +); + + +-- +-- Name: NUCLEUS_TABLES; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "NUCLEUS_TABLES" ( + "CLASS_NAME" character varying(128) NOT NULL, + "TABLE_NAME" character varying(128) NOT NULL, + "TYPE" character varying(4) NOT NULL, + "OWNER" character varying(2) NOT NULL, + "VERSION" character varying(20) NOT NULL, + "INTERFACE_NAME" character varying(255) DEFAULT NULL::character varying +); + + +-- +-- Name: PARTITIONS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PARTITIONS" ( + "PART_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "LAST_ACCESS_TIME" bigint NOT NULL, + "PART_NAME" character varying(767) DEFAULT NULL::character varying, + "SD_ID" bigint, + "TBL_ID" bigint +); + + +-- +-- Name: PARTITION_EVENTS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PARTITION_EVENTS" ( + "PART_NAME_ID" bigint NOT NULL, + "CAT_NAME" character varying(256), + "DB_NAME" character varying(128), + "EVENT_TIME" bigint NOT NULL, + "EVENT_TYPE" integer NOT NULL, + "PARTITION_NAME" character varying(767), + "TBL_NAME" character varying(256) +); + + +-- +-- Name: PARTITION_KEYS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PARTITION_KEYS" ( + "TBL_ID" bigint NOT NULL, + "PKEY_COMMENT" character varying(4000) DEFAULT NULL::character varying, + "PKEY_NAME" character varying(128) NOT NULL, + "PKEY_TYPE" character varying(767) NOT NULL, + "INTEGER_IDX" bigint NOT NULL +); + + +-- +-- Name: PARTITION_KEY_VALS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PARTITION_KEY_VALS" ( + "PART_ID" bigint NOT NULL, + "PART_KEY_VAL" character varying(256) DEFAULT NULL::character varying, + "INTEGER_IDX" bigint NOT NULL +); + + +-- +-- Name: PARTITION_PARAMS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PARTITION_PARAMS" ( + "PART_ID" bigint NOT NULL, + "PARAM_KEY" character varying(256) NOT NULL, + "PARAM_VALUE" character varying(4000) DEFAULT NULL::character varying +); + + +-- +-- Name: PART_COL_PRIVS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PART_COL_PRIVS" ( + "PART_COLUMN_GRANT_ID" bigint NOT NULL, + "COLUMN_NAME" character varying(767) DEFAULT NULL::character varying, + "CREATE_TIME" bigint NOT NULL, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PART_ID" bigint, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "PART_COL_PRIV" character varying(128) DEFAULT NULL::character varying, + "AUTHORIZER" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: PART_PRIVS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PART_PRIVS" ( + "PART_GRANT_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PART_ID" bigint, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "PART_PRIV" character varying(128) DEFAULT NULL::character varying, + "AUTHORIZER" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: ROLES; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "ROLES" ( + "ROLE_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "OWNER_NAME" character varying(128) DEFAULT NULL::character varying, + "ROLE_NAME" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: ROLE_MAP; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "ROLE_MAP" ( + "ROLE_GRANT_ID" bigint NOT NULL, + "ADD_TIME" bigint NOT NULL, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "ROLE_ID" bigint +); + + +-- +-- Name: SDS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "SDS" ( + "SD_ID" bigint NOT NULL, + "INPUT_FORMAT" character varying(4000) DEFAULT NULL::character varying, + "IS_COMPRESSED" boolean NOT NULL, + "LOCATION" character varying(4000) DEFAULT NULL::character varying, + "NUM_BUCKETS" bigint NOT NULL, + "OUTPUT_FORMAT" character varying(4000) DEFAULT NULL::character varying, + "SERDE_ID" bigint, + "CD_ID" bigint, + "IS_STOREDASSUBDIRECTORIES" boolean NOT NULL +); + + +-- +-- Name: SD_PARAMS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "SD_PARAMS" ( + "SD_ID" bigint NOT NULL, + "PARAM_KEY" character varying(256) NOT NULL, + "PARAM_VALUE" text DEFAULT NULL +); + + +-- +-- Name: SEQUENCE_TABLE; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "SEQUENCE_TABLE" ( + "SEQUENCE_NAME" character varying(255) NOT NULL, + "NEXT_VAL" bigint NOT NULL +); + +INSERT INTO "SEQUENCE_TABLE" ("SEQUENCE_NAME", "NEXT_VAL") VALUES ('org.apache.hadoop.hive.metastore.model.MNotificationLog', 1); + +-- +-- Name: SERDES; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "SERDES" ( + "SERDE_ID" bigint NOT NULL, + "NAME" character varying(128) DEFAULT NULL::character varying, + "SLIB" character varying(4000) DEFAULT NULL::character varying, + "DESCRIPTION" varchar(4000), + "SERIALIZER_CLASS" varchar(4000), + "DESERIALIZER_CLASS" varchar(4000), + "SERDE_TYPE" integer +); + + +-- +-- Name: SERDE_PARAMS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "SERDE_PARAMS" ( + "SERDE_ID" bigint NOT NULL, + "PARAM_KEY" character varying(256) NOT NULL, + "PARAM_VALUE" text DEFAULT NULL +); + + +-- +-- Name: SORT_COLS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "SORT_COLS" ( + "SD_ID" bigint NOT NULL, + "COLUMN_NAME" character varying(767) DEFAULT NULL::character varying, + "ORDER" bigint NOT NULL, + "INTEGER_IDX" bigint NOT NULL +); + + +-- +-- Name: TABLE_PARAMS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "TABLE_PARAMS" ( + "TBL_ID" bigint NOT NULL, + "PARAM_KEY" character varying(256) NOT NULL, + "PARAM_VALUE" text DEFAULT NULL +); + + +-- +-- Name: TBLS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "TBLS" ( + "TBL_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "DB_ID" bigint, + "LAST_ACCESS_TIME" bigint NOT NULL, + "OWNER" character varying(767) DEFAULT NULL::character varying, + "OWNER_TYPE" character varying(10) DEFAULT NULL::character varying, + "RETENTION" bigint NOT NULL, + "SD_ID" bigint, + "TBL_NAME" character varying(256) DEFAULT NULL::character varying, + "TBL_TYPE" character varying(128) DEFAULT NULL::character varying, + "VIEW_EXPANDED_TEXT" text, + "VIEW_ORIGINAL_TEXT" text, + "IS_REWRITE_ENABLED" boolean NOT NULL DEFAULT false +); + +-- +-- Name: MV_CREATION_METADATA; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "MV_CREATION_METADATA" ( + "MV_CREATION_METADATA_ID" bigint NOT NULL, + "CAT_NAME" character varying(256) NOT NULL, + "DB_NAME" character varying(128) NOT NULL, + "TBL_NAME" character varying(256) NOT NULL, + "TXN_LIST" text, + "MATERIALIZATION_TIME" bigint NOT NULL +); + +-- +-- Name: MV_TABLES_USED; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "MV_TABLES_USED" ( + "MV_CREATION_METADATA_ID" bigint NOT NULL, + "TBL_ID" bigint NOT NULL +); + +-- +-- Name: TBL_COL_PRIVS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "TBL_COL_PRIVS" ( + "TBL_COLUMN_GRANT_ID" bigint NOT NULL, + "COLUMN_NAME" character varying(767) DEFAULT NULL::character varying, + "CREATE_TIME" bigint NOT NULL, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "TBL_COL_PRIV" character varying(128) DEFAULT NULL::character varying, + "TBL_ID" bigint, + "AUTHORIZER" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: TBL_PRIVS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "TBL_PRIVS" ( + "TBL_GRANT_ID" bigint NOT NULL, + "CREATE_TIME" bigint NOT NULL, + "GRANT_OPTION" smallint NOT NULL, + "GRANTOR" character varying(128) DEFAULT NULL::character varying, + "GRANTOR_TYPE" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_NAME" character varying(128) DEFAULT NULL::character varying, + "PRINCIPAL_TYPE" character varying(128) DEFAULT NULL::character varying, + "TBL_PRIV" character varying(128) DEFAULT NULL::character varying, + "TBL_ID" bigint, + "AUTHORIZER" character varying(128) DEFAULT NULL::character varying +); + + +-- +-- Name: TYPES; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "TYPES" ( + "TYPES_ID" bigint NOT NULL, + "TYPE_NAME" character varying(128) DEFAULT NULL::character varying, + "TYPE1" character varying(767) DEFAULT NULL::character varying, + "TYPE2" character varying(767) DEFAULT NULL::character varying +); + + +-- +-- Name: TYPE_FIELDS; Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "TYPE_FIELDS" ( + "TYPE_NAME" bigint NOT NULL, + "COMMENT" character varying(256) DEFAULT NULL::character varying, + "FIELD_NAME" character varying(128) NOT NULL, + "FIELD_TYPE" character varying(767) NOT NULL, + "INTEGER_IDX" bigint NOT NULL +); + +CREATE TABLE "SKEWED_STRING_LIST" ( + "STRING_LIST_ID" bigint NOT NULL +); + +CREATE TABLE "SKEWED_STRING_LIST_VALUES" ( + "STRING_LIST_ID" bigint NOT NULL, + "STRING_LIST_VALUE" character varying(256) DEFAULT NULL::character varying, + "INTEGER_IDX" bigint NOT NULL +); + +CREATE TABLE "SKEWED_COL_NAMES" ( + "SD_ID" bigint NOT NULL, + "SKEWED_COL_NAME" character varying(256) DEFAULT NULL::character varying, + "INTEGER_IDX" bigint NOT NULL +); + +CREATE TABLE "SKEWED_COL_VALUE_LOC_MAP" ( + "SD_ID" bigint NOT NULL, + "STRING_LIST_ID_KID" bigint NOT NULL, + "LOCATION" character varying(4000) DEFAULT NULL::character varying +); + +CREATE TABLE "SKEWED_VALUES" ( + "SD_ID_OID" bigint NOT NULL, + "STRING_LIST_ID_EID" bigint NOT NULL, + "INTEGER_IDX" bigint NOT NULL +); + + +-- +-- Name: TAB_COL_STATS Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "MASTER_KEYS" +( + "KEY_ID" SERIAL, + "MASTER_KEY" varchar(767) NULL, + PRIMARY KEY ("KEY_ID") +); + +CREATE TABLE "DELEGATION_TOKENS" +( + "TOKEN_IDENT" varchar(767) NOT NULL, + "TOKEN" varchar(767) NULL, + PRIMARY KEY ("TOKEN_IDENT") +); + +CREATE TABLE "TAB_COL_STATS" ( + "CS_ID" bigint NOT NULL, + "CAT_NAME" character varying(256) DEFAULT NULL::character varying, + "DB_NAME" character varying(128) DEFAULT NULL::character varying, + "TABLE_NAME" character varying(256) DEFAULT NULL::character varying, + "COLUMN_NAME" character varying(767) DEFAULT NULL::character varying, + "COLUMN_TYPE" character varying(128) DEFAULT NULL::character varying, + "TBL_ID" bigint NOT NULL, + "LONG_LOW_VALUE" bigint, + "LONG_HIGH_VALUE" bigint, + "DOUBLE_LOW_VALUE" double precision, + "DOUBLE_HIGH_VALUE" double precision, + "BIG_DECIMAL_LOW_VALUE" character varying(4000) DEFAULT NULL::character varying, + "BIG_DECIMAL_HIGH_VALUE" character varying(4000) DEFAULT NULL::character varying, + "NUM_NULLS" bigint NOT NULL, + "NUM_DISTINCTS" bigint, + "BIT_VECTOR" bytea, + "AVG_COL_LEN" double precision, + "MAX_COL_LEN" bigint, + "NUM_TRUES" bigint, + "NUM_FALSES" bigint, + "LAST_ANALYZED" bigint NOT NULL +); + +-- +-- Table structure for VERSION +-- +CREATE TABLE "VERSION" ( + "VER_ID" bigint, + "SCHEMA_VERSION" character varying(127) NOT NULL, + "VERSION_COMMENT" character varying(255) NOT NULL +); + +-- +-- Name: PART_COL_STATS Type: TABLE; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE TABLE "PART_COL_STATS" ( + "CS_ID" bigint NOT NULL, + "CAT_NAME" character varying(256) DEFAULT NULL::character varying, + "DB_NAME" character varying(128) DEFAULT NULL::character varying, + "TABLE_NAME" character varying(256) DEFAULT NULL::character varying, + "PARTITION_NAME" character varying(767) DEFAULT NULL::character varying, + "COLUMN_NAME" character varying(767) DEFAULT NULL::character varying, + "COLUMN_TYPE" character varying(128) DEFAULT NULL::character varying, + "PART_ID" bigint NOT NULL, + "LONG_LOW_VALUE" bigint, + "LONG_HIGH_VALUE" bigint, + "DOUBLE_LOW_VALUE" double precision, + "DOUBLE_HIGH_VALUE" double precision, + "BIG_DECIMAL_LOW_VALUE" character varying(4000) DEFAULT NULL::character varying, + "BIG_DECIMAL_HIGH_VALUE" character varying(4000) DEFAULT NULL::character varying, + "NUM_NULLS" bigint NOT NULL, + "NUM_DISTINCTS" bigint, + "BIT_VECTOR" bytea, + "AVG_COL_LEN" double precision, + "MAX_COL_LEN" bigint, + "NUM_TRUES" bigint, + "NUM_FALSES" bigint, + "LAST_ANALYZED" bigint NOT NULL +); + +-- +-- Table structure for FUNCS +-- +CREATE TABLE "FUNCS" ( + "FUNC_ID" BIGINT NOT NULL, + "CLASS_NAME" VARCHAR(4000), + "CREATE_TIME" INTEGER NOT NULL, + "DB_ID" BIGINT, + "FUNC_NAME" VARCHAR(128), + "FUNC_TYPE" INTEGER NOT NULL, + "OWNER_NAME" VARCHAR(128), + "OWNER_TYPE" VARCHAR(10), + PRIMARY KEY ("FUNC_ID") +); + +-- +-- Table structure for FUNC_RU +-- +CREATE TABLE "FUNC_RU" ( + "FUNC_ID" BIGINT NOT NULL, + "RESOURCE_TYPE" INTEGER NOT NULL, + "RESOURCE_URI" VARCHAR(4000), + "INTEGER_IDX" INTEGER NOT NULL, + PRIMARY KEY ("FUNC_ID", "INTEGER_IDX") +); + +CREATE TABLE "NOTIFICATION_LOG" +( + "NL_ID" BIGINT NOT NULL, + "EVENT_ID" BIGINT NOT NULL, + "EVENT_TIME" INTEGER NOT NULL, + "EVENT_TYPE" VARCHAR(32) NOT NULL, + "CAT_NAME" VARCHAR(256), + "DB_NAME" VARCHAR(128), + "TBL_NAME" VARCHAR(256), + "MESSAGE" text, + "MESSAGE_FORMAT" VARCHAR(16), + PRIMARY KEY ("NL_ID") +); + +CREATE TABLE "NOTIFICATION_SEQUENCE" +( + "NNI_ID" BIGINT NOT NULL, + "NEXT_EVENT_ID" BIGINT NOT NULL, + PRIMARY KEY ("NNI_ID") +); + +INSERT INTO "NOTIFICATION_SEQUENCE" ("NNI_ID", "NEXT_EVENT_ID") SELECT 1,1 WHERE NOT EXISTS ( SELECT "NEXT_EVENT_ID" FROM "NOTIFICATION_SEQUENCE"); + +CREATE TABLE "KEY_CONSTRAINTS" +( + "CHILD_CD_ID" BIGINT, + "CHILD_INTEGER_IDX" BIGINT, + "CHILD_TBL_ID" BIGINT, + "PARENT_CD_ID" BIGINT, + "PARENT_INTEGER_IDX" BIGINT NOT NULL, + "PARENT_TBL_ID" BIGINT NOT NULL, + "POSITION" BIGINT NOT NULL, + "CONSTRAINT_NAME" VARCHAR(400) NOT NULL, + "CONSTRAINT_TYPE" SMALLINT NOT NULL, + "UPDATE_RULE" SMALLINT, + "DELETE_RULE" SMALLINT, + "ENABLE_VALIDATE_RELY" SMALLINT NOT NULL, + "DEFAULT_VALUE" VARCHAR(400), + PRIMARY KEY ("CONSTRAINT_NAME", "POSITION") +) ; + +--- +--- Table structure for METASTORE_DB_PROPERTIES +--- +CREATE TABLE "METASTORE_DB_PROPERTIES" +( + "PROPERTY_KEY" VARCHAR(255) NOT NULL, + "PROPERTY_VALUE" VARCHAR(1000) NOT NULL, + "DESCRIPTION" VARCHAR(1000) +); + + +CREATE TABLE "WM_RESOURCEPLAN" ( + "RP_ID" bigint NOT NULL, + "NAME" character varying(128) NOT NULL, + "QUERY_PARALLELISM" integer, + "STATUS" character varying(20) NOT NULL, + "DEFAULT_POOL_ID" bigint +); + +CREATE TABLE "WM_POOL" ( + "POOL_ID" bigint NOT NULL, + "RP_ID" bigint NOT NULL, + "PATH" character varying(1024) NOT NULL, + "ALLOC_FRACTION" double precision, + "QUERY_PARALLELISM" integer, + "SCHEDULING_POLICY" character varying(1024) +); + +CREATE TABLE "WM_TRIGGER" ( + "TRIGGER_ID" bigint NOT NULL, + "RP_ID" bigint NOT NULL, + "NAME" character varying(128) NOT NULL, + "TRIGGER_EXPRESSION" character varying(1024) DEFAULT NULL::character varying, + "ACTION_EXPRESSION" character varying(1024) DEFAULT NULL::character varying, + "IS_IN_UNMANAGED" smallint NOT NULL DEFAULT 0 +); + +CREATE TABLE "WM_POOL_TO_TRIGGER" ( + "POOL_ID" bigint NOT NULL, + "TRIGGER_ID" bigint NOT NULL +); + +CREATE TABLE "WM_MAPPING" ( + "MAPPING_ID" bigint NOT NULL, + "RP_ID" bigint NOT NULL, + "ENTITY_TYPE" character varying(128) NOT NULL, + "ENTITY_NAME" character varying(128) NOT NULL, + "POOL_ID" bigint, + "ORDERING" integer +); + +-- +-- Name: BUCKETING_COLS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "BUCKETING_COLS" + ADD CONSTRAINT "BUCKETING_COLS_pkey" PRIMARY KEY ("SD_ID", "INTEGER_IDX"); + + +-- +-- Name: CDS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "CDS" + ADD CONSTRAINT "CDS_pkey" PRIMARY KEY ("CD_ID"); + + +-- +-- Name: COLUMNS_V2_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "COLUMNS_V2" + ADD CONSTRAINT "COLUMNS_V2_pkey" PRIMARY KEY ("CD_ID", "COLUMN_NAME"); + + +-- +-- Name: DATABASE_PARAMS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "DATABASE_PARAMS" + ADD CONSTRAINT "DATABASE_PARAMS_pkey" PRIMARY KEY ("DB_ID", "PARAM_KEY"); + + +-- +-- Name: DBPRIVILEGEINDEX; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "DB_PRIVS" + ADD CONSTRAINT "DBPRIVILEGEINDEX" UNIQUE ("AUTHORIZER", "DB_ID", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "DB_PRIV", "GRANTOR", "GRANTOR_TYPE"); + + +-- +-- Name: DBS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "DBS" + ADD CONSTRAINT "DBS_pkey" PRIMARY KEY ("DB_ID"); + + +-- +-- Name: DB_PRIVS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "DB_PRIVS" + ADD CONSTRAINT "DB_PRIVS_pkey" PRIMARY KEY ("DB_GRANT_ID"); + + +-- +-- Name: GLOBALPRIVILEGEINDEX; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "GLOBAL_PRIVS" + ADD CONSTRAINT "GLOBALPRIVILEGEINDEX" UNIQUE ("AUTHORIZER", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "USER_PRIV", "GRANTOR", "GRANTOR_TYPE"); + + +-- +-- Name: GLOBAL_PRIVS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "GLOBAL_PRIVS" + ADD CONSTRAINT "GLOBAL_PRIVS_pkey" PRIMARY KEY ("USER_GRANT_ID"); + + +-- +-- Name: IDXS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "IDXS" + ADD CONSTRAINT "IDXS_pkey" PRIMARY KEY ("INDEX_ID"); + + +-- +-- Name: INDEX_PARAMS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "INDEX_PARAMS" + ADD CONSTRAINT "INDEX_PARAMS_pkey" PRIMARY KEY ("INDEX_ID", "PARAM_KEY"); + + +-- +-- Name: NUCLEUS_TABLES_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "NUCLEUS_TABLES" + ADD CONSTRAINT "NUCLEUS_TABLES_pkey" PRIMARY KEY ("CLASS_NAME"); + + +-- +-- Name: PARTITIONS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PARTITIONS" + ADD CONSTRAINT "PARTITIONS_pkey" PRIMARY KEY ("PART_ID"); + + +-- +-- Name: PARTITION_EVENTS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PARTITION_EVENTS" + ADD CONSTRAINT "PARTITION_EVENTS_pkey" PRIMARY KEY ("PART_NAME_ID"); + + +-- +-- Name: PARTITION_KEYS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PARTITION_KEYS" + ADD CONSTRAINT "PARTITION_KEYS_pkey" PRIMARY KEY ("TBL_ID", "PKEY_NAME"); + + +-- +-- Name: PARTITION_KEY_VALS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PARTITION_KEY_VALS" + ADD CONSTRAINT "PARTITION_KEY_VALS_pkey" PRIMARY KEY ("PART_ID", "INTEGER_IDX"); + + +-- +-- Name: PARTITION_PARAMS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PARTITION_PARAMS" + ADD CONSTRAINT "PARTITION_PARAMS_pkey" PRIMARY KEY ("PART_ID", "PARAM_KEY"); + + +-- +-- Name: PART_COL_PRIVS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PART_COL_PRIVS" + ADD CONSTRAINT "PART_COL_PRIVS_pkey" PRIMARY KEY ("PART_COLUMN_GRANT_ID"); + + +-- +-- Name: PART_PRIVS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PART_PRIVS" + ADD CONSTRAINT "PART_PRIVS_pkey" PRIMARY KEY ("PART_GRANT_ID"); + + +-- +-- Name: ROLEENTITYINDEX; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "ROLES" + ADD CONSTRAINT "ROLEENTITYINDEX" UNIQUE ("ROLE_NAME"); + + +-- +-- Name: ROLES_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "ROLES" + ADD CONSTRAINT "ROLES_pkey" PRIMARY KEY ("ROLE_ID"); + + +-- +-- Name: ROLE_MAP_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "ROLE_MAP" + ADD CONSTRAINT "ROLE_MAP_pkey" PRIMARY KEY ("ROLE_GRANT_ID"); + + +-- +-- Name: SDS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "SDS" + ADD CONSTRAINT "SDS_pkey" PRIMARY KEY ("SD_ID"); + + +-- +-- Name: SD_PARAMS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "SD_PARAMS" + ADD CONSTRAINT "SD_PARAMS_pkey" PRIMARY KEY ("SD_ID", "PARAM_KEY"); + + +-- +-- Name: SEQUENCE_TABLE_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "SEQUENCE_TABLE" + ADD CONSTRAINT "SEQUENCE_TABLE_pkey" PRIMARY KEY ("SEQUENCE_NAME"); + + +-- +-- Name: SERDES_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "SERDES" + ADD CONSTRAINT "SERDES_pkey" PRIMARY KEY ("SERDE_ID"); + + +-- +-- Name: SERDE_PARAMS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "SERDE_PARAMS" + ADD CONSTRAINT "SERDE_PARAMS_pkey" PRIMARY KEY ("SERDE_ID", "PARAM_KEY"); + + +-- +-- Name: SORT_COLS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "SORT_COLS" + ADD CONSTRAINT "SORT_COLS_pkey" PRIMARY KEY ("SD_ID", "INTEGER_IDX"); + + +-- +-- Name: TABLE_PARAMS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TABLE_PARAMS" + ADD CONSTRAINT "TABLE_PARAMS_pkey" PRIMARY KEY ("TBL_ID", "PARAM_KEY"); + + +-- +-- Name: TBLS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TBLS" + ADD CONSTRAINT "TBLS_pkey" PRIMARY KEY ("TBL_ID"); + + +-- +-- Name: TBL_COL_PRIVS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TBL_COL_PRIVS" + ADD CONSTRAINT "TBL_COL_PRIVS_pkey" PRIMARY KEY ("TBL_COLUMN_GRANT_ID"); + + +-- +-- Name: TBL_PRIVS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TBL_PRIVS" + ADD CONSTRAINT "TBL_PRIVS_pkey" PRIMARY KEY ("TBL_GRANT_ID"); + + +-- +-- Name: TYPES_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TYPES" + ADD CONSTRAINT "TYPES_pkey" PRIMARY KEY ("TYPES_ID"); + + +-- +-- Name: TYPE_FIELDS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TYPE_FIELDS" + ADD CONSTRAINT "TYPE_FIELDS_pkey" PRIMARY KEY ("TYPE_NAME", "FIELD_NAME"); + +ALTER TABLE ONLY "SKEWED_STRING_LIST" + ADD CONSTRAINT "SKEWED_STRING_LIST_pkey" PRIMARY KEY ("STRING_LIST_ID"); + +ALTER TABLE ONLY "SKEWED_STRING_LIST_VALUES" + ADD CONSTRAINT "SKEWED_STRING_LIST_VALUES_pkey" PRIMARY KEY ("STRING_LIST_ID", "INTEGER_IDX"); + + +ALTER TABLE ONLY "SKEWED_COL_NAMES" + ADD CONSTRAINT "SKEWED_COL_NAMES_pkey" PRIMARY KEY ("SD_ID", "INTEGER_IDX"); + +ALTER TABLE ONLY "SKEWED_COL_VALUE_LOC_MAP" + ADD CONSTRAINT "SKEWED_COL_VALUE_LOC_MAP_pkey" PRIMARY KEY ("SD_ID", "STRING_LIST_ID_KID"); + +ALTER TABLE ONLY "SKEWED_VALUES" + ADD CONSTRAINT "SKEWED_VALUES_pkey" PRIMARY KEY ("SD_ID_OID", "INTEGER_IDX"); + +-- +-- Name: TAB_COL_STATS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- +ALTER TABLE ONLY "TAB_COL_STATS" ADD CONSTRAINT "TAB_COL_STATS_pkey" PRIMARY KEY("CS_ID"); + +-- +-- Name: PART_COL_STATS_pkey; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- +ALTER TABLE ONLY "PART_COL_STATS" ADD CONSTRAINT "PART_COL_STATS_pkey" PRIMARY KEY("CS_ID"); + +-- +-- Name: UNIQUEINDEX; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "IDXS" + ADD CONSTRAINT "UNIQUEINDEX" UNIQUE ("INDEX_NAME", "ORIG_TBL_ID"); + + +-- +-- Name: UNIQUEPARTITION; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "PARTITIONS" + ADD CONSTRAINT "UNIQUEPARTITION" UNIQUE ("PART_NAME", "TBL_ID"); + + +-- +-- Name: UNIQUETABLE; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TBLS" + ADD CONSTRAINT "UNIQUETABLE" UNIQUE ("TBL_NAME", "DB_ID"); + + +-- +-- Name: UNIQUE_DATABASE; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "DBS" + ADD CONSTRAINT "UNIQUE_DATABASE" UNIQUE ("NAME", "CTLG_NAME"); + + +-- +-- Name: UNIQUE_TYPE; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "TYPES" + ADD CONSTRAINT "UNIQUE_TYPE" UNIQUE ("TYPE_NAME"); + + +-- +-- Name: USERROLEMAPINDEX; Type: CONSTRAINT; Schema: public; Owner: hiveuser; Tablespace: +-- + +ALTER TABLE ONLY "ROLE_MAP" + ADD CONSTRAINT "USERROLEMAPINDEX" UNIQUE ("PRINCIPAL_NAME", "ROLE_ID", "GRANTOR", "GRANTOR_TYPE"); + +ALTER TABLE ONLY "METASTORE_DB_PROPERTIES" + ADD CONSTRAINT "PROPERTY_KEY_PK" PRIMARY KEY ("PROPERTY_KEY"); + + +-- Resource plan: Primary key and unique key constraints. +ALTER TABLE ONLY "WM_RESOURCEPLAN" + ADD CONSTRAINT "WM_RESOURCEPLAN_pkey" PRIMARY KEY ("RP_ID"); + +ALTER TABLE ONLY "WM_RESOURCEPLAN" + ADD CONSTRAINT "UNIQUE_WM_RESOURCEPLAN" UNIQUE ("NAME"); + +ALTER TABLE ONLY "WM_POOL" + ADD CONSTRAINT "WM_POOL_pkey" PRIMARY KEY ("POOL_ID"); + +ALTER TABLE ONLY "WM_POOL" + ADD CONSTRAINT "UNIQUE_WM_POOL" UNIQUE ("RP_ID", "PATH"); + +ALTER TABLE ONLY "WM_TRIGGER" + ADD CONSTRAINT "WM_TRIGGER_pkey" PRIMARY KEY ("TRIGGER_ID"); + +ALTER TABLE ONLY "WM_TRIGGER" + ADD CONSTRAINT "UNIQUE_WM_TRIGGER" UNIQUE ("RP_ID", "NAME"); + +ALTER TABLE ONLY "WM_POOL_TO_TRIGGER" + ADD CONSTRAINT "WM_POOL_TO_TRIGGER_pkey" PRIMARY KEY ("POOL_ID", "TRIGGER_ID"); + +ALTER TABLE ONLY "WM_MAPPING" + ADD CONSTRAINT "WM_MAPPING_pkey" PRIMARY KEY ("MAPPING_ID"); + +ALTER TABLE ONLY "WM_MAPPING" + ADD CONSTRAINT "UNIQUE_WM_MAPPING" UNIQUE ("RP_ID", "ENTITY_TYPE", "ENTITY_NAME"); + +-- +-- Name: BUCKETING_COLS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "BUCKETING_COLS_N49" ON "BUCKETING_COLS" USING btree ("SD_ID"); + + +-- +-- Name: DATABASE_PARAMS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "DATABASE_PARAMS_N49" ON "DATABASE_PARAMS" USING btree ("DB_ID"); + + +-- +-- Name: DB_PRIVS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "DB_PRIVS_N49" ON "DB_PRIVS" USING btree ("DB_ID"); + + +-- +-- Name: IDXS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "IDXS_N49" ON "IDXS" USING btree ("ORIG_TBL_ID"); + + +-- +-- Name: IDXS_N50; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "IDXS_N50" ON "IDXS" USING btree ("INDEX_TBL_ID"); + + +-- +-- Name: IDXS_N51; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "IDXS_N51" ON "IDXS" USING btree ("SD_ID"); + + +-- +-- Name: INDEX_PARAMS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "INDEX_PARAMS_N49" ON "INDEX_PARAMS" USING btree ("INDEX_ID"); + + +-- +-- Name: PARTITIONCOLUMNPRIVILEGEINDEX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITIONCOLUMNPRIVILEGEINDEX" ON "PART_COL_PRIVS" USING btree ("AUTHORIZER", "PART_ID", "COLUMN_NAME", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "PART_COL_PRIV", "GRANTOR", "GRANTOR_TYPE"); + + +-- +-- Name: PARTITIONEVENTINDEX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITIONEVENTINDEX" ON "PARTITION_EVENTS" USING btree ("PARTITION_NAME"); + + +-- +-- Name: PARTITIONS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITIONS_N49" ON "PARTITIONS" USING btree ("TBL_ID"); + + +-- +-- Name: PARTITIONS_N50; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITIONS_N50" ON "PARTITIONS" USING btree ("SD_ID"); + + +-- +-- Name: PARTITION_KEYS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITION_KEYS_N49" ON "PARTITION_KEYS" USING btree ("TBL_ID"); + + +-- +-- Name: PARTITION_KEY_VALS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITION_KEY_VALS_N49" ON "PARTITION_KEY_VALS" USING btree ("PART_ID"); + + +-- +-- Name: PARTITION_PARAMS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTITION_PARAMS_N49" ON "PARTITION_PARAMS" USING btree ("PART_ID"); + + +-- +-- Name: PARTPRIVILEGEINDEX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PARTPRIVILEGEINDEX" ON "PART_PRIVS" USING btree ("AUTHORIZER", "PART_ID", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "PART_PRIV", "GRANTOR", "GRANTOR_TYPE"); + + +-- +-- Name: PART_COL_PRIVS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PART_COL_PRIVS_N49" ON "PART_COL_PRIVS" USING btree ("PART_ID"); + + +-- +-- Name: PART_PRIVS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PART_PRIVS_N49" ON "PART_PRIVS" USING btree ("PART_ID"); + + +-- +-- Name: PCS_STATS_IDX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PCS_STATS_IDX" ON "PART_COL_STATS" USING btree ("CAT_NAME", "DB_NAME","TABLE_NAME","COLUMN_NAME","PARTITION_NAME"); + + +-- +-- Name: ROLE_MAP_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "ROLE_MAP_N49" ON "ROLE_MAP" USING btree ("ROLE_ID"); + + +-- +-- Name: SDS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "SDS_N49" ON "SDS" USING btree ("SERDE_ID"); + + +-- +-- Name: SD_PARAMS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "SD_PARAMS_N49" ON "SD_PARAMS" USING btree ("SD_ID"); + + +-- +-- Name: SERDE_PARAMS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "SERDE_PARAMS_N49" ON "SERDE_PARAMS" USING btree ("SERDE_ID"); + + +-- +-- Name: SORT_COLS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "SORT_COLS_N49" ON "SORT_COLS" USING btree ("SD_ID"); + + +-- +-- Name: TABLECOLUMNPRIVILEGEINDEX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TABLECOLUMNPRIVILEGEINDEX" ON "TBL_COL_PRIVS" USING btree ("AUTHORIZER", "TBL_ID", "COLUMN_NAME", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "TBL_COL_PRIV", "GRANTOR", "GRANTOR_TYPE"); + + +-- +-- Name: TABLEPRIVILEGEINDEX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TABLEPRIVILEGEINDEX" ON "TBL_PRIVS" USING btree ("AUTHORIZER", "TBL_ID", "PRINCIPAL_NAME", "PRINCIPAL_TYPE", "TBL_PRIV", "GRANTOR", "GRANTOR_TYPE"); + + +-- +-- Name: TABLE_PARAMS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TABLE_PARAMS_N49" ON "TABLE_PARAMS" USING btree ("TBL_ID"); + + +-- +-- Name: TBLS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TBLS_N49" ON "TBLS" USING btree ("DB_ID"); + + +-- +-- Name: TBLS_N50; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TBLS_N50" ON "TBLS" USING btree ("SD_ID"); + + +-- +-- Name: TBL_COL_PRIVS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TBL_COL_PRIVS_N49" ON "TBL_COL_PRIVS" USING btree ("TBL_ID"); + +-- +-- Name: TAB_COL_STATS_IDX; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TAB_COL_STATS_IDX" ON "TAB_COL_STATS" USING btree ("CAT_NAME", "DB_NAME","TABLE_NAME","COLUMN_NAME"); + +-- +-- Name: TBL_PRIVS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TBL_PRIVS_N49" ON "TBL_PRIVS" USING btree ("TBL_ID"); + + +-- +-- Name: TYPE_FIELDS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TYPE_FIELDS_N49" ON "TYPE_FIELDS" USING btree ("TYPE_NAME"); + +-- +-- Name: TAB_COL_STATS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "TAB_COL_STATS_N49" ON "TAB_COL_STATS" USING btree ("TBL_ID"); + +-- +-- Name: PART_COL_STATS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "PART_COL_STATS_N49" ON "PART_COL_STATS" USING btree ("PART_ID"); + +-- +-- Name: UNIQUEFUNCTION; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE UNIQUE INDEX "UNIQUEFUNCTION" ON "FUNCS" ("FUNC_NAME", "DB_ID"); + +-- +-- Name: FUNCS_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "FUNCS_N49" ON "FUNCS" ("DB_ID"); + +-- +-- Name: FUNC_RU_N49; Type: INDEX; Schema: public; Owner: hiveuser; Tablespace: +-- + +CREATE INDEX "FUNC_RU_N49" ON "FUNC_RU" ("FUNC_ID"); + +CREATE INDEX "CONSTRAINTS_PARENT_TBLID_INDEX" ON "KEY_CONSTRAINTS" USING BTREE ("PARENT_TBL_ID"); + +CREATE INDEX "CONSTRAINTS_CONSTRAINT_TYPE_INDEX" ON "KEY_CONSTRAINTS" USING BTREE ("CONSTRAINT_TYPE"); + +ALTER TABLE ONLY "SKEWED_STRING_LIST_VALUES" + ADD CONSTRAINT "SKEWED_STRING_LIST_VALUES_fkey" FOREIGN KEY ("STRING_LIST_ID") REFERENCES "SKEWED_STRING_LIST"("STRING_LIST_ID") DEFERRABLE; + + +ALTER TABLE ONLY "SKEWED_COL_NAMES" + ADD CONSTRAINT "SKEWED_COL_NAMES_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +ALTER TABLE ONLY "SKEWED_COL_VALUE_LOC_MAP" + ADD CONSTRAINT "SKEWED_COL_VALUE_LOC_MAP_fkey1" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + +ALTER TABLE ONLY "SKEWED_COL_VALUE_LOC_MAP" + ADD CONSTRAINT "SKEWED_COL_VALUE_LOC_MAP_fkey2" FOREIGN KEY ("STRING_LIST_ID_KID") REFERENCES "SKEWED_STRING_LIST"("STRING_LIST_ID") DEFERRABLE; + +ALTER TABLE ONLY "SKEWED_VALUES" + ADD CONSTRAINT "SKEWED_VALUES_fkey1" FOREIGN KEY ("STRING_LIST_ID_EID") REFERENCES "SKEWED_STRING_LIST"("STRING_LIST_ID") DEFERRABLE; + +ALTER TABLE ONLY "SKEWED_VALUES" + ADD CONSTRAINT "SKEWED_VALUES_fkey2" FOREIGN KEY ("SD_ID_OID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: BUCKETING_COLS_SD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "BUCKETING_COLS" + ADD CONSTRAINT "BUCKETING_COLS_SD_ID_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: COLUMNS_V2_CD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "COLUMNS_V2" + ADD CONSTRAINT "COLUMNS_V2_CD_ID_fkey" FOREIGN KEY ("CD_ID") REFERENCES "CDS"("CD_ID") DEFERRABLE; + + +-- +-- Name: DATABASE_PARAMS_DB_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "DATABASE_PARAMS" + ADD CONSTRAINT "DATABASE_PARAMS_DB_ID_fkey" FOREIGN KEY ("DB_ID") REFERENCES "DBS"("DB_ID") DEFERRABLE; + + +-- +-- Name: DB_PRIVS_DB_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "DB_PRIVS" + ADD CONSTRAINT "DB_PRIVS_DB_ID_fkey" FOREIGN KEY ("DB_ID") REFERENCES "DBS"("DB_ID") DEFERRABLE; + + +-- +-- Name: IDXS_INDEX_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "IDXS" + ADD CONSTRAINT "IDXS_INDEX_TBL_ID_fkey" FOREIGN KEY ("INDEX_TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: IDXS_ORIG_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "IDXS" + ADD CONSTRAINT "IDXS_ORIG_TBL_ID_fkey" FOREIGN KEY ("ORIG_TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: IDXS_SD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "IDXS" + ADD CONSTRAINT "IDXS_SD_ID_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: INDEX_PARAMS_INDEX_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "INDEX_PARAMS" + ADD CONSTRAINT "INDEX_PARAMS_INDEX_ID_fkey" FOREIGN KEY ("INDEX_ID") REFERENCES "IDXS"("INDEX_ID") DEFERRABLE; + + +-- +-- Name: PARTITIONS_SD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PARTITIONS" + ADD CONSTRAINT "PARTITIONS_SD_ID_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: PARTITIONS_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PARTITIONS" + ADD CONSTRAINT "PARTITIONS_TBL_ID_fkey" FOREIGN KEY ("TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: PARTITION_KEYS_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PARTITION_KEYS" + ADD CONSTRAINT "PARTITION_KEYS_TBL_ID_fkey" FOREIGN KEY ("TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: PARTITION_KEY_VALS_PART_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PARTITION_KEY_VALS" + ADD CONSTRAINT "PARTITION_KEY_VALS_PART_ID_fkey" FOREIGN KEY ("PART_ID") REFERENCES "PARTITIONS"("PART_ID") DEFERRABLE; + + +-- +-- Name: PARTITION_PARAMS_PART_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PARTITION_PARAMS" + ADD CONSTRAINT "PARTITION_PARAMS_PART_ID_fkey" FOREIGN KEY ("PART_ID") REFERENCES "PARTITIONS"("PART_ID") DEFERRABLE; + + +-- +-- Name: PART_COL_PRIVS_PART_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PART_COL_PRIVS" + ADD CONSTRAINT "PART_COL_PRIVS_PART_ID_fkey" FOREIGN KEY ("PART_ID") REFERENCES "PARTITIONS"("PART_ID") DEFERRABLE; + + +-- +-- Name: PART_PRIVS_PART_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "PART_PRIVS" + ADD CONSTRAINT "PART_PRIVS_PART_ID_fkey" FOREIGN KEY ("PART_ID") REFERENCES "PARTITIONS"("PART_ID") DEFERRABLE; + + +-- +-- Name: ROLE_MAP_ROLE_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "ROLE_MAP" + ADD CONSTRAINT "ROLE_MAP_ROLE_ID_fkey" FOREIGN KEY ("ROLE_ID") REFERENCES "ROLES"("ROLE_ID") DEFERRABLE; + + +-- +-- Name: SDS_CD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "SDS" + ADD CONSTRAINT "SDS_CD_ID_fkey" FOREIGN KEY ("CD_ID") REFERENCES "CDS"("CD_ID") DEFERRABLE; + + +-- +-- Name: SDS_SERDE_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "SDS" + ADD CONSTRAINT "SDS_SERDE_ID_fkey" FOREIGN KEY ("SERDE_ID") REFERENCES "SERDES"("SERDE_ID") DEFERRABLE; + + +-- +-- Name: SD_PARAMS_SD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "SD_PARAMS" + ADD CONSTRAINT "SD_PARAMS_SD_ID_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: SERDE_PARAMS_SERDE_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "SERDE_PARAMS" + ADD CONSTRAINT "SERDE_PARAMS_SERDE_ID_fkey" FOREIGN KEY ("SERDE_ID") REFERENCES "SERDES"("SERDE_ID") DEFERRABLE; + + +-- +-- Name: SORT_COLS_SD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "SORT_COLS" + ADD CONSTRAINT "SORT_COLS_SD_ID_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: TABLE_PARAMS_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "TABLE_PARAMS" + ADD CONSTRAINT "TABLE_PARAMS_TBL_ID_fkey" FOREIGN KEY ("TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: TBLS_DB_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "TBLS" + ADD CONSTRAINT "TBLS_DB_ID_fkey" FOREIGN KEY ("DB_ID") REFERENCES "DBS"("DB_ID") DEFERRABLE; + + +-- +-- Name: TBLS_SD_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "TBLS" + ADD CONSTRAINT "TBLS_SD_ID_fkey" FOREIGN KEY ("SD_ID") REFERENCES "SDS"("SD_ID") DEFERRABLE; + + +-- +-- Name: TBL_COL_PRIVS_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "TBL_COL_PRIVS" + ADD CONSTRAINT "TBL_COL_PRIVS_TBL_ID_fkey" FOREIGN KEY ("TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: TBL_PRIVS_TBL_ID_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "TBL_PRIVS" + ADD CONSTRAINT "TBL_PRIVS_TBL_ID_fkey" FOREIGN KEY ("TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: TYPE_FIELDS_TYPE_NAME_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- + +ALTER TABLE ONLY "TYPE_FIELDS" + ADD CONSTRAINT "TYPE_FIELDS_TYPE_NAME_fkey" FOREIGN KEY ("TYPE_NAME") REFERENCES "TYPES"("TYPES_ID") DEFERRABLE; + +-- +-- Name: TAB_COL_STATS_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- +ALTER TABLE ONLY "TAB_COL_STATS" ADD CONSTRAINT "TAB_COL_STATS_fkey" FOREIGN KEY("TBL_ID") REFERENCES "TBLS"("TBL_ID") DEFERRABLE; + + +-- +-- Name: PART_COL_STATS_fkey; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +-- +ALTER TABLE ONLY "PART_COL_STATS" ADD CONSTRAINT "PART_COL_STATS_fkey" FOREIGN KEY("PART_ID") REFERENCES "PARTITIONS"("PART_ID") DEFERRABLE; + +ALTER TABLE "DBS" ADD CONSTRAINT "DBS_FK1" FOREIGN KEY ("CTLG_NAME") REFERENCES "CTLGS" ("NAME"); + +ALTER TABLE ONLY "VERSION" ADD CONSTRAINT "VERSION_pkey" PRIMARY KEY ("VER_ID"); + +-- Name: FUNCS_FK1; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +ALTER TABLE ONLY "FUNCS" + ADD CONSTRAINT "FUNCS_FK1" FOREIGN KEY ("DB_ID") REFERENCES "DBS" ("DB_ID") DEFERRABLE; + +-- Name: FUNC_RU_FK1; Type: FK CONSTRAINT; Schema: public; Owner: hiveuser +ALTER TABLE ONLY "FUNC_RU" + ADD CONSTRAINT "FUNC_RU_FK1" FOREIGN KEY ("FUNC_ID") REFERENCES "FUNCS" ("FUNC_ID") DEFERRABLE; + +-- Resource plan FK constraints. + +ALTER TABLE ONLY "WM_POOL" + ADD CONSTRAINT "WM_POOL_FK1" FOREIGN KEY ("RP_ID") REFERENCES "WM_RESOURCEPLAN" ("RP_ID") DEFERRABLE; + +ALTER TABLE ONLY "WM_RESOURCEPLAN" + ADD CONSTRAINT "WM_RESOURCEPLAN_FK1" FOREIGN KEY ("DEFAULT_POOL_ID") REFERENCES "WM_POOL" ("POOL_ID") DEFERRABLE; + +ALTER TABLE ONLY "WM_TRIGGER" + ADD CONSTRAINT "WM_TRIGGER_FK1" FOREIGN KEY ("RP_ID") REFERENCES "WM_RESOURCEPLAN" ("RP_ID") DEFERRABLE; + +ALTER TABLE ONLY "WM_POOL_TO_TRIGGER" + ADD CONSTRAINT "WM_POOL_TO_TRIGGER_FK1" FOREIGN KEY ("POOL_ID") REFERENCES "WM_POOL" ("POOL_ID") DEFERRABLE; + +ALTER TABLE ONLY "WM_POOL_TO_TRIGGER" + ADD CONSTRAINT "WM_POOL_TO_TRIGGER_FK2" FOREIGN KEY ("TRIGGER_ID") REFERENCES "WM_TRIGGER" ("TRIGGER_ID") DEFERRABLE; + +ALTER TABLE ONLY "WM_MAPPING" + ADD CONSTRAINT "WM_MAPPING_FK1" FOREIGN KEY ("RP_ID") REFERENCES "WM_RESOURCEPLAN" ("RP_ID") DEFERRABLE; + +ALTER TABLE ONLY "WM_MAPPING" + ADD CONSTRAINT "WM_MAPPING_FK2" FOREIGN KEY ("POOL_ID") REFERENCES "WM_POOL" ("POOL_ID") DEFERRABLE; + +ALTER TABLE ONLY "MV_CREATION_METADATA" + ADD CONSTRAINT "MV_CREATION_METADATA_PK" PRIMARY KEY ("MV_CREATION_METADATA_ID"); + +CREATE INDEX "MV_UNIQUE_TABLE" + ON "MV_CREATION_METADATA" USING btree ("TBL_NAME", "DB_NAME"); + +ALTER TABLE ONLY "MV_TABLES_USED" + ADD CONSTRAINT "MV_TABLES_USED_FK1" FOREIGN KEY ("MV_CREATION_METADATA_ID") REFERENCES "MV_CREATION_METADATA" ("MV_CREATION_METADATA_ID") DEFERRABLE; + +ALTER TABLE ONLY "MV_TABLES_USED" + ADD CONSTRAINT "MV_TABLES_USED_FK2" FOREIGN KEY ("TBL_ID") REFERENCES "TBLS" ("TBL_ID") DEFERRABLE; + +-- +-- Name: public; Type: ACL; Schema: -; Owner: hiveuser +-- + +REVOKE ALL ON SCHEMA public FROM PUBLIC; +GRANT ALL ON SCHEMA public TO PUBLIC; + +-- +-- PostgreSQL database dump complete +-- + +------------------------------ +-- Transaction and lock tables +------------------------------ +CREATE TABLE TXNS ( + TXN_ID bigint PRIMARY KEY, + TXN_STATE char(1) NOT NULL, + TXN_STARTED bigint NOT NULL, + TXN_LAST_HEARTBEAT bigint NOT NULL, + TXN_USER varchar(128) NOT NULL, + TXN_HOST varchar(128) NOT NULL, + TXN_AGENT_INFO varchar(128), + TXN_META_INFO varchar(128), + TXN_HEARTBEAT_COUNT integer, + TXN_TYPE integer +); + +CREATE TABLE TXN_COMPONENTS ( + TC_TXNID bigint NOT NULL REFERENCES TXNS (TXN_ID), + TC_DATABASE varchar(128) NOT NULL, + TC_TABLE varchar(128), + TC_PARTITION varchar(767) DEFAULT NULL, + TC_OPERATION_TYPE char(1) NOT NULL, + TC_WRITEID bigint +); + +CREATE INDEX TC_TXNID_INDEX ON TXN_COMPONENTS USING hash (TC_TXNID); + +CREATE TABLE COMPLETED_TXN_COMPONENTS ( + CTC_TXNID bigint NOT NULL, + CTC_DATABASE varchar(128) NOT NULL, + CTC_TABLE varchar(256), + CTC_PARTITION varchar(767), + CTC_TIMESTAMP timestamp DEFAULT CURRENT_TIMESTAMP NOT NULL, + CTC_WRITEID bigint, + CTC_UPDATE_DELETE char(1) NOT NULL +); + +CREATE INDEX COMPLETED_TXN_COMPONENTS_INDEX ON COMPLETED_TXN_COMPONENTS USING btree (CTC_DATABASE, CTC_TABLE, CTC_PARTITION); + +CREATE TABLE NEXT_TXN_ID ( + NTXN_NEXT bigint NOT NULL +); +INSERT INTO NEXT_TXN_ID VALUES(1); + +CREATE TABLE HIVE_LOCKS ( + HL_LOCK_EXT_ID bigint NOT NULL, + HL_LOCK_INT_ID bigint NOT NULL, + HL_TXNID bigint NOT NULL, + HL_DB varchar(128) NOT NULL, + HL_TABLE varchar(128), + HL_PARTITION varchar(767) DEFAULT NULL, + HL_LOCK_STATE char(1) NOT NULL, + HL_LOCK_TYPE char(1) NOT NULL, + HL_LAST_HEARTBEAT bigint NOT NULL, + HL_ACQUIRED_AT bigint, + HL_USER varchar(128) NOT NULL, + HL_HOST varchar(128) NOT NULL, + HL_HEARTBEAT_COUNT integer, + HL_AGENT_INFO varchar(128), + HL_BLOCKEDBY_EXT_ID bigint, + HL_BLOCKEDBY_INT_ID bigint, + PRIMARY KEY(HL_LOCK_EXT_ID, HL_LOCK_INT_ID) +); + +CREATE INDEX HL_TXNID_INDEX ON HIVE_LOCKS USING hash (HL_TXNID); + +CREATE TABLE NEXT_LOCK_ID ( + NL_NEXT bigint NOT NULL +); +INSERT INTO NEXT_LOCK_ID VALUES(1); + +CREATE TABLE COMPACTION_QUEUE ( + CQ_ID bigint PRIMARY KEY, + CQ_DATABASE varchar(128) NOT NULL, + CQ_TABLE varchar(128) NOT NULL, + CQ_PARTITION varchar(767), + CQ_STATE char(1) NOT NULL, + CQ_TYPE char(1) NOT NULL, + CQ_TBLPROPERTIES varchar(2048), + CQ_WORKER_ID varchar(128), + CQ_START bigint, + CQ_RUN_AS varchar(128), + CQ_HIGHEST_WRITE_ID bigint, + CQ_META_INFO bytea, + CQ_HADOOP_JOB_ID varchar(32) +); + +CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( + NCQ_NEXT bigint NOT NULL +); +INSERT INTO NEXT_COMPACTION_QUEUE_ID VALUES(1); + +CREATE TABLE COMPLETED_COMPACTIONS ( + CC_ID bigint PRIMARY KEY, + CC_DATABASE varchar(128) NOT NULL, + CC_TABLE varchar(128) NOT NULL, + CC_PARTITION varchar(767), + CC_STATE char(1) NOT NULL, + CC_TYPE char(1) NOT NULL, + CC_TBLPROPERTIES varchar(2048), + CC_WORKER_ID varchar(128), + CC_START bigint, + CC_END bigint, + CC_RUN_AS varchar(128), + CC_HIGHEST_WRITE_ID bigint, + CC_META_INFO bytea, + CC_HADOOP_JOB_ID varchar(32) +); + +CREATE TABLE AUX_TABLE ( + MT_KEY1 varchar(128) NOT NULL, + MT_KEY2 bigint NOT NULL, + MT_COMMENT varchar(255), + PRIMARY KEY(MT_KEY1, MT_KEY2) +); + +CREATE TABLE WRITE_SET ( + WS_DATABASE varchar(128) NOT NULL, + WS_TABLE varchar(128) NOT NULL, + WS_PARTITION varchar(767), + WS_TXNID bigint NOT NULL, + WS_COMMIT_ID bigint NOT NULL, + WS_OPERATION_TYPE char(1) NOT NULL +); + +CREATE TABLE TXN_TO_WRITE_ID ( + T2W_TXNID bigint NOT NULL, + T2W_DATABASE varchar(128) NOT NULL, + T2W_TABLE varchar(256) NOT NULL, + T2W_WRITEID bigint NOT NULL +); + +CREATE UNIQUE INDEX TBL_TO_TXN_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_TXNID); +CREATE UNIQUE INDEX TBL_TO_WRITE_ID_IDX ON TXN_TO_WRITE_ID (T2W_DATABASE, T2W_TABLE, T2W_WRITEID); + +CREATE TABLE NEXT_WRITE_ID ( + NWI_DATABASE varchar(128) NOT NULL, + NWI_TABLE varchar(256) NOT NULL, + NWI_NEXT bigint NOT NULL +); + +CREATE UNIQUE INDEX NEXT_WRITE_ID_IDX ON NEXT_WRITE_ID (NWI_DATABASE, NWI_TABLE); + +CREATE TABLE MIN_HISTORY_LEVEL ( + MHL_TXNID bigint NOT NULL, + MHL_MIN_OPEN_TXNID bigint NOT NULL, + PRIMARY KEY(MHL_TXNID) +); + +CREATE INDEX MIN_HISTORY_LEVEL_IDX ON MIN_HISTORY_LEVEL (MHL_MIN_OPEN_TXNID); + +CREATE TABLE MATERIALIZATION_REBUILD_LOCKS ( + MRL_TXN_ID bigint NOT NULL, + MRL_DB_NAME varchar(128) NOT NULL, + MRL_TBL_NAME varchar(256) NOT NULL, + MRL_LAST_HEARTBEAT bigint NOT NULL, + PRIMARY KEY(MRL_TXN_ID) +); + +CREATE TABLE "I_SCHEMA" ( + "SCHEMA_ID" bigint primary key, + "SCHEMA_TYPE" integer not null, + "NAME" varchar(256) unique, + "DB_ID" bigint references "DBS" ("DB_ID"), + "COMPATIBILITY" integer not null, + "VALIDATION_LEVEL" integer not null, + "CAN_EVOLVE" boolean not null, + "SCHEMA_GROUP" varchar(256), + "DESCRIPTION" varchar(4000) +); + +CREATE TABLE "SCHEMA_VERSION" ( + "SCHEMA_VERSION_ID" bigint primary key, + "SCHEMA_ID" bigint references "I_SCHEMA" ("SCHEMA_ID"), + "VERSION" integer not null, + "CREATED_AT" bigint not null, + "CD_ID" bigint references "CDS" ("CD_ID"), + "STATE" integer not null, + "DESCRIPTION" varchar(4000), + "SCHEMA_TEXT" text, + "FINGERPRINT" varchar(256), + "SCHEMA_VERSION_NAME" varchar(256), + "SERDE_ID" bigint references "SERDES" ("SERDE_ID"), + unique ("SCHEMA_ID", "VERSION") +); + +CREATE TABLE REPL_TXN_MAP ( + RTM_REPL_POLICY varchar(256) NOT NULL, + RTM_SRC_TXN_ID bigint NOT NULL, + RTM_TARGET_TXN_ID bigint NOT NULL, + PRIMARY KEY (RTM_REPL_POLICY, RTM_SRC_TXN_ID) +); + + +CREATE TABLE RUNTIME_STATS ( + RS_ID bigint primary key, + CREATE_TIME bigint NOT NULL, + WEIGHT bigint NOT NULL, + PAYLOAD bytea +); + +CREATE INDEX IDX_RUNTIME_STATS_CREATE_TIME ON RUNTIME_STATS(CREATE_TIME); + +CREATE TABLE "TXN_WRITE_NOTIFICATION_LOG" ( + "WNL_ID" bigint NOT NULL, + "WNL_TXNID" bigint NOT NULL, + "WNL_WRITEID" bigint NOT NULL, + "WNL_DATABASE" varchar(128) NOT NULL, + "WNL_TABLE" varchar(128) NOT NULL, + "WNL_PARTITION" varchar(767) NOT NULL, + "WNL_TABLE_OBJ" text NOT NULL, + "WNL_PARTITION_OBJ" text, + "WNL_FILES" text, + "WNL_EVENT_TIME" integer NOT NULL, + PRIMARY KEY ("WNL_TXNID", "WNL_DATABASE", "WNL_TABLE", "WNL_PARTITION") +); + +INSERT INTO "SEQUENCE_TABLE" ("SEQUENCE_NAME", "NEXT_VAL") VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- ----------------------------------------------------------------- +-- Record schema version. Should be the last step in the init script +-- ----------------------------------------------------------------- +INSERT INTO "VERSION" ("VER_ID", "SCHEMA_VERSION", "VERSION_COMMENT") VALUES (1, '3.2.0', 'Hive release version 3.2.0'); diff --git a/standalone-metastore/src/main/sql/postgres/upgrade-3.1.0-to-3.2.0.postgres.sql b/standalone-metastore/src/main/sql/postgres/upgrade-3.1.0-to-3.2.0.postgres.sql new file mode 100644 index 0000000000..6a035103bb --- /dev/null +++ b/standalone-metastore/src/main/sql/postgres/upgrade-3.1.0-to-3.2.0.postgres.sql @@ -0,0 +1,21 @@ +SELECT 'Upgrading MetaStore schema from 3.1.0 to 3.2.0'; + +-- HIVE-19267 +CREATE TABLE "TXN_WRITE_NOTIFICATION_LOG" ( + "WNL_ID" bigint NOT NULL, + "WNL_TXNID" bigint NOT NULL, + "WNL_WRITEID" bigint NOT NULL, + "WNL_DATABASE" varchar(128) NOT NULL, + "WNL_TABLE" varchar(128) NOT NULL, + "WNL_PARTITION" varchar(767) NOT NULL, + "WNL_TABLE_OBJ" text NOT NULL, + "WNL_PARTITION_OBJ" text, + "WNL_FILES" text, + "WNL_EVENT_TIME" integer NOT NULL, + PRIMARY KEY ("WNL_TXNID", "WNL_DATABASE", "WNL_TABLE", "WNL_PARTITION") +); +INSERT INTO "SEQUENCE_TABLE" ("SEQUENCE_NAME", "NEXT_VAL") VALUES ('org.apache.hadoop.hive.metastore.model.MTxnWriteNotificationLog', 1); + +-- These lines need to be last. Insert any changes above. +UPDATE "VERSION" SET "SCHEMA_VERSION"='3.2.0', "VERSION_COMMENT"='Hive release version 3.2.0' where "VER_ID"=1; +SELECT 'Finished upgrading MetaStore schema from 3.1.0 to 3.2.0'; diff --git a/standalone-metastore/src/main/sql/postgres/upgrade.order.postgres b/standalone-metastore/src/main/sql/postgres/upgrade.order.postgres index f43da9af19..e6eb71ae56 100644 --- a/standalone-metastore/src/main/sql/postgres/upgrade.order.postgres +++ b/standalone-metastore/src/main/sql/postgres/upgrade.order.postgres @@ -15,3 +15,4 @@ 2.2.0-to-2.3.0 2.3.0-to-3.0.0 3.0.0-to-3.1.0 +3.1.0-to-3.2.0 diff --git a/standalone-metastore/src/main/thrift/hive_metastore.thrift b/standalone-metastore/src/main/thrift/hive_metastore.thrift index ad1dc1f769..8965059cf6 100644 --- a/standalone-metastore/src/main/thrift/hive_metastore.thrift +++ b/standalone-metastore/src/main/thrift/hive_metastore.thrift @@ -867,6 +867,18 @@ struct AbortTxnsRequest { struct CommitTxnRequest { 1: required i64 txnid, 2: optional string replPolicy, + // Information related to write operations done in this transaction. + 3: optional list writeEventInfos, +} + +struct WriteEventInfo { + 1: required i64 writeId, + 2: required string database, + 3: required string table, + 4: required string files, + 5: optional string partition, + 6: optional string tableObj, // repl txn task does not need table object for commit + 7: optional string partitionObj, } struct ReplTblWriteIdStateRequest { @@ -1103,6 +1115,8 @@ struct InsertEventRequestData { 2: required list filesAdded, // Checksum of files (hex string of checksum byte payload) 3: optional list filesAddedChecksum, + // Used by acid operation to create the sub directory + 4: optional list subDirectoryList, } union FireEventRequestData { @@ -1123,7 +1137,20 @@ struct FireEventRequest { struct FireEventResponse { // NOP for now, this is just a place holder for future responses } - + +struct WriteNotificationLogRequest { + 1: required i64 txnId, + 2: required i64 writeId, + 3: required string db, + 4: required string table, + 5: required InsertEventRequestData fileInfo, + 6: optional list partitionVals, +} + +struct WriteNotificationLogResponse { + // NOP for now, this is just a place holder for future responses +} + struct MetadataPpdResult { 1: optional binary metadata, 2: optional binary includeBitset @@ -2102,6 +2129,7 @@ service ThriftHiveMetastore extends fb303.FacebookService NotificationEventsCountResponse get_notification_events_count(1:NotificationEventsCountRequest rqst) FireEventResponse fire_listener_event(1:FireEventRequest rqst) void flushCache() + WriteNotificationLogResponse add_write_notification_log(WriteNotificationLogRequest rqst) // Repl Change Management api CmRecycleResponse cm_recycle(1:CmRecycleRequest request) throws(1:MetaException o1) diff --git a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java index bf00d2ecdf..e57fcf2d9e 100644 --- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java +++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java @@ -85,6 +85,7 @@ import org.apache.hadoop.hive.metastore.api.UnknownTableException; import org.apache.hadoop.hive.metastore.api.WMMapping; import org.apache.hadoop.hive.metastore.api.WMPool; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.ColStatsObjWithSourceInfo; import org.apache.thrift.TException; @@ -1187,6 +1188,16 @@ public int deleteRuntimeStats(int maxRetainSecs) throws MetaException { return objectStore.deleteRuntimeStats(maxRetainSecs); } + @Override + public void cleanWriteNotificationEvents(int olderThan) { + objectStore.cleanWriteNotificationEvents(olderThan); + } + + @Override + public List getAllWriteEventInfo(long txnId, String dbName, String tableName) throws MetaException { + return objectStore.getAllWriteEventInfo(txnId, dbName, tableName); + } + @Override public List getTableNamesWithStats() throws MetaException, NoSuchObjectException { diff --git a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreForJdoConnection.java b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreForJdoConnection.java index 7ea09a05f4..46548e5a92 100644 --- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreForJdoConnection.java +++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/DummyRawStoreForJdoConnection.java @@ -83,6 +83,7 @@ import org.apache.hadoop.hive.metastore.api.UnknownTableException; import org.apache.hadoop.hive.metastore.api.WMMapping; import org.apache.hadoop.hive.metastore.api.WMPool; +import org.apache.hadoop.hive.metastore.api.WriteEventInfo; import org.apache.hadoop.hive.metastore.conf.MetastoreConf; import org.apache.hadoop.hive.metastore.partition.spec.PartitionSpecProxy; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; @@ -1192,4 +1193,13 @@ public int deleteRuntimeStats(int maxRetainSecs) throws MetaException { NoSuchObjectException { return null; } + + @Override + public void cleanWriteNotificationEvents(int olderThan) { + } + + @Override + public List getAllWriteEventInfo(long txnId, String dbName, String tableName) throws MetaException { + return null; + } } diff --git a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java index 2d57cfb10f..d91f737450 100644 --- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java +++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClientPreCatalog.java @@ -2228,10 +2228,8 @@ public void commitTxn(long txnid) } @Override - public void replCommitTxn(long srcTxnId, String replPolicy) + public void replCommitTxn(CommitTxnRequest rqst) throws NoSuchTxnException, TxnAbortedException, TException { - CommitTxnRequest rqst = new CommitTxnRequest(srcTxnId); - rqst.setReplPolicy(replPolicy); client.commit_txn(rqst); } @@ -2472,6 +2470,12 @@ public FireEventResponse fireListenerEvent(FireEventRequest rqst) throws TExcept return client.fire_listener_event(rqst); } + @InterfaceAudience.LimitedPrivate({"Apache Hive, HCatalog"}) + @Override + public void addWriteNotificationLog(WriteNotificationLogRequest rqst) throws TException { + client.add_write_notification_log(rqst); + } + /** * Creates a synchronized wrapper for any {@link IMetaStoreClient}. * This may be used by multi-threaded applications until we have