commit 9f5dafa2d2a4eafe311b0460dae045be5f1d6e64 Author: Alan Gates Date: Fri Jan 8 14:54:39 2016 -0800 HIVE-12816 Thrift and schema changes for HIVE-11685 diff --git metastore/if/hive_metastore.thrift metastore/if/hive_metastore.thrift index 4ec44af..c5d4c0a 100755 --- metastore/if/hive_metastore.thrift +++ metastore/if/hive_metastore.thrift @@ -660,6 +660,7 @@ struct ShowCompactResponseElement { 9: optional i64 hightestTxnId, // Highest Txn ID handled by this compaction 10: optional string metaInfo, 11: optional i64 endTime, + 12: optional string hadoopJobId = "None", } struct ShowCompactResponse { diff --git metastore/scripts/upgrade/derby/025-HIVE-12816.derby.sql metastore/scripts/upgrade/derby/025-HIVE-12816.derby.sql new file mode 100644 index 0000000..2a78bf1 --- /dev/null +++ metastore/scripts/upgrade/derby/025-HIVE-12816.derby.sql @@ -0,0 +1 @@ +ALTER TABLE "COMPACTION_QUEUE" ADD "CQ_HADOOP_JOB_ID" VARCHAR(32); diff --git metastore/scripts/upgrade/derby/hive-txn-schema-1.3.0.derby.sql metastore/scripts/upgrade/derby/hive-txn-schema-1.3.0.derby.sql index 085bddc..0d3e168 100644 --- metastore/scripts/upgrade/derby/hive-txn-schema-1.3.0.derby.sql +++ metastore/scripts/upgrade/derby/hive-txn-schema-1.3.0.derby.sql @@ -78,7 +78,8 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_START bigint, CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID bigint, - CQ_META_INFO varchar(2048) for bit data + CQ_META_INFO varchar(2048) for bit data, + CQ_HADOOP_JOB_ID varchar(32) ); CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/derby/hive-txn-schema-2.0.0.derby.sql metastore/scripts/upgrade/derby/hive-txn-schema-2.0.0.derby.sql index 085bddc..0d3e168 100644 --- metastore/scripts/upgrade/derby/hive-txn-schema-2.0.0.derby.sql +++ metastore/scripts/upgrade/derby/hive-txn-schema-2.0.0.derby.sql @@ -78,7 +78,8 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_START bigint, CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID bigint, - CQ_META_INFO varchar(2048) for bit data + CQ_META_INFO varchar(2048) for bit data, + CQ_HADOOP_JOB_ID varchar(32) ); CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/derby/upgrade-1.2.0-to-1.3.0.derby.sql metastore/scripts/upgrade/derby/upgrade-1.2.0-to-1.3.0.derby.sql index f199d23..d0d0dc0 100644 --- metastore/scripts/upgrade/derby/upgrade-1.2.0-to-1.3.0.derby.sql +++ metastore/scripts/upgrade/derby/upgrade-1.2.0-to-1.3.0.derby.sql @@ -2,5 +2,6 @@ RUN '021-HIVE-11970.derby.sql'; RUN '023-HIVE-12807.derby.sql'; RUN '024-HIVE-12814.derby.sql'; +RUN '025-HIVE-12816.derby.sql'; UPDATE "APP".VERSION SET SCHEMA_VERSION='1.3.0', VERSION_COMMENT='Hive release version 1.3.0' where VER_ID=1; diff --git metastore/scripts/upgrade/derby/upgrade-1.2.0-to-2.0.0.derby.sql metastore/scripts/upgrade/derby/upgrade-1.2.0-to-2.0.0.derby.sql index 17a116e..f3f8ab2 100644 --- metastore/scripts/upgrade/derby/upgrade-1.2.0-to-2.0.0.derby.sql +++ metastore/scripts/upgrade/derby/upgrade-1.2.0-to-2.0.0.derby.sql @@ -2,5 +2,6 @@ RUN '021-HIVE-11970.derby.sql'; RUN '023-HIVE-12807.derby.sql'; RUN '024-HIVE-12814.derby.sql'; +RUN '025-HIVE-12816.derby.sql'; UPDATE "APP".VERSION SET SCHEMA_VERSION='2.0.0', VERSION_COMMENT='Hive release version 2.0.0' where VER_ID=1; diff --git metastore/scripts/upgrade/mssql/010-HIVE-12816.mssql.sql metastore/scripts/upgrade/mssql/010-HIVE-12816.mssql.sql new file mode 100644 index 0000000..0c0342f --- /dev/null +++ metastore/scripts/upgrade/mssql/010-HIVE-12816.mssql.sql @@ -0,0 +1 @@ +ALTER TABLE COMPACTION_QUEUE ADD CQ_HADOOP_JOB_ID nvarchar(32) NULL; diff --git metastore/scripts/upgrade/mssql/hive-schema-1.3.0.mssql.sql metastore/scripts/upgrade/mssql/hive-schema-1.3.0.mssql.sql index 787bfec..b97b344 100644 --- metastore/scripts/upgrade/mssql/hive-schema-1.3.0.mssql.sql +++ metastore/scripts/upgrade/mssql/hive-schema-1.3.0.mssql.sql @@ -868,6 +868,7 @@ CREATE TABLE COMPACTION_QUEUE( CQ_RUN_AS nvarchar(128) NULL, CQ_HIGHEST_TXN_ID bigint NULL, CQ_META_INFO varbinary(2048) NULL, + CQ_HADOOP_JOB_ID nvarchar(128) NULL, PRIMARY KEY CLUSTERED ( CQ_ID ASC diff --git metastore/scripts/upgrade/mssql/hive-schema-2.0.0.mssql.sql metastore/scripts/upgrade/mssql/hive-schema-2.0.0.mssql.sql index 46ba0a4..282f616 100644 --- metastore/scripts/upgrade/mssql/hive-schema-2.0.0.mssql.sql +++ metastore/scripts/upgrade/mssql/hive-schema-2.0.0.mssql.sql @@ -868,6 +868,7 @@ CREATE TABLE COMPACTION_QUEUE( CQ_RUN_AS nvarchar(128) NULL, CQ_HIGHEST_TXN_ID bigint NULL, CQ_META_INFO varbinary(2048) NULL, + CQ_HADOOP_JOB_ID nvarchar(128) NULL, PRIMARY KEY CLUSTERED ( CQ_ID ASC diff --git metastore/scripts/upgrade/mssql/hive-schema-2.1.0.mssql.sql metastore/scripts/upgrade/mssql/hive-schema-2.1.0.mssql.sql index a9cef1c..6427287 100644 --- metastore/scripts/upgrade/mssql/hive-schema-2.1.0.mssql.sql +++ metastore/scripts/upgrade/mssql/hive-schema-2.1.0.mssql.sql @@ -867,6 +867,8 @@ CREATE TABLE COMPACTION_QUEUE( CQ_START bigint NULL, CQ_RUN_AS nvarchar(128) NULL, CQ_HIGHEST_TXN_ID bigint NULL, + CQ_META_INFO varbinary(2048) NULL, + CQ_HADOOP_JOB_ID nvarchar(128) NULL, PRIMARY KEY CLUSTERED ( CQ_ID ASC diff --git metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-1.3.0.mssql.sql metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-1.3.0.mssql.sql index c3ebd98..355c341 100644 --- metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-1.3.0.mssql.sql +++ metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-1.3.0.mssql.sql @@ -3,6 +3,7 @@ SELECT 'Upgrading MetaStore schema from 1.2.0 to 1.3.0' AS MESSAGE; :r 007-HIVE-11970.mssql.sql; :r 008-HIVE-12807.mssql.sql; :r 009-HIVE-12814.mssql.sql; +:r 010-HIVE-12816.mssql.sql; UPDATE VERSION SET SCHEMA_VERSION='1.3.0', VERSION_COMMENT='Hive release version 1.3.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 1.3.0' AS MESSAGE; diff --git metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-2.0.0.mssql.sql metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-2.0.0.mssql.sql index 9956839..a1530a9 100644 --- metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-2.0.0.mssql.sql +++ metastore/scripts/upgrade/mssql/upgrade-1.2.0-to-2.0.0.mssql.sql @@ -3,6 +3,7 @@ SELECT 'Upgrading MetaStore schema from 1.2.0 to 2.0.0' AS MESSAGE; :r 007-HIVE-11970.mssql.sql; :r 008-HIVE-12807.mssql.sql; :r 009-HIVE-12814.mssql.sql; +:r 010-HIVE-12816.mssql.sql; UPDATE VERSION SET SCHEMA_VERSION='2.0.0', VERSION_COMMENT='Hive release version 2.0.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 2.0.0' AS MESSAGE; diff --git metastore/scripts/upgrade/mysql/025-HIVE-12816.mysql.sql metastore/scripts/upgrade/mysql/025-HIVE-12816.mysql.sql new file mode 100644 index 0000000..87e48b4 --- /dev/null +++ metastore/scripts/upgrade/mysql/025-HIVE-12816.mysql.sql @@ -0,0 +1 @@ +ALTER TABLE `COMPACTION_QUEUE` ADD `CQ_HADOOP_JOB_ID` varchar(32); diff --git metastore/scripts/upgrade/mysql/hive-txn-schema-1.3.0.mysql.sql metastore/scripts/upgrade/mysql/hive-txn-schema-1.3.0.mysql.sql index 73d2a91..50b54a9 100644 --- metastore/scripts/upgrade/mysql/hive-txn-schema-1.3.0.mysql.sql +++ metastore/scripts/upgrade/mysql/hive-txn-schema-1.3.0.mysql.sql @@ -82,6 +82,7 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID bigint, CQ_META_INFO varbinary(2048), + CQ_HADOOP_JOB_ID varchar(32), ) ENGINE=InnoDB DEFAULT CHARSET=latin1; CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/mysql/hive-txn-schema-2.0.0.mysql.sql metastore/scripts/upgrade/mysql/hive-txn-schema-2.0.0.mysql.sql index 73d2a91..50b54a9 100644 --- metastore/scripts/upgrade/mysql/hive-txn-schema-2.0.0.mysql.sql +++ metastore/scripts/upgrade/mysql/hive-txn-schema-2.0.0.mysql.sql @@ -82,6 +82,7 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID bigint, CQ_META_INFO varbinary(2048), + CQ_HADOOP_JOB_ID varchar(32), ) ENGINE=InnoDB DEFAULT CHARSET=latin1; CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-1.3.0.mysql.sql metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-1.3.0.mysql.sql index 697cf1d..45f8886 100644 --- metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-1.3.0.mysql.sql +++ metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-1.3.0.mysql.sql @@ -3,6 +3,7 @@ SOURCE 021-HIVE-7018.mysql.sql; SOURCE 022-HIVE-11970.mysql.sql; SOURCE 023-HIVE-12807.mysql.sql; SOURCE 024-HIVE-12814.mysql.sql; +SOURCE 025-HIVE-12816.mysql.sql; UPDATE VERSION SET SCHEMA_VERSION='1.3.0', VERSION_COMMENT='Hive release version 1.3.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 1.3.0' AS ' '; diff --git metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-2.0.0.mysql.sql metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-2.0.0.mysql.sql index 29c923d..95b0e91 100644 --- metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-2.0.0.mysql.sql +++ metastore/scripts/upgrade/mysql/upgrade-1.2.0-to-2.0.0.mysql.sql @@ -3,6 +3,7 @@ SOURCE 021-HIVE-7018.mysql.sql; SOURCE 022-HIVE-11970.mysql.sql; SOURCE 023-HIVE-12807.mysql.sql; SOURCE 024-HIVE-12814.mysql.sql; +SOURCE 025-HIVE-12816.mysql.sql; UPDATE VERSION SET SCHEMA_VERSION='2.0.0', VERSION_COMMENT='Hive release version 2.0.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 2.0.0' AS ' '; diff --git metastore/scripts/upgrade/oracle/025-HIVE-12816.oracle.sql metastore/scripts/upgrade/oracle/025-HIVE-12816.oracle.sql new file mode 100644 index 0000000..9a96d8c --- /dev/null +++ metastore/scripts/upgrade/oracle/025-HIVE-12816.oracle.sql @@ -0,0 +1 @@ +ALTER TABLE COMPACTION_QUEUE ADD CQ_HADOOP_JOB_ID varchar2(32); diff --git metastore/scripts/upgrade/oracle/hive-txn-schema-1.3.0.oracle.sql metastore/scripts/upgrade/oracle/hive-txn-schema-1.3.0.oracle.sql index d32dce3..a600b6e 100644 --- metastore/scripts/upgrade/oracle/hive-txn-schema-1.3.0.oracle.sql +++ metastore/scripts/upgrade/oracle/hive-txn-schema-1.3.0.oracle.sql @@ -80,6 +80,7 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID NUMBER(19), CQ_META_INFO BLOB, + CQ_HADOOP_JOB_ID varchar2(32) ) ROWDEPENDENCIES; CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/oracle/hive-txn-schema-2.0.0.oracle.sql metastore/scripts/upgrade/oracle/hive-txn-schema-2.0.0.oracle.sql index d32dce3..a600b6e 100644 --- metastore/scripts/upgrade/oracle/hive-txn-schema-2.0.0.oracle.sql +++ metastore/scripts/upgrade/oracle/hive-txn-schema-2.0.0.oracle.sql @@ -80,6 +80,7 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID NUMBER(19), CQ_META_INFO BLOB, + CQ_HADOOP_JOB_ID varchar2(32) ) ROWDEPENDENCIES; CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-1.3.0.oracle.sql metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-1.3.0.oracle.sql index 4161b3b..0f20fe6 100644 --- metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-1.3.0.oracle.sql +++ metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-1.3.0.oracle.sql @@ -3,6 +3,7 @@ SELECT 'Upgrading MetaStore schema from 1.2.0 to 1.3.0' AS Status from dual; @022-HIVE-11970.oracle.sql; @023-HIVE-12807.oracle.sql; @024-HIVE-12814.oracle.sql; +@025-HIVE-12816.oracle.sql; UPDATE VERSION SET SCHEMA_VERSION='1.3.0', VERSION_COMMENT='Hive release version 1.3.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 1.3.0' AS Status from dual; diff --git metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-2.0.0.oracle.sql metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-2.0.0.oracle.sql index 0b835d0..5f75e6e 100644 --- metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-2.0.0.oracle.sql +++ metastore/scripts/upgrade/oracle/upgrade-1.2.0-to-2.0.0.oracle.sql @@ -3,6 +3,7 @@ SELECT 'Upgrading MetaStore schema from 1.2.0 to 2.0.0' AS Status from dual; @022-HIVE-11970.oracle.sql; @023-HIVE-12807.oracle.sql; @024-HIVE-12814.oracle.sql; +@025-HIVE-12816.oracle.sql; UPDATE VERSION SET SCHEMA_VERSION='2.0.0', VERSION_COMMENT='Hive release version 2.0.0' where VER_ID=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 2.0.0' AS Status from dual; diff --git metastore/scripts/upgrade/postgres/024-HIVE-12816.postgres.sql metastore/scripts/upgrade/postgres/024-HIVE-12816.postgres.sql new file mode 100644 index 0000000..a5190ac --- /dev/null +++ metastore/scripts/upgrade/postgres/024-HIVE-12816.postgres.sql @@ -0,0 +1 @@ +ALTER TABLE "COMPACTION_QUEUE" ADD COLUMN "CQ_HADOOP_JOB_ID" varchar(32); diff --git metastore/scripts/upgrade/postgres/hive-txn-schema-1.3.0.postgres.sql metastore/scripts/upgrade/postgres/hive-txn-schema-1.3.0.postgres.sql index 33ffaf1..b26e975 100644 --- metastore/scripts/upgrade/postgres/hive-txn-schema-1.3.0.postgres.sql +++ metastore/scripts/upgrade/postgres/hive-txn-schema-1.3.0.postgres.sql @@ -79,7 +79,8 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_START bigint, CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID bigint, - CQ_META_INFO bytea + CQ_META_INFO bytea, + CQ_HADOOP_JOB_ID varchar(32) ); CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/postgres/hive-txn-schema-2.0.0.postgres.sql metastore/scripts/upgrade/postgres/hive-txn-schema-2.0.0.postgres.sql index 33ffaf1..b26e975 100644 --- metastore/scripts/upgrade/postgres/hive-txn-schema-2.0.0.postgres.sql +++ metastore/scripts/upgrade/postgres/hive-txn-schema-2.0.0.postgres.sql @@ -79,7 +79,8 @@ CREATE TABLE COMPACTION_QUEUE ( CQ_START bigint, CQ_RUN_AS varchar(128), CQ_HIGHEST_TXN_ID bigint, - CQ_META_INFO bytea + CQ_META_INFO bytea, + CQ_HADOOP_JOB_ID varchar(32) ); CREATE TABLE NEXT_COMPACTION_QUEUE_ID ( diff --git metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-1.3.0.postgres.sql metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-1.3.0.postgres.sql index 85b4e78..975d2ee 100644 --- metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-1.3.0.postgres.sql +++ metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-1.3.0.postgres.sql @@ -3,6 +3,7 @@ SELECT 'Upgrading MetaStore schema from 1.2.0 to 1.3.0'; \i 021-HIVE-11970.postgres.sql; \i 022-HIVE-12807.postgres.sql; \i 023-HIVE-12814.postgres.sql; +\i 024-HIVE-12816.postgres.sql; UPDATE "VERSION" SET "SCHEMA_VERSION"='1.3.0', "VERSION_COMMENT"='Hive release version 1.3.0' where "VER_ID"=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 1.3.0'; diff --git metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-2.0.0.postgres.sql metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-2.0.0.postgres.sql index 895505d..9386002 100644 --- metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-2.0.0.postgres.sql +++ metastore/scripts/upgrade/postgres/upgrade-1.2.0-to-2.0.0.postgres.sql @@ -3,6 +3,7 @@ SELECT 'Upgrading MetaStore schema from 1.2.0 to 2.0.0'; \i 021-HIVE-11970.postgres.sql; \i 022-HIVE-12807.postgres.sql; \i 023-HIVE-12814.postgres.sql; +\i 024-HIVE-12816.postgres.sql; UPDATE "VERSION" SET "SCHEMA_VERSION"='2.0.0', "VERSION_COMMENT"='Hive release version 2.0.0' where "VER_ID"=1; SELECT 'Finished upgrading MetaStore schema from 1.2.0 to 2.0.0'; diff --git metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp index 188c3ab..0edc951 100644 --- metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp +++ metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp @@ -12579,6 +12579,11 @@ void ShowCompactResponseElement::__set_endTime(const int64_t val) { __isset.endTime = true; } +void ShowCompactResponseElement::__set_hadoopJobId(const std::string& val) { + this->hadoopJobId = val; +__isset.hadoopJobId = true; +} + uint32_t ShowCompactResponseElement::read(::apache::thrift::protocol::TProtocol* iprot) { apache::thrift::protocol::TInputRecursionTracker tracker(*iprot); @@ -12694,6 +12699,14 @@ uint32_t ShowCompactResponseElement::read(::apache::thrift::protocol::TProtocol* xfer += iprot->skip(ftype); } break; + case 12: + if (ftype == ::apache::thrift::protocol::T_STRING) { + xfer += iprot->readString(this->hadoopJobId); + this->__isset.hadoopJobId = true; + } else { + xfer += iprot->skip(ftype); + } + break; default: xfer += iprot->skip(ftype); break; @@ -12770,6 +12783,11 @@ uint32_t ShowCompactResponseElement::write(::apache::thrift::protocol::TProtocol xfer += oprot->writeI64(this->endTime); xfer += oprot->writeFieldEnd(); } + if (this->__isset.hadoopJobId) { + xfer += oprot->writeFieldBegin("hadoopJobId", ::apache::thrift::protocol::T_STRING, 12); + xfer += oprot->writeString(this->hadoopJobId); + xfer += oprot->writeFieldEnd(); + } xfer += oprot->writeFieldStop(); xfer += oprot->writeStructEnd(); return xfer; @@ -12788,6 +12806,7 @@ void swap(ShowCompactResponseElement &a, ShowCompactResponseElement &b) { swap(a.hightestTxnId, b.hightestTxnId); swap(a.metaInfo, b.metaInfo); swap(a.endTime, b.endTime); + swap(a.hadoopJobId, b.hadoopJobId); swap(a.__isset, b.__isset); } @@ -12803,6 +12822,7 @@ ShowCompactResponseElement::ShowCompactResponseElement(const ShowCompactResponse hightestTxnId = other557.hightestTxnId; metaInfo = other557.metaInfo; endTime = other557.endTime; + hadoopJobId = other557.hadoopJobId; __isset = other557.__isset; } ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowCompactResponseElement& other558) { @@ -12817,6 +12837,7 @@ ShowCompactResponseElement& ShowCompactResponseElement::operator=(const ShowComp hightestTxnId = other558.hightestTxnId; metaInfo = other558.metaInfo; endTime = other558.endTime; + hadoopJobId = other558.hadoopJobId; __isset = other558.__isset; return *this; } @@ -12834,6 +12855,7 @@ void ShowCompactResponseElement::printTo(std::ostream& out) const { out << ", " << "hightestTxnId="; (__isset.hightestTxnId ? (out << to_string(hightestTxnId)) : (out << "")); out << ", " << "metaInfo="; (__isset.metaInfo ? (out << to_string(metaInfo)) : (out << "")); out << ", " << "endTime="; (__isset.endTime ? (out << to_string(endTime)) : (out << "")); + out << ", " << "hadoopJobId="; (__isset.hadoopJobId ? (out << to_string(hadoopJobId)) : (out << "")); out << ")"; } diff --git metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h index d11cca9..45209f0 100644 --- metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h +++ metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h @@ -5119,7 +5119,7 @@ inline std::ostream& operator<<(std::ostream& out, const ShowCompactRequest& obj } typedef struct _ShowCompactResponseElement__isset { - _ShowCompactResponseElement__isset() : partitionname(false), workerid(false), start(false), runAs(false), hightestTxnId(false), metaInfo(false), endTime(false) {} + _ShowCompactResponseElement__isset() : partitionname(false), workerid(false), start(false), runAs(false), hightestTxnId(false), metaInfo(false), endTime(false), hadoopJobId(true) {} bool partitionname :1; bool workerid :1; bool start :1; @@ -5127,6 +5127,7 @@ typedef struct _ShowCompactResponseElement__isset { bool hightestTxnId :1; bool metaInfo :1; bool endTime :1; + bool hadoopJobId :1; } _ShowCompactResponseElement__isset; class ShowCompactResponseElement { @@ -5134,7 +5135,7 @@ class ShowCompactResponseElement { ShowCompactResponseElement(const ShowCompactResponseElement&); ShowCompactResponseElement& operator=(const ShowCompactResponseElement&); - ShowCompactResponseElement() : dbname(), tablename(), partitionname(), type((CompactionType::type)0), state(), workerid(), start(0), runAs(), hightestTxnId(0), metaInfo(), endTime(0) { + ShowCompactResponseElement() : dbname(), tablename(), partitionname(), type((CompactionType::type)0), state(), workerid(), start(0), runAs(), hightestTxnId(0), metaInfo(), endTime(0), hadoopJobId("None") { } virtual ~ShowCompactResponseElement() throw(); @@ -5149,6 +5150,7 @@ class ShowCompactResponseElement { int64_t hightestTxnId; std::string metaInfo; int64_t endTime; + std::string hadoopJobId; _ShowCompactResponseElement__isset __isset; @@ -5174,6 +5176,8 @@ class ShowCompactResponseElement { void __set_endTime(const int64_t val); + void __set_hadoopJobId(const std::string& val); + bool operator == (const ShowCompactResponseElement & rhs) const { if (!(dbname == rhs.dbname)) @@ -5212,6 +5216,10 @@ class ShowCompactResponseElement { return false; else if (__isset.endTime && !(endTime == rhs.endTime)) return false; + if (__isset.hadoopJobId != rhs.__isset.hadoopJobId) + return false; + else if (__isset.hadoopJobId && !(hadoopJobId == rhs.hadoopJobId)) + return false; return true; } bool operator != (const ShowCompactResponseElement &rhs) const { diff --git metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponseElement.java metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponseElement.java index 91918fd..212ecd6 100644 --- metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponseElement.java +++ metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ShowCompactResponseElement.java @@ -49,6 +49,7 @@ private static final org.apache.thrift.protocol.TField HIGHTEST_TXN_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("hightestTxnId", org.apache.thrift.protocol.TType.I64, (short)9); private static final org.apache.thrift.protocol.TField META_INFO_FIELD_DESC = new org.apache.thrift.protocol.TField("metaInfo", org.apache.thrift.protocol.TType.STRING, (short)10); private static final org.apache.thrift.protocol.TField END_TIME_FIELD_DESC = new org.apache.thrift.protocol.TField("endTime", org.apache.thrift.protocol.TType.I64, (short)11); + private static final org.apache.thrift.protocol.TField HADOOP_JOB_ID_FIELD_DESC = new org.apache.thrift.protocol.TField("hadoopJobId", org.apache.thrift.protocol.TType.STRING, (short)12); private static final Map, SchemeFactory> schemes = new HashMap, SchemeFactory>(); static { @@ -67,6 +68,7 @@ private long hightestTxnId; // optional private String metaInfo; // optional private long endTime; // optional + private String hadoopJobId; // optional /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { @@ -84,7 +86,8 @@ RUN_AS((short)8, "runAs"), HIGHTEST_TXN_ID((short)9, "hightestTxnId"), META_INFO((short)10, "metaInfo"), - END_TIME((short)11, "endTime"); + END_TIME((short)11, "endTime"), + HADOOP_JOB_ID((short)12, "hadoopJobId"); private static final Map byName = new HashMap(); @@ -121,6 +124,8 @@ public static _Fields findByThriftId(int fieldId) { return META_INFO; case 11: // END_TIME return END_TIME; + case 12: // HADOOP_JOB_ID + return HADOOP_JOB_ID; default: return null; } @@ -165,7 +170,7 @@ public String getFieldName() { private static final int __HIGHTESTTXNID_ISSET_ID = 1; private static final int __ENDTIME_ISSET_ID = 2; private byte __isset_bitfield = 0; - private static final _Fields optionals[] = {_Fields.PARTITIONNAME,_Fields.WORKERID,_Fields.START,_Fields.RUN_AS,_Fields.HIGHTEST_TXN_ID,_Fields.META_INFO,_Fields.END_TIME}; + private static final _Fields optionals[] = {_Fields.PARTITIONNAME,_Fields.WORKERID,_Fields.START,_Fields.RUN_AS,_Fields.HIGHTEST_TXN_ID,_Fields.META_INFO,_Fields.END_TIME,_Fields.HADOOP_JOB_ID}; public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); @@ -191,11 +196,15 @@ public String getFieldName() { new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); tmpMap.put(_Fields.END_TIME, new org.apache.thrift.meta_data.FieldMetaData("endTime", org.apache.thrift.TFieldRequirementType.OPTIONAL, new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I64))); + tmpMap.put(_Fields.HADOOP_JOB_ID, new org.apache.thrift.meta_data.FieldMetaData("hadoopJobId", org.apache.thrift.TFieldRequirementType.OPTIONAL, + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); metaDataMap = Collections.unmodifiableMap(tmpMap); org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(ShowCompactResponseElement.class, metaDataMap); } public ShowCompactResponseElement() { + this.hadoopJobId = "None"; + } public ShowCompactResponseElement( @@ -243,6 +252,9 @@ public ShowCompactResponseElement(ShowCompactResponseElement other) { this.metaInfo = other.metaInfo; } this.endTime = other.endTime; + if (other.isSetHadoopJobId()) { + this.hadoopJobId = other.hadoopJobId; + } } public ShowCompactResponseElement deepCopy() { @@ -265,6 +277,8 @@ public void clear() { this.metaInfo = null; setEndTimeIsSet(false); this.endTime = 0; + this.hadoopJobId = "None"; + } public String getDbname() { @@ -525,6 +539,29 @@ public void setEndTimeIsSet(boolean value) { __isset_bitfield = EncodingUtils.setBit(__isset_bitfield, __ENDTIME_ISSET_ID, value); } + public String getHadoopJobId() { + return this.hadoopJobId; + } + + public void setHadoopJobId(String hadoopJobId) { + this.hadoopJobId = hadoopJobId; + } + + public void unsetHadoopJobId() { + this.hadoopJobId = null; + } + + /** Returns true if field hadoopJobId is set (has been assigned a value) and false otherwise */ + public boolean isSetHadoopJobId() { + return this.hadoopJobId != null; + } + + public void setHadoopJobIdIsSet(boolean value) { + if (!value) { + this.hadoopJobId = null; + } + } + public void setFieldValue(_Fields field, Object value) { switch (field) { case DBNAME: @@ -615,6 +652,14 @@ public void setFieldValue(_Fields field, Object value) { } break; + case HADOOP_JOB_ID: + if (value == null) { + unsetHadoopJobId(); + } else { + setHadoopJobId((String)value); + } + break; + } } @@ -653,6 +698,9 @@ public Object getFieldValue(_Fields field) { case END_TIME: return getEndTime(); + case HADOOP_JOB_ID: + return getHadoopJobId(); + } throw new IllegalStateException(); } @@ -686,6 +734,8 @@ public boolean isSet(_Fields field) { return isSetMetaInfo(); case END_TIME: return isSetEndTime(); + case HADOOP_JOB_ID: + return isSetHadoopJobId(); } throw new IllegalStateException(); } @@ -802,6 +852,15 @@ public boolean equals(ShowCompactResponseElement that) { return false; } + boolean this_present_hadoopJobId = true && this.isSetHadoopJobId(); + boolean that_present_hadoopJobId = true && that.isSetHadoopJobId(); + if (this_present_hadoopJobId || that_present_hadoopJobId) { + if (!(this_present_hadoopJobId && that_present_hadoopJobId)) + return false; + if (!this.hadoopJobId.equals(that.hadoopJobId)) + return false; + } + return true; } @@ -864,6 +923,11 @@ public int hashCode() { if (present_endTime) list.add(endTime); + boolean present_hadoopJobId = true && (isSetHadoopJobId()); + list.add(present_hadoopJobId); + if (present_hadoopJobId) + list.add(hadoopJobId); + return list.hashCode(); } @@ -985,6 +1049,16 @@ public int compareTo(ShowCompactResponseElement other) { return lastComparison; } } + lastComparison = Boolean.valueOf(isSetHadoopJobId()).compareTo(other.isSetHadoopJobId()); + if (lastComparison != 0) { + return lastComparison; + } + if (isSetHadoopJobId()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.hadoopJobId, other.hadoopJobId); + if (lastComparison != 0) { + return lastComparison; + } + } return 0; } @@ -1094,6 +1168,16 @@ public String toString() { sb.append(this.endTime); first = false; } + if (isSetHadoopJobId()) { + if (!first) sb.append(", "); + sb.append("hadoopJobId:"); + if (this.hadoopJobId == null) { + sb.append("null"); + } else { + sb.append(this.hadoopJobId); + } + first = false; + } sb.append(")"); return sb.toString(); } @@ -1243,6 +1327,14 @@ public void read(org.apache.thrift.protocol.TProtocol iprot, ShowCompactResponse org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } break; + case 12: // HADOOP_JOB_ID + if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { + struct.hadoopJobId = iprot.readString(); + struct.setHadoopJobIdIsSet(true); + } else { + org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); + } + break; default: org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -1319,6 +1411,13 @@ public void write(org.apache.thrift.protocol.TProtocol oprot, ShowCompactRespons oprot.writeI64(struct.endTime); oprot.writeFieldEnd(); } + if (struct.hadoopJobId != null) { + if (struct.isSetHadoopJobId()) { + oprot.writeFieldBegin(HADOOP_JOB_ID_FIELD_DESC); + oprot.writeString(struct.hadoopJobId); + oprot.writeFieldEnd(); + } + } oprot.writeFieldStop(); oprot.writeStructEnd(); } @@ -1362,7 +1461,10 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponse if (struct.isSetEndTime()) { optionals.set(6); } - oprot.writeBitSet(optionals, 7); + if (struct.isSetHadoopJobId()) { + optionals.set(7); + } + oprot.writeBitSet(optionals, 8); if (struct.isSetPartitionname()) { oprot.writeString(struct.partitionname); } @@ -1384,6 +1486,9 @@ public void write(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponse if (struct.isSetEndTime()) { oprot.writeI64(struct.endTime); } + if (struct.isSetHadoopJobId()) { + oprot.writeString(struct.hadoopJobId); + } } @Override @@ -1397,7 +1502,7 @@ public void read(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponseE struct.setTypeIsSet(true); struct.state = iprot.readString(); struct.setStateIsSet(true); - BitSet incoming = iprot.readBitSet(7); + BitSet incoming = iprot.readBitSet(8); if (incoming.get(0)) { struct.partitionname = iprot.readString(); struct.setPartitionnameIsSet(true); @@ -1426,6 +1531,10 @@ public void read(org.apache.thrift.protocol.TProtocol prot, ShowCompactResponseE struct.endTime = iprot.readI64(); struct.setEndTimeIsSet(true); } + if (incoming.get(7)) { + struct.hadoopJobId = iprot.readString(); + struct.setHadoopJobIdIsSet(true); + } } } diff --git metastore/src/gen/thrift/gen-php/metastore/Types.php metastore/src/gen/thrift/gen-php/metastore/Types.php index 5873ba8..ea8f0f3 100644 --- metastore/src/gen/thrift/gen-php/metastore/Types.php +++ metastore/src/gen/thrift/gen-php/metastore/Types.php @@ -12511,6 +12511,10 @@ class ShowCompactResponseElement { * @var int */ public $endTime = null; + /** + * @var string + */ + public $hadoopJobId = "None"; public function __construct($vals=null) { if (!isset(self::$_TSPEC)) { @@ -12559,6 +12563,10 @@ class ShowCompactResponseElement { 'var' => 'endTime', 'type' => TType::I64, ), + 12 => array( + 'var' => 'hadoopJobId', + 'type' => TType::STRING, + ), ); } if (is_array($vals)) { @@ -12595,6 +12603,9 @@ class ShowCompactResponseElement { if (isset($vals['endTime'])) { $this->endTime = $vals['endTime']; } + if (isset($vals['hadoopJobId'])) { + $this->hadoopJobId = $vals['hadoopJobId']; + } } } @@ -12694,6 +12705,13 @@ class ShowCompactResponseElement { $xfer += $input->skip($ftype); } break; + case 12: + if ($ftype == TType::STRING) { + $xfer += $input->readString($this->hadoopJobId); + } else { + $xfer += $input->skip($ftype); + } + break; default: $xfer += $input->skip($ftype); break; @@ -12762,6 +12780,11 @@ class ShowCompactResponseElement { $xfer += $output->writeI64($this->endTime); $xfer += $output->writeFieldEnd(); } + if ($this->hadoopJobId !== null) { + $xfer += $output->writeFieldBegin('hadoopJobId', TType::STRING, 12); + $xfer += $output->writeString($this->hadoopJobId); + $xfer += $output->writeFieldEnd(); + } $xfer += $output->writeFieldStop(); $xfer += $output->writeStructEnd(); return $xfer; diff --git metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py index 479bf92..a58cb58 100644 --- metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py +++ metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py @@ -8737,6 +8737,7 @@ class ShowCompactResponseElement: - hightestTxnId - metaInfo - endTime + - hadoopJobId """ thrift_spec = ( @@ -8752,9 +8753,10 @@ class ShowCompactResponseElement: (9, TType.I64, 'hightestTxnId', None, None, ), # 9 (10, TType.STRING, 'metaInfo', None, None, ), # 10 (11, TType.I64, 'endTime', None, None, ), # 11 + (12, TType.STRING, 'hadoopJobId', None, "None", ), # 12 ) - def __init__(self, dbname=None, tablename=None, partitionname=None, type=None, state=None, workerid=None, start=None, runAs=None, hightestTxnId=None, metaInfo=None, endTime=None,): + def __init__(self, dbname=None, tablename=None, partitionname=None, type=None, state=None, workerid=None, start=None, runAs=None, hightestTxnId=None, metaInfo=None, endTime=None, hadoopJobId=thrift_spec[12][4],): self.dbname = dbname self.tablename = tablename self.partitionname = partitionname @@ -8766,6 +8768,7 @@ def __init__(self, dbname=None, tablename=None, partitionname=None, type=None, s self.hightestTxnId = hightestTxnId self.metaInfo = metaInfo self.endTime = endTime + self.hadoopJobId = hadoopJobId def read(self, iprot): if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None: @@ -8831,6 +8834,11 @@ def read(self, iprot): self.endTime = iprot.readI64() else: iprot.skip(ftype) + elif fid == 12: + if ftype == TType.STRING: + self.hadoopJobId = iprot.readString() + else: + iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() @@ -8885,6 +8893,10 @@ def write(self, oprot): oprot.writeFieldBegin('endTime', TType.I64, 11) oprot.writeI64(self.endTime) oprot.writeFieldEnd() + if self.hadoopJobId is not None: + oprot.writeFieldBegin('hadoopJobId', TType.STRING, 12) + oprot.writeString(self.hadoopJobId) + oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() @@ -8913,6 +8925,7 @@ def __hash__(self): value = (value * 31) ^ hash(self.hightestTxnId) value = (value * 31) ^ hash(self.metaInfo) value = (value * 31) ^ hash(self.endTime) + value = (value * 31) ^ hash(self.hadoopJobId) return value def __repr__(self): diff --git metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb index 0e041c5..725a264 100644 --- metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb +++ metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb @@ -2006,6 +2006,7 @@ class ShowCompactResponseElement HIGHTESTTXNID = 9 METAINFO = 10 ENDTIME = 11 + HADOOPJOBID = 12 FIELDS = { DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbname'}, @@ -2018,7 +2019,8 @@ class ShowCompactResponseElement RUNAS => {:type => ::Thrift::Types::STRING, :name => 'runAs', :optional => true}, HIGHTESTTXNID => {:type => ::Thrift::Types::I64, :name => 'hightestTxnId', :optional => true}, METAINFO => {:type => ::Thrift::Types::STRING, :name => 'metaInfo', :optional => true}, - ENDTIME => {:type => ::Thrift::Types::I64, :name => 'endTime', :optional => true} + ENDTIME => {:type => ::Thrift::Types::I64, :name => 'endTime', :optional => true}, + HADOOPJOBID => {:type => ::Thrift::Types::STRING, :name => 'hadoopJobId', :default => %q"None", :optional => true} } def struct_fields; FIELDS; end