Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
3.0.0
-
None
Description
Resource plan with multiple pools and trigger set to move job across those pools seems to be failing to do so
Resource plan:
1: jdbc:hive2://ctr-e137-1514896590304-51538-> show resource plan plan_2; INFO : Compiling command(queryId=hive_20180202220823_2fb8bca7-5b7a-48cf-8ff9-8d5f3548d334): show resource plan plan_2 INFO : Semantic Analysis Completed INFO : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:line, type:string, comment:from deserializer)], properties:null) INFO : Completed compiling command(queryId=hive_20180202220823_2fb8bca7-5b7a-48cf-8ff9-8d5f3548d334); Time taken: 0.008 seconds INFO : Executing command(queryId=hive_20180202220823_2fb8bca7-5b7a-48cf-8ff9-8d5f3548d334): show resource plan plan_2 INFO : Starting task [Stage-0:DDL] in serial mode INFO : Completed executing command(queryId=hive_20180202220823_2fb8bca7-5b7a-48cf-8ff9-8d5f3548d334); Time taken: 0.196 seconds INFO : OK +----------------------------------------------------+ | line | +----------------------------------------------------+ | plan_2[status=ACTIVE,parallelism=null,defaultPool=pool2] | | + pool2[allocFraction=0.5,schedulingPolicy=default,parallelism=3] | | | trigger too_large_write_triger: if (HDFS_BYTES_WRITTEN > 10kb) { MOVE TO pool1 } | | | mapped for default | | + pool1[allocFraction=0.3,schedulingPolicy=default,parallelism=5] | | | trigger slow_pool_trigger: if (ELAPSED_TIME > 30000) { MOVE TO pool3 } | | + pool3[allocFraction=0.2,schedulingPolicy=default,parallelism=3] | | + default[allocFraction=0.0,schedulingPolicy=null,parallelism=4] | +----------------------------------------------------+ 8 rows selected (0.25 seconds)
Workload Manager Events Summary from query run:
INFO : { "queryId" : "hive_20180202213425_9633d7af-4242-4e95-a391-2cd3823e3eac", "queryStartTime" : 1517607265395, "queryEndTime" : 1517607321648, "queryCompleted" : true, "queryWmEvents" : [ { "wmTezSessionInfo" : { "sessionId" : "21f8a4ab-511e-4828-a2dd-1d5f2932c492", "poolName" : "pool2", "clusterPercent" : 50.0 }, "eventStartTimestamp" : 1517607269660, "eventEndTimestamp" : 1517607269661, "eventType" : "GET", "elapsedTime" : 1 }, { "wmTezSessionInfo" : { "sessionId" : "21f8a4ab-511e-4828-a2dd-1d5f2932c492", "poolName" : null, "clusterPercent" : 0.0 }, "eventStartTimestamp" : 1517607321663, "eventEndTimestamp" : 1517607321663, "eventType" : "RETURN", "elapsedTime" : 0 } ], "appliedTriggers" : [ { "name" : "too_large_write_triger", "expression" : { "counterLimit" : { "limit" : 10240, "name" : "HDFS_BYTES_WRITTEN" }, "predicate" : "GREATER_THAN" }, "action" : { "type" : "MOVE_TO_POOL", "poolName" : "pool1" }, "violationMsg" : null } ], "subscribedCounters" : [ "HDFS_BYTES_WRITTEN" ], "currentCounters" : { "HDFS_BYTES_WRITTEN" : 33306829 }, "elapsedTime" : 56284 }
From the Workload Manager Event Summary it could seen that the 'MOVE' event didn't happen though the limit for counter (10240) HDFS_BYTES_WRITTEN was exceeded