Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-14220 Build and test Spark against Scala 2.12
  3. SPARK-25029

Scala 2.12 issues: TaskNotSerializable and Janino "Two non-abstract methods ..." errors

    XMLWordPrintableJSON

Details

    • Sub-task
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • 2.4.0
    • 2.4.0
    • Build
    • None
    • Hide
      Release Notes text:

      Because of differences in how Scala 2.12 serializes closures, you may find that some jobs now fail with "TaskNotSerializableException". This typically occurs when a closure actually captures references to objects it does not need, and which now can't be automatically removed by the closure cleaner. The solution is to change the declaration of the closure to avoid this capture. Often, moving it outside a method or declaring class can fix it.
      Show
      Release Notes text: Because of differences in how Scala 2.12 serializes closures, you may find that some jobs now fail with "TaskNotSerializableException". This typically occurs when a closure actually captures references to objects it does not need, and which now can't be automatically removed by the closure cleaner. The solution is to change the declaration of the closure to avoid this capture. Often, moving it outside a method or declaring class can fix it.

    Description

      We actually still have some test failures in the Scala 2.12 build. There seem to be two types. First are that some tests fail with "TaskNotSerializable" because some code construct now captures a reference to scalatest's AssertionHelper. Example:

      - LegacyAccumulatorWrapper with AccumulatorParam that has no equals/hashCode *** FAILED *** java.io.NotSerializableException: org.scalatest.Assertions$AssertionsHelper Serialization stack: - object not serializable (class: org.scalatest.Assertions$AssertionsHelper, value: org.scalatest.Assertions$AssertionsHelper@3bc5fc8f)

      These seem generally easy to fix by tweaking the test code. It's not clear if something about closure cleaning in 2.12 could be improved to detect this situation automatically; given that yet only a handful of tests fail for this reason, it's unlikely to be a systemic problem.

       

      The other error is curioser. Janino fails to compile generate code in many cases with errors like:

      - encode/decode for seq of string: List(abc, xyz) *** FAILED ***
      java.lang.RuntimeException: Error while encoding: org.codehaus.janino.InternalCompilerException: failed to compile: org.codehaus.janino.InternalCompilerException: Compiling "GeneratedClass": Two non-abstract methods "public int scala.collection.TraversableOnce.size()" have the same parameter types, declaring type and return type

       

      I include the full generated code that failed in one case below. There is no size() in the generated code. It's got to be down to some difference in Scala 2.12, potentially even a Janino problem.

       

      Caused by: org.codehaus.janino.InternalCompilerException: Compiling "GeneratedClass": Two non-abstract methods "public int scala.collection.TraversableOnce.size()" have the same parameter types, declaring type and return type
      
      at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:361)
      
      at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:234)
      
      at org.codehaus.janino.SimpleCompiler.compileToClassLoader(SimpleCompiler.java:446)
      
      at org.codehaus.janino.ClassBodyEvaluator.compileToClass(ClassBodyEvaluator.java:313)
      
      at org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:235)
      
      at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:204)
      
      at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:80)
      
      at org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.org$apache$spark$sql$catalyst$expressions$codegen$CodeGenerator$$doCompile(CodeGenerator.scala:1342)
      
      ... 30 more
      
      Caused by: org.codehaus.janino.InternalCompilerException: Two non-abstract methods "public int scala.collection.TraversableOnce.size()" have the same parameter types, declaring type and return type
      
      at org.codehaus.janino.UnitCompiler.findMostSpecificIInvocable(UnitCompiler.java:9112)
      
      at org.codehaus.janino.UnitCompiler.findMostSpecificIInvocable(UnitCompiler.java:8888)
      
      at org.codehaus.janino.UnitCompiler.findIMethod(UnitCompiler.java:8770)
      
      at org.codehaus.janino.UnitCompiler.findIMethod(UnitCompiler.java:8672)
      
      at org.codehaus.janino.UnitCompiler.compileGet2(UnitCompiler.java:4737)
      
      at org.codehaus.janino.UnitCompiler.access$8300(UnitCompiler.java:212)
      
      at org.codehaus.janino.UnitCompiler$12.visitMethodInvocation(UnitCompiler.java:4097)
      
      at org.codehaus.janino.UnitCompiler$12.visitMethodInvocation(UnitCompiler.java:4070)
      
      at org.codehaus.janino.Java$MethodInvocation.accept(Java.java:4902)
      
      at org.codehaus.janino.UnitCompiler.compileGet(UnitCompiler.java:4070)
      
      at org.codehaus.janino.UnitCompiler.compileGetValue(UnitCompiler.java:5253)
      
      at org.codehaus.janino.UnitCompiler.compileGet2(UnitCompiler.java:4391)
      
      at org.codehaus.janino.UnitCompiler.access$8000(UnitCompiler.java:212)
      
      at org.codehaus.janino.UnitCompiler$12.visitConditionalExpression(UnitCompiler.java:4094)
      
      at org.codehaus.janino.UnitCompiler$12.visitConditionalExpression(UnitCompiler.java:4070)
      
      at org.codehaus.janino.Java$ConditionalExpression.accept(Java.java:4344)
      
      at org.codehaus.janino.UnitCompiler.compileGet(UnitCompiler.java:4070)
      
      at org.codehaus.janino.UnitCompiler.compileGetValue(UnitCompiler.java:5253)
      
      at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2559)
      
      at org.codehaus.janino.UnitCompiler.access$2700(UnitCompiler.java:212)
      
      at org.codehaus.janino.UnitCompiler$6.visitLocalVariableDeclarationStatement(UnitCompiler.java:1482)
      
      at org.codehaus.janino.UnitCompiler$6.visitLocalVariableDeclarationStatement(UnitCompiler.java:1466)
      
      at org.codehaus.janino.Java$LocalVariableDeclarationStatement.accept(Java.java:3351)
      ...

       

       

      /* 001 */ public java.lang.Object generate(Object[] references) {
      /* 002 */ return new SpecificUnsafeProjection(references);
      /* 003 */ }
      /* 004 */
      /* 005 */ class SpecificUnsafeProjection extends org.apache.spark.sql.catalyst.expressions.UnsafeProjection {
      /* 006 */
      /* 007 */ private Object[] references;
      /* 008 */ private scala.collection.immutable.Map MapObjects_loopValue140;
      /* 009 */ private int ExternalMapToCatalyst_key8;
      /* 010 */ private int ExternalMapToCatalyst_value8;
      /* 011 */ private boolean globalIsNull_0;
      /* 012 */ private boolean MapObjects_loopIsNull140;
      /* 013 */ private boolean globalIsNull_1;
      /* 014 */ private boolean globalIsNull_2;
      /* 015 */ private boolean globalIsNull_3;
      /* 016 */ private org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter[] mutableStateArray_0 = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter[1];
      /* 017 */ private org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter[] mutableStateArray_1 = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter[6];
      /* 018 */
      /* 019 */ public SpecificUnsafeProjection(Object[] references) {
      /* 020 */ this.references = references;
      /* 021 */
      /* 022 */ mutableStateArray_0[0] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter(2, 64);
      /* 023 */ mutableStateArray_1[0] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter(mutableStateArray_0[0], 8);
      /* 024 */ mutableStateArray_1[1] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter(mutableStateArray_1[0], 4);
      /* 025 */ mutableStateArray_1[2] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter(mutableStateArray_1[0], 4);
      /* 026 */ mutableStateArray_1[3] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter(mutableStateArray_0[0], 8);
      /* 027 */ mutableStateArray_1[4] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter(mutableStateArray_1[3], 4);
      /* 028 */ mutableStateArray_1[5] = new org.apache.spark.sql.catalyst.expressions.codegen.UnsafeArrayWriter(mutableStateArray_1[3], 4);
      /* 029 */
      /* 030 */ }
      /* 031 */
      /* 032 */ public void initialize(int partitionIndex) {
      /* 033 */
      /* 034 */ }
      /* 035 */
      /* 036 */ // Scala.Function1 need this
      /* 037 */ public java.lang.Object apply(java.lang.Object row) {
      /* 038 */ return apply((InternalRow) row);
      /* 039 */ }
      /* 040 */
      /* 041 */ public UnsafeRow apply(InternalRow i) {
      /* 042 */ mutableStateArray_0[0].reset();
      /* 043 */
      /* 044 */
      /* 045 */ mutableStateArray_0[0].zeroOutNullBytes();
      /* 046 */ writeFields_0_0(i);
      /* 047 */ writeFields_0_1(i);
      /* 048 */ return (mutableStateArray_0[0].getRow());
      /* 049 */ }
      /* 050 */
      /* 051 */
      /* 052 */ private void writeFields_0_1(InternalRow i) {
      /* 053 */
      /* 054 */ ArrayData value_9 = MapObjects_1(i);
      /* 055 */ if (globalIsNull_3) {
      /* 056 */ mutableStateArray_0[0].setNullAt(1);
      /* 057 */ } else {
      /* 058 */ // Remember the current cursor so that we can calculate how many bytes are
      /* 059 */ // written later.
      /* 060 */ final int previousCursor_2 = mutableStateArray_0[0].cursor();
      /* 061 */
      /* 062 */ final ArrayData tmpInput_4 = value_9;
      /* 063 */ if (tmpInput_4 instanceof UnsafeArrayData) {
      /* 064 */ mutableStateArray_0[0].write((UnsafeArrayData) tmpInput_4);
      /* 065 */ } else {
      /* 066 */ final int numElements_3 = tmpInput_4.numElements();
      /* 067 */ mutableStateArray_1[3].initialize(numElements_3);
      /* 068 */
      /* 069 */ for (int index_5 = 0; index_5 < numElements_3; index_5++) {
      /* 070 */ if (tmpInput_4.isNullAt(index_5)) {
      /* 071 */ mutableStateArray_1[3].setNull8Bytes(index_5);
      /* 072 */ } else {
      /* 073 */
      /* 074 */ final MapData tmpInput_5 = tmpInput_4.getMap(index_5);
      /* 075 */ if (tmpInput_5 instanceof UnsafeMapData) {
      /* 076 */ mutableStateArray_1[3].write(index_5, (UnsafeMapData) tmpInput_5);
      /* 077 */ } else {
      /* 078 */ // Remember the current cursor so that we can calculate how many bytes are
      /* 079 */ // written later.
      /* 080 */ final int previousCursor_3 = mutableStateArray_1[3].cursor();
      /* 081 */
      /* 082 */ // preserve 8 bytes to write the key array numBytes later.
      /* 083 */ mutableStateArray_1[3].grow(8);
      /* 084 */ mutableStateArray_1[3].increaseCursor(8);
      /* 085 */
      /* 086 */ // Remember the current cursor so that we can write numBytes of key array later.
      /* 087 */ final int tmpCursor_1 = mutableStateArray_1[3].cursor();
      /* 088 */
      /* 089 */
      /* 090 */ final ArrayData tmpInput_6 = tmpInput_5.keyArray();
      /* 091 */ if (tmpInput_6 instanceof UnsafeArrayData) {
      /* 092 */ mutableStateArray_1[3].write((UnsafeArrayData) tmpInput_6);
      /* 093 */ } else {
      /* 094 */ final int numElements_4 = tmpInput_6.numElements();
      /* 095 */ mutableStateArray_1[4].initialize(numElements_4);
      /* 096 */
      /* 097 */ for (int index_6 = 0; index_6 < numElements_4; index_6++) {
      /* 098 */ if (tmpInput_6.isNullAt(index_6)) {
      /* 099 */ mutableStateArray_1[4].setNull4Bytes(index_6);
      /* 100 */ } else {
      /* 101 */ mutableStateArray_1[4].write(index_6, tmpInput_6.getInt(index_6));
      /* 102 */ }
      /* 103 */ }
      /* 104 */ }
      /* 105 */
      /* 106 */
      /* 107 */ // Write the numBytes of key array into the first 8 bytes.
      /* 108 */ Platform.putLong(
      /* 109 */ mutableStateArray_1[3].getBuffer(),
      /* 110 */ tmpCursor_1 - 8,
      /* 111 */ mutableStateArray_1[3].cursor() - tmpCursor_1);
      /* 112 */
      /* 113 */
      /* 114 */ final ArrayData tmpInput_7 = tmpInput_5.valueArray();
      /* 115 */ if (tmpInput_7 instanceof UnsafeArrayData) {
      /* 116 */ mutableStateArray_1[3].write((UnsafeArrayData) tmpInput_7);
      /* 117 */ } else {
      /* 118 */ final int numElements_5 = tmpInput_7.numElements();
      /* 119 */ mutableStateArray_1[5].initialize(numElements_5);
      /* 120 */
      /* 121 */ for (int index_7 = 0; index_7 < numElements_5; index_7++) {
      /* 122 */ if (tmpInput_7.isNullAt(index_7)) {
      /* 123 */ mutableStateArray_1[5].setNull4Bytes(index_7);
      /* 124 */ } else {
      /* 125 */ mutableStateArray_1[5].write(index_7, tmpInput_7.getInt(index_7));
      /* 126 */ }
      /* 127 */ }
      /* 128 */ }
      /* 129 */
      /* 130 */ mutableStateArray_1[3].setOffsetAndSizeFromPreviousCursor(index_5, previousCursor_3);
      /* 131 */ }
      /* 132 */
      /* 133 */ }
      /* 134 */ }
      /* 135 */ }
      /* 136 */
      /* 137 */ mutableStateArray_0[0].setOffsetAndSizeFromPreviousCursor(1, previousCursor_2);
      /* 138 */ }
      /* 139 */
      /* 140 */ }
      /* 141 */
      /* 142 */
      /* 143 */ private MapData ExternalMapToCatalyst_1(InternalRow i) {
      /* 144 */ MapData value_7 = null;
      /* 145 */ if (!MapObjects_loopIsNull140) {
      /* 146 */ final int length_1 = MapObjects_loopValue140.size();
      /* 147 */ final Object[] convertedKeys_1 = new Object[length_1];
      /* 148 */ final Object[] convertedValues_1 = new Object[length_1];
      /* 149 */ int index_1 = 0;
      /* 150 */ final scala.collection.Iterator entries_1 = MapObjects_loopValue140.iterator();
      /* 151 */ while(entries_1.hasNext()) {
      /* 152 */
      /* 153 */ final scala.Tuple2 entry_1 = (scala.Tuple2) entries_1.next();
      /* 154 */ ExternalMapToCatalyst_key8 = (Integer) entry_1._1();
      /* 155 */ ExternalMapToCatalyst_value8 = (Integer) entry_1._2();
      /* 156 */
      /* 157 */
      /* 158 */
      /* 159 */
      /* 160 */
      /* 161 */ if (false) {
      /* 162 */ throw new RuntimeException("Cannot use null as map key!");
      /* 163 */ } else {
      /* 164 */ convertedKeys_1[index_1] = (Integer) ExternalMapToCatalyst_key8;
      /* 165 */ }
      /* 166 */
      /* 167 */
      /* 168 */ if (false) {
      /* 169 */ convertedValues_1[index_1] = null;
      /* 170 */ } else {
      /* 171 */ convertedValues_1[index_1] = (Integer) ExternalMapToCatalyst_value8;
      /* 172 */ }
      /* 173 */
      /* 174 */ index_1++;
      /* 175 */ }
      /* 176 */
      /* 177 */ value_7 = new org.apache.spark.sql.catalyst.util.ArrayBasedMapData(new org.apache.spark.sql.catalyst.util.GenericArrayData(convertedKeys_1), new org.apache.spark.sql.catalyst.util.GenericArrayData(convertedValues_1));
      /* 178 */ }
      /* 179 */ globalIsNull_2 = MapObjects_loopIsNull140;
      /* 180 */ return value_7;
      /* 181 */ }
      /* 182 */
      /* 183 */
      /* 184 */ private ArrayData MapObjects_1(InternalRow i) {
      /* 185 */ boolean isNull_4 = i.isNullAt(0);
      /* 186 */ scala.collection.immutable.Map[] value_6 = isNull_4 ?
      /* 187 */ null : ((scala.collection.immutable.Map[])i.get(0, null));
      /* 188 */ ArrayData value_5 = null;
      /* 189 */
      /* 190 */ if (!isNull_4) {
      /* 191 */
      /* 192 */ int dataLength_1 = value_6.length;
      /* 193 */
      /* 194 */ MapData[] convertedArray_1 = null;
      /* 195 */ convertedArray_1 = new MapData[dataLength_1];
      /* 196 */
      /* 197 */
      /* 198 */ int loopIndex_1 = 0;
      /* 199 */
      /* 200 */ while (loopIndex_1 < dataLength_1) {
      /* 201 */ MapObjects_loopValue140 = (scala.collection.immutable.Map) (value_6[loopIndex_1]);
      /* 202 */ MapObjects_loopIsNull140 = MapObjects_loopValue140 == null;
      /* 203 */
      /* 204 */ MapData value_8 = ExternalMapToCatalyst_1(i);
      /* 205 */ if (globalIsNull_2) {
      /* 206 */ convertedArray_1[loopIndex_1] = null;
      /* 207 */ } else {
      /* 208 */ convertedArray_1[loopIndex_1] = value_8 instanceof UnsafeMapData? value_8.copy() : value_8;
      /* 209 */ }
      /* 210 */
      /* 211 */ loopIndex_1 += 1;
      /* 212 */ }
      /* 213 */
      /* 214 */ value_5 = new org.apache.spark.sql.catalyst.util.GenericArrayData(convertedArray_1);
      /* 215 */ }
      /* 216 */ globalIsNull_3 = isNull_4;
      /* 217 */ return value_5;
      /* 218 */ }
      /* 219 */
      /* 220 */
      /* 221 */ private void writeFields_0_0(InternalRow i) {
      /* 222 */
      /* 223 */ ArrayData value_4 = MapObjects_0(i);
      /* 224 */ if (globalIsNull_1) {
      /* 225 */ mutableStateArray_0[0].setNullAt(0);
      /* 226 */ } else {
      /* 227 */ // Remember the current cursor so that we can calculate how many bytes are
      /* 228 */ // written later.
      /* 229 */ final int previousCursor_0 = mutableStateArray_0[0].cursor();
      /* 230 */
      /* 231 */ final ArrayData tmpInput_0 = value_4;
      /* 232 */ if (tmpInput_0 instanceof UnsafeArrayData) {
      /* 233 */ mutableStateArray_0[0].write((UnsafeArrayData) tmpInput_0);
      /* 234 */ } else {
      /* 235 */ final int numElements_0 = tmpInput_0.numElements();
      /* 236 */ mutableStateArray_1[0].initialize(numElements_0);
      /* 237 */
      /* 238 */ for (int index_2 = 0; index_2 < numElements_0; index_2++) {
      /* 239 */ if (tmpInput_0.isNullAt(index_2)) {
      /* 240 */ mutableStateArray_1[0].setNull8Bytes(index_2);
      /* 241 */ } else {
      /* 242 */
      /* 243 */ final MapData tmpInput_1 = tmpInput_0.getMap(index_2);
      /* 244 */ if (tmpInput_1 instanceof UnsafeMapData) {
      /* 245 */ mutableStateArray_1[0].write(index_2, (UnsafeMapData) tmpInput_1);
      /* 246 */ } else {
      /* 247 */ // Remember the current cursor so that we can calculate how many bytes are
      /* 248 */ // written later.
      /* 249 */ final int previousCursor_1 = mutableStateArray_1[0].cursor();
      /* 250 */
      /* 251 */ // preserve 8 bytes to write the key array numBytes later.
      /* 252 */ mutableStateArray_1[0].grow(8);
      /* 253 */ mutableStateArray_1[0].increaseCursor(8);
      /* 254 */
      /* 255 */ // Remember the current cursor so that we can write numBytes of key array later.
      /* 256 */ final int tmpCursor_0 = mutableStateArray_1[0].cursor();
      /* 257 */
      /* 258 */
      /* 259 */ final ArrayData tmpInput_2 = tmpInput_1.keyArray();
      /* 260 */ if (tmpInput_2 instanceof UnsafeArrayData) {
      /* 261 */ mutableStateArray_1[0].write((UnsafeArrayData) tmpInput_2);
      /* 262 */ } else {
      /* 263 */ final int numElements_1 = tmpInput_2.numElements();
      /* 264 */ mutableStateArray_1[1].initialize(numElements_1);
      /* 265 */
      /* 266 */ for (int index_3 = 0; index_3 < numElements_1; index_3++) {
      /* 267 */ if (tmpInput_2.isNullAt(index_3)) {
      /* 268 */ mutableStateArray_1[1].setNull4Bytes(index_3);
      /* 269 */ } else {
      /* 270 */ mutableStateArray_1[1].write(index_3, tmpInput_2.getInt(index_3));
      /* 271 */ }
      /* 272 */ }
      /* 273 */ }
      /* 274 */
      /* 275 */
      /* 276 */ // Write the numBytes of key array into the first 8 bytes.
      /* 277 */ Platform.putLong(
      /* 278 */ mutableStateArray_1[0].getBuffer(),
      /* 279 */ tmpCursor_0 - 8,
      /* 280 */ mutableStateArray_1[0].cursor() - tmpCursor_0);
      /* 281 */
      /* 282 */
      /* 283 */ final ArrayData tmpInput_3 = tmpInput_1.valueArray();
      /* 284 */ if (tmpInput_3 instanceof UnsafeArrayData) {
      /* 285 */ mutableStateArray_1[0].write((UnsafeArrayData) tmpInput_3);
      /* 286 */ } else {
      /* 287 */ final int numElements_2 = tmpInput_3.numElements();
      /* 288 */ mutableStateArray_1[2].initialize(numElements_2);
      /* 289 */
      /* 290 */ for (int index_4 = 0; index_4 < numElements_2; index_4++) {
      /* 291 */ if (tmpInput_3.isNullAt(index_4)) {
      /* 292 */ mutableStateArray_1[2].setNull4Bytes(index_4);
      /* 293 */ } else {
      /* 294 */ mutableStateArray_1[2].write(index_4, tmpInput_3.getInt(index_4));
      /* 295 */ }
      /* 296 */ }
      /* 297 */ }
      /* 298 */
      /* 299 */ mutableStateArray_1[0].setOffsetAndSizeFromPreviousCursor(index_2, previousCursor_1);
      /* 300 */ }
      /* 301 */
      /* 302 */ }
      /* 303 */ }
      /* 304 */ }
      /* 305 */
      /* 306 */ mutableStateArray_0[0].setOffsetAndSizeFromPreviousCursor(0, previousCursor_0);
      /* 307 */ }
      /* 308 */
      /* 309 */ }
      /* 310 */
      /* 311 */
      /* 312 */ private MapData ExternalMapToCatalyst_0(InternalRow i) {
      /* 313 */ MapData value_2 = null;
      /* 314 */ if (!MapObjects_loopIsNull140) {
      /* 315 */ final int length_0 = MapObjects_loopValue140.size();
      /* 316 */ final Object[] convertedKeys_0 = new Object[length_0];
      /* 317 */ final Object[] convertedValues_0 = new Object[length_0];
      /* 318 */ int index_0 = 0;
      /* 319 */ final scala.collection.Iterator entries_0 = MapObjects_loopValue140.iterator();
      /* 320 */ while(entries_0.hasNext()) {
      /* 321 */
      /* 322 */ final scala.Tuple2 entry_0 = (scala.Tuple2) entries_0.next();
      /* 323 */ ExternalMapToCatalyst_key8 = (Integer) entry_0._1();
      /* 324 */ ExternalMapToCatalyst_value8 = (Integer) entry_0._2();
      /* 325 */
      /* 326 */
      /* 327 */
      /* 328 */
      /* 329 */
      /* 330 */ if (false) {
      /* 331 */ throw new RuntimeException("Cannot use null as map key!");
      /* 332 */ } else {
      /* 333 */ convertedKeys_0[index_0] = (Integer) ExternalMapToCatalyst_key8;
      /* 334 */ }
      /* 335 */
      /* 336 */
      /* 337 */ if (false) {
      /* 338 */ convertedValues_0[index_0] = null;
      /* 339 */ } else {
      /* 340 */ convertedValues_0[index_0] = (Integer) ExternalMapToCatalyst_value8;
      /* 341 */ }
      /* 342 */
      /* 343 */ index_0++;
      /* 344 */ }
      /* 345 */
      /* 346 */ value_2 = new org.apache.spark.sql.catalyst.util.ArrayBasedMapData(new org.apache.spark.sql.catalyst.util.GenericArrayData(convertedKeys_0), new org.apache.spark.sql.catalyst.util.GenericArrayData(convertedValues_0));
      /* 347 */ }
      /* 348 */ globalIsNull_0 = MapObjects_loopIsNull140;
      /* 349 */ return value_2;
      /* 350 */ }
      /* 351 */
      /* 352 */
      /* 353 */ private ArrayData MapObjects_0(InternalRow i) {
      /* 354 */ boolean isNull_1 = i.isNullAt(0);
      /* 355 */ scala.collection.immutable.Map[] value_1 = isNull_1 ?
      /* 356 */ null : ((scala.collection.immutable.Map[])i.get(0, null));
      /* 357 */ ArrayData value_0 = null;
      /* 358 */
      /* 359 */ if (!isNull_1) {
      /* 360 */
      /* 361 */ int dataLength_0 = value_1.length;
      /* 362 */
      /* 363 */ MapData[] convertedArray_0 = null;
      /* 364 */ convertedArray_0 = new MapData[dataLength_0];
      /* 365 */
      /* 366 */
      /* 367 */ int loopIndex_0 = 0;
      /* 368 */
      /* 369 */ while (loopIndex_0 < dataLength_0) {
      /* 370 */ MapObjects_loopValue140 = (scala.collection.immutable.Map) (value_1[loopIndex_0]);
      /* 371 */ MapObjects_loopIsNull140 = MapObjects_loopValue140 == null;
      /* 372 */
      /* 373 */ MapData value_3 = ExternalMapToCatalyst_0(i);
      /* 374 */ if (globalIsNull_0) {
      /* 375 */ convertedArray_0[loopIndex_0] = null;
      /* 376 */ } else {
      /* 377 */ convertedArray_0[loopIndex_0] = value_3 instanceof UnsafeMapData? value_3.copy() : value_3;
      /* 378 */ }
      /* 379 */
      /* 380 */ loopIndex_0 += 1;
      /* 381 */ }
      /* 382 */
      /* 383 */ value_0 = new org.apache.spark.sql.catalyst.util.GenericArrayData(convertedArray_0);
      /* 384 */ }
      /* 385 */ globalIsNull_1 = isNull_1;
      /* 386 */ return value_0;
      /* 387 */ }
      /* 388 */
      /* 389 */ }

      Attachments

        Issue Links

          Activity

            People

              srowen Sean R. Owen
              srowen Sean R. Owen
              Votes:
              1 Vote for this issue
              Watchers:
              11 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: