Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-20666

Flaky test - SparkListenerBus randomly failing java.lang.IllegalAccessError

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Critical
    • Resolution: Fixed
    • 2.2.0
    • 2.2.0
    • ML, Spark Core, SparkR
    • None

    Description

      seeing quite a bit of this on AppVeyor, aka Windows only,-> seems like in other test runs too, always only when running ML tests, it seems

      Exception in thread "SparkListenerBus" java.lang.IllegalAccessError: Attempted to access garbage collected accumulator 159454
      	at org.apache.spark.util.AccumulatorContext$$anonfun$get$1.apply(AccumulatorV2.scala:265)
      	at org.apache.spark.util.AccumulatorContext$$anonfun$get$1.apply(AccumulatorV2.scala:261)
      	at scala.Option.map(Option.scala:146)
      	at org.apache.spark.util.AccumulatorContext$.get(AccumulatorV2.scala:261)
      	at org.apache.spark.util.AccumulatorV2.name(AccumulatorV2.scala:88)
      	at org.apache.spark.sql.execution.metric.SQLMetric.toInfo(SQLMetrics.scala:67)
      	at org.apache.spark.sql.execution.ui.SQLListener$$anonfun$onTaskEnd$1.apply(SQLListener.scala:216)
      	at org.apache.spark.sql.execution.ui.SQLListener$$anonfun$onTaskEnd$1.apply(SQLListener.scala:216)
      	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
      	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
      	at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
      	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
      	at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
      	at scala.collection.AbstractTraversable.map(Traversable.scala:104)
      	at org.apache.spark.sql.execution.ui.SQLListener.onTaskEnd(SQLListener.scala:216)
      	at org.apache.spark.scheduler.SparkListenerBus$class.doPostEvent(SparkListenerBus.scala:45)
      	at org.apache.spark.scheduler.LiveListenerBus.doPostEvent(LiveListenerBus.scala:36)
      	at org.apache.spark.scheduler.LiveListenerBus.doPostEvent(LiveListenerBus.scala:36)
      	at org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:63)
      	at org.apache.spark.scheduler.LiveListenerBus.postToAll(LiveListenerBus.scala:36)
      	at org.apache.spark.scheduler.LiveListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(LiveListenerBus.scala:94)
      	at org.apache.spark.scheduler.LiveListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(LiveListenerBus.scala:79)
      	at org.apache.spark.scheduler.LiveListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(LiveListenerBus.scala:79)
      	at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
      	at org.apache.spark.scheduler.LiveListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(LiveListenerBus.scala:78)
      	at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1268)
      	at org.apache.spark.scheduler.LiveListenerBus$$anon$1.run(LiveListenerBus.scala:77)
      1
      MLlib recommendation algorithms: Spark package found in SPARK_HOME: C:\projects\spark\bin\..
      
      
      java.lang.IllegalStateException: SparkContext has been shutdown
      	at org.apache.spark.SparkContext.runJob(SparkContext.scala:2015)
      	at org.apache.spark.SparkContext.runJob(SparkContext.scala:2044)
      	at org.apache.spark.SparkContext.runJob(SparkContext.scala:2063)
      	at org.apache.spark.SparkContext.runJob(SparkContext.scala:2088)
      	at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936)
      	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
      	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
      	at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
      	at org.apache.spark.rdd.RDD.collect(RDD.scala:935)
      	at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:275)
      	at org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$collectFromPlan(Dataset.scala:2923)
      	at org.apache.spark.sql.Dataset$$anonfun$collect$1.apply(Dataset.scala:2474)
      	at org.apache.spark.sql.Dataset$$anonfun$collect$1.apply(Dataset.scala:2474)
      	at org.apache.spark.sql.Dataset$$anonfun$57.apply(Dataset.scala:2907)
      	at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65)
      	at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2906)
      	at org.apache.spark.sql.Dataset.collect(Dataset.scala:2474)
      	at org.apache.spark.sql.api.r.SQLUtils$.dfToCols(SQLUtils.scala:173)
      	at org.apache.spark.sql.api.r.SQLUtils.dfToCols(SQLUtils.scala)
      	at sun.reflect.GeneratedMethodAccessor104.invoke(Unknown Source)
      	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
      	at java.lang.reflect.Method.invoke(Method.java:498)
      	at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
      	at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
      	at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
      	at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
      	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
      	at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
      	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
      	at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
      	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
      	at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
      	at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
      	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
      	at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
      	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
      	at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
      	at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
      	at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
      	at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
      	at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
      	at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
      	at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
      	at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
      	at java.lang.Thread.run(Thread.java:745)
      1: expect_equivalent(expected_predictions, collect(predict(model, new_data))) at C:/projects/spark/R/lib/SparkR/tests/testthat/test_mllib_fpm.R:63
      

      Attachments

        Activity

          People

            cloud_fan Wenchen Fan
            felixcheung Felix Cheung
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: