Uploaded image for project: 'Kylin'
  1. Kylin
  2. KYLIN-3931

Caused by: java.lang.RuntimeException: Could not create interface org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactory Is the hadoop compatibility jar on the classpath?

    XMLWordPrintableJSON

Details

    • Test
    • Status: Resolved
    • Critical
    • Resolution: Cannot Reproduce
    • v2.5.2
    • None
    • Environment
    • None
    • apache-hive-2.3.4-bin
      hbase-1.4.9
      hadoop-3.1.0
      apache-kylin-2.5.2-bin-hbase1x

    Description

      I am getting the below error. please find the environment details.
      apache-hive-2.3.4-bin
      hbase-1.4.9
      hadoop-3.1.0
      apache-kylin-2.5.2-bin-hbase1x.tar.gz
      Engine Type: Spark

      Kylin 63% completed after that i am getting this below error.

      WARN TaskSetManager: Lost task 0.0 in stage 1.0 (TID 2, server.com, executor 1): java.lang.ExceptionInInitializerError
      at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.getNewWriter(HFileOutputFormat2.java:305)
      at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.write(HFileOutputFormat2.java:229)
      at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.write(HFileOutputFormat2.java:167)
      at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12$$anonfun$apply$4.apply$mcV$sp(PairRDDFunctions.scala:1125)
      at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12$$anonfun$apply$4.apply(PairRDDFunctions.scala:1123)
      at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12$$anonfun$apply$4.apply(PairRDDFunctions.scala:1123)
      at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1353)
      at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12.apply(PairRDDFunctions.scala:1131)
      at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12.apply(PairRDDFunctions.scala:1102)
      at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
      at org.apache.spark.scheduler.Task.run(Task.scala:99)
      at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:325)
      at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
      at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
      at java.lang.Thread.run(Thread.java:748)
      Caused by: java.lang.RuntimeException: Could not create interface org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactory Is the hadoop compatibility jar on the classpath?
      at org.apache.hadoop.hbase.CompatibilitySingletonFactory.getInstance(CompatibilitySingletonFactory.java:73)
      at org.apache.hadoop.hbase.io.MetricsIO.<init>(MetricsIO.java:31)
      at org.apache.hadoop.hbase.io.hfile.HFile.<clinit>(HFile.java:191)
      ... 15 more
      Caused by: java.util.NoSuchElementException
      at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:365)
      at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
      at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
      at org.apache.hadoop.hbase.CompatibilitySingletonFactory.getInstance(CompatibilitySingletonFactory.java:59)
      ... 17 more

      19/04/01 11:47:55 INFO TaskSetManager: Starting task 0.1 in stage 1.0 (TID 3, server.com, executor 1, partition 0, NODE_LOCAL, 5842 bytes)
      19/04/01 11:47:55 WARN TaskSetManager: Lost task 0.1 in stage 1.0 (TID 3, server.com, executor 1): java.lang.NoClassDefFoundError: Could not initialize class org.apache.hadoop.hbase.io.hfile.HFile

      Attachments

        Activity

          People

            Unassigned Unassigned
            praveenece praveenece
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: