Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-34754

sparksql 'add jar' not support hdfs ha mode in k8s

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Open
    • Major
    • Resolution: Unresolved
    • 2.4.7
    • None
    • Kubernetes
    • None

    Description

      Submit app to K8S,  the executors meet exception  "java.net.UnknownHostException: xx". 

      The udf jar uri using hdfs ha style, but the exception stack show  "...createNonHAProxy..."

       

      hql: 

      // code placeholder
      
      add jar hdfs://xx/test.jar;
      create temporary function test_udf as 'com.xxx.xxx';
      
      create table test.test_udf as 
      select test_udf('1') name_1;
       

       

       

      exception:

      // code placeholder
       TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, 172.30.89.44, executor 1): java.lang.IllegalArgumentException: java.net.UnknownHostException: xx
              at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:439)
              at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:321)
              at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176)
              at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:696)
              at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:636)
              at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:160)
              at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2796)
              at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:99)
              at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2830)
              at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2812)
              at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:390)
              at org.apache.spark.util.Utils$.getHadoopFileSystem(Utils.scala:1866)
              at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:721)
              at org.apache.spark.util.Utils$.fetchFile(Utils.scala:496)
              at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:816)
              at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:808)
              at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
              at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
              at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
              at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
              at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
              at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
              at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
              at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:808)
              at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:375)
              at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
              at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
              at java.lang.Thread.run(Thread.java:748)
      Caused by: java.net.UnknownHostException: xx
              ... 28 more
      
      

       

      Attachments

        Activity

          People

            Unassigned Unassigned
            lithiumlee-_- lithiumlee-_-
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

            Dates

              Created:
              Updated: