Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-23420

Datasource loading not handling paths with regex chars.

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Incomplete
    • 2.2.1
    • None
    • Input/Output

    Description

      Greetings, during some recent testing I ran across an issue attempting to load files with regex chars like []()* etc. in them. The files are valid in the various storages and the normal hadoop APIs all function properly accessing them.

      When my code is executed, I get the following stack trace.

      8/02/14 04:52:46 ERROR yarn.ApplicationMaster: User class threw exception: java.io.IOException: Illegal file pattern: Unmatched closing ')' near index 130 A_VERY_LONG_DIRECTORY_FOLDER_THAT_INCLUDES_MULTIBYTE_AND_SPECIAL_CHARACTERS_abcdefghijklmnopqrst_0123456789_@#\$%^&()=+[(?:]);',\.??????????????????????????????? ^ java.io.IOException: Illegal file pattern: Unmatched closing ')' near index 130 A_VERY_LONG_DIRECTORY_FOLDER_THAT_INCLUDES_MULTIBYTE_AND_SPECIAL_CHARACTERS_abcdefghijklmnopqrst_0123456789@#\$%^&()=+[(?:]);',\.??????????????????????????????? ^ at org.apache.hadoop.fs.GlobFilter.init(GlobFilter.java:71) at org.apache.hadoop.fs.GlobFilter.<init>(GlobFilter.java:50) at org.apache.hadoop.fs.Globber.doGlob(Globber.java:210) at org.apache.hadoop.fs.Globber.glob(Globber.java:149) at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:1955) at org.apache.hadoop.fs.s3a.S3AFileSystem.globStatus(S3AFileSystem.java:2477) at org.apache.spark.deploy.SparkHadoopUtil.globPath(SparkHadoopUtil.scala:234) at org.apache.spark.deploy.SparkHadoopUtil.globPathIfNecessary(SparkHadoopUtil.scala:244) at org.apache.spark.sql.execution.datasources.DataSource$.org$apache$spark$sql$execution$datasources$DataSource$$checkAndGlobPathIfNecessary(DataSource.scala:618) at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$14.apply(DataSource.scala:350) at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$14.apply(DataSource.scala:350) at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241) at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241) at scala.collection.immutable.List.foreach(List.scala:381) at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241) at scala.collection.immutable.List.flatMap(List.scala:344) at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:349) at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:178) at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:533) at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:412) at com.sap.profile.SparkProfileTask.main(SparkProfileTask.java:95) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:635) Caused by: java.util.regex.PatternSyntaxException: Unmatched closing ')' near index 130 A_VERY_LONG_DIRECTORY_FOLDER_THAT_INCLUDES_MULTIBYTE_AND_SPECIAL_CHARACTERS_abcdefghijklmnopqrst_0123456789@#\$%^&()=+[(?:]);',\.??????????????????????????????? ^ at java.util.regex.Pattern.error(Pattern.java:1955) at java.util.regex.Pattern.compile(Pattern.java:1700) at java.util.regex.Pattern.<init>(Pattern.java:1351) at java.util.regex.Pattern.compile(Pattern.java:1054) at org.apache.hadoop.fs.GlobPattern.set(GlobPattern.java:156) at org.apache.hadoop.fs.GlobPattern.<init>(GlobPattern.java:42) at org.apache.hadoop.fs.GlobFilter.init(GlobFilter.java:67) ... 25 more 18/02/14 04:52:46 INFO yarn.ApplicationMaster: Final app status: FAILED, exitCode: 15, (reason: User class threw exception: java.io.IOException: Illegal file pattern: Unmatched closing ')' near index 130 A_VERY_LONG_DIRECTORY_FOLDER_THAT_INCLUDES_MULTIBYTE_AND_SPECIAL_CHARACTERS_abcdefghijklmnopqrst_0123456789@#\$%^&()=+[(?:]);',\.???????????????????_???????????? ^) 18/02/14 04:52:46 INFO spark.SparkContext: Invoking stop() from shutdown hook

       

      Code is as follows ...

      Dataset<Row> input = sqlContext.read().option("header", "true").option("sep", ",").option("quote", "\"").option("charset", "utf8").option("escape", "
      ").csv("s3a://myBucket/A_VERY_LONG_DIRECTORY_FOLDER_THAT_INCLUDES_MULTIBYTE_AND_SPECIAL_CHARACTERS_abcdefghijklmnopqrst_0123456789_~@#$%^&()-_=+[\{]};',._鞍鞍亜_Белебей_鞍鞍めぐみ林原ぐみ林原めぐみ솅ᄌ종대왕_ไชยแม็คอินमाधु/COLUMN_HEADER_PRESENT_a_longer_file_name_with_different_types_of_characters_including_numbers_upper_case_and_lower_case_鞍鞍亜_Белебей_鞍鞍_林原めぐみ林原めぐみ林原めぐみ솅_CSV_PIPE_DELIM.csv"));

      Attachments

        Activity

          People

            Unassigned Unassigned
            mitch.gudmundson@sap.com Mitchell Gudmundson
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: