Uploaded image for project: 'Apache Arrow'
  1. Apache Arrow
  2. ARROW-9226

[Python] pyarrow.fs.HadoopFileSystem - retrieve options from core-site.xml or hdfs-site.xml if available

    XMLWordPrintableJSON

Details

    Description

      'Legacy' pyarrow.hdfs.connect was somehow able to get the namenode info from the hadoop configuration files.

      The new pyarrow.fs.HadoopFileSystem requires the host to be specified.

      Inferring this info from "the environment" makes it easier to deploy pipelines.

      But more important, for HA namenodes it is almost impossible to know for sure what to specify. If a rolling restart is ongoing, the namenode is changing. There is no guarantee on which will be active in a HA setup.

      I tried connecting to the standby namenode. The connection gets established, but when writing a file an error is raised that standby namenodes are not allowed to write to.

       

      Attachments

        Issue Links

          Activity

            People

              itamarst Itamar Turner-Trauring
              bquinart Bruno Quinart
              Votes:
              0 Vote for this issue
              Watchers:
              8 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved:

                Time Tracking

                  Estimated:
                  Original Estimate - Not Specified
                  Not Specified
                  Remaining:
                  Remaining Estimate - 0h
                  0h
                  Logged:
                  Time Spent - 1h 40m
                  1h 40m