Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-14821

Executing the command 'hdfs -Dhadoop.security.credential.provider.path=file1.jceks,file2.jceks' fails if permission is denied to some files

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Open
    • Critical
    • Resolution: Unresolved
    • 2.8.0
    • None
    • fs/s3, hdfs-client, security

    Description

      =======
      Request Use Case:
      UC1:
      The customer has the path to a directory and subdirectories full of keys. The customer knows that he does not have the access to all the keys, but ignoring this problem, the customer makes a list of the keys.

      UC1.2:
      The customer in a FIFO manner, try his access to the key provided on the list. If the access is granted locally then he can try the login on the s3a.

      UC1.2:
      The customer in a FIFO manner, try his access to the key provided on the list. If the access is not granted locally then he will skip the login on the s3a and try the next key on the list.
      ===========

      For now, the UC1.2 fails with below exception and does not try the next key:

      $ hdfs  --loglevel DEBUG dfs -Dhadoop.security.credential.provider.path=jceks://hdfs/tmp/aws.jceks,jceks://hdfs/tmp/awst.jceks -ls s3a://av-dl-hwx-nprod-anhffpoc-enriched/hive/e_ceod/
      
      Not retrying because try once and fail.
      org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=502549376, access=READ, inode="/tmp/aws.jceks":admin:hdfs:-rwx------
      

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              emattosHWX Ernani Pereira de Mattos Junior
              Votes:
              0 Vote for this issue
              Watchers:
              6 Start watching this issue

              Dates

                Created:
                Updated: