Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-930

Add support for reading regular (non-block-based) files from S3 in S3FileSystem



    • New Feature
    • Status: Closed
    • Major
    • Resolution: Fixed
    • 0.10.1
    • 0.18.0
    • fs
    • None
    • Reviewed
    • Added support for reading and writing native S3 files. Native S3 files are referenced using s3n URIs. See http://wiki.apache.org/hadoop/AmazonS3 for more details.


      People often have input data on S3 that they want to use for a Map Reduce job and the current S3FileSystem implementation cannot read it since it assumes a block-based format.

      We would add the following metadata to files written by S3FileSystem: an indication that it is block oriented ("S3FileSystem.type=block") and a filesystem version number ("S3FileSystem.version=1.0"). Regular S3 files would not have the type metadata so S3FileSystem would not try to interpret them as inodes.

      An extension to write regular files to S3 would not be covered by this change - we could do this as a separate piece of work (we still need to decide whether to introduce another scheme - e.g. rename block-based S3 to "s3fs" and call regular S3 "s3" - or whether to just use a configuration property to control block-based vs. regular writes).


        1. jets3t-0.6.0.jar
          282 kB
          Thomas White
        2. hadoop-930-v5.patch
          99 kB
          Thomas White
        3. hadoop-930-v4.patch
          99 kB
          Thomas White
        4. hadoop-930-v3.patch
          99 kB
          Thomas White
        5. hadoop-930-v2.patch
          98 kB
          Thomas White
        6. hadoop-930.patch
          96 kB
          Thomas White

        Issue Links



              tomwhite Thomas White
              tomwhite Thomas White
              0 Vote for this issue
              2 Start watching this issue