Details
-
Sub-task
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
None
Description
Currently, AWS region is either constructed via the endpoint URL, by making an assumption that the 2nd component after delimiter "." is the region in endpoint URL, which doesn't work for private links and sets the default to us-east-1 thus causing authorization issue w.r.t the private link.
The option fs.s3a.endpoint.region allows this to be explicitly set
how to set the s3 region on older hadoop releases
For anyone who needs to set the signing region on older versions of the s3a client you do not need this festure. instead just provide a custom endpoint to region mapping json file
- Download the default region mapping file awssdk_config_default.json
- Add a new regular expression to map the endpoint/hostname to the target region
- Save the file as /etc/hadoop/conf/awssdk_config_override.json
- verify basic hadop fs -ls commands work
- copy to the rest of the cluster.
- There should be no need to restart any services
Attachments
Issue Links
- is depended upon by
-
HADOOP-15603 S3A to support configuring various AWS S3 client extended options
- Open
-
HADOOP-17677 Distcp is unable to determine region with S3 PrivateLink endpoints
- Resolved
- is required by
-
HADOOP-18338 Unable to access data from S3 bucket over a vpc endpoint - 400 bad request
- Resolved
- relates to
-
HADOOP-18448 s3a endpoint per bucket configuration in pyspark is ignored
- Resolved
- links to