Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
None
-
None
-
Reviewed
Description
The attached script simulates a process opening ~50 files via webhdfs and performing random reads. Note that there are at most 50 concurrent reads, and all webhdfs sessions are kept open. Each read is ~64k at a random position.
The script periodically (once per second) shells into the NameNode and produces a summary of the socket states. For my test cluster with 5 nodes, it took ~30 seconds for the NameNode to have ~25000 active connections and fails.
It appears that each request to the webhdfs client is opening a new connection to the NameNode and keeping it open after the request is complete. If the process continues to run, eventually (~30-60 seconds), all of the open connections are closed and the NameNode recovers.
This smells like SoftReference reaping. Are we using SoftReferences in the webhdfs client to cache NameNode connections but never re-using them?
Attachments
Attachments
Issue Links
- is blocked by
-
HADOOP-12424 Add a function to build unique cache key for Token.
- Resolved
- is duplicated by
-
HDFS-9370 TestDataNodeUGIProvider fails intermittently due to non-deterministic cache expiry.
- Resolved
- is related to
-
HADOOP-13436 RPC connections are leaking due to not overriding hashCode and equals
- Open