Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-4630

Datanode is going OOM due to small files in hdfs

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Blocker
    • Resolution: Invalid
    • 2.0.0-alpha
    • None
    • datanode, namenode
    • None
    • Ubuntu, Java 1.6

    Description

      Hi,

      We have very small files(size ranging 10KB-1MB) in our hdfs and no of files are in tens of millions. Due to this namenode and datanode both going out of memory very frequently. When we analyse the head dump of datanode most of the memory was used by ReplicaMap.

      Can we use EhCache or other to not to store all the data in memory?

      Thanks
      Ankush

      Attachments

        Activity

          People

            Unassigned Unassigned
            ankushb Ankush Bhatiya
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: