Uploaded image for project: 'Hadoop Map/Reduce'
  1. Hadoop Map/Reduce
  2. MAPREDUCE-931

rumen should use its own interpolation classes to create runtimes for simulated tasks

    Details

    • Type: Improvement
    • Status: Closed
    • Priority: Minor
    • Resolution: Fixed
    • Affects Version/s: None
    • Fix Version/s: 0.21.0
    • Component/s: tools/rumen
    • Labels:
      None
    • Hadoop Flags:
      Reviewed

      Description

      Currently, when a simulator or benchmark is running and simulating hadoop jobs using rumen data, and rumen's runtime system is used to get execution times for the tasks in the simulated jobs, rumen would use some ad hoc code, despite the fact that rumen has a perfectly good interpolation framework to generate random variables that fit discrete CDFs.

      We should use the interpolation framework.

        Attachments

          Activity

            People

            • Assignee:
              dking Dick King
              Reporter:
              dking Dick King
            • Votes:
              0 Vote for this issue
              Watchers:
              0 Start watching this issue

              Dates

              • Created:
                Updated:
                Resolved: