Uploaded image for project: 'Mahout'
  1. Mahout
  2. MAHOUT-482

Defaulting $HADOOP_CONF_DIR to $HADOOP_HOME/conf

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Closed
    • Trivial
    • Resolution: Fixed
    • None
    • 0.4
    • None
    • None

    Description

      Hi all,

      while trying to understand the clusterdump utility, I encountered a message "no HADOOP_CONF_DIR or HADOOP_HOME set, running locally". As per some of the discussions in Apr'10 (http://lucene.472066.n3.nabble.com/Dealing-with-kmean-and-meanshift-output-td708824.html#a709066), it would be good to default the hadoop conf directory to hadoop_home/conf.

      So I've changed the script to accommodate this.
      Existing code:
      if [ "$HADOOP_CONF_DIR" = "" ] || [ "$HADOOP_HOME" = "" ] || [ "$MAHOUT_LOCAL" != "" ] ; then
      if [ "$HADOOP_CONF_DIR" = "" ] || [ "$HADOOP_HOME" = "" ] ; then
      echo "no HADOOP_CONF_DIR or HADOOP_HOME set, running locally"
      elif [ "$MAHOUT_LOCAL" != "" ] ; then
      echo "MAHOUT_LOCAL is set, running locally"
      fi
      exec "$JAVA" $JAVA_HEAP_MAX $MAHOUT_OPTS -classpath "$CLASSPATH" $CLASS "$@"
      else
      echo "running on hadoop, using HADOOP_HOME=$HADOOP_HOME and HADOOP_CONF_DIR=$HADOOP_CONF_DIR"

      New code:
      if [ "$HADOOP_HOME" = "" ] || [ "$MAHOUT_LOCAL" != "" ] ; then
      if [ "$HADOOP_HOME" = "" ] ; then
      echo "no HADOOP_HOME set, running locally"
      elif [ "$MAHOUT_LOCAL" != "" ] ; then
      echo "MAHOUT_LOCAL is set, running locally"
      fi
      exec "$JAVA" $JAVA_HEAP_MAX $MAHOUT_OPTS -classpath "$CLASSPATH" $CLASS "$@"
      else
      if [ "$HADOOP_CONF_DIR" = "" ] ; then
      HADOOP_CONF_DIR=$HADOOP_HOME/conf
      echo "no HADOOP_CONF_DIR set. so defaulting HADOOP_CONF_DIR to $HADOOP_HOME/conf "
      fi
      echo "running on hadoop, using HADOOP_HOME=$HADOOP_HOME and HADOOP_CONF_DIR=$HADOOP_CONF_DIR"

      In my local machine, I've just set HADOOP_HOME and the script defaults the $HADOOP_CONF_DIR directory and runs on hadoop.
      If this issue / fix looks valid, I'll open a JIRA issue and attach the patch.

      Appreciate your thoughts / feedbacks,
      Joe.

      Drew Farris to dev
      show details 8:53 AM (12 hours ago)
      On Tue, Aug 17, 2010 at 12:26 AM, Joe Kumar <joekumar@gmail.com> wrote:
      > Hi all,
      >
      > while trying to understand the clusterdump utility, I encountered a message
      > "no HADOOP_CONF_DIR or HADOOP_HOME set, running locally". As per some of the
      > discussions in Apr'10 (
      > http://lucene.472066.n3.nabble.com/Dealing-with-kmean-and-meanshift-output-td708824.html#a709066),
      > it would be good to default the hadoop conf directory to hadoop_home/conf.

      Agreed, and your solution looks good. I'll keep an eye out for the patch.

      Jake Mannix to dev
      show details 3:46 PM (5 hours ago)
      +1

      Good idea. I thought we already did this...

      -jake

      Attachments

        1. MAHOUT-482.patch
          1 kB
          Joe Prasanna Kumar

        Activity

          People

            drew.farris Drew Farris
            joekumar Joe Prasanna Kumar
            Votes:
            0 Vote for this issue
            Watchers:
            0 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Time Tracking

                Estimated:
                Original Estimate - 20m
                20m
                Remaining:
                Remaining Estimate - 20m
                20m
                Logged:
                Time Spent - Not Specified
                Not Specified