Index: build-common.xml =================================================================== --- build-common.xml (revision 763135) +++ build-common.xml (working copy) @@ -20,6 +20,10 @@ + + + + @@ -48,6 +52,12 @@ + + + + + @@ -297,4 +307,29 @@ + + + + + + + + + + + + + + + + Index: hwi/src/java/org/apache/hadoop/hive/hwi/HWISessionManager.java =================================================================== --- hwi/src/java/org/apache/hadoop/hive/hwi/HWISessionManager.java (revision 763135) +++ hwi/src/java/org/apache/hadoop/hive/hwi/HWISessionManager.java (working copy) @@ -119,10 +119,10 @@ * because we need to set parameters the client is not aware of. One such * parameter is the command line arguments the server was started with. * - * @param sessionname + * @param a + * Authenticated user + * @param sessionName * Represents the session name - * @param sessionpass - * the session password, can be "" * @return a new SessionItem or null if a session with that name already * exists */ @@ -201,4 +201,4 @@ return this.items.get(auth); } -} \ No newline at end of file +} Index: hwi/src/java/org/apache/hadoop/hive/hwi/HWISessionItem.java =================================================================== --- hwi/src/java/org/apache/hadoop/hive/hwi/HWISessionItem.java (revision 763135) +++ hwi/src/java/org/apache/hadoop/hive/hwi/HWISessionItem.java (working copy) @@ -196,7 +196,7 @@ * * @param wanted * a ConfVar - * @return + * @return Value of the configuration variable. */ public String getHiveConfVar(HiveConf.ConfVars wanted) throws HWIException { String result = null; Index: hwi/build.xml =================================================================== --- hwi/build.xml (revision 763135) +++ hwi/build.xml (working copy) @@ -34,4 +34,30 @@ + + + + + + + + + + + + + + + + + Index: docs/changes/ChangesFancyStyle.css =================================================================== --- docs/changes/ChangesFancyStyle.css (revision 0) +++ docs/changes/ChangesFancyStyle.css (revision 0) @@ -0,0 +1,170 @@ +/* +* Licensed to the Apache Software Foundation (ASF) under one or more +* contributor license agreements. See the NOTICE file distributed with +* this work for additional information regarding copyright ownership. +* The ASF licenses this file to You under the Apache License, Version 2.0 +* (the "License"); you may not use this file except in compliance with +* the License. You may obtain a copy of the License at +* +* http://www.apache.org/licenses/LICENSE-2.0 +* +* Unless required by applicable law or agreed to in writing, software +* distributed under the License is distributed on an "AS IS" BASIS, +* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +* See the License for the specific language governing permissions and +* limitations under the License. +*/ +/** + * General + */ + +img { border: 0; } + +#content table { + border: 0; + width: 100%; +} +/*Hack to get IE to render the table at 100%*/ +* html #content table { margin-left: -3px; } + +#content th, +#content td { + margin: 0; + padding: 0; + vertical-align: top; +} + +.clearboth { + clear: both; +} + +.note, .warning, .fixme { + border: solid black 1px; + margin: 1em 3em; +} + +.note .label { + background: #369; + color: white; + font-weight: bold; + padding: 5px 10px; +} +.note .content { + background: #F0F0FF; + color: black; + line-height: 120%; + font-size: 90%; + padding: 5px 10px; +} +.warning .label { + background: #C00; + color: white; + font-weight: bold; + padding: 5px 10px; +} +.warning .content { + background: #FFF0F0; + color: black; + line-height: 120%; + font-size: 90%; + padding: 5px 10px; +} +.fixme .label { + background: #C6C600; + color: black; + font-weight: bold; + padding: 5px 10px; +} +.fixme .content { + padding: 5px 10px; +} + +/** + * Typography + */ + +body { + font-family: verdana, "Trebuchet MS", arial, helvetica, sans-serif; + font-size: 100%; +} + +#content { + font-family: Georgia, Palatino, Times, serif; + font-size: 95%; +} +#tabs { + font-size: 70%; +} +#menu { + font-size: 80%; +} +#footer { + font-size: 70%; +} + +h1, h2, h3, h4, h5, h6 { + font-family: "Trebuchet MS", verdana, arial, helvetica, sans-serif; + font-weight: bold; + margin-top: 1em; + margin-bottom: .5em; +} + +h1 { + margin-top: 0; + margin-bottom: 1em; + font-size: 1.4em; + background-color: 73CAFF +} +#content h1 { + font-size: 160%; + margin-bottom: .5em; +} +#menu h1 { + margin: 0; + padding: 10px; + background: #336699; + color: white; +} +h2 { + font-size: 120%; + background-color: 73CAFF +} +h3 { font-size: 100%; } +h4 { font-size: 90%; } +h5 { font-size: 80%; } +h6 { font-size: 75%; } + +p { + line-height: 120%; + text-align: left; + margin-top: .5em; + margin-bottom: 1em; +} + +#content li, +#content th, +#content td, +#content li ul, +#content li ol{ + margin-top: .5em; + margin-bottom: .5em; +} + + +#content li li, +#minitoc-area li{ + margin-top: 0em; + margin-bottom: 0em; +} + +#content .attribution { + text-align: right; + font-style: italic; + font-size: 85%; + margin-top: 1em; +} + +.codefrag { + font-family: "Courier New", Courier, monospace; + font-size: 110%; +} Index: docs/changes/changes2html.pl =================================================================== --- docs/changes/changes2html.pl (revision 0) +++ docs/changes/changes2html.pl (revision 0) @@ -0,0 +1,282 @@ +#!/usr/bin/perl +# +# Transforms Lucene Java's CHANGES.txt into Changes.html +# +# Input is on STDIN, output is to STDOUT +# +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +use strict; +use warnings; + +my $jira_url_prefix = 'http://issues.apache.org/jira/browse/'; +my $title = undef; +my $release = undef; +my $sections = undef; +my $items = undef; +my $first_relid = undef; +my $second_relid = undef; +my @releases = (); + +my @lines = <>; # Get all input at once + +# +# Parse input and build hierarchical release structure in @releases +# +for (my $line_num = 0 ; $line_num <= $#lines ; ++$line_num) { + $_ = $lines[$line_num]; + next unless (/\S/); # Skip blank lines + + unless ($title) { + if (/\S/) { + s/^\s+//; # Trim leading whitespace + s/\s+$//; # Trim trailing whitespace + } + $title = $_; + next; + } + + if (/^(Release)|(Trunk)/) { # Release headings + $release = $_; + $sections = []; + push @releases, [ $release, $sections ]; + ($first_relid = lc($release)) =~ s/\s+/_/g if ($#releases == 0); + ($second_relid = lc($release)) =~ s/\s+/_/g if ($#releases == 1); + $items = undef; + next; + } + + # Section heading: 2 leading spaces, words all capitalized + if (/^ ([A-Z]+)\s*/) { + my $heading = $_; + $items = []; + push @$sections, [ $heading, $items ]; + next; + } + + # Handle earlier releases without sections - create a headless section + unless ($items) { + $items = []; + push @$sections, [ undef, $items ]; + } + + my $type; + if (@$items) { # A list item has been encountered in this section before + $type = $items->[0]; # 0th position of items array is list type + } else { + $type = get_list_type($_); + push @$items, $type; + } + + if ($type eq 'numbered') { # The modern items list style + # List item boundary is another numbered item or an unindented line + my $line; + my $item = $_; + $item =~ s/^(\s{0,2}\d+\.\s*)//; # Trim the leading item number + my $leading_ws_width = length($1); + $item =~ s/\s+$//; # Trim trailing whitespace + $item .= "\n"; + + while ($line_num < $#lines + and ($line = $lines[++$line_num]) !~ /^(?:\s{0,2}\d+\.\s*\S|\S)/) { + $line =~ s/^\s{$leading_ws_width}//; # Trim leading whitespace + $line =~ s/\s+$//; # Trim trailing whitespace + $item .= "$line\n"; + } + $item =~ s/\n+\Z/\n/; # Trim trailing blank lines + push @$items, $item; + --$line_num unless ($line_num == $#lines); + } elsif ($type eq 'paragraph') { # List item boundary is a blank line + my $line; + my $item = $_; + $item =~ s/^(\s+)//; + my $leading_ws_width = defined($1) ? length($1) : 0; + $item =~ s/\s+$//; # Trim trailing whitespace + $item .= "\n"; + + while ($line_num < $#lines and ($line = $lines[++$line_num]) =~ /\S/) { + $line =~ s/^\s{$leading_ws_width}//; # Trim leading whitespace + $line =~ s/\s+$//; # Trim trailing whitespace + $item .= "$line\n"; + } + push @$items, $item; + --$line_num unless ($line_num == $#lines); + } else { # $type is one of the bulleted types + # List item boundary is another bullet or a blank line + my $line; + my $item = $_; + $item =~ s/^(\s*$type\s*)//; # Trim the leading bullet + my $leading_ws_width = length($1); + $item =~ s/\s+$//; # Trim trailing whitespace + $item .= "\n"; + + while ($line_num < $#lines + and ($line = $lines[++$line_num]) !~ /^\s*(?:$type|\Z)/) { + $line =~ s/^\s{$leading_ws_width}//; # Trim leading whitespace + $line =~ s/\s+$//; # Trim trailing whitespace + $item .= "$line\n"; + } + push @$items, $item; + --$line_num unless ($line_num == $#lines); + } +} + +# +# Print HTML-ified version to STDOUT +# +print<<"__HTML_HEADER__"; + + + + $title + + + + + + + +Hadoop +

$title

+ +__HTML_HEADER__ + +my $heading; +my $relcnt = 0; +my $header = 'h2'; +for my $rel (@releases) { + if (++$relcnt == 3) { + $header = 'h3'; + print "

"; + print "Older Releases"; + print "

\n"; + print "\n" if ($relcnt > 3); +print "\n\n"; + + +# +# Subroutine: get_list_type +# +# Takes one parameter: +# +# - The first line of a sub-section/point +# +# Returns one scalar: +# +# - The list type: 'numbered'; or one of the bulleted types '-', or '.' or +# 'paragraph'. +# +sub get_list_type { + my $first_list_item_line = shift; + my $type = 'paragraph'; # Default to paragraph type + + if ($first_list_item_line =~ /^\s{0,2}\d+\.\s+\S+/) { + $type = 'numbered'; + } elsif ($first_list_item_line =~ /^\s*([-.])\s+\S+/) { + $type = $1; + } + return $type; +} + +1; Property changes on: docs/changes/changes2html.pl ___________________________________________________________________ Added: svn:executable + * Index: docs/changes/ChangesSimpleStyle.css =================================================================== --- docs/changes/ChangesSimpleStyle.css (revision 0) +++ docs/changes/ChangesSimpleStyle.css (revision 0) @@ -0,0 +1,49 @@ +/* +* Licensed to the Apache Software Foundation (ASF) under one or more +* contributor license agreements. See the NOTICE file distributed with +* this work for additional information regarding copyright ownership. +* The ASF licenses this file to You under the Apache License, Version 2.0 +* (the "License"); you may not use this file except in compliance with +* the License. You may obtain a copy of the License at +* +* http://www.apache.org/licenses/LICENSE-2.0 +* +* Unless required by applicable law or agreed to in writing, software +* distributed under the License is distributed on an "AS IS" BASIS, +* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +* See the License for the specific language governing permissions and +* limitations under the License. +*/ + +body { + font-family: Courier New, monospace; + font-size: 10pt; +} + +h1 { + font-family: Courier New, monospace; + font-size: 10pt; +} + +h2 { + font-family: Courier New, monospace; + font-size: 10pt; +} + +h3 { + font-family: Courier New, monospace; + font-size: 10pt; +} + +a:link { + color: blue; +} + +a:visited { + color: purple; +} + +li { + margin-top: 1em; + margin-bottom: 1em; +} Index: build.xml =================================================================== --- build.xml (revision 763135) +++ build.xml (working copy) @@ -20,6 +20,12 @@ + + + + + + @@ -30,8 +36,49 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + @@ -156,7 +203,7 @@ - + @@ -212,6 +259,9 @@ + + + @@ -252,4 +302,86 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Index: service/build.xml =================================================================== --- service/build.xml (revision 763135) +++ service/build.xml (working copy) @@ -56,4 +56,32 @@ + + + + + + + + + + + + + + + + + + + Index: jdbc/build.xml =================================================================== --- jdbc/build.xml (revision 763135) +++ jdbc/build.xml (working copy) @@ -48,4 +48,31 @@ + + + + + + + + + + + + + + + + + + Index: serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyLong.java =================================================================== --- serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyLong.java (revision 763135) +++ serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyLong.java (working copy) @@ -159,8 +159,8 @@ * a variable. That's one of the reasons that we don't make radix a * parameter here. * + * @param out the outputstream to write to * @param i an int to write out - * @return String the representation of the argument * @throws IOException */ public static void writeUTF8(OutputStream out, long i) throws IOException { Index: serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyMap.java =================================================================== --- serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyMap.java (revision 763135) +++ serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyMap.java (working copy) @@ -283,7 +283,7 @@ * @param itemSeparator The separator between different entries. * @param keyValueSeparator The separator between key and value. * @param nullSequence The byte sequence representing NULL. - * @return + * @return the map object */ public Map getMap(byte itemSeparator, byte keyValueSeparator, Text nullSequence) { Index: serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyInteger.java =================================================================== --- serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyInteger.java (revision 763135) +++ serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyInteger.java (working copy) @@ -159,8 +159,8 @@ * a variable. That's one of the reasons that we don't make radix a * parameter here. * + * @param out the outputstream to write to * @param i an int to write out - * @return String the representation of the argument * @throws IOException */ public static void writeUTF8(OutputStream out, int i) throws IOException { Index: serde/src/java/org/apache/hadoop/hive/serde2/MetadataTypedColumnsetSerDe.java =================================================================== --- serde/src/java/org/apache/hadoop/hive/serde2/MetadataTypedColumnsetSerDe.java (revision 763135) +++ serde/src/java/org/apache/hadoop/hive/serde2/MetadataTypedColumnsetSerDe.java (working copy) @@ -125,7 +125,7 @@ /** * Split the row into columns. * @param limit up to limit columns will be produced (the last column takes all the rest), -1 for unlimited. - * @return + * @return The ColumnSet object * @throws Exception */ public static Object deserialize(ColumnSet c, String row, String sep, Index: serde/build.xml =================================================================== --- serde/build.xml (revision 763135) +++ serde/build.xml (working copy) @@ -79,4 +79,31 @@ + + + + + + + + + + + + + + + + + + Index: metastore/src/model/org/apache/hadoop/hive/metastore/model/MFieldSchema.java =================================================================== --- metastore/src/model/org/apache/hadoop/hive/metastore/model/MFieldSchema.java (revision 763135) +++ metastore/src/model/org/apache/hadoop/hive/metastore/model/MFieldSchema.java (working copy) @@ -71,7 +71,7 @@ return type; } /** - * @param type the type to set + * @param field the type to set */ public void setType(String field) { this.type = field; Index: metastore/src/java/org/apache/hadoop/hive/metastore/RWTable.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/RWTable.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/RWTable.java (working copy) @@ -104,7 +104,6 @@ * delete the data, but not the schema * Can be applied on a partition by partition basis * - * @param partition partition in that table or "" or null * @exception MetaException if any problems instantiating this object * */ @@ -135,8 +134,7 @@ * * Add column names to a column set ser de table. * - * @param tableName the name of the table to alter - * @param columns the name of the columns + * @param schema the property value pairs for the table * @exception MetaException if any problems altering the table * */ Index: metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreClient.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreClient.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreClient.java (working copy) @@ -453,7 +453,7 @@ /** * This operation is unsupported in this metastore. - * @see MetaStoreClient#dropTable(String, String) + * @see IMetaStoreClient#dropTable(String, String, boolean, boolean) */ public void dropTable(String dbname, String name, boolean deleteData, boolean ignoreUknownTab) throws ExistingDependentsException, MetaException, TException, NoSuchObjectException { Index: metastore/src/java/org/apache/hadoop/hive/metastore/RawStore.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/RawStore.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/RawStore.java (working copy) @@ -43,7 +43,7 @@ /** * if this is the commit of the first open call then an actual commit is called. - * @return + * @return true or false */ public abstract boolean commitTransaction(); Index: metastore/src/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java (working copy) @@ -58,7 +58,7 @@ /** * Drop the table. - * @param dbName The database for this table + * @param dbname The database for this table * @param tableName The table to drop * @throws MetaException Could not drop table properly. * @throws NoSuchObjectException The table wasn't found. @@ -98,10 +98,10 @@ throws MetaException, TException, NoSuchObjectException; /** - * @param table_name - * @param db_name - * @param part_vals - * @return + * @param tableName + * @param dbName + * @param partVals + * @return the partition object * @throws InvalidObjectException * @throws AlreadyExistsException * @throws MetaException @@ -125,13 +125,13 @@ MetaException, TException; /** - * @param tbl_name - * @param db_name - * @param part_vals - * @return + * @param tblName + * @param dbName + * @param partVals + * @return the partition object * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#lookup_partition(java.lang.String, java.lang.String, java.util.List) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_partition(java.lang.String, java.lang.String, java.util.List) */ public Partition getPartition(String tblName, String dbName, List partVals) throws MetaException, TException ; @@ -140,11 +140,10 @@ * @param tbl_name * @param db_name * @param max_parts - * @return + * @return the list of partitions * @throws NoSuchObjectException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#list_partitions(java.lang.String, java.lang.String, short) */ public List listPartitions(String db_name, String tbl_name, short max_parts) throws NoSuchObjectException, MetaException, TException; @@ -153,7 +152,6 @@ throws MetaException, TException; /** * @param tbl - * @return * @throws AlreadyExistsException * @throws InvalidObjectException * @throws MetaException @@ -173,12 +171,12 @@ * @param tbl_name * @param part_vals * @param deleteData delete the underlying data or just delete the table in metadata - * @return + * @return true or false * @throws NoSuchObjectException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_partition(java.lang.String, java.lang.String, java.util.List) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_partition(java.lang.String, java.lang.String, java.util.List, boolean) */ public boolean dropPartition(String db_name, String tbl_name, List part_vals, boolean deleteData) throws NoSuchObjectException, MetaException, TException; -} \ No newline at end of file +} Index: metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java (working copy) @@ -198,7 +198,7 @@ /** * @param new_part - * @return + * @return the added partition * @throws InvalidObjectException * @throws AlreadyExistsException * @throws MetaException @@ -214,7 +214,7 @@ * @param table_name * @param db_name * @param part_vals - * @return + * @return the appended partition * @throws InvalidObjectException * @throws AlreadyExistsException * @throws MetaException @@ -229,7 +229,7 @@ /** * @param name * @param location_uri - * @return + * @return true or false * @throws AlreadyExistsException * @throws MetaException * @throws TException @@ -254,7 +254,7 @@ /** * @param type - * @return + * @return true or false * @throws AlreadyExistsException * @throws InvalidObjectException * @throws MetaException @@ -268,7 +268,7 @@ /** * @param name - * @return + * @return true or false * @throws MetaException * @throws TException * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_database(java.lang.String) @@ -281,7 +281,7 @@ * @param tbl_name * @param db_name * @param part_vals - * @return + * @return true or false * @throws NoSuchObjectException * @throws MetaException * @throws TException @@ -297,11 +297,11 @@ * @param tbl_name * @param part_vals * @param deleteData delete the underlying data or just delete the table in metadata - * @return + * @return true or false * @throws NoSuchObjectException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_partition(java.lang.String, java.lang.String, java.util.List) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_partition(java.lang.String, java.lang.String, java.util.List, boolean) */ public boolean dropPartition(String db_name, String tbl_name, List part_vals, boolean deleteData) throws NoSuchObjectException, MetaException, TException { @@ -311,12 +311,11 @@ /** * @param name * @param dbname - * @return * @throws NoSuchObjectException * @throws ExistingDependentsException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_table(java.lang.String, java.lang.String) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_table(java.lang.String, java.lang.String, boolean) */ public void dropTable(String dbname, String name) throws NoSuchObjectException, ExistingDependentsException, MetaException, TException { @@ -327,12 +326,11 @@ * @param dbname * @param name * @param deleteData delete the underlying data or just delete the table in metadata - * @return * @throws NoSuchObjectException * @throws ExistingDependentsException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_table(java.lang.String, java.lang.String) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_table(java.lang.String, java.lang.String, boolean) */ public void dropTable(String dbname, String name, boolean deleteData, boolean ignoreUknownTab) throws ExistingDependentsException, MetaException, TException, NoSuchObjectException { @@ -347,7 +345,7 @@ /** * @param type - * @return + * @return true if the type is dropped * @throws MetaException * @throws TException * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#drop_type(java.lang.String) @@ -358,7 +356,7 @@ /** * @param name - * @return + * @return map of types * @throws MetaException * @throws TException * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_type_all(java.lang.String) @@ -368,10 +366,10 @@ } /** - * @return + * @return the list of databases * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#list_databases() + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_databases() */ public List getDatabases() throws MetaException, TException { return client.get_databases(); @@ -381,11 +379,10 @@ * @param tbl_name * @param db_name * @param max_parts - * @return + * @return list of partitions * @throws NoSuchObjectException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#list_partitions(java.lang.String, java.lang.String, short) */ public List listPartitions(String db_name, String tbl_name, short max_parts) throws NoSuchObjectException, MetaException, TException { @@ -394,11 +391,11 @@ /** * @param name - * @return + * @return the database * @throws NoSuchObjectException * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#lookup_database(java.lang.String) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_database(java.lang.String) */ public Database getDatabase(String name) throws NoSuchObjectException, MetaException, TException { @@ -409,10 +406,10 @@ * @param tbl_name * @param db_name * @param part_vals - * @return + * @return the partition * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#lookup_partition(java.lang.String, java.lang.String, java.util.List) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_partition(java.lang.String, java.lang.String, java.util.List) */ public Partition getPartition(String db_name, String tbl_name, List part_vals) throws MetaException, TException { @@ -422,12 +419,12 @@ /** * @param name * @param dbname - * @return + * @return the table * @throws NoSuchObjectException * @throws MetaException * @throws TException * @throws NoSuchObjectException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#lookup_table(java.lang.String, java.lang.String) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_table(java.lang.String, java.lang.String) */ public Table getTable(String dbname, String name) throws MetaException, TException, NoSuchObjectException { return client.get_table(dbname, name); @@ -435,10 +432,10 @@ /** * @param name - * @return + * @return the type * @throws MetaException * @throws TException - * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#lookup_type(java.lang.String) + * @see org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore.Iface#get_type(java.lang.String) */ public Type getType(String name) throws MetaException, TException { return client.get_type(name); Index: metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java (working copy) @@ -221,7 +221,7 @@ /** * if this is the commit of the first open call then an actual commit is called. - * @return + * @return Always returns true */ @SuppressWarnings("nls") public boolean commitTransaction() { Index: metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreUtils.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreUtils.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreUtils.java (working copy) @@ -138,9 +138,8 @@ * * Get the Deserializer for a table given its name and properties. * - * @param name the name of the table - * @param conf - hadoop config - * @param p - the properties to use to instantiate the schema + * @param conf hadoop config + * @param schema the properties to use to instantiate the deserializer * @return the Deserializer * @exception MetaException if any problems instantiating the Deserializer * @@ -163,11 +162,10 @@ /** * getDeserializer * - * Get the Deserializer for a table given its name and properties. + * Get the Deserializer for a table. * - * @param name the name of the table * @param conf - hadoop config - * @param p - SerDe info + * @param table the table * @return the Deserializer * @exception MetaException if any problems instantiating the Deserializer * @@ -237,8 +235,8 @@ * Checks the name conforms to our standars which are: "[a-zA-z_0-9]+". * checks this is just characters and numbers and _ * - * @param tableName the name to validate - * @return none + * @param name the name to validate + * @return true or false depending on conformance * @exception MetaException if it doesn't match the pattern. */ static public boolean validateName(String name) { @@ -541,7 +539,7 @@ /** * @param tableName * @param deserializer - * @return + * @return the list of fields * @throws SerDeException * @throws MetaException */ Index: metastore/src/java/org/apache/hadoop/hive/metastore/ROTable.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/ROTable.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/ROTable.java (working copy) @@ -99,7 +99,6 @@ * Scan the file system and find all the partitions of this table * Not recursive right now - needs to be! * - * @param the table name * @return a list of partitions - not full paths * @exception MetaException if gneneral problem or this table does not exist. */ Index: metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreServer.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreServer.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreServer.java (working copy) @@ -186,9 +186,8 @@ * * return a table's partitions * - * @param tableName - a valid table name - * @param low - low index - * @param high - high index + * @param dbName the database name + * @param tableName a valid table name * @return a string (including '\n's) of the rows * @exception TException if thrift problem * @exception MetaException if internal problem or bad input @@ -235,8 +234,9 @@ * * For a columnsetSerDe table, add column names to it * - * @param tableName - a valid existing table name - * @param columns - ordered list of column names + * @param dbName the database + * @param tableName a valid existing table name + * @param schema The schema information for the table * @exception TException if thrift problem * @exception MetaException if internal problem or bad input * @exception UnknownTableException if table does not exist already @@ -267,8 +267,8 @@ * * Create names columns for a columnset type table * + * @param dbName a valid database name * @param tableName - a valid table name - * @param columns - ordered list of column names * @exception TException if thrift problem * @exception MetaException if internal problem or bad input */ @@ -364,8 +364,8 @@ * * drop a table * - * @param tableName - a valid existing table name - * @param delete_data - should the store auto delete the data. + * @param dbName the name of a database + * @param tableName a valid existing table name * @exception TException if thrift problem * @exception MetaException if internal problem or bad input * @exception UnknownTableException if table does not exist already @@ -391,8 +391,8 @@ * * drop a table * - * @param tableName - a valid existing table name - * @param delete_data - should the store auto delete the data. + * @param dbName a valid database name + * @param tableName a valid existing table name * @exception TException if thrift problem * @exception MetaException if internal problem or bad input * @exception UnknownTableException if table does not exist already @@ -418,7 +418,7 @@ * * Gets the (opaque) schema which is currently represented as a key=>value map. * - * @param name - the name of the table + * @param tableName - the name of the table * @return the key/value of the opaque schema * @exception MetaException if internal problem * @exception UnknownTableException if the table doesn't exist Index: metastore/src/java/org/apache/hadoop/hive/metastore/FileStore.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/FileStore.java (revision 763135) +++ metastore/src/java/org/apache/hadoop/hive/metastore/FileStore.java (working copy) @@ -178,7 +178,7 @@ * * Looks at metastore directories * - * @param db + * @param parent * @param tablePattern * @return the list of tables * @exception MetaException Index: metastore/build.xml =================================================================== --- metastore/build.xml (revision 763135) +++ metastore/build.xml (working copy) @@ -90,4 +90,33 @@ + + + + + + + + + + + + + + + + + + + + Index: ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java (working copy) @@ -83,7 +83,7 @@ * * @param name the name of this table in the metadb * @param schema an object that represents the schema that this SerDe must know - * @param serDe a Class to be used for serializing and deserializing the data + * @param deserializer a Class to be used for deserializing the data * @param dataLocation where is the table ? (e.g., dfs://hadoop001.sf2p.facebook.com:9000/user/facebook/warehouse/example) NOTE: should not be hardcoding this, but ok for now * * @exception HiveException on internal error. Note not possible now, but in the future reserve the right to throw an exception @@ -317,7 +317,7 @@ } /** - * @param serDe the serDe to set + * @param deserializer the deserializer to set */ public void setDeserializer(Deserializer deserializer) { this.deserializer = deserializer; @@ -501,7 +501,7 @@ } /** - * @return + * @return The owner of the table. * @see org.apache.hadoop.hive.metastore.api.Table#getOwner() */ public String getOwner() { @@ -509,7 +509,7 @@ } /** - * @return + * @return The table parameters. * @see org.apache.hadoop.hive.metastore.api.Table#getParameters() */ public Map getParameters() { @@ -517,7 +517,7 @@ } /** - * @return + * @return The retention on the table. * @see org.apache.hadoop.hive.metastore.api.Table#getRetention() */ public int getRetention() { Index: ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java (working copy) @@ -62,7 +62,7 @@ private LinkedHashMap spec; /** - * @return + * @return The values of the partition * @see org.apache.hadoop.hive.metastore.api.Partition#getValues() */ public List getValues() { Index: ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (working copy) @@ -87,11 +87,11 @@ }; /** - * Returns hive object for the current thread. If one is not initialized then a new one is created + * Gets hive object for the current thread. If one is not initialized then a new one is created * If the new configuration is different in metadata conf vars then a new one is created. * @param c new Hive Configuration * @return Hive object for current thread - * @exception + * @throws HiveException * */ public static Hive get(HiveConf c) throws HiveException { @@ -114,7 +114,7 @@ * get a connection to metastore. see get(HiveConf) function for comments * @param c new conf * @param needsRefresh if true then creates a new one - * @return + * @return The connection to the metastore * @throws HiveException */ public static Hive get(HiveConf c, boolean needsRefresh) throws HiveException { @@ -471,8 +471,8 @@ /** * @param name - * @param location_uri - * @return + * @param locationUri + * @return true or false * @throws AlreadyExistsException * @throws MetaException * @throws TException @@ -485,7 +485,7 @@ /** * @param name - * @return + * @return true or false * @throws MetaException * @throws TException * @see org.apache.hadoop.hive.metastore.HiveMetaStoreClient#dropDatabase(java.lang.String) @@ -500,7 +500,6 @@ * - If he partition does not exist - one is created * - files in loadPath are moved into Hive. But the directory itself is not removed. * - * @param jc Job configuration * @param loadPath Directory containing files to load into Table * @param tableName name of table to be loaded. * @param partSpec defines which partition needs to be loaded @@ -524,7 +523,6 @@ * - If table does not exist - an exception is thrown * - files in loadPath are moved into Hive. But the directory itself is not removed. * - * @param jc Job configuration * @param loadPath Directory containing files to load into Table * @param tableName name of table to be loaded. * @param replace if true - replace files in the table, otherwise add files to table @@ -583,7 +581,7 @@ /** * Returns partition metadata - * @param tableName name of the partition's table + * @param tbl the partition's table * @param partSpec partition keys and values * @param forceCreate if this is true and partition doesn't exist then a partition is created * @return result partition object or null if there is no partition Index: ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRProcContext.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRProcContext.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMRProcContext.java (working copy) @@ -103,8 +103,6 @@ * @param rootTasks root tasks for the plan * @param mvTask the final move task * @param scratchDir directory for temp destinations - * @param randomId identifier used for temp destinations - * @param pathId identifier used for temp destinations * @param mapCurrCtx operator to task mappings */ public GenMRProcContext ( @@ -240,7 +238,7 @@ } /** - * @param randomId identifier used for temp destinations + * @param randomid identifier used for temp destinations */ public void setRandomId(int randomid) { this.randomid = randomid; Index: ql/src/java/org/apache/hadoop/hive/ql/optimizer/unionproc/UnionProcessor.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/optimizer/unionproc/UnionProcessor.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/optimizer/unionproc/UnionProcessor.java (working copy) @@ -53,7 +53,7 @@ /** * Transform the query tree. For each union, store the fact whether both the * sub-queries are map-only - * @param pactx the current parse context + * @param pCtx the current parse context */ public ParseContext transform(ParseContext pCtx) throws SemanticException { // create a walker which walks the tree in a DFS manner while maintaining the operator stack. Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerInfo.java (working copy) @@ -126,7 +126,7 @@ /** * Returns true if the specified expression is pushdown candidate else false * @param expr - * @return + * @return true or false */ public boolean isCandidate(exprNodeDesc expr) { ExprInfo ei = exprInfoMap.get(expr); @@ -151,7 +151,7 @@ /** * Returns the alias of the specified expr * @param expr - * @return + * @return The alias of the expression */ public String getAlias(exprNodeDesc expr) { ExprInfo ei = exprInfoMap.get(expr); Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/ExprWalkerProcFactory.java (working copy) @@ -226,7 +226,7 @@ * @param opContext operator context used for resolving column references * @param op operator of the predicates being processed * @param preds - * @return + * @return The expression walker information * @throws SemanticException */ public static ExprWalkerInfo extractPushdownPreds(OpWalkerInfo opContext, Index: ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/ppd/OpProcFactory.java (working copy) @@ -279,7 +279,7 @@ /** * @param nd - * @param extractPushdownPreds + * @param ewi */ protected void logExpr(Node nd, ExprWalkerInfo ewi) { for (Entry> e : ewi.getFinalCandidates().entrySet()) { Index: ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java (working copy) @@ -187,7 +187,7 @@ /** * get hiveHitsory object which does structured logging - * @return + * @return The hive history object */ public HiveHistory getHiveHistory(){ return hiveHist; Index: ql/src/java/org/apache/hadoop/hive/ql/exec/UDF.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/exec/UDF.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/exec/UDF.java (working copy) @@ -60,7 +60,7 @@ /** * Sets the resolver * - * @param The method resolver to use for method resolution. + * @param rslv The method resolver to use for method resolution. */ public void setResolver(UDFMethodResolver rslv) { this.rslv = rslv; Index: ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/exec/Operator.java (working copy) @@ -307,7 +307,7 @@ * jobClose is called from the jobclient side once the job has completed * * @param conf Configuration with with which job was submitted - * @param succes whether the job was completed successfully or not + * @param success whether the job was completed successfully or not */ public void jobClose(Configuration conf, boolean success) throws HiveException { if(childOperators == null) Index: ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java (working copy) @@ -277,7 +277,7 @@ * Get the UDF method for the name and argumentClasses. * @param name the name of the UDF * @param argumentClasses - * @return + * @return The UDF method */ public static Method getUDFMethod(String name, List> argumentClasses) { Class udf = getUDFClass(name); @@ -298,7 +298,7 @@ * Get the UDAF evaluator for the name and argumentClasses. * @param name the name of the UDAF * @param argumentClasses - * @return + * @return The UDAF evaluator */ public static Class getUDAFEvaluator(String name, List> argumentClasses) { Class udf = getUDAF(name); Index: ql/src/java/org/apache/hadoop/hive/ql/exec/UDAF.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/exec/UDAF.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/exec/UDAF.java (working copy) @@ -81,7 +81,7 @@ /** * Sets the resolver * - * @param The method resolver to use for method resolution. + * @param rslv The method resolver to use for method resolution. */ public void setResolver(UDAFEvaluatorResolver rslv) { this.rslv = rslv; Index: ql/src/java/org/apache/hadoop/hive/ql/lib/Dispatcher.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/lib/Dispatcher.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/lib/Dispatcher.java (working copy) @@ -31,7 +31,7 @@ /** * Dispatcher function. * @param nd operator to process. - * @param Stack operator stack to process. + * @param stack operator stack to process. * @param nodeOutputs The argument list of outputs from processing other nodes that are * passed to this dispatcher from the walker. * @return Object The return object from the processing call. Index: ql/src/java/org/apache/hadoop/hive/ql/lib/PreOrderWalker.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/lib/PreOrderWalker.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/lib/PreOrderWalker.java (working copy) @@ -42,7 +42,6 @@ /** * Constructor - * @param ctx graph of operators to walk * @param disp dispatcher to call for each op encountered */ public PreOrderWalker(Dispatcher disp) { Index: ql/src/java/org/apache/hadoop/hive/ql/lib/DefaultGraphWalker.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/lib/DefaultGraphWalker.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/lib/DefaultGraphWalker.java (working copy) @@ -43,7 +43,6 @@ /** * Constructor - * @param ctx graph of operators to walk * @param disp dispatcher to call for each op encountered */ public DefaultGraphWalker(Dispatcher disp) { @@ -67,8 +66,8 @@ /** * Dispatch the current operator - * @param op operator being walked - * @param opStack stack of operators encountered + * @param nd node being walked + * @param ndStack stack of nodes encountered * @throws SemanticException */ public void dispatch(Node nd, Stack ndStack) throws SemanticException { Index: ql/src/java/org/apache/hadoop/hive/ql/io/FlatFileInputFormat.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/io/FlatFileInputFormat.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/io/FlatFileInputFormat.java (working copy) @@ -46,7 +46,7 @@ import org.apache.hadoop.io.compress.CompressionCodec; import org.apache.hadoop.util.ReflectionUtils; -/** An {@link InputFormat} for Plain files with {@link Deserializer} records */ +/** An {@link org.apache.hadoop.mapred.InputFormat} for Plain files with {@link Deserializer} records */ public class FlatFileInputFormat extends FileInputFormat> { /** Index: ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java (working copy) @@ -374,7 +374,7 @@ * Get the exprNodeDesc * @param name * @param children - * @return + * @return The expression node descriptor */ public static exprNodeDesc getFuncExprNodeDesc(String name, exprNodeDesc... children) { return getFuncExprNodeDesc(name, Arrays.asList(children)); Index: ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseNumericOp.java (working copy) @@ -30,7 +30,7 @@ /** * Constructor. * This constructor sets the resolver to be used for comparison operators. - * See {@link UDFMethodResolver} + * See {@link org.apache.hadoop.hive.ql.exec.UDFMethodResolver} */ public UDFBaseNumericOp() { super(null); Index: ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateSub.java (working copy) @@ -46,7 +46,8 @@ * NOTE: This is a subset of what MySQL offers as: * http://dev.mysql.com/doc/refman/5.1/en/date-and-time-functions.html#function_date-sub * - * @param date1 the date string in the format of "yyyy-MM-dd HH:mm:ss" or "yyyy-MM-dd". + * @param dateString1 the date string in the format of "yyyy-MM-dd HH:mm:ss" or "yyyy-MM-dd". + * @param days the number of days to subtract. * @return the date in the format of "yyyy-MM-dd". */ public String evaluate(String dateString1, int days) { Index: ql/src/java/org/apache/hadoop/hive/ql/udf/UDFJson.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/udf/UDFJson.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/udf/UDFJson.java (working copy) @@ -55,7 +55,7 @@ * Syntax not supported that's worth noticing: * '' : Zero length string as key * .. : Recursive descent - * @ : Current object/element + * &#064; : Current object/element * () : Script expression * ?() : Filter (script) expression. * [,] : Union operator Index: ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseCompare.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseCompare.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseCompare.java (working copy) @@ -25,7 +25,7 @@ /** * This constructor sets the resolver to be used for comparison operators. - * See {@link UDFMethodResolver} + * See {@link org.apache.hadoop.hive.ql.exec.UDFMethodResolver} */ public UDFBaseCompare() { super(null); Index: ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateAdd.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateAdd.java (revision 763135) +++ ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDateAdd.java (working copy) @@ -46,7 +46,8 @@ * NOTE: This is a subset of what MySQL offers as: * http://dev.mysql.com/doc/refman/5.1/en/date-and-time-functions.html#function_date-add * - * @param date1 the date string in the format of "yyyy-MM-dd HH:mm:ss" or "yyyy-MM-dd". + * @param dateString1 the date string in the format of "yyyy-MM-dd HH:mm:ss" or "yyyy-MM-dd". + * @param days The number of days to add. * @return the date in the format of "yyyy-MM-dd". */ public String evaluate(String dateString1, int days) { Index: ql/build.xml =================================================================== --- ql/build.xml (revision 763135) +++ ql/build.xml (working copy) @@ -186,5 +186,30 @@ todir="${deploy.dir}"/> + + + + + + + + + + + + + + + +