Index: src/docbkx/book.xml =================================================================== --- src/docbkx/book.xml (revision 1170801) +++ src/docbkx/book.xml (working copy) @@ -1703,6 +1703,23 @@ General + When should I use HBase? + + + Anybody can download and give HBase a spin, even on a laptop. The scope of this answer is when + would it be best to use HBase in a real deployment. + + First, make sure you have enough hardware. Even HDFS doesn't do well with anything less than + 5 DataNodes (due to things such as HDFS block replication which has a default of 3), plus a NameNode. + Second, make sure you have enough data. HBase isn't suitable for every problem. If you have + hundreds of millions or billions of rows, then HBase is a good candidate. If you only have a few + thousand/million rows, then using a traditional RDBMS might be a better choice due to the + fact that all of your data might wind up on a single node (or two) and the rest of the cluster may + be sitting idle. + + + + Are there other HBase FAQs? @@ -1738,18 +1755,6 @@ - - Why are logs flooded with '2011-01-10 12:40:48,407 INFO org.apache.hadoop.io.compress.CodecPool: Got - brand-new compressor' messages? - - - Because we are not using the native versions of compression - libraries. See HBASE-1900 Put back native support when hadoop 0.21 is released. - Copy the native libs from hadoop under hbase lib dir or - symlink them into place and the message should go away. - - - EC2 @@ -1796,6 +1801,18 @@ + + Why are logs flooded with '2011-01-10 12:40:48,407 INFO org.apache.hadoop.io.compress.CodecPool: Got + brand-new compressor' messages? + + + Because we are not using the native versions of compression + libraries. See HBASE-1900 Put back native support when hadoop 0.21 is released. + Copy the native libs from hadoop under hbase lib dir or + symlink them into place and the message should go away. + + + How do I...?