Another couple of miscellaneous install/run instructions
Run time instructions:
a) We need to export HADOOP_HOME before we can run hcat.
b) We need to export HIVE_CONF_DIR to /etc/hcatalog (or wherever we've installed hcat's config before we can run the 'hive' commandline ('hcat' cmdline works okay)
a) DB connector jar : we need a jdbc connector jar that we use for connecting to whatever db we're using, and we also have to make sure a database server is externally set up for use. This needs to be reflected in hcat-env.sh (typically in /etc/hcatalog/) in the DBROOT parameter. ( for eg:, I have DBROOT=/home/sush/opt/mysql-connector-java-3.1.14 , and I installed it there from the tarball provided by mysql. No official rpm provided, and also license incompatible if provided for distributing, but one can set that up themselves to whatever db they desire)
b) cp /etc/hcatalog/proto-hive-site.xml /etc/hcatalog/hive-site.xml , hand-edit and replace parameters in allcaps to whatever value makes sense in that installation (DBHOSTNAME,PASSWORD,WAREHOUSE_DIR,KEYTAB_PATH) and/or hand-edit other parameters that are desired (jdbc connection url, sasl_enabled, etc)
c) verify HADOOP_HOME,DBROOT and USER in /etc/hcat-env.sh is to deployer's satisfaction.