Details
-
Sub-task
-
Status: Done
-
Major
-
Resolution: Done
-
None
-
None
-
None
Description
When attempting to deploy Metron in Ambari 2.7.3 with HDP 3.1 the following error occurs when deploying Metron on CentOS 7.
$ cd metron-deployment/development/centos7/ $ vagrant up ... TASK [ambari_config : Deploy cluster with Ambari; http://node1:8080] *********** The full traceback is: File "/tmp/ansible_LTBkrV/ansible_module_ambari_cluster_state.py", line 211, in main if not blueprint_exists(ambari_url, username, password, blueprint_name): File "/tmp/ansible_LTBkrV/ansible_module_ambari_cluster_state.py", line 334, in blueprint_exists blueprints = get_blueprints(ambari_url, user, password) File "/tmp/ansible_LTBkrV/ansible_module_ambari_cluster_state.py", line 315, in get_blueprints services = json.loads(r.content) File "/usr/lib64/python2.7/json/__init__.py", line 338, in loads return _default_decoder.decode(s) File "/usr/lib64/python2.7/json/decoder.py", line 366, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib64/python2.7/json/decoder.py", line 384, in raw_decode raise ValueError("No JSON object could be decoded") fatal: [node1]: FAILED! => { "changed": false, "invocation": { "module_args": { "blueprint_name": "metron_blueprint", "blueprint_var": { "groups": [ { "cardinality": 1, "components": [ { "name": "NAMENODE" }, { "name": "SECONDARY_NAMENODE" }, { "name": "RESOURCEMANAGER" }, { "name": "HISTORYSERVER" }, { "name": "ZOOKEEPER_SERVER" }, { "name": "NIMBUS" }, { "name": "STORM_UI_SERVER" }, { "name": "DRPC_SERVER" }, { "name": "HBASE_MASTER" }, { "name": "HBASE_CLIENT" }, { "name": "APP_TIMELINE_SERVER" }, { "name": "DATANODE" }, { "name": "HDFS_CLIENT" }, { "name": "NODEMANAGER" }, { "name": "YARN_CLIENT" }, { "name": "MAPREDUCE2_CLIENT" }, { "name": "ZOOKEEPER_CLIENT" }, { "name": "SUPERVISOR" }, { "name": "KAFKA_BROKER" }, { "name": "HBASE_REGIONSERVER" }, { "name": "KIBANA_MASTER" }, { "name": "METRON_INDEXING" }, { "name": "METRON_PROFILER" }, { "name": "METRON_PCAP" }, { "name": "METRON_ENRICHMENT_MASTER" }, { "name": "METRON_PARSERS" }, { "name": "METRON_REST" }, { "name": "METRON_MANAGEMENT_UI" }, { "name": "METRON_ALERTS_UI" }, { "name": "ES_MASTER" } ], "configurations": [], "name": "host_group_1" } ], "required_configurations": [ { "metron-env": { "es_hosts": "node1", "solr_zookeeper_url": "node1:9983", "storm_rest_addr": "http://node1:8744", "zeppelin_server_url": "node1:9995" } }, { "metron-rest-env": { "metron_jdbc_driver": "org.h2.Driver", "metron_jdbc_password": "root", "metron_jdbc_platform": "h2", "metron_jdbc_url": "jdbc:h2:file:~/metrondb", "metron_jdbc_username": "root" } }, { "kibana-env": { "kibana_default_application": "dashboard/AV-YpDmwdXwc6Ua9Muh9", "kibana_es_url": "http://node1:9200", "kibana_log_dir": "/var/log/kibana", "kibana_pid_dir": "/var/run/kibana", "kibana_server_host": "0.0.0.0", "kibana_server_port": 5000 } } ], "stack_name": "HDP", "stack_version": "3.1" }, "cluster_name": "metron_cluster", "cluster_state": "present", "configurations": [ { "zoo.cfg": { "dataDir": "/hadoop/zookeeper" } }, { "hadoop-env": { "dtnode_heapsize": 512, "hadoop_heapsize": 1024, "namenode_heapsize": 2048, "namenode_opt_permsize": "128m" } }, { "hbase-env": { "hbase_master_heapsize": 512, "hbase_regionserver_heapsize": 512, "hbase_regionserver_xmn_max": 512 } }, { "hdfs-site": { "dfs.datanode.data.dir": "/hadoop/hdfs/data", "dfs.journalnode.edits.dir": "/hadoop/hdfs/journalnode", "dfs.namenode.checkpoint.dir": "/hadoop/hdfs/namesecondary", "dfs.namenode.name.dir": "/hadoop/hdfs/namenode", "dfs.replication": 1 } }, { "yarn-env": { "apptimelineserver_heapsize": 512, "min_user_id": 500, "nodemanager_heapsize": 512, "resourcemanager_heapsize": 1024, "yarn_heapsize": 512 } }, { "mapred-env": { "jobhistory_heapsize": 256 } }, { "mapred-site": { "mapreduce.jobhistory.recovery.store.leveldb.path": "/hadoop/mapreduce/jhs", "mapreduce.map.java.opts": "-Xmx1024m", "mapreduce.map.memory.mb": 1229, "mapreduce.reduce.java.opts": "-Xmx1024m", "mapreduce.reduce.memory.mb": 1229 } }, { "yarn-site": { "yarn.nodemanager.local-dirs": "/hadoop/yarn/local", "yarn.nodemanager.log-dirs": "/hadoop/yarn/log", "yarn.nodemanager.resource.memory-mb": 4096, "yarn.timeline-service.leveldb-state-store.path": "/hadoop/yarn/timeline", "yarn.timeline-service.leveldb-timeline-store.path": "/hadoop/yarn/timeline" } }, { "storm-site": { "nimbus.childopts": "-Xmx1024m _JAAS_PLACEHOLDER", "storm.cluster.metrics.consumer.register": "[{\"class\": \"org.apache.storm.metric.LoggingMetricsConsumer\"}]", "storm.local.dir": "/hadoop/storm", "supervisor.childopts": "-Xmx256m _JAAS_PLACEHOLDER", "supervisor.slots.ports": "[6700, 6701, 6702, 6703, 6704, 6705]", "topology.classpath": "/etc/hbase/conf:/etc/hadoop/conf", "topology.metrics.consumer.register": "[{\"class\": \"org.apache.storm.metric.LoggingMetricsConsumer\", \"parallelism.hint\": 1, \"whitelist\": [\"kafkaOffset\\\\..+/\", \"__complete-latency\", \"__process-latency\", \"__receive\\\\.population$\", \"__sendqueue\\\\.population$\", \"__execute-count\", \"__emit-count\", \"__ack-count\", \"__fail-count\", \"memory/heap\\\\.usedBytes$\", \"memory/nonHeap\\\\.usedBytes$\", \"GC/.+\\\\.count$\", \"GC/.+\\\\.timeMs$\"]}]", "worker.childopts": "-Xmx768m _JAAS_PLACEHOLDER" } }, { "kafka-env": { "content": "\n#!/bin/bash\n\n# Set KAFKA specific environment variables here.\n\n# The java implementation to use.\nexport KAFKA_HEAP_OPTS=\"-Xms256M -Xmx256M\"\nexport KAFKA_JVM_PERFORMANCE_OPTS=\"-server -XX:+UseG1GC -XX:+DisableExplicitGC -Djava.awt.headless=true\"\nexport JAVA_HOME={{java64_home}}\nexport PATH=$PATH:$JAVA_HOME/bin\nexport PID_DIR={{kafka_pid_dir}}\nexport LOG_DIR={{kafka_log_dir}}\nexport KAFKA_KERBEROS_PARAMS={{kafka_kerberos_params}}\n# Add kafka sink to classpath and related depenencies\nif [ -e \"/usr/lib/ambari-metrics-kafka-sink/ambari-metrics-kafka-sink.jar\" ]; then\n export CLASSPATH=$CLASSPATH:/usr/lib/ambari-metrics-kafka-sink/ambari-metrics-kafka-sink.jar\n export CLASSPATH=$CLASSPATH:/usr/lib/ambari-metrics-kafka-sink/lib/*\nfi\nif [ -f /etc/kafka/conf/kafka-ranger-env.sh ]; then\n . /etc/kafka/conf/kafka-ranger-env.sh\nfi" } }, { "kafka-broker": { "delete.topic.enable": "true", "log.dirs": "/kafka-log", "offsets.topic.replication.factor": 1 } }, { "metron-rest-env": { "metron_spring_profiles_active": "dev" } }, { "metron-parsers-env": { "parsers": "\"bro,snort,yaf\"" } }, { "elastic-site": { "gateway_recover_after_data_nodes": 1, "index_number_of_replicas": 0, "index_number_of_shards": 1, "masters_also_are_datanodes": "1", "network_host": "[ _local_, _site_ ]", "zen_discovery_ping_unicast_hosts": "[ node1 ]" } } ], "host": "node1", "password": "admin", "port": 8080, "username": "admin", "wait_for_complete": true } }, "msg": "Ambari client exception occurred: No JSON object could be decoded" } to retry, use: --limit @/Users/nallen/Development/metron/metron-deployment/development/centos7/ansible/playbook.retry PLAY RECAP ********************************************************************* node1 : ok=83 changed=16 unreachable=0 failed=1 [WARNING]: Module did not set no_log for password Ansible failed to complete successfully. Any error output should be visible above. Please fix these errors and try again.
Attachments
Issue Links
- links to