These scripts help to collect Hadoop jmx and evently sent the metrics to stdout or Kafka.
-
edit the configuration file
cronus.ini
(json file) -
run the scripts
python hadoop_jmx_kafka.py > 1.txt
python cronus_hadoop_jmx.py
-
input: "port" defines the hadoop service port, such as 50070 => "namenode", 60010 => "hbase master".
-
filter: "monitoring.group.selected" can filter out beans which we care about.
-
output: if we left it empty, then the output is stdout by default.
"output": {}
It also supports Kafka as its output.
"output": { "kafka": { "topic": "apollo-phx_cronus_nn_jmx", "brokerList": [ "druid-test-host1-556191.slc01.dev.ebayc3.com:9092", "druid-test-host2-550753.slc01.dev.ebayc3.com:9092", "druid-test-host3-550755.slc01.dev.ebayc3.com:9092"] } }
{
"env": {
"site": "apollo-phx",
"cluster": "apollo",
"datacenter": "phx"
},
"input": {
"port": "50070",
"https": false
},
"filter": {
"monitoring.group.selected": ["hadoop"]
},
"output": {
}
}