-
Notifications
You must be signed in to change notification settings - Fork 812
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add HDFS check using snakebite #551
Changes from 2 commits
d767032
6ff3a94
d2ae1e2
bc904f3
1e40685
1165b07
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,46 @@ | ||
from checks import AgentCheck | ||
|
||
import snakebite.client | ||
|
||
class HDFSCheck(AgentCheck): | ||
"""Report on free space and space used in HDFS. | ||
""" | ||
|
||
def check(self, instance): | ||
if 'namenode' not in instance: | ||
self.log.info('Missing key \'namenode\' in HDFSCheck config') | ||
return | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. instead of ```python
Doing so will display this message in the info page. |
||
|
||
hostport = instance['namenode'] | ||
if ':' in hostport: | ||
host, _, port = hostport.partition(':') | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. string.partition is not 2.4 compatible. Wouldn't it be easier to have a dedicated field in the config for the port ? |
||
port = int(port) | ||
else: | ||
host = hostport | ||
port = 8020 | ||
|
||
hdfs = snakebite.client.Client(host, port) | ||
stats = hdfs.df() | ||
# {'used': 2190859321781L, | ||
# 'capacity': 76890897326080L, | ||
# 'under_replicated': 0L, | ||
# 'missing_blocks': 0L, | ||
# 'filesystem': 'hdfs://hostname:port', | ||
# 'remaining': 71186818453504L, | ||
# 'corrupt_blocks': 0L} | ||
|
||
self.gauge('hdfs.used', stats['used']) | ||
self.gauge('hdfs.free', stats['remaining']) | ||
self.gauge('hdfs.capacity', stats['capacity']) | ||
self.gauge('hdfs.in_use', float(stats['used']) / float(stats['capacity'])) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. should this be surrounded by a try except to catch a possible ZeroDivisionError or the capacity will never be 0 ? (I don't really know hdfs) There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I can do that if you like. It's very unlikely for the capacity (in bytes) to be 0 -- that would probably only happen if the hadoop cluster were seriously broken. |
||
self.gauge('hdfs.under_replicated', stats['under_replicated']) | ||
self.gauge('hdfs.missing_blocks', stats['missing_blocks']) | ||
self.gauge('hdfs.corrupt_blocks', stats['corrupt_blocks']) | ||
|
||
if __name__ == '__main__': | ||
check, instances = HDFSCheck.from_yaml('./hdfs.yaml') | ||
for instance in instances: | ||
check.check(instance) | ||
print "Events: %r" % check.get_events() | ||
print "Metrics: %r" % check.get_metrics() | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,6 @@ | ||
init_config: | ||
# HDFS check does not require any init_config | ||
|
||
instances: | ||
# Each instance requires a namenode hostname:port | ||
- namenode: namenode.example.com:8020 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you make this import within the check method so it won't crash if the module is missing ?
Something like