Hi, author of Fifemon here. Do you really want to run Supervisor and the probes as root? We run as much as possible in user space. See http://supervisord.org/running.html#runtime-security.
If you don't want to deal with supervisor, you can instead run the probe with cron by passing the '--once' flag to 'condor_probe.py'. Or, here's a simple systemd unit file you could use (untested; change user, group, and command as appropriate): https://gist.github.com/retzkek/0dcd7b19548de96531bd5362c51ee2a6 Init script is left as an exercise
Please let me know if you have any other questions or issues!
Regards, Kevin Retzke From: HTCondor-users <htcondor-users-bounces@xxxxxxxxxxx> on behalf of Uchenna Ojiaku - NOAA Affiliate <uchenna.ojiaku@xxxxxxxx>
Sent: Friday, February 24, 2017 1:20 PM To: htcondor-users@xxxxxxxxxxx Subject: [HTCondor-users] Grafana + HTCondor Job Metrics Hello,Has anyone configured grafana to display htcondor job/slot metrics? I found this online. But my supervisord isn't working. I get this is error "Supervisord is running as root and it is searching" So I try killing any running supervisord processes but I still
get the same error. Then I tried unlinking with '
FifemonCollect HTCondor statistics and report into time-series database. All modules support Graphite, and there is some support for InfluxDB. Additionally report select job and slot Classads into Elasticsearch via Logstash. Note: this is a fork of the scripts used for monitoring the HTCondor pools at Fermilab, and while generally intended to be "generic" for any pool still may require some tweaking to work well for your pool. Copyright Fermi National Accelerator Laboratory (FNAL/Fermilab). See LICENSE.txt. Requirements
For current job and slot state:
InstallationAssuming HTCondor and Python virtualenv packages are already installed:
Optionally, for crash mails:
ConfigurationCondor metrics probeExample probe config is in
SupervisorExample supervisor config is in Job and slot stateThe scripts that collect raw job and slot records into elasticsearch are much simpler than the metrics probe - simply point at your pool with --pool and JSON records are output to stdout. We use logstash to pipe the output to Elasticsearch; see
RunningUsing supervisor:
If using influxdb:
Start supervisor:
|