Skip to content

Latest commit

 

History

History
172 lines (124 loc) · 6.34 KB

README.md

File metadata and controls

172 lines (124 loc) · 6.34 KB

This is a fork of rochaporto's version, which seems to have been abandoned around 2015, with plenty of bugs and pull requests lying around. I have tried to incorporate most of the outstanding pulls and some of the interesting looking but not overly conflicting forks.

If you have a clean pull request I will eventually try to merge in. I don't promise to fix issues but if anyone do I'll merge it.

I don't really want to maintain this but I don't want to let it lying bug-ridden, I want to use it.

collectd-ceph

Overview

A set of collectd plugins monitoring and publishing metrics for Ceph components.

Screenshots

Sample Ceph Overview dashboard displaying common metrics from the plugins.

image

image

image

Sample Ceph Pool Capacity dashboard, if you are using pool quota this sample may be usefull.

image

Check here for the Ceph Overview dashboard definition.

Check here for the Ceph Pool Capacity dashboard definition.

Plugins and Metrics

There are several plugins, usually mapping to the ceph command line tools.

Find below a list of the available plugins and the metrics they publish.

  • ceph_monitor_plugin
    • ceph-<cluster>.mon.gauge.number (total number of monitors)
    • ceph-<cluster>.mon.gauge.quorum (number of monitors in quorum)
  • ceph_osd_plugin
    • ceph-<cluster>.osd.gauge.up (number of osds 'up')
    • ceph-<cluster>.osd.gauge.down (number of osds 'down')
    • ceph-<cluster>.osd.gauge.in (number of osds 'in')
    • ceph-<cluster>.osd.gauge.out (number of osds 'out')
  • ceph_pool_plugin
    • ceph-<cluster>.pool-<name>.gauge.size (per pool size)
    • ceph-<cluster>.pool-<name>.gauge.min_size (per pool min_size)
    • ceph-<cluster>.pool-<name>.gauge.pg_num (per pool pg_num)
    • ceph-<cluster>.pool-<name>.gauge.pgp_num (per pool pg_placement_num)
    • ceph-<cluster>.pool-<name>.gauge.quota_max_bytes (per pool quota_max_bytes)
    • ceph-<cluster>.pool-<name>.gauge.quota_max_objects (per pool quota_max_objects)
    • ceph-<cluster>.pool-<name>.gauge.max_avail (per pool max_available)
    • ceph-<cluster>.pool-<name>.gauge.objects (per pool objects number)
    • ceph-<cluster>.pool-<name>.gauge.objects (per pool objects number)
    • ceph-<cluster>.pool-<name>.gauge.read_bytes_sec (per pool read bytes/sec)
    • ceph-<cluster>.pool-<name>.gauge.write_bytes_sec (per pool write bytes/sec)
    • ceph-<cluster>.pool-<name>.gauge.op_per_sec (per pool iops)
    • ceph-<cluster>.pool-<name>.gauge.bytes_used (per pool bytes used)
    • ceph-<cluster>.pool-<name>.gauge.kb_used (per pool KBytes used)
    • ceph-<cluster>.pool-<name>.gauge.objects (per pool number of objects) min_size
    • ceph-<cluster>.cluster.gauge.total_avail (cluster space available)
    • ceph-<cluster>.cluster.gauge.total_space (cluster total raw space)
    • ceph-<cluster>.cluster.gauge.total_used (cluster raw space used)
  • ceph_pg_plugin
    • ceph-<cluster>.pg.gauge.<state> (number of pgs in <state>)
    • ceph-<cluster>.osd-<id>.gauge.fs_commit_latency (fs commit latency for osd)
    • ceph-<cluster>.osd-<id>.gauge.apply_commit_latency (apply commit latency for osd)
    • ceph-<cluster>.osd-<id>.gauge.kb_used (kb used by osd)
    • ceph-<cluster>.osd-<id>.gauge.kb (total space of osd)
  • ceph_latency_plugin
    • ceph-<cluster>.cluster.gauge.avg_latency (avg cluster latency)
    • ceph-<cluster>.cluster.gauge.max_latency (max cluster latency)
    • ceph-<cluster>.cluster.gauge.min_latency (min cluster latency)
    • ceph-<cluster>.cluster.gauge.stddev_latency (stddev of cluster latency)

Requirements

It assumes an existing installation of collectd - check docs for details.

If you want to publish to graphite, configure the write_graphite collectd plugin.

And you might want the awesome grafana too, which provides awesome displays.

Setup and Configuration

The example configuration(s) below assume the plugins to be located under /usr/lib/collectd/plugins/ceph.

If you're under ubuntu, consider installing from this ppa.

Each plugin should have its own config file, under /etc/collectd/conf.d/<pluginname>.conf, which should follow something similar to:

# cat /etc/collectd/conf.d/ceph_pool.conf

<LoadPlugin "python">
    Globals true
</LoadPlugin>

<Plugin "python">
    ModulePath "/usr/lib/collectd/plugins/ceph"

    Import "ceph_pool_plugin"

    <Module "ceph_pool_plugin">
        Verbose "True"
        Cluster "ceph"
        Interval "60"
        TestPool "test"
    </Module>
</Plugin>

Puppet

If you use puppet for configuration, then try this excelent collectd module.

It has plenty of docs on how to use it, but for our specific plugins:

  collectd::plugin::python { 'ceph_pool':
    modulepath => '/usr/lib/collectd/plugins/ceph',
    module     => 'ceph_pool_plugin',
    config     => {
      'Verbose'  => 'true',
      'Cluster'  => 'ceph',
      'Interval' => 60,
      'TestPool' => 'test',
    },
  }

Docker

Check this repo for a nice docker setup to run collectd-ceph (thanks to Ian Babrou).

Limitations

The debian packaging files are provided, but not yet available in the official repos.

Development

All contributions more than welcome, just send pull requests.

License

GPLv2 (check LICENSE).

Contributors

Support

Please log tickets and issues at the github home.

Additional Notes

Some handy instructions on how to build for ubuntu.