An open monitoring distribution (OMD) puppet configuration module
This module is intended to configure, install and manage an OMD distribution.
A note on the installation: the default is now to directly download a Check-mk Raw Edition from checkmk.com (the consol.labs editions seem to have dropped check_mk support). Consol.labs repos configs is left in the modules if you want to use them.
This module borrows files from check_mk for the agent setup : those file license is under manifests/check_mk/agent .
Last note about OMD upgrades :
- if not using the repositories, a file is downloaded in the puppet cache dir for package install. Remember to cleanup old files, (but keep the current one)
- after omd updates, an "omd site upgrade" is necessary : this is not managed by this module
include omd
class {omd: with_repos => true}
omd::site {'test': }
omd::site {'irfu': manage_crontabs=> true}
This will install the 'test' site and change the default http (htaccess) password for the user
omd::check_mk::multisite::site { 'remotehost.example.org': site => 'irfu', alias=>'remote1' }
The % char actually tells the defined resource to strip everything after it, so that the omdadmin user, which is the real username can be used for both OMD sites. This unfortunately is required because of puppet not allowing duplicate defined resources titles
omd::site::user {'omdadmin': site=>'test', ensure=>absent}
omd::site::user {'omdadmin%2': site=>'irfu', ensure=>absent}
This use will be added with a custom password to the htaccess and given admin rights in multisite (password : "changeme")
omd::site::user {'irfu': site=>'irfu', privileges=>'admin', pass_md5=>'HVSJ.h631LfUw'}
Please note this requires you manage the apache module with -say- puppetlabs-apache
class { omd::ssl::vhost: sites=>[test], port => 443}
This will import everything for the OMD site irfu + all globally explorted resources
omd::check_mk::import_resources{'irfu':}
####Create custom check_mk variables
they are added before any inventory action, but beware that some variables only influence further inventories
omd::check_mk::var::set {ignored_checktypes: site=>'test', content=>'[ "ipmi_sensors", "if", "if64" ,"nfsmounts"]'}
omd::check_mk::var::append {ignored_checktypes: site=>'test', content=>'["ipmi"]'}
omd::check_mk::var::set {inventory_df_exclude_fs: site=>'test', content=>'["nfs", "fuse"]'}
omd::check_mk::var::set {'filesystem_default_levels#levels': site=>'test', content=>'(80, 90)'}
omd::check_mk::var::set {'filesystem_default_levels#levels_low': site=>'test', content=>'(60,75)'}
omd::check_mk::var::set {'filesystem_default_levels#magic': site=>'test', content=>'0.5' }
omd::check_mk::var::set {'ntp_default_levels': site=>'test', content=>'(10, 200.0, 1000.0)'}
check parameters...
#FS check
omd::check_mk::var::set {check_parameters: site=>'test', content=>'[( (85,90,0.5), ALL_HOSTS, [ "fs_/var","fs_/tmp","fs_/home",])]'}
omd::check_mk::var::append {'check_parameters|1': site=>'test', content=>'[( (90,95,1) , ALL_HOSTS, [ "fs_/$",])]'}
#specific setup for wn /var
omd::check_mk::var::append {'check_parameters|2': site=>'test', content=>'[( (90,95,0.5), [ "wn", ], ALL_HOSTS, [ "fs_/var"])]'}
#MEMORY usage over 110/150% of RAM
omd::check_mk::var::append {'check_parameters|3': site=>'test', content=>'[( (110.0, 150.0),ALL_HOSTS, [ "mem.used",] )]'}
#don't use the deprecated network check
omd::check_mk::var::set {linux_nic_check: site=>'test', content=>'"lnx_if"'}
Performance data...
#these are related, so make sure the first one is created before the second one
omd::check_mk::var::set {'_disabled_graphs': site=>'test', content=>'["fs_","Kernel","NTP","Number of threads","Uptime"]' , concat_order=> '009'}
omd::check_mk::var::set {"extra_service_conf#process_perf_data": site=>'test', content=>'[ ( "0", ALL_HOSTS, _disabled_graphs ), ]'}
host groups
omd::check_mk::var::set {define_hostgroups: site=>'test', content=>"'true'"}
omd::check_mk::var::set {host_groups: site=>'test', content=>"[( 'wn', [ 'wn' ], ALL_HOSTS )]"} #association des tags à des hostgroups
omd::check_mk::var::append {'host_groups|1': site=>'test', content=>"[( 'se', [ 'dpm_disk' ], ALL_HOSTS )]"}
omd::check_mk::var::append {'host_groups|2': site=>'test', content=>"[( 'se', [ 'dpm_head' ], ALL_HOSTS )]"}
omd::check_mk::var::append {'host_groups|3': site=>'test', content=>"[( 'other', [ '!wn', '!dpm_disk' ], ALL_HOSTS )]"}
Manually add hosts to the config :
omd::check_mk::addhost{'10.2.5.8': site=> test, tags=>'snmp|Xtreme|BD-8810'}
append to arbitrary var
omd::check_mk::var::set {'_toto': site=>'test', content=>'{"a" : 1}' }
omd::check_mk::var::append {'_toto#a|1': site=>'test', content=>'1' }
Add host static aliases
# we have to set the variable empty first, in order to avoid breaking check_mk - until it's used, it does not exist and we can't append.
# or we could set it for the forst host, and append for others.
# or we could set everything in one line (or not)
omd::check_mk::var::set {'extra_host_conf#alias': site=>'test', content=>'[]' }
omd::check_mk::var::append {'extra_host_conf#alias|1': site=>'test', content=>'[("myalias" , ["my real hostname"])]' }
create a legacy nagios/check_mk service adding it to host tags
before cmk 1.6 :
omd::check_mk::legacy::service {'ldap': site=> 'irfu', command =>'check_tcp!2170', mk_tags=>['ce','cream','dpm_head','bdii'], perfdata=>true}
Now with cmk 1.6 :
omd::check_mk::legacy::service {'ldap': site=> 'irfu', command =>'check_tcp -H $HOSTADDRESS$ -p 2170', mk_tags=>['bdii']}
create a nagios service adding it to a specific host
omd::check_mk::legacy::service {'backbone': site=> 'irfu', command =>'check_snmp -P 2c -C public -o ifHCInOctets.3003,ifHCOutOctets.3003 -u InBytes,OutBytes -l bandwidth', mk_hosts=>['10.2.5.8'],}
omd::check_mk::server::plugins::puppet { 'irfu': }
If you have nodes that should not be monitored anymore using the puppetdb (removed nodes), you can remove them from the monitoring using this command :
puppet node clean --unexport <nodename>
puppet node deactivate <nodename>
If you happen to be using foreman for managing your nodes, take a look at the puppetdb_foreman plugin (for foreman), which is aslo shiped as an RPM (ruby193-rubygem-puppetdb_foreman) by the foreman team : https://github.com/theforeman/puppetdb_foreman
Ldap connections in check_mk allow a lot of different configurations. Thus, setting up ldap requires the user to know exactly what he needs. This is an example of how to set up a connection to a ActiveDirectory Server using the standard port and Version 3:
First, we need to define the server address & login credentials. The name needs to be the site name (at moment, only one ldap connection is allowed with check_mk:
omd::site::ldap { 'irfu':
server => 'intra.irfu.com',
bind_dn => 'admin@intra.irfu.com',
bind_pw => 'password',
A user base dn, user filter and user scope can be used to get the right users. 'uid_umlauts' defines, how umlauts should be handled:
user_dn => 'OU=Users,DC=intra,DC=irfu,DC=com',
user_filter => '(&(objectclass=user)(objectcategory=person))',
user_scope => 'sub', # sub means the whole subtree, 'base' (only this dn) and 'one' (go one more down) are also available
uid_umlauts => 'replace', # Default setting here is 'keep'
The next thing is to activate define the base dn for users & groups and the scope. If required, a group-filter can also be used
group_dn => 'OU=Groups,DC=intra,DC=irfu,DC=com',
group_filter => '(&(objectclass=group)(CN=monitoring*))',
group_scope => 'sub', # sub is the default, so this setting wouldn't be needed. Same for user_scope
Additional settings can be applied like default roles & contact groups:
default_user_roles => ['guest'],
default_user_cgroups => ['first_level'],
Defining the active plugins is a little more tricky. The wato allows various different settings, so this is handled in only one big string. The following example manually defines the alias and maps groups to roles:
active_plugins => " 'alias' : { 'attr' : 'userPrincipalName' },
'groups_to_toles' : {
'admin' : u'CN=monitoring-admins,OU=Groups,DC=intra,DC=irfu,DC=com',
'user' : u'CN=monitoring-users,OU=Groups,DC=intra,DC=irfu,DC=com',
'nested' : True, # This sets nested groups (local-global) to true
},",
If ssl is required, add the following:
use_ssl => true,
Install the check_mk agent. This will start/enable the xinetd check_mk service :
#using hiera for specifying the IP whitelist
include omd::check_mk::agent
Add a check_mk tag "nagios" on a host. You can now specify the monitoring network that should be used when exporting the node ipaddress to check_mk, in order to avoid using the ::ipaddress fact in case multiple networks are configured on a host (you should probably use hiera for that...) .
omd::check_mk::addtag{nagios:}
or
ensure_resource( 'omd::check_mk::addtag' ,'nagios',{})
or
omd::check_mk::addtag{multi_networks: monitoring_network=> '10.0.0.0', monitoring_netmask: '255.255.255.0' }
or add in hierra :
# omd::monitoring_network: 10.0.0.0
# omd::monitoring_netmask: 255.255.255.0
Add an MRPE test that will be automatically inventoried ( and exporting a check_mk tag) :
omd::check_mk::mrpe::check{'hung_nrpe': command=>'check_procs -w :5 -C nrpe'}
Make sure your host which doesn't have DNS resolution but is reachable is monitored correctly :
# helps find out the host monitoring address using its network address
omd::monitoring_network: 1.2.3.0
omd::monitoring_netmask: 255.255.255.128
# use the omd discovered "override_ip"
omd::brokendns: true
include omd::check_mk::plugins::puppet
- use the check_mk web api to create hosts, which will allow for wato usage
- setup github CI/CD for module validation
- refactor/rework
- make all classes and files "puppet5 native", with params types, and remove legacy still working things.
- see if we can remove the binary "waitmax" from the module
This implements exported resources and collectors for check_mk tags as explained in this blog, but adapted to OMD : http://blog.matsharpe.com/2013/01/puppet-checkmk.html
If you limit cron usage, please make sure to allow the OMD users, for instance, add :
cron::allow {irfu: } -> omd::site {'irfu': }
And probably also , since puppet now defaults to purging cron entries (unless_* params do not seem to work at all) :
resources{'cron':
purge => false,
}
Working on : CentOS 7