Skip to content

Puppet module to manage repositories that are needed by Icinga and other base class classes like icinga::redis.

License

Notifications You must be signed in to change notification settings

voxpupuli/puppet-icinga

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

icinga

Build Status Release Puppet Forge puppet integration Apache-2.0 License Donated by Icinga Sponsored by NETWAYS

Icinga Logo

Table of Contents

  1. Description
  2. Setup - The basics of getting started with icinga
  3. Usage - Configuration options and additional functionality
  4. Reference

Description

This module provides several non private helper classes for the other official Icinga modules:

How to use the classes for Icinga Web or any database use on Ubuntu Noble

To get Icinga Web 2 running on Ubutunt Noble use puppet-php >=8.3.0 and set:

php::globals::php_version: '8.3'

The current MariaDB logs to syslog by default so set:

mysql::server::override_options:
  mysqld:
    log-error: ~

This disables the logging to file and the requirement and management of an existing directory /var/log/mysql.

If using PostgreSQL you have to set the version to '16':

---
postgresql::globals::version: '16'

How to use the classes for Icinga Web or databases with MariaDB on Debian Bookwork

To get Icinga Web 2 running on Debian Bookworm use puppet-php >=8.2.0 (no longer necessary if puppet-php >= 10.2.0 is used) and set:

php::globals::php_version: '8.2'

The current MariaDB logs to syslog by default so set:

mysql::server::override_options:
  mysqld:
    log-error: ~

This disables the logging to file and the requirement and management of an existing directory /var/log/mysql.

Setup

What the Icinga Puppet module supports

  • [icinga::repos] involves the needed repositories to install icinga2, icingadb and icingaweb2:
    • The Icinga Project repository for the stages: stable, testing or nightly builds
    • EPEL repository for RHEL simular platforms
    • Backports repository for Debian and Ubuntu
    • NETWAYS extras repository for Icinga Web 2
    • NETWAYS plugins repository with some additional monitoring plugins
  • Classes to manage and setup an Icinga environment much easier:
    • [icinga::server] setups an Icinga 2 including CA, config server, zones and workers aka satellites
    • [icinga::worker] installs an Icinga 2 worker aka satellite
    • [icinga::ido] configures the IDO backend including the database
    • [icinga::web] manages Icinga Web 2, an Apache and a PHP-FPM

Setup Requirements

The requirements depend on the class to be used.

Beginning with icinga

Add this declaration to your Puppetfile:

mod 'icinga',
  :git => 'https://github.com/icinga/puppet-icinga.git',
  :tag => 'v2.5.0'

Then run:

bolt puppetfile install

Or do a git clone by hand into your modules directory:

git clone https://github.com/icinga/puppet-icinga.git icinga

Change to icinga directory and check out your desired version:

cd icinga
git checkout v2.5.0

Usage

icinga::repos

The class supports:

  • [puppet] >= 7.9.0 < 9.0.0

And requires:

  • [puppetlabs/stdlib] >= 6.6.0 < 10.0.0
  • [puppetlabs/apt] >= 9.2.0 < 10.0.0
  • [puppet/zypprepo] >= 4.0.0 < 6.0.0
  • [puppetlabs/yumrepo_core] >= 1.1.0 < 3.0.0

By default the upstream Icinga repository for stable release are involved.

include icinga::repos

To setup the testing repository for release candidates use instead:

class { 'icinga::repos':
  manage_stable  => false,
  manage_testing => true,
}

Or the nightly builds:

class { 'icinga::repos':
  manage_stable  => false,
  manage_nightly => true,
}

Other possible needed repositories like EPEL on RHEL or the Backports on Debian can also be involved:

class { 'icinga::repos':
  manage_epel         => true,
  configure_backports => true,
}

The prefix configure means that the repository is not manageable by the module. But backports can be configured by the class apt::backports, that is used by this module.

Enable and Disable Repositories

When manage is set to true for a repository the ressource is managed and the repository is enabled by default. To switch off a repository again, it still has to be managed and the corresponding parameter has to set via hiera. The module does a deep merge lookup for a hash named icinga::repos. Allowed keys are:

  • icinga-stable-release
  • icinga-testing-builds
  • icinga-snapshot-builds
  • epel (only on RHEL platforms)
  • powertools (only RHEL 8 platforms)
  • crb (only RHEL 9 platforms)
  • netways-plugins
  • netways-extras

An example for Yum or Zypper based platforms to change from stable to testing repo:

---
icinga::repos::manage_testing: true
icinga::repos:
  icinga-stable-release:
    enabled: 0

Or on Apt based platforms:

---
icinga::repos::manage_testing: true
icinga::repos:
  icinga-stable-release:
    ensure: absent

Configure Icinga subscription repositories

For some time now, access to current RPM packages on Icinga has required a paid subscription. Unfortunately, using older package versions for an Icinga server is not provided for in this project.

A subscription is required, it is configured as follows, e.g. in hiera:

---
icinga::repos:
  icinga-stable-release:
    baseurl: 'https://packages.icinga.com/subscription/rhel/$releasever/release/'
    username: <username>
    password: <password>

Installing from Non-Upstream Repositories

To change to a non upstream repository, e.g. a local mirror, the repos can be customized via hiera. The module does a deep merge lookup for a hash named icinga::repos. Allowed keys are:

  • icinga-stable-release
  • icinga-testing-builds
  • icinga-snapshot-builds
  • epel (only on RHEL Enterprise platforms)
  • powertools (only RHEL 8 platforms)
  • crb (only RHEL 9 platforms)
  • netways-plugins
  • netways-extras

An example to configure a local mirror of the stable release:

---
icinga::repos:
  icinga-stable-release:
    baseurl: 'https://repo.example.com/icinga/epel/$releasever/release/'
    gpgkey: https://repo.example.com/icinga/icinga.key

IMPORTANT: The configuration hash depends on the platform an requires one of the following resources:

Also the Backports repo on Debian can be configured like the apt class of course, see https://forge.puppet.com/puppetlabs/apt to configure the class apt::backports via Hiera.

As an example, how you configure backports on a debian squeeze. For squeeze the repository is already moved to the unsupported archive:

---
apt::confs:
  no-check-valid-until:
    content: 'Acquire::Check-Valid-Until no;'
    priority: 99
    notify_update: true
apt::backports::location: 'https://archive.debian.org/debian'

Setting up Server or Worker and Agent

The class supports:

  • [puppet] >= 7.9.0 < 9.0

And requires:

  • [icinga/icinga2] >= 3.1.0 < 7.0.0

Setting up a Icinga Server with a CA and to store configuration:

class { 'icinga::server':
  ca            => true,
  ticket_salt   => Sensitive('supersecret'),
  config_server => true,
  workers       => { 'dmz' => { 'endpoints' => { 'worker.example.org' => { 'host' => '172.16.2.11' }}, }},
  global_zones  => [ 'global-templates', 'linux-commands', 'windows-commands' ],
}

Addtition a connection to a worker is configured. By default the zone for the server is named main. When config_server is enabled directories are managed for all zones, including the worker and global zones.

IMPORTANT: A alpha numeric String has to be set to ticket_salt in Hiera to protect the CA! An alternative is to set icinga::ticket_salt in a hiera common section for all agents, workers and servers.

The associated worker could look like this:

class { 'icinga::worker':
  ca_server        => '172.16.1.11',
  zone             => 'dmz',
  parent_endpoints => { 'server.example.org' => { 'host' => '172.16.1.11', }, },
  global_zones     => [ 'global-templates', 'linux-commands', 'windows-commands' ],
}

If the worker doesn't have a certificate, it sends a certificate request to the CA on the host ca_server. The default parent zone is main. Thus, only the associated endpoint has to be defined.

If icinga::ticket_salt is also set in Hiera for the worker, he's automatically sent a certificate. Otherwise the request will be saved on the CA server and must be signed manually.

Both, server and workers, can operated with a parnter in the same zone to share load. The endpoint of the respective partner is specified as an Icinga object in colocation_endpoints.

colocation_endpoints => { 'server2.example.org' => { 'host' => '172.16.1.12', } },

Of course, the second endpoint must also be specified in the respective parent_endpoints of the worker or agent.

An agent is very similar to a worker, only it has no parameter colocation_endpoints:

class { 'icinga::agent':
  ca_server        => '172.16.1.11',
  parent_endpoints => { 'worker.example.org' => { 'host' => '172.16.2.11', }, } },
  global_zones     => [ 'linux-commands' ],
}

NOTICE: To switch off the package installation via chocolatey on windows, icinga2::manage_packgaes must be set to false for the corresponding hosts in Hiera. That works only on Windows, on Linux package installation is always used.

icinga::db

The class supports:

  • [puppet] >= 7.9.0 < 9.0

Ands requires:

  • [puppetlabs/mysql] >= 10.9.0 < 16.0.0
  • [puppetlabs/postgresql] >= 7.0.0 < 11.0.0
  • [icinga/icinga2] >= 3.1.0 < 7.0.0
  • [icinga/icingadb] >= 2.1.0 < 4.0.0

To activate and configure the IcingaDB (usally on a server) do:

class { 'icinga::db':
  db_type         => 'pgsql',
  db_host         => 'localhost',
  db_pass         => Sensitive('icingadb'),
  manage_database => true,
  manage_redis    => true,
  manage_feature  => true,
}

Setting manage_database to true also setups a database as specified in db_type including database for the IcingaDB. The same applies to manage_redis and the required Redis cache. With manage_feature the Icinga 2 feature for the IcingaDB is additionally activated. The latter two are switched on by default.

icinga::ido

The class supports:

  • [puppet] >= 7.9.0 < 9.0

Ands requires:

  • [puppetlabs/mysql] >= 10.9.0 < 17.0.0
  • [puppetlabs/postgresql] >= 7.0.0 < 11.0.0
  • [icinga/icinga2] >= 3.1.0 < 7.0.0

To activate and configure the IDO feature (usally on a server) do:

class { 'icinga::ido':
  db_type         => 'pgsql',
  db_host         => 'localhost',
  db_pass         => Sensitive('icinga2'),
  manage_database => true,
}

Setting manage_database to true also setups a database as specified in db_type including database for the IDO. Supported are pgsql for PostgreSQL und maysql for MariaDB. By default the database name is set to icinga2 and the user to icinga2.

icinga::web

The class supports:

  • [puppet] >= 7.9.0 < 9.0

And requires:

  • [puppetlabs/mysql] >= 10.9.0 < 17.0.0
  • [puppetlabs/postgresql] >= 7.0.0 < 11.0.0
  • [icinga/icingaweb2] >= 3.7.0 < 6.0.0
  • [icinga/icinga2] >= 3.1.0 < 7.0.0
  • [puppetlabs/apache] >= 5.8.0 < 13.0.0
  • [puppet/php] >= 8.0.0 < 11.0.0

A Icinga Web 2 with an Apache and PHP-FPM can be managed as follows:

class { 'icinga::web':
  db_type         => 'pgsql',
  db_host         => 'localhost',
  db_pass         => Sensitive('supersecret'),
  manage_database => true,
  api_pass        => $icinga::server::web_api_pass,
}

Setting manage_database to true, a database of the specified type is also installed here. It is used to save user settings for the users of the Icinga Web 2 and serves as a backend for managing Icinga Web 2 users and user groups.

IMPORTANT: If you plan tu use icingacli as plugin, e.g. director health checks, businessprocess checks or vspheredb checks, set the parameter run_web => true for icinga::server on the same host icinga::web is declared. That put the Icinga user to the group icingaweb2 and restart the icinga2 process if necessary.

icinga::web::icingadb

If the Icinga Web 2 is operated on the same host as the IcingaDB, the required user credentials can be accessed, otherwise they must be specified explicitly.

class { 'icinga::web::icingadb':
  db_type => $icinga::db::db_type,
  db_host => $icinga::db::db_host,
  db_name => $icinga::db::db_name,
  db_user => $icinga::db::db_user,
  db_pass => $icinga::db::db_pass,
}

IMPORTANT: Must be declared on the same host as icinga::web.

icinga::web::monitoring

If the Icinga Web 2 is operated on the same host as the IDO, the required user credentials can be accessed, otherwise they must be specified explicitly.

class { 'icinga::web::monitoring':
  db_type => $icinga::ido::db_type,
  db_host => $icinga::ido::db_host,
  db_pass => $icinga::ido::db_pass,
}

IMPORTANT: Must be declareid on the same host as icinga::web.

icinga::web::director

Install and manage the famous Icinga Director and the required database. A graphical addon to manage your monitoring environment, the hosts, services, notifications etc.

Here an example with an PostgreSQL database on the same host:

class { 'icinga::web::director':
  db_type         => 'pgsql',
  db_host         => 'localhost',
  db_pass         => Sensitive('supersecret'),
  manage_database => true,
  endpoint        => $facts['networking']['fqdn'],
  api_host        => 'localhost',
  api_pass        => $icinga::server::director_api_pass,
}

In this example the Icinga server is running on the same Host like the web and the director.

icinga::web::vspheredb

The following example sets up the vspheredb Icinga Web 2 module and the required database. At this time only MySQL/MariaDB is support by the Icinga team, so this class also supports only mysql.

class { 'icinga::web::vspheredb':
  db_type         => 'mysql',
  db_host         => 'localhost',
  db_pass         => Sensitive('vspheredb'),
  manage_database => true,
}

icinga::web::reporting

The class supports:

  • [puppet] >= 7.9.0 < 9.0

And required in addition to icinga::web::icingadb or icinga::web::monitoring:

  • [puppetlabs/mysql] >= 10.9.0 < 17.0.0
  • [puppetlabs/postgresql] >= 7.0.0 < 11.0.0
  • [icinga/icingaweb2] >= 3.7.0 < 6.0.0

An example to setup reporting and the required database:

class { 'icinga::web::reporting':
  db_type         => 'pqsql',
  db_host         => 'localhost',
  db_pass         => Sensitive('reporting'),
  manage_database => true,
}

If icinga::web::monitoring is declared before, the required module idoreports for IDO is declared automatically.

Using exported resources for auto discovery

As expected with exported resources, the use of an OpenVoxDB/PuppetDB in the environment is a requirement here. The method described here is also designed exclusively for use in Hiera, as icinga::objects is a parameter of a private class. This means that the various components can be set up as described in the previous section.

The following assumes a Hiera configuration like this:

---
version: 5
hierarchy:
  - name: "Per-node data (yaml version)"
    path: "nodes/%{::trusted.certname}.yaml"
  - name: "role data (yaml version)"
    path: "roles/%{trusted.extensions.pp_role}.yaml"
  - name: 'Operating System'
    path: 'os/%{facts.os.family}.yaml'
  - name: "Other YAML hierarchy levels"
    paths:
      - "common.yaml"

data/common.yaml

In addition to the icinga::ticket and icinga::agent settings required for an Icinga environment with agents, the following parameters are added to common.yaml:

---
icinga::config_server: server.icinga.example.org

icinga::objects:
  Host:
    "%{lookup('icinga::cert_name')}":
      import: ['generic-host']
      display_name: "%{facts.networking.hostname}"
      address: "%{facts.networking.ip}"
  Service:
    ping4:
      host_name: "%{lookup('icinga::cert_name')}"
      import: ['generic-service']
      check_command: ping4
    icinga:
      host_name: "%{lookup('icinga::cert_name')}"
      import: ['generic-service']
      check_command: icinga
      command_endpoint: host_name
    cluster zone:
      host_name: "%{lookup('icinga::cert_name')}"
      import: ['generic-service']
      check_command: cluster-zone

Here is defined that all Icinga hosts get an host object with three services by default. Important and required for such an dscovery setup, icinga::config_server has to be set to the Icinga used certname of the Icinga config server.

data/roles/monitor::server.yaml

The host intended for the Icinga server is therefore assigned the role monitor::server in csr_attributes.yaml. Additional to the regular Icinga server stuff some basic Icinga objects are defined:

icinga::objects:
  Host:
    generic-host:
      template: true
      check_command: hostalive
      check_interval: 1m
      retry_interval: 30s
      max_check_attempts: 3
      target: /etc/icinga2/zones.d/global-templates/auto.conf
  Service:
    generic-service:
      template: true
      check_interval: 1m
      retry_interval: 30s
      max_check_attempts: 5
      target: /etc/icinga2/zones.d/global-templates/auto.conf
    icingadb:
      host_name: "%{lookup('icinga::cert_name')}"
      import: ['generic-service']
      check_command: icingadb
      command_endpoint: host_name
    cluster zone:
      ensure: absent

Both, generic-service and generic-host are local stored templates on the Icinga server and they are used by the the other hosts and services. Another service only for the server is icingadb to monitor the IcingaDB behavior. The defined service cluster zone for all hosts is disabled here because a Icinga server doesn't connect to itself.

data/roles/monitor::worker.yaml

The host intended for an Icinga worker is therefore assigned the role monitor::worker in csr_attributes.yaml. Here also the regular stuff for a Icinga worker aren't descibed:

icinga::worker::parent_export:
  log_duration: 120m

icinga::objects:
  Host:
    "%{lookup('icinga::cert_name')}":
      vars:
        cluster_zone: "%{lookup('icinga::worker::zone')}"

If icinga::config_server is set (done in data/common.yaml) the worker class exports a zone and endpoint object that is collected later by the Icinga Config Server. The enpoint object can overriden or extended with icinga::worker::parent_export. Done here to lower the logging buffer.

Override parameters of Icinga objects

Because icinga::objects is deeply merged thru all hiera layers (e.g. node, os or network layer), any parameter of all objects can be overriden.

---
icinga::objects:
  Service:
    ping4:
      max_check_attempts: 4
      vars:
        ping_packets: 6
        ping_wrta: 200
        ping_crta: 400

Inject Icinga Objects by Manifest

The class icinga::configi can also be used to inject Icinga objects from a manifest.

class profile::icinga (
  Enum['agent', 'worker', 'server'] $type = 'agent',
) {
  case $type {
    'agent': {
      if $facts['kernel'] != 'windows' {
        include icinga::repos
      }
      include icinga::agent
    } # agent

    'worker': {
      include icinga::repos
      include icinga::worker
    } # worker

    'server': {
      class { 'icinga::repos':
        manage_extras => true,
      }

      include profile::icinga::server
    } # server

    default: {}
  }

  class { 'icinga::config':
    objects => {
      'Service' => profile::icinga::disks($icinga::cert_name),
    },
  }
}

Here to add services for mounted file system. The services are returned dynamically via a self-written function, e.g.:

# @summary
#   Discover mounted disks.
#
# @return
#   Disks with parameter to monitor.
#
function profile::icinga::disks(
  String         $hostname,
  Array[String]  $filesystems = ['xfs', 'ext4', 'vfat'],
  Array[String]  $drivetypes  = ['Fixed'],
) >> Hash {
  # @param hostname
  #   Icinga object host_name
  #
  # @param filesystems
  #   List of Linux file system types to discover
  #
  # @param drivetypes
  #   List of Windows drive types to discover
  #
  if $facts['kernel'] == 'windows' {
    $facts['drives'].filter |$keys, $values| { $values['drivetype'] in $drivetypes }.keys.reduce( {} ) |$memo, $disk| {
      $memo + {
        "disk ${disk}" => {
          import           => ['generic-service'],
          host_name        => $hostname,
          command_endpoint => 'host_name',
          check_command    => 'disk-windows',
          vars             => { disk_win_path => $facts['drives'][$disk]['root'] },
        }
      }
    }
  } else {
    $facts['mountpoints'].filter |$keys, $values| { $values['filesystem'] in $filesystems }.keys.reduce( {} ) |$memo, $disk| {
      $memo + {
        "disk ${disk}" => {
          import           => ['generic-service'],
          host_name        => $hostname,
          command_endpoint => 'host_name',
          check_command    => 'disk',
          vars             => { disk_partitions => $disk },
        }
      }
    }
  }
}

Override parameters of this disk Icinga objects use also icinga::objects with the correct service object name of course!

Reference

See REFERENCE.md

Transfer Notice

This plugin was originally authored by Icinga. The maintainer preferred that Vox Pupuli take ownership of the module for future improvement and maintenance. Existing pull requests and issues were transferred over, please fork and continue to contribute here instead of Icinga.

Previously: https://github.com/icinga/puppet-icinga

About

Puppet module to manage repositories that are needed by Icinga and other base class classes like icinga::redis.

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Sponsor this project

  •  

Packages

No packages published

Contributors 11