linux_course_doc/modules/qualifying/learning_centralized_system...

20 KiB

Centralized System Monitoring

When you have multiple servers or client computers under your control it's nice to have a centralized way of monitoring them for health or unexpected errors. There are a few different solutions out there but we'll have a go at installing and configuring munin.

Munin

I set one up on this domain for demonstration purposes. You can log in with the account details I'll hand out in class.

munin overview

munin zoom

munin details

Architecture

Munin has a server client architecture where they refer to the server as master and the clients as nodes. When monitoring for example ten workstations or servers you will have one master and on each workstation or server you'll install one node but in total you'll have ten nodes connected to your single master.

munin architecture

The master

First we'll install one munin master for everybody in the classroom. The installation documentation on the official website is very good so we'll follow that one. The overall processes has three components.

  1. install munin
  2. configure the master to query some future nodes
  3. install a web server to serve the monitor pages

Installing

Installing can be done through the Debian package manager. Do a tab-complete to see all available packages and note that on the master you need the munin package.

sudo apt install munin

It installs and configures itself as a systemd service we can start, stop or inspect with the following commands.

sudo systemctl start munin.service
sudo systemctl stop munin.service
sudo systemctl status munin.service

If you used tab-complete, as you should, you probably noticed that there are two munin services running. munin.service is the master and munin-node.service is the node. Why does the master also install the node? Well, just so it can monitor itself and display it in the web gui. This gives you already a good idea of what will need to be installed and configured on the nodes.

Configuring

The main configuration file for munin is located at /etc/munin/munin.conf. Open up your favorite text editor and browse through it. There is should be an entry to just one node that is your localhost. This is the server monitoring itself. On my demonstration server the configuration is as follows.

# a simple host tree
[muninserver.86thumbs.net]
    address 127.0.0.1
    use_node_name yes

[vps.86thumbs.net]
    address 10.8.0.1
    use_node_name yes

[bigone.86thumbs.net]
    address 10.8.0.2
    use_node_name yes

[hugeone.86thumbs.net]
    address 192.168.0.236
    use_node_name yes

[yacyone.86thumbs.net]
    address 10.8.0.4
    use_node_name yes

For now we'll just leave it as is and move on to installing the web server.

The web server

Here you have a choice to make but any way you go, you can't really go wrong. My web server of choice is often nginx and we'll install that one together because I'm most comfortable with it. When you install your own master maybe try out a different one. The documentation has templates for the following we servers.

Nginx is really easy to install on Debian. Just type sudo apt install nginx and when you get your prompt back, you'll have a web server running. To verify it's working we can use any web browser we want.

it's alive!

The main configuration directory for nginx can be found at /etc/nginx. Nginx is a bit different compared to most configuration we've done up until now. The server itself is configured via the /etc/nginx/nginx.conf file and the /etc/nginx/conf.d/ folder but the websites your server serves to the outside world are stored in /etc/nginx/sites-available and /etc/nginx/sites-enabled. Each site should have an actual configuration file in the /etc/nginx/sites-available folder and if you want the lite to be live you should create a symbolic link to this site in the /etc/nginx/sites-enabled folder.

First have a look at the /etc/nginx/sitex-available/default file to understand how the configuration works.

Can you change something on the default webpage please?

We can add a new file in the /etc/nginx/sites-available directory called munin. You can name this any which way you want. For example, on my vps that runs our domain I have the following files which offer websites:

➜  sites-enabled ls -l
total 36
-rw-r--r-- 1 root root 2049 Mar 25 20:56 86thumbs.net
-rw-r--r-- 1 root root 1813 Mar 25 12:42 gitea.86thumbs.net
-rw-r--r-- 1 root root 1646 May 24 12:00 ipcdb.86thumbs.net
lrwxrwxrwx 1 root root   50 Mar  4 22:20 jitsi.86thumbs.net.conf -> /etc/nginx/sites-available/jitsi.86thumbs.net.conf
-rw-r--r-- 1 root root 1907 Mar 25 20:55 matrix.86thumbs.net
-rw-r--r-- 1 root root 1720 Jul  8 16:49 munin.86thumbs.net
-rw-r--r-- 1 root root 1758 Mar 11 16:41 radio.86thumbs.net
-rw-r--r-- 1 root root 1881 Mar 11 16:41 riot.86thumbs.net
-rw-r--r-- 1 root root 1930 Mar 11 19:07 taskjuggler.86thumbs.net
-rw-r--r-- 1 root root 2200 Jul  7 18:50 yacy.86thumbs.net
➜  sites-enabled 

As you can see, I did not follow the proper rules and only one file is a symbolic link towards sites-available. Shame on me! The documentation gives us a good idea of how to configure the website but it's not fully complete. The following configuration should give you a fully functional monitoring page.

server {
	listen 80;
	listen [::]:80;
	
	location /munin/static/ {
    		alias /etc/munin/static/;
    		expires modified +1w;
	}

	location /munin {
    		alias /var/cache/munin/www/;
    		expires modified +310s;
	}
}

It's even more basic than the one supplied by munin itself just to illustrate a bare bones website. If you now go to the server you should see the munin monitoring page with one node available, itself.

basic munin

I urge you to read up a bit on nginx configuration and to look at the folders on your system that are used by nginx to serve websites. Web servers are an essential part of any Linux system administrator's life.

Can you figure out how to password protect this folder so it behaves like my munin server?

The first node

Let's add some nodes! The nodes have to be in the same network in order to be contactable by the master. There are nifty ways to overcome this problem but that is for a later date. If you do feel like diving deeper into it, have a look here.

On any virtual machine you have around you can install the node by simply doing sudo apt install munin-node. The node is now up, running and listening for incoming connections on port 4949 (by default). It is however quite picky in accepting incoming connections! The configuration file for the node is located at /etc/munin/munin-node.conf. Open it up and have a read. You should see two lines very similar to the one below.

allow ^127\.0\.0\.1$
allow ^::1$

Can you tell me what those are? Yes, that's a regex! By default the node only accepts incoming connections from the localhost. We can add as many masters as we like to this list but we must respect the regex syntax. I'll leave that up to you to figure out.

Plugins

Munin works with plugins to supply the master with data to plot. You can install extra plugins or even write your own in bash or in python. To have your node auto detect which plugins should be used you can execute the following command as root (not sudo).

munin-node-configure --shell --families=contrib,auto | sh -x

Other operating systems

Because the munin server client model is so basic you can find node software for most operating systems. I've run it successfully on freeNAS but you can try a windows node as well.

When I say simple I mean really simple. For those who like digging a bit deeper, you can connect to the munin node over telnet. From a terminal on the master, or any IP address that is in the allow list of the node, you can do the following.

waldek@muninserver:~$ telnet 10.8.0.1 4949
Trying 10.8.0.1...
Connected to 10.8.0.1.
Escape character is '^]'.
# munin node at vps-42975ad1.vps.ovh.net
help
# Unknown command. Try cap, list, nodes, config, fetch, version or quit
list
cpu df df_inode entropy fail2ban forks fw_conntrack fw_forwarded_local fw_packets if_err_eth0 if_err_tun0 if_eth0 if_tun0 interrupts irqstats load memory netstat ntp_kernel_err ntp_kernel_pll_freq ntp_kernel_pll_off ntp_offset open_files open_inodes proc_pri processes swap systemd-status threads uptime users vmstat
fetch cpu
user.value 21547781
nice.value 22016
system.value 8773929
idle.value 895574842
iowait.value 449848
irq.value 0
softirq.value 773399
steal.value 540782
guest.value 0
.
quit
Connection closed by foreign host.
waldek@muninserver:~$ 

Notes and hints

You might feel munin is too simple or basic because it does not offer any tweaking via it's web GUI but that's the main reason why I like and use it so much. Because it's so simple, it's not a real security risk. I would however not expose a master to the internet without SSL and a good htpasswd. There are some known exploits in the CGI part of munin.

  • the Debian system administrator handbook has a good section on munin
  • if your dynamic zoom is not working have a look here

Alternatives

There are quite a few alternatives out there to choose from. One of the better known is nagios which can be installed on Debian as well. There are ways to have munin and nagios work together that are described here.

Cacti

Cacti is a very similar monitoring system compared to munin. Where munin shines in it's simplicity, cacti can be fine-tuned host by host, monitor even more diverse clients such as Cisco routers and switches. I personally prefer munin mostly because I know it better but let's have a go at installing cacti to see if we like it better or not!

Architecture

The architecture for cacti is very similar to munin and can be seen below. They serve the same purpose but do two main things differently.

  1. how cacti fetches the data
  2. how cacti stores the data

cacti architecture

Fetching data

Munin fetches it's data via a simple telnet connection. To restrict access it defaults to blacklisting all IP's except the ones we allow in the configuration file located at /etc/munin/munin-node.conf. Cacti relies on the more elaborated SNMP protocol. As it's a full blown protocol it can do a lot more than just fetch data and hence is more prone to exploits.

Storing data

Munin stores it's data as simple binary files located at /var/lib/munin/$DOMAIN_NAME. It's not the quickest nor the most compact way of storing data but as it's not a high demand service it can get away with it. Cacti on the other hand uses an SQL database to store it's information which is a more mature way of dealing with data availability but comes at the cost of an other server, a mysql in this case, and more configuration and potential security threats.

Installing the master

Debian has a good all-in-one package to install cacti. It takes care of all the heavy lifting for you and will prompt you with some questions. You install it with the good old package manager sudo apt install cacti.

First you'll be prompted with a choice of web server to install. You can choose any which one you like, or choose none if you want to use nginx for example. Depending on your choice, apt will fetch the necessary packages.

web server

Once that's all done you'll be prompted to configure the database. Here I highly advise you to choose yes as sql databases can be a bit of a pain to set up.

db choice

If you chose yes you'll immediately be prompted to set a password for the cacti user in the database. This will also be the admin password you'll use to log into the web GUI. For testing purposes you can choose a simple one but if you think you'll use the server for anything serious, you need a serious password!

setting the password

That's it, your cacti server is now up and running! You can visit it's login page by going to the IP address of your cacti server. If you see the default apache2 page as seen below, append /cacti after your IP. You need to use the password you set during the installation, but the username for the default user is admin.

apache2 default page

cacti login page

Configuring the nodes

Now that we have a server up and running it's time to add some clients to it. Contrary to munin, cacti does not install it's node/client component on the server in order for you to be able to monitor itself. Now, which package do we need to install on a client in order for it to be able to submit data to our server? The answer lies in the protocol used by cacti. The following command will install the SNMP daemon onto our first client, our cacti server.

sudo apt install snmpd

This is a quick install and the server should now be ready to report to itself. Let's try to add it as a client. In the web GUI you need to go to the create->new device tab.

new device

In the description you can put anything you want, but the hostname field needs to be the IP address of your server, which in our case is the loopback device 127.0.0.1. You'll see the following screen if your device was successfully added, which should be the case if you followed the isntructions.

successfully added new device

You can now inspect the data collected in the graph tab of the web GUI. It should look similar to the screenshot below. It's very possible that the graphs are not yet showing. This depends a bit on whether the poller has already finished polling the SNMP daemon but this should be a one time problem fixed by waiting a couple of minutes.

graph

The procedure above is how you add clients to the server. It's basically a two step process, but we'll see in a minute we need one extra half step to make it work with clients that are not the server itself.

  1. install snmpd on the client
    • configure which IP the client listens on
  2. add the client IP to the server

So, let's add an other client to the master. Take any Linux machine you have and install the snmpd package. Once installed try to add it's IP address to the web GUI of cacti. You'll be confronted with an error screen like the one below. The client was added but there is an error message in red. What's happening here?

error adding client

System SNMP error - SNMP::get(): No response from 192.168.0.222 means that the client is not listening for incoming connections. We can fix this by configuring the client to listen on all interfaces. By default the SNMP daemon only listens on the loopback device but it's an easy fix. Open the following file /etc/snmp/snmp.conf with your favorite text editor and the problem is sitting right in front of your nose.

#
#  AGENT BEHAVIOUR
#

#  Listen for connections from the local system only
agentAddress  udp:127.0.0.1:161
#  Listen for connections on all interfaces (both IPv4 *and* IPv6)
#agentAddress udp:161,udp6:[::1]:161

You should comment the active agentAddress in exchange for the second one. The second one makes the daemon listen on all interfaces for both ipv4 and ipv6. If for some reason you configured your Linux kernel to only do ipv4, you'll need to remove the ipv6 entry as follows agentAddress udp:161 otherwise the daemon will fail to restart. If you restart the service with sudo systemctl restart snmpd.service you're client should be contactable by your server!

success

Securing the connection to the client

There are different versions of SNMP and it is highly recommended to run only the v3. It offers encryption and authentication for servers who try to connect to clients. In order to run v3 you need to install additional packages. The main tool you need to add users is called net-snmp-create-v3-user and it can be installed by running the following command sudo apt install libsnmp-dev. Once this command completes you'll have the necessary programs to add user accounts to your snmpd client. You can only run this program when the snmpd service is stopped.

➜  ~ git:(master) ✗ sudo net-snmp-create-v3-user 
Apparently at least one snmpd demon is already running.
You must stop them in order to use this command.
➜  ~ git:(master) ✗ sudo systemctl stop snmpd.service     
➜  ~ git:(master) ✗ sudo net-snmp-create-v3-user     
Enter a SNMPv3 user name to create: 
student
Enter authentication pass-phrase: 
student2021
Enter encryption pass-phrase: 
  [press return to reuse the authentication pass-phrase]
student2021
adding the following line to /var/lib/snmp/snmpd.conf:
   createUser student MD5 "student2021" DES student2021
adding the following line to /usr/share/snmp/snmpd.conf:
   rwuser student
➜  ~ git:(master) ✗ sudo systemctl start snmpd.service     

Now we need to change the authentication settings in the web GUI of cacti. Go to the Management->devices list and choose your client. Change the version to version 3, put in the login and password and you're good to go. If all of this is working it's time to stop offering version 1 and version 2. This is again done on the client in the /etc/snmp/snmpd.conf file. Locate the line starting with rocommunity and comment them out. Restart the service and now you're only offering v3 connections!

Additional security with iptables

We have not seen iptables yet, but it's the main program in Linux to control incoming and outgoing connections. Finally, we could ensure that no one except us can access SNMP form outside. The simplest way to achieve this is to add some firewall rules with iptables. To ensure the iptable configuration will be loaded automatically install the following package in addition: apt-get install iptables-persistent This ensures that the iptable rules are automatically loaded after a reboot of the system, the rules will be loaded from a persistent stored file. To trigger an update of the currently used iptables of the system run one of the following commands:

iptables-save > /etc/iptables/rules.v4
ip6tables-save > /etc/iptables/rules.v6

Now we could add 4 new iptables entries to allow only access from our external system and block all other ones. Do not forget to replace 11.11.11.11 with your ip address or range.

iptables -A INPUT -s 11.11.11.11 -p udp -m udp --dport 161 -j ACCEPT
iptables -A INPUT -s 11.11.11.11 -p udp -m udp --dport 162 -j ACCEPT
iptables -A INPUT -p udp -m udp --dport 161 -j DROP
iptables -A INPUT -p udp -m udp --dport 162 -j DROP