I came across a project at work recently which was developing some code to audit the state of an environment. At the same time I was starting to play around with Ansible at home to define my personal systems desired state as code. Being new to Ansible, I wondered what it would take to audit my home systems using Ansible. I thought I’d have a go at writing the same thing out of hours and compare how the two solutions differed once complete.
From a beginners perspective, it seemed like it would be pretty easy. The basic playbooks I had previously built returned OK’s, CHANGED and FAILED. I figured an auditing script only needed to return OK’s and FAIL’s… at this point I hadn’t thought of the issue I was about to run into.
The default way Ansible operates is that it will stop processing the playbooks when an error is returned from a task. This allows you to fix the problem and restart where you left off. As it’s perfectly fine for an audit script to have failed tests, the default way Ansible worked was no good for my purpose here.
There was a way around this. ‘ignore_errors: yes’ in a task allowed an Ansible playbook to continue processing if it found errors. The only problem for me was that this resulted in any failed tasks being reported as OK in the end of playbook statistics.
The other thing that was throwing out the end of playbook statistics was that when using shell or command modules in the task, Ansible counts this as changed (system state was changed) and reports this in the statistics. Using ‘changed_when: false’ helped on command/shell tasks, however, the ignore_errors issue still existed.
I really like Ansible… how could I work around the issue I was facing?
Develop a plugin
As Ansible was returning OK for tasks that FAILED when ignored_errors was set, the statistics were skewed at the end of the playbook. I decided to persevere with Ansible and wondered how to work around this. I delved into the Ansible Architecture overview / developer guide. It seemed like I’d need to store my own statistics to overcome the hurdle of ignore_failed statistics.
I came across Ansible Modules and Plugins.
Modules are essentially telling Ansible how to connecting to your node/device and configure it’s desired state.
Plugins allow you to alter how Ansible works.
From the Developer Guide:
|The following types of plugins are available:
The Callback plugins allowed me to hook into Ansible events for the events I wanted to track. For example:It looked like developing a Callback plugin was my best bet to capture OK/FAILED/SKIPPED/UNREACHABLE tasks and store statistics about them.
def runner_on_failed(self, host, data, ignore_errors=False):
"""Routine for handling runner (task) failures"""
def runner_on_ok(self, host, data):
"""Routine for handling runner (task) successes"""
def runner_on_skipped(self, host, item=None):
"""Routine for handling skipped tasks"""
def runner_on_unreachable(self, host, data):
"""Routine for handling unreachable hosts"""
These would allow me to keep track of the number of times any of these events fired and which host they were for.
Preparing to run the Ansible audit
To use the audit callback plugin, you’ll need:
- Ansible 22.214.171.124 & 126.96.36.199 (tested)
- Other versions may work
- Python 2.7 (tested)
Then, download the callback plugin from: https://github.com/Im0/ansible-audit
Either write a sample playbook, or, download an example from here: https://github.com/Im0/ansible-audit-playbook-example
Running the Ansible audit script
The easiest way to run an audit would be to clone the audit git repository and copy the audit.py and *.jinja files into the ansible plugins directory. In my case the plugins directory resides here: /usr/local/lib/python2.7/dist-packages/ansible/plugins/callback/
The plugins directories may reside in other places as well. For example, the default configuration file for Ansible (although commented out) refers to /usr/share/ansible/plugins/
git clone https://github.com/Im0/ansible-audit.git
sudo cp ansible-audit/audit* /usr/local/lib/python2.7/dist-packages/ansible/plugins/callback/
git clone https://github.com/Im0/ansible-audit-playbook-example.git
AUDIT_NAME='My Audit' CUSTOMER="Test Customer" ANSIBLE_CALLBACK_WHITELIST=audit ansible-playbook --ask-sudo-pass ansible-audit-playbook-example/site.yml -k
Checking the output
By default the plugin creates and outputs a zip file into: /var/log/ansible/audits/
The zip file contains a JSON file and two HTML files which contain the results.
Fruit salad output
More detail output
There are a few things I want to tidy up, such as:
- Colour on UNREACHABLE hosts on the fruit salad output needs to be red.
- Always output the zip file full path on exit
- Add some error handling around the output directory creation
- Update instructions to include output directory requirements
- Tweak output filename
- Consider sending output via email
- Add options to select method of delivering output (ie. send via email, to zip file or just JSON output)
- Windows examples in example playbook