# How to process Kippo output in an ELK stack (Note: work in progress, instructions are not verified) ## Prerequisites * Working Kippo installation * Kippo JSON log file (enable database json in kippo.cfg) ## Installation * Install logstash, elasticsearch and kibana ``` apt-get install logstash apt-get install elasticsearch ```` * Install Kibana This may be different depending on your operating system. Kibana will need additional components such as a web server ## ElasticSearch Configuration TBD ## Logstash Configuration * Download GeoIP data ``` wget http://geolite.maxmind.com/download/geoip/database/GeoLiteCity.dat.gz wget http://download.maxmind.com/download/geoip/database/asnum/GeoIPASNum.dat.gz ``` * Place these somewhere in your filesystem. * Configure logstash ``` cp logstash-kippo.conf /etc/logstash/conf.d ``` * Make sure the configuration file is correct. Check the input section (path), filter (geoip databases) and output (elasticsearch hostname) ``` service logstash restart ``` * By default the logstash is creating debug logs in /tmp. * To test whether logstash is working correctly, check the file in /tmp ``` tail /tmp/kippo-logstash.log ``` * To test whether data is loaded into ElasticSearch, run the following query: ``` http://:9200/_search?q=kippo&size=5 ``` * If this gives output, your data is correctly loaded into ElasticSearch