Browse Source

Added ansible playbook to remote networks to ship logs to logz.io

pull/1471/head
Greg Szabo 6 years ago
parent
commit
089ce6744c
6 changed files with 419 additions and 1 deletions
  1. +14
    -1
      networks/remote/ansible/README.rst
  2. +13
    -0
      networks/remote/ansible/logzio.yml
  3. +15
    -0
      networks/remote/ansible/roles/logzio/files/journalbeat.service
  4. +8
    -0
      networks/remote/ansible/roles/logzio/handlers/main.yml
  5. +27
    -0
      networks/remote/ansible/roles/logzio/tasks/main.yml
  6. +342
    -0
      networks/remote/ansible/roles/logzio/templates/journalbeat.yml.j2

+ 14
- 1
networks/remote/ansible/README.rst View File

@ -17,7 +17,6 @@ Prerequisites
- Create SSH keys
- Install the python dopy package (for the digital_ocean.py script)
Build
-----
@ -37,3 +36,17 @@ Build
# Reconfig the sentry nodes with a new BINARY and the configuration files from the build folder:
ansible-playbook -i inventory/digital_ocean.py -l remotenet config.yml -e BINARY=`pwd`/build/tendermint -e CONFIGDIR=`pwd`/build
Shipping logs to logz.io
------------------------
Logz.io is an Elastic stack (Elastic search, Logstash and Kibana) service provider. You can set up your nodes to log there automatically. Create an account and get your API key from the notes on `this page <https://app.logz.io/#/dashboard/data-sources/Filebeat>`__.
::
yum install systemd-devel || echo "This will only work on RHEL-based systems."
apt-get install libsystemd-dev || echo "This will only work on Debian-based systems."
go get github.com/mheese/journalbeat
ansible-playbook -i inventory/digital_ocean.py -l remotenet logzio.yml -e LOGZIO_TOKEN=ABCDEFGHIJKLMNOPQRSTUVWXYZ012345

+ 13
- 0
networks/remote/ansible/logzio.yml View File

@ -0,0 +1,13 @@
---
- hosts: all
user: root
any_errors_fatal: true
gather_facts: no
vars:
- service: tendermint
- JOURNALBEAT_BINARY: "{{lookup('env', 'GOPATH')}}/bin/journalbeat"
# - LOGZIO_TOKEN: ABCDEFGHIJKLMNOPQRSTUVWXYZ012345
roles:
- logzio

+ 15
- 0
networks/remote/ansible/roles/logzio/files/journalbeat.service View File

@ -0,0 +1,15 @@
[Unit]
Description=journalbeat
#propagates activation, deactivation and activation fails.
Requires=network-online.target
After=network-online.target
[Service]
Restart=on-failure
ExecStart=/usr/bin/journalbeat -c /etc/journalbeat/journalbeat.yml -path.home /usr/share/journalbeat -path.config /etc/journalbeat -path.data /var/lib/journalbeat -path.logs /var/log/journalbeat
Restart=always
[Install]
WantedBy=multi-user.target

+ 8
- 0
networks/remote/ansible/roles/logzio/handlers/main.yml View File

@ -0,0 +1,8 @@
---
- name: reload daemon
command: "systemctl daemon-reload"
- name: restart journalbeat
service: name=journalbeat state=restarted

+ 27
- 0
networks/remote/ansible/roles/logzio/tasks/main.yml View File

@ -0,0 +1,27 @@
---
- name: Copy journalbeat binary
copy: src="{{JOURNALBEAT_BINARY}}" dest=/usr/bin/journalbeat mode=0755
notify: restart journalbeat
- name: Create folders
file: "path={{item}} state=directory recurse=yes"
with_items:
- /etc/journalbeat
- /etc/pki/tls/certs
- /usr/share/journalbeat
- /var/log/journalbeat
- name: Copy journalbeat config
template: src=journalbeat.yml.j2 dest=/etc/journalbeat/journalbeat.yml mode=0600
notify: restart journalbeat
- name: Get server certificate for Logz.io
get_url: "url=https://raw.githubusercontent.com/logzio/public-certificates/master/COMODORSADomainValidationSecureServerCA.crt force=yes dest=/etc/pki/tls/certs/COMODORSADomainValidationSecureServerCA.crt"
- name: Copy journalbeat service config
copy: src=journalbeat.service dest=/etc/systemd/system/journalbeat.service
notify:
- reload daemon
- restart journalbeat

+ 342
- 0
networks/remote/ansible/roles/logzio/templates/journalbeat.yml.j2 View File

@ -0,0 +1,342 @@
#======================== Journalbeat Configuration ============================
journalbeat:
# What position in journald to seek to at start up
# options: cursor, tail, head (defaults to tail)
#seek_position: tail
# If seek_position is set to cursor and seeking to cursor fails
# fall back to this method. If set to none will it will exit
# options: tail, head, none (defaults to tail)
#cursor_seek_fallback: tail
# Store the cursor of the successfully published events
#write_cursor_state: true
# Path to the file to store the cursor (defaults to ".journalbeat-cursor-state")
#cursor_state_file: .journalbeat-cursor-state
# How frequently should we save the cursor to disk (defaults to 5s)
#cursor_flush_period: 5s
# Path to the file to store the queue of events pending (defaults to ".journalbeat-pending-queue")
#pending_queue.file: .journalbeat-pending-queue
# How frequently should we save the queue to disk (defaults to 1s).
# Pending queue represents the WAL of events queued to be published
# or being published and waiting for acknowledgement. In case of a
# regular restart of journalbeat all the events not yet acknowledged
# will be flushed to disk during the shutdown.
# In case of disaster most probably journalbeat won't get a chance to shutdown
# itself gracefully and this flush period option will serve you as a
# backup creation frequency option.
#pending_queue.flush_period: 1s
# Lowercase and remove leading underscores, e.g. "_MESSAGE" -> "message"
# (defaults to false)
#clean_field_names: false
# All journal entries are strings by default. You can try to convert them to numbers.
# (defaults to false)
#convert_to_numbers: false
# Store all the fields of the Systemd Journal entry under this field
# Can be almost any string suitable to be a field name of an ElasticSearch document.
# Dots can be used to create nested fields.
# Two exceptions:
# - no repeated dots;
# - no trailing dots, e.g. "journal..field_name." will fail
# (defaults to "" hence stores on the upper level of the event)
#move_metadata_to_field: ""
# Specific units to monitor.
units: ["{{service}}.service"]
# Specify Journal paths to open. You can pass an array of paths to Systemd Journal paths.
# If you want to open Journal from directory just pass an array consisting of one element
# representing the path. See: https://www.freedesktop.org/software/systemd/man/sd_journal_open.html
# By default this setting is empty thus journalbeat will attempt to find all journal files automatically
#journal_paths: ["/var/log/journal"]
#default_type: journal
#================================ General ======================================
# The name of the shipper that publishes the network data. It can be used to group
# all the transactions sent by a single shipper in the web interface.
# If this options is not defined, the hostname is used.
#name: journalbeat
# The tags of the shipper are included in their own field with each
# transaction published. Tags make it easy to group servers by different
# logical properties.
tags: ["{{service}}"]
# Optional fields that you can specify to add additional information to the
# output. Fields can be scalar values, arrays, dictionaries, or any nested
# combination of these.
fields:
logzio_codec: plain
token: {{LOGZIO_TOKEN}}
# If this option is set to true, the custom fields are stored as top-level
# fields in the output document instead of being grouped under a fields
# sub-dictionary. Default is false.
fields_under_root: true
# Internal queue size for single events in processing pipeline
#queue_size: 1000
# The internal queue size for bulk events in the processing pipeline.
# Do not modify this value.
#bulk_queue_size: 0
# Sets the maximum number of CPUs that can be executing simultaneously. The
# default is the number of logical CPUs available in the system.
#max_procs:
#================================ Processors ===================================
# Processors are used to reduce the number of fields in the exported event or to
# enhance the event with external metadata. This section defines a list of
# processors that are applied one by one and the first one receives the initial
# event:
#
# event -> filter1 -> event1 -> filter2 ->event2 ...
#
# The supported processors are drop_fields, drop_event, include_fields, and
# add_cloud_metadata.
#
# For example, you can use the following processors to keep the fields that
# contain CPU load percentages, but remove the fields that contain CPU ticks
# values:
#
processors:
#- include_fields:
# fields: ["cpu"]
- drop_fields:
fields: ["beat.name", "beat.version", "logzio_codec", "SYSLOG_IDENTIFIER", "SYSLOG_FACILITY", "PRIORITY"]
#
# The following example drops the events that have the HTTP response code 200:
#
#processors:
#- drop_event:
# when:
# equals:
# http.code: 200
#
# The following example enriches each event with metadata from the cloud
# provider about the host machine. It works on EC2, GCE, and DigitalOcean.
#
#processors:
#- add_cloud_metadata:
#
#================================ Outputs ======================================
# Configure what outputs to use when sending the data collected by the beat.
# Multiple outputs may be used.
#----------------------------- Logstash output ---------------------------------
output.logstash:
# Boolean flag to enable or disable the output module.
enabled: true
# The Logstash hosts
hosts: ["listener.logz.io:5015"]
# Number of workers per Logstash host.
#worker: 1
# Set gzip compression level.
#compression_level: 3
# Optional load balance the events between the Logstash hosts
#loadbalance: true
# Number of batches to be send asynchronously to logstash while processing
# new batches.
#pipelining: 0
# Optional index name. The default index name is set to name of the beat
# in all lowercase.
#index: 'beatname'
# SOCKS5 proxy server URL
#proxy_url: socks5://user:password@socks5-server:2233
# Resolve names locally when using a proxy server. Defaults to false.
#proxy_use_local_resolver: false
# Enable SSL support. SSL is automatically enabled, if any SSL setting is set.
ssl.enabled: true
# Configure SSL verification mode. If `none` is configured, all server hosts
# and certificates will be accepted. In this mode, SSL based connections are
# susceptible to man-in-the-middle attacks. Use only for testing. Default is
# `full`.
ssl.verification_mode: full
# List of supported/valid TLS versions. By default all TLS versions 1.0 up to
# 1.2 are enabled.
#ssl.supported_protocols: [TLSv1.0, TLSv1.1, TLSv1.2]
# Optional SSL configuration options. SSL is off by default.
# List of root certificates for HTTPS server verifications
ssl.certificate_authorities: ["/etc/pki/tls/certs/COMODORSADomainValidationSecureServerCA.crt"]
# Certificate for SSL client authentication
#ssl.certificate: "/etc/pki/client/cert.pem"
# Client Certificate Key
#ssl.key: "/etc/pki/client/cert.key"
# Optional passphrase for decrypting the Certificate Key.
#ssl.key_passphrase: ''
# Configure cipher suites to be used for SSL connections
#ssl.cipher_suites: []
# Configure curve types for ECDHE based cipher suites
#ssl.curve_types: []
#------------------------------- File output -----------------------------------
#output.file:
# Boolean flag to enable or disable the output module.
#enabled: true
# Path to the directory where to save the generated files. The option is
# mandatory.
#path: "/tmp/beatname"
# Name of the generated files. The default is `beatname` and it generates
# files: `beatname`, `beatname.1`, `beatname.2`, etc.
#filename: beatname
# Maximum size in kilobytes of each file. When this size is reached, and on
# every beatname restart, the files are rotated. The default value is 10240
# kB.
#rotate_every_kb: 10000
# Maximum number of files under path. When this number of files is reached,
# the oldest file is deleted and the rest are shifted from last to first. The
# default is 7 files.
#number_of_files: 7
#----------------------------- Console output ---------------------------------
#output.console:
# Boolean flag to enable or disable the output module.
#enabled: true
# Pretty print json event
#pretty: false
#================================= Paths ======================================
# The home path for the beatname installation. This is the default base path
# for all other path settings and for miscellaneous files that come with the
# distribution (for example, the sample dashboards).
# If not set by a CLI flag or in the configuration file, the default for the
# home path is the location of the binary.
#path.home:
# The configuration path for the beatname installation. This is the default
# base path for configuration files, including the main YAML configuration file
# and the Elasticsearch template file. If not set by a CLI flag or in the
# configuration file, the default for the configuration path is the home path.
#path.config: ${path.home}
# The data path for the beatname installation. This is the default base path
# for all the files in which beatname needs to store its data. If not set by a
# CLI flag or in the configuration file, the default for the data path is a data
# subdirectory inside the home path.
#path.data: ${path.home}/data
# The logs path for a beatname installation. This is the default location for
# the Beat's log files. If not set by a CLI flag or in the configuration file,
# the default for the logs path is a logs subdirectory inside the home path.
#path.logs: ${path.home}/logs
#============================== Dashboards =====================================
# These settings control loading the sample dashboards to the Kibana index. Loading
# the dashboards is disabled by default and can be enabled either by setting the
# options here, or by using the `-setup` CLI flag.
#dashboards.enabled: false
# The URL from where to download the dashboards archive. By default this URL
# has a value which is computed based on the Beat name and version. For released
# versions, this URL points to the dashboard archive on the artifacts.elastic.co
# website.
#dashboards.url:
# The directory from where to read the dashboards. It is used instead of the URL
# when it has a value.
#dashboards.directory:
# The file archive (zip file) from where to read the dashboards. It is used instead
# of the URL when it has a value.
#dashboards.file:
# If this option is enabled, the snapshot URL is used instead of the default URL.
#dashboards.snapshot: false
# The URL from where to download the snapshot version of the dashboards. By default
# this has a value which is computed based on the Beat name and version.
#dashboards.snapshot_url
# In case the archive contains the dashboards from multiple Beats, this lets you
# select which one to load. You can load all the dashboards in the archive by
# setting this to the empty string.
#dashboards.beat: beatname
# The name of the Kibana index to use for setting the configuration. Default is ".kibana"
#dashboards.kibana_index: .kibana
# The Elasticsearch index name. This overwrites the index name defined in the
# dashboards and index pattern. Example: testbeat-*
#dashboards.index:
#================================ Logging ======================================
# There are three options for the log output: syslog, file, stderr.
# Under Windows systems, the log files are per default sent to the file output,
# under all other system per default to syslog.
# Sets log level. The default log level is info.
# Available log levels are: critical, error, warning, info, debug
#logging.level: info
# Enable debug output for selected components. To enable all selectors use ["*"]
# Other available selectors are "beat", "publish", "service"
# Multiple selectors can be chained.
#logging.selectors: [ ]
# Send all logging output to syslog. The default is false.
#logging.to_syslog: true
# If enabled, beatname periodically logs its internal metrics that have changed
# in the last period. For each metric that changed, the delta from the value at
# the beginning of the period is logged. Also, the total values for
# all non-zero internal metrics are logged on shutdown. The default is true.
#logging.metrics.enabled: true
# The period after which to log the internal metrics. The default is 30s.
#logging.metrics.period: 30s
# Logging to rotating files files. Set logging.to_files to false to disable logging to
# files.
logging.to_files: true
logging.files:
# Configure the path where the logs are written. The default is the logs directory
# under the home path (the binary location).
#path: /var/log/beatname
# The name of the files where the logs are written to.
#name: beatname
# Configure log file size limit. If limit is reached, log file will be
# automatically rotated
#rotateeverybytes: 10485760 # = 10MB
# Number of rotated log files to keep. Oldest files will be deleted first.
#keepfiles: 7

Loading…
Cancel
Save