diff --git a/SDP/SDP/SDP.py b/SDP/SDP/SDP.py index ee98bab36fa7cef38131f8c607bc46cbfeb652ac..88408be0444a98218baa6e73e2e0a43891396561 100644 --- a/SDP/SDP/SDP.py +++ b/SDP/SDP/SDP.py @@ -26,9 +26,11 @@ import numpy from wrappers import only_in_states, only_when_on, fault_on_error from opcua_connection import OPCUAConnection +from lofar_logging import device_logging_to_python, log_exceptions __all__ = ["SDP", "main"] +@device_logging_to_python({"device": "SDP"}) class SDP(Device): """ @@ -192,6 +194,8 @@ class SDP(Device): self.info_stream("Mapping OPC-UA MP/CP to attributes done.") + @log_exceptions + @DebugIt() def init_device(self): """ Instantiates the device in the OFF state. """ @@ -200,6 +204,7 @@ class SDP(Device): self.set_state(DevState.OFF) + @log_exceptions def initialise(self): """Initialises the attributes and properties of the SDP.""" @@ -475,7 +480,6 @@ class SDP(Device): """ self.set_state(DevState.FAULT) - # ---------- # Run server # ---------- diff --git a/devices/src/hardware_device.py b/devices/src/hardware_device.py index 29636f5d05148e3f40873a8e97987d68d2a535c7..b94484ac281b9aa8ceadfc49389ac5d54ba4b13f 100644 --- a/devices/src/hardware_device.py +++ b/devices/src/hardware_device.py @@ -17,6 +17,7 @@ from tango import DevState, DebugIt # Additional import from src.attribute_wrapper import * +from src.lofar_logging import device_logging_to_python, log_exceptions __all__ = ["hardware_device"] @@ -24,6 +25,7 @@ __all__ = ["hardware_device"] from src.wrappers import only_in_states +@device_logging_to_python({"device": "SDP"}) class hardware_device(Device): """ @@ -59,6 +61,7 @@ class hardware_device(Device): self.value_dict = {i: i.initial_value() for i in self.attr_list()} + @log_exceptions() def init_device(self): """ Instantiates the device in the OFF state. """ @@ -162,6 +165,7 @@ class hardware_device(Device): """Method always executed before any TANGO command is executed.""" pass + @log_exceptions() def delete_device(self): """Hook to delete resources allocated in init_device. diff --git a/devices/src/lofar_logging.py b/devices/src/lofar_logging.py new file mode 100644 index 0000000000000000000000000000000000000000..aa7d3633138679c63fd1934cf8d5638df7b1cedf --- /dev/null +++ b/devices/src/lofar_logging.py @@ -0,0 +1,71 @@ +import logging +from functools import wraps + +# Always also log the hostname because it makes the origin of the log clear. +import socket +hostname = socket.gethostname() + +def configure_logger(logger: logging.Logger, log_extra=None): + logger.setLevel(logging.DEBUG) + + try: + from logstash_async.handler import AsynchronousLogstashHandler, LogstashFormatter + + # log to the tcp_input of logstash in our ELK stack + handler = AsynchronousLogstashHandler("elk", 5959, database_path='pending_log_messages.db') + + # configure log messages + formatter = LogstashFormatter(extra=log_extra, tags=["python", "lofar"]) + handler.setFormatter(formatter) + + # install the handler + logger.addHandler(handler) + + # for now, also log to stderr + # Set up logging in a way that it can be understood by a human reader, be + # easily grep'ed, be parsed with a couple of shell commands and + # easily fed into an Kibana/Elastic search system. + handler = logging.StreamHandler() + formatter = logging.Formatter(fmt = '%(asctime)s.%(msecs)d %(levelname)s - HOST="{}" PID="%(process)d" TNAME="%(threadName)s" TID="%(thread)d" FILE="%(pathname)s" LINE="%(lineno)d" FUNC="%(funcName)s" MSG="%(message)s"'.format(hostname), datefmt = '%Y-%m-%dT%H:%M:%S') + handler.setFormatter(formatter) + logger.addHandler(handler) + except Exception: + logger.exception("Cannot import or configure logstash_async module, not forwarding logs to ELK stack.") + + return logger + +def device_logging_to_python(log_extra: dict = None): + """ Call this on a Tango Device instance or class to have your Tango Device log to python instead. """ + + def inner(cls): + # Create a logger that logs to ELK, dedicated for this class + logger = logging.getLogger(cls.__name__) + configure_logger(logger, log_extra) + + # Monkey patch the python logger to replace the tango logger + cls.debug_stream = logger.debug + cls.info_stream = logger.info + cls.warn_stream = logger.warning + cls.warning_stream = logger.warning + cls.error_stream = logger.error + cls.fatal_stream = logger.fatal + cls.critical_stream = logger.critical + return cls + + return inner + +def log_exceptions(): + """ Decorator that logs all exceptions that the function raises. """ + + def wrapper(func): + @wraps(func) + def inner(self, *args, **kwargs): + try: + return func(self, *args, **kwargs) + except Exception as e: + self.error_stream("Caught exception: %s: %s", e.__class__.__name__, e, exc_info=1) + raise e + + return inner + + return wrapper diff --git a/docker-compose/Makefile b/docker-compose/Makefile index 65de8951a33821b2f06c3b530b11b2f3c1fbc355..693bc26cbe78a45ead288c49ed2929ca7944c920 100644 --- a/docker-compose/Makefile +++ b/docker-compose/Makefile @@ -82,7 +82,7 @@ endif DOCKER_COMPOSE_ARGS := DISPLAY=$(DISPLAY) XAUTHORITY=$(XAUTHORITY) TANGO_HOST=$(TANGO_HOST) \ NETWORK_MODE=$(NETWORK_MODE) XAUTHORITY_MOUNT=$(XAUTHORITY_MOUNT) TANGO_SKA_CONTAINER_MOUNT=$(TANGO_SKA_CONTAINER_MOUNT) TANGO_LOFAR_CONTAINER_MOUNT=$(TANGO_LOFAR_CONTAINER_MOUNT) TANGO_LOFAR_CONTAINER_DIR=${TANGO_LOFAR_CONTAINER_DIR} MYSQL_HOST=$(MYSQL_HOST) \ - CONTAINER_NAME_PREFIX=$(CONTAINER_NAME_PREFIX) COMPOSE_IGNORE_ORPHANS=true + CONTAINER_NAME_PREFIX=$(CONTAINER_NAME_PREFIX) COMPOSE_IGNORE_ORPHANS=true CONTAINER_EXECUTION_UID=$(shell id -u) .PHONY: up down minimal start stop status clean pull help diff --git a/docker-compose/archiver.yml b/docker-compose/archiver.yml index e423c79ead2b599338a89e16311a78c88e1c4d33..38245d4933515d92ec2da41167f2404453e3a957 100644 --- a/docker-compose/archiver.yml +++ b/docker-compose/archiver.yml @@ -67,4 +67,5 @@ services: - ${TANGO_LOFAR_CONTAINER_MOUNT} - ${HOME}:/hosthome - ../docker/tango/tango-archiver:/tango-archiver + restart: on-failure diff --git a/docker-compose/device-pcc.yml b/docker-compose/device-pcc.yml index 96eb6901583216b69035b1d2e906a2c1744e6644..a7c6bad9d1edc07807864b986120f90662b0733a 100644 --- a/docker-compose/device-pcc.yml +++ b/docker-compose/device-pcc.yml @@ -21,11 +21,12 @@ services: - ${TANGO_LOFAR_CONTAINER_MOUNT} environment: - TANGO_HOST=${TANGO_HOST} + user: ${CONTAINER_EXECUTION_UID} entrypoint: - /usr/local/bin/wait-for-it.sh - ${TANGO_HOST} - --timeout=30 - --strict - -- - - python3 -u ${TANGO_LOFAR_CONTAINER_DIR}/PCC/PCC LTS -v + - python3 -u ${TANGO_LOFAR_CONTAINER_DIR}/devices/PCC.py LTS -v restart: on-failure diff --git a/docker-compose/device-sdp.yml b/docker-compose/device-sdp.yml index 8fbc3cbc536360658446d2f21c57231d29b8d6ca..653e0d61b8115ade7d1e3587096d6b10ab178377 100644 --- a/docker-compose/device-sdp.yml +++ b/docker-compose/device-sdp.yml @@ -21,11 +21,12 @@ services: - ${TANGO_LOFAR_CONTAINER_MOUNT} environment: - TANGO_HOST=${TANGO_HOST} + user: ${CONTAINER_EXECUTION_UID} entrypoint: - /usr/local/bin/wait-for-it.sh - ${TANGO_HOST} - --timeout=30 - --strict - -- - - python3 -u ${TANGO_LOFAR_CONTAINER_DIR}/SDP/SDP LTS -v + - python3 -u ${TANGO_LOFAR_CONTAINER_DIR}/devices/SDP.py LTS -v restart: on-failure diff --git a/docker-compose/elk-configure-host/Dockerfile b/docker-compose/elk-configure-host/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..269032484ab8d9a03e1ef4ce006fd046eec1d203 --- /dev/null +++ b/docker-compose/elk-configure-host/Dockerfile @@ -0,0 +1,4 @@ +FROM ubuntu:20.04 + +# Expand VM settings as required by ELK +CMD bash -c "if [ -w /proc/sys/vm/max_map_count ] && [ $(cat /proc/sys/vm/max_map_count) -lt 262144 ]; then echo 262144 > /proc/sys/vm/max_map_count; fi" diff --git a/docker-compose/elk.yml b/docker-compose/elk.yml new file mode 100644 index 0000000000000000000000000000000000000000..50060566f3824203c4b91e0332f853861ef6c7a4 --- /dev/null +++ b/docker-compose/elk.yml @@ -0,0 +1,38 @@ +# +# Docker compose file that launches an ELK stack. +# See https://elk-docker.readthedocs.io/ +# +# Defines: +# - elk-configure-host: Configures the hosts's kernel to be able to use the ELK stack +# - elk: ELK stack +# +version: '2' + +volumes: + elk-data: {} + +services: + elk-configure-host: + image: elk-configure-host + build: + context: elk-configure-host + container_name: ${CONTAINER_NAME_PREFIX}elk-configure-host + network_mode: ${NETWORK_MODE} + privileged: true + + elk: + image: elk + build: + context: elk + container_name: ${CONTAINER_NAME_PREFIX}elk + network_mode: ${NETWORK_MODE} + volumes: + - elk-data:/var/lib/elasticsearch + ports: + - "5601:5601" # kibana + - "9200:9200" # elasticsearch + - "5044:5044" # logstash beats input + - "1514:1514" # logstash syslog input + - "5959:5959" # logstash tcp json input + depends_on: + - elk-configure-host diff --git a/docker-compose/elk/Dockerfile b/docker-compose/elk/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..fda7ed40e1deafcb1ef144f63b6ba985f83430bf --- /dev/null +++ b/docker-compose/elk/Dockerfile @@ -0,0 +1,18 @@ +FROM sebp/elk + +# Give more time for ElasticSearch to startup on our poor dev laptops +ENV ES_CONNECT_RETRY=60 + +# Provide our logstash config +ADD logstash /etc/logstash/ + +# Provide our kibana config +# +# For now: +# 1. Start the container, +# 2. Make sure there are log messages in the database (to populate their patterns), +# 3. Run in this directory +# $ curl -X POST http://localhost:5601/api/saved_objects/_import -H "kbn-xsrf: true" --form file=@kibana/default-objects.ndjson +# +# See also https://www.elastic.co/guide/en/kibana/7.x/saved-objects-api-import.html +# and https://github.com/Bitergia/archimedes diff --git a/docker-compose/elk/kibana/default-objects.ndjson b/docker-compose/elk/kibana/default-objects.ndjson new file mode 100755 index 0000000000000000000000000000000000000000..3c9f118b21075ce52ecf859cda87c66c09fc8222 --- /dev/null +++ b/docker-compose/elk/kibana/default-objects.ndjson @@ -0,0 +1,3 @@ +{"attributes":{"fieldAttrs":"{\"level\":{\"count\":1},\"message\":{\"count\":1},\"program\":{\"count\":2}}","fields":"[]","timeFieldName":"@timestamp","title":"logstash-*"},"id":"5340ad20-923d-11eb-9cc8-2fc0b321a697","migrationVersion":{"index-pattern":"7.11.0"},"references":[],"type":"index-pattern","updated_at":"2021-03-31T18:16:04.716Z","version":"WzE0NiwxXQ=="} +{"attributes":{"columns":["extra.device","level","message"],"description":"","hits":0,"kibanaSavedObjectMeta":{"searchSourceJSON":"{\"highlightAll\":true,\"version\":true,\"query\":{\"query\":\"\",\"language\":\"kuery\"},\"filter\":[],\"indexRefName\":\"kibanaSavedObjectMeta.searchSourceJSON.index\"}"},"sort":[],"title":"Last hour","version":1},"id":"cf4947d0-924f-11eb-9cc8-2fc0b321a697","migrationVersion":{"search":"7.9.3"},"references":[{"id":"5340ad20-923d-11eb-9cc8-2fc0b321a697","name":"kibanaSavedObjectMeta.searchSourceJSON.index","type":"index-pattern"}],"type":"search","updated_at":"2021-03-31T18:35:04.269Z","version":"WzE2NywxXQ=="} +{"exportedCount":2,"missingRefCount":0,"missingReferences":[]} \ No newline at end of file diff --git a/docker-compose/elk/logstash/conf.d/02-beats-input.conf b/docker-compose/elk/logstash/conf.d/02-beats-input.conf new file mode 100644 index 0000000000000000000000000000000000000000..4ab52b37081aa46fdda7edd82a0395b9f73a3705 --- /dev/null +++ b/docker-compose/elk/logstash/conf.d/02-beats-input.conf @@ -0,0 +1,8 @@ +input { + beats { + port => 5044 + ssl => true + ssl_certificate => "/etc/pki/tls/certs/logstash-beats.crt" + ssl_key => "/etc/pki/tls/private/logstash-beats.key" + } +} diff --git a/docker-compose/elk/logstash/conf.d/03-syslog-input.conf b/docker-compose/elk/logstash/conf.d/03-syslog-input.conf new file mode 100644 index 0000000000000000000000000000000000000000..b859a357d505d9ea8d59eb3cf39d2af97b76a119 --- /dev/null +++ b/docker-compose/elk/logstash/conf.d/03-syslog-input.conf @@ -0,0 +1,5 @@ +input { + syslog { + port => 1514 + } +} diff --git a/docker-compose/elk/logstash/conf.d/04-tcp-input.conf b/docker-compose/elk/logstash/conf.d/04-tcp-input.conf new file mode 100644 index 0000000000000000000000000000000000000000..67def0887aef58ab0fd5d2d6cb65d5700a4088c7 --- /dev/null +++ b/docker-compose/elk/logstash/conf.d/04-tcp-input.conf @@ -0,0 +1,6 @@ +input { + tcp { + port => 5959 + codec => json + } +} diff --git a/docker-compose/elk/logstash/conf.d/10-syslog.conf b/docker-compose/elk/logstash/conf.d/10-syslog.conf new file mode 100644 index 0000000000000000000000000000000000000000..acce463cd42543ce87c5a51496aec25b06d21fdd --- /dev/null +++ b/docker-compose/elk/logstash/conf.d/10-syslog.conf @@ -0,0 +1,13 @@ +filter { + if [type] == "syslog" { + grok { + match => { "message" => "%{SYSLOGTIMESTAMP:syslog_timestamp} %{SYSLOGHOST:syslog_hostname} %{DATA:syslog_program}(?:\[%{POSINT:syslog_pid}\])?: %{GREEDYDATA:syslog_message}" } + add_field => [ "received_at", "%{@timestamp}" ] + add_field => [ "received_from", "%{host}" ] + } + syslog_pri { } + date { + match => [ "syslog_timestamp", "MMM d HH:mm:ss", "MMM dd HH:mm:ss" ] + } + } +} diff --git a/docker-compose/elk/logstash/conf.d/11-nginx.conf b/docker-compose/elk/logstash/conf.d/11-nginx.conf new file mode 100644 index 0000000000000000000000000000000000000000..d4a45db2d8454d982d52056cd035a9c8ef865389 --- /dev/null +++ b/docker-compose/elk/logstash/conf.d/11-nginx.conf @@ -0,0 +1,7 @@ +filter { + if [type] == "nginx-access" { + grok { + match => { "message" => "%{NGINXACCESS}" } + } + } +} diff --git a/docker-compose/elk/logstash/conf.d/30-output.conf b/docker-compose/elk/logstash/conf.d/30-output.conf new file mode 100644 index 0000000000000000000000000000000000000000..1893dd990ff62ce1a6b165fb449dc5afc4442268 --- /dev/null +++ b/docker-compose/elk/logstash/conf.d/30-output.conf @@ -0,0 +1,7 @@ +output { + elasticsearch { + hosts => ["localhost"] + manage_template => false + index => "logstash-%{+YYYY.MM.dd}" + } +} diff --git a/docker-compose/itango.yml b/docker-compose/itango.yml index d131d405371bc6ff7e59892587b2bd2aba9d1fd6..4e1e20ed9f4d8bda0bf1fffa2850ba8d53c9eee1 100644 --- a/docker-compose/itango.yml +++ b/docker-compose/itango.yml @@ -28,6 +28,7 @@ services: - TANGO_HOST=${TANGO_HOST} - XAUTHORITY=${XAUTHORITY} - DISPLAY=${DISPLAY} + user: ${CONTAINER_EXECUTION_UID} stdin_open: true tty: true entrypoint: diff --git a/docker-compose/itango/Dockerfile b/docker-compose/itango/Dockerfile index 3e044037cc333e6042a707ba652b32a32713589d..147ad6f346f8d467dd4c738b8dcd9fd97be6bc14 100644 --- a/docker-compose/itango/Dockerfile +++ b/docker-compose/itango/Dockerfile @@ -8,8 +8,8 @@ RUN sudo apt-get -y update && \ sudo apt-get -y install apt-file apt-transport-https apt-utils aptitude && \ sudo aptitude -y install htop iftop iproute2 mc most net-tools tcpdump telnet tmux traceroute vim xterm && \ sudo aptitude clean && \ - sudo aptitude autoclean && \ - sudo pip3 install "opcua >= 0.98.9" asyncua astropy && \ - sudo pip3 uninstall -y jedi parso && \ - sudo pip3 install "parso == 0.7.1" "jedi == 0.17.2" && \ + sudo aptitude autoclean + +COPY lofar-requirements.txt /lofar-requirements.txt +RUN sudo pip3 install -r /lofar-requirements.txt && \ sudo pip3 cache purge diff --git a/docker-compose/itango/lofar-requirements.txt b/docker-compose/itango/lofar-requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..375287a0cbc6e4668a1d72162e376194222a834d --- /dev/null +++ b/docker-compose/itango/lofar-requirements.txt @@ -0,0 +1,5 @@ +parso == 0.7.1 +jedi == 0.17.2 +opcua >= 0.98.9 +asyncua +astropy diff --git a/docker-compose/jupyter.yml b/docker-compose/jupyter.yml index f1970916f7e9facc25908c08c63f0938be0c4eb7..71daf48e08afed3ae2ba8c873b216561b9bde213 100644 --- a/docker-compose/jupyter.yml +++ b/docker-compose/jupyter.yml @@ -26,6 +26,7 @@ services: - DISPLAY=${DISPLAY} ports: - "8888:8888" + user: ${CONTAINER_EXECUTION_UID} working_dir: /jupyter-notebooks entrypoint: - /usr/local/bin/wait-for-it.sh @@ -33,5 +34,5 @@ services: - --timeout=30 - --strict - -- - - /usr/bin/tini -- jupyter notebook --port=8888 --no-browser --ip=0.0.0.0 --allow-root --NotebookApp.token= --NotebookApp.password= + - /usr/bin/tini -- /usr/local/bin/jupyter-notebook --port=8888 --no-browser --ip=0.0.0.0 --allow-root --NotebookApp.token= --NotebookApp.password= restart: on-failure diff --git a/docker-compose/jupyter/Dockerfile b/docker-compose/jupyter/Dockerfile index 336179e73d9320d6f37edf6cdd13231ef291bcec..b9812188296627e9a5e2662007efa71ff78da90e 100644 --- a/docker-compose/jupyter/Dockerfile +++ b/docker-compose/jupyter/Dockerfile @@ -1,22 +1,27 @@ ARG VERSION=latest FROM nexus.engageska-portugal.pt/ska-docker/tango-itango:${VERSION} -RUN pip3 install jupyter -RUN pip3 install ipykernel -RUN pip3 install jupyter_bokeh +RUN sudo pip3 install jupyter +RUN sudo pip3 install ipykernel +RUN sudo pip3 install jupyter_bokeh # Configure jupyter_bokeh RUN sudo mkdir -p /usr/share/jupyter /usr/etc RUN sudo chmod a+rwx /usr/share/jupyter /usr/etc -RUN jupyter nbextension install --sys-prefix --symlink --py jupyter_bokeh -RUN jupyter nbextension enable jupyter_bokeh --py --sys-prefix +RUN sudo jupyter nbextension install --sys-prefix --symlink --py jupyter_bokeh +RUN sudo jupyter nbextension enable jupyter_bokeh --py --sys-prefix # Install profiles for ipython & jupyter COPY ipython-profiles /opt/ipython-profiles/ RUN sudo chown tango.tango -R /opt/ipython-profiles COPY jupyter-kernels /usr/local/share/jupyter/kernels/ +# Install patched jupyter executable +RUN sudo pip3 install python-logstash-async +COPY jupyter-notebook /usr/local/bin/jupyter-notebook + # Add Tini. Tini operates as a process subreaper for jupyter. This prevents kernel crashes. ENV TINI_VERSION v0.6.0 +ENV JUPYTER_RUNTIME_DIR=/tmp ADD https://github.com/krallin/tini/releases/download/${TINI_VERSION}/tini /usr/bin/tini RUN sudo chmod +x /usr/bin/tini diff --git a/docker-compose/jupyter/jupyter-notebook b/docker-compose/jupyter/jupyter-notebook new file mode 100755 index 0000000000000000000000000000000000000000..59613a137cc1bb5c86b4cd7c82f3a2cb1f9abde3 --- /dev/null +++ b/docker-compose/jupyter/jupyter-notebook @@ -0,0 +1,28 @@ +#!/usr/bin/python3 +# -*- coding: utf-8 -*- +# An adjustment of the `jupyter-notebook' executable patched to: +# - log to the ELK stack +# +# We go straight for the notebook executable here, as the "jupyter" command +# execvp's into the requested notebook subcommand, erasing all configuration +# we set here. +import re +import sys + +from notebook.notebookapp import main + +from logstash_async.handler import AsynchronousLogstashHandler, LogstashFormatter +import logging + +if __name__ == '__main__': + # log to the tcp_input of logstash in our ELK stack + handler = AsynchronousLogstashHandler("elk", 5959, database_path='/tmp/pending_log_messages.db') + + # add to logger of Jupyter traitlets Application. As that logger is configured not to propagate + # messages upward, we need to configure it directly. + logger = logging.getLogger("NotebookApp") + logger.addHandler(handler) + logger.setLevel(logging.DEBUG) + + sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0]) + sys.exit(main()) diff --git a/docker-compose/lofar-device-base/Dockerfile b/docker-compose/lofar-device-base/Dockerfile index cd6b59511f7e0a3e515ab282775edeadec6b9aaf..85326698052c8e9e48b1270377a513e3dbcda5d3 100644 --- a/docker-compose/lofar-device-base/Dockerfile +++ b/docker-compose/lofar-device-base/Dockerfile @@ -1,4 +1,7 @@ FROM nexus.engageska-portugal.pt/ska-docker/tango-itango:latest COPY lofar-requirements.txt /lofar-requirements.txt -RUN pip3 install -r /lofar-requirements.txt +RUN sudo pip3 install -r /lofar-requirements.txt + +ENV TANGO_LOG_PATH=/var/log/tango +RUN sudo mkdir -p /var/log/tango && sudo chmod a+rwx /var/log/tango diff --git a/docker-compose/lofar-device-base/lofar-requirements.txt b/docker-compose/lofar-device-base/lofar-requirements.txt index 90d21efe0e5ac0601204d4f05ce7efcd16dca2de..7ed18f76527891ca48953150977b2f23703c9baa 100644 --- a/docker-compose/lofar-device-base/lofar-requirements.txt +++ b/docker-compose/lofar-device-base/lofar-requirements.txt @@ -1,2 +1,3 @@ opcua >= 0.98.9 -astropy +astropy +python-logstash-async