# This file is part of cloud-init. See LICENSE file for license information. import abc import fcntl import json import six import os import re import struct import threading import time from cloudinit import log as logging from cloudinit.registry import DictRegistry from cloudinit import (url_helper, util) from datetime import datetime if six.PY2: from multiprocessing.queues import JoinableQueue as JQueue else: from queue import Queue as JQueue LOG = logging.getLogger(__name__) class ReportException(Exception): pass @six.add_metaclass(abc.ABCMeta) class ReportingHandler(object): """Base class for report handlers. Implement :meth:`~publish_event` for controlling what the handler does with an event. """ @abc.abstractmethod def publish_event(self, event): """Publish an event.""" def flush(self): """Ensure ReportingHandler has published all events""" pass class LogHandler(ReportingHandler): """Publishes events to the cloud-init log at the ``DEBUG`` log level.""" def __init__(self, level="DEBUG"): super(LogHandler, self).__init__() if isinstance(level, int): pass else: input_level = level try: level = getattr(logging, level.upper()) except Exception: LOG.warning("invalid level '%s', using WARN", input_level) level = logging.WARN self.level = level def publish_event(self, event): logger = logging.getLogger( '.'.join(['cloudinit', 'reporting', event.event_type, event.name])) logger.log(self.level, event.as_string()) class PrintHandler(ReportingHandler): """Print the event as a string.""" def publish_event(self, event): print(event.as_string()) class WebHookHandler(ReportingHandler): def __init__(self, endpoint, consumer_key=None, token_key=None, token_secret=None, consumer_secret=None, timeout=None, retries=None): super(WebHookHandler, self).__init__() if any([consumer_key, token_key, token_secret, consumer_secret]): self.oauth_helper = url_helper.OauthUrlHelper( consumer_key=consumer_key, token_key=token_key, token_secret=token_secret, consumer_secret=consumer_secret) else: self.oauth_helper = None self.endpoint = endpoint self.timeout = timeout self.retries = retries self.ssl_details = util.fetch_ssl_details() def publish_event(self, event): if self.oauth_helper: readurl = self.oauth_helper.readurl else: readurl = url_helper.readurl try: return readurl( self.endpoint, data=json.dumps(event.as_dict()), timeout=self.timeout, retries=self.retries, ssl_details=self.ssl_details) except Exception: LOG.warning("failed posting event: %s", event.as_string()) class HyperVKvpReportingHandler(ReportingHandler): """ Reports events to a Hyper-V host using Key-Value-Pair exchange protocol and can be used to obtain high level diagnostic information from the host. To use this facility, the KVP user-space daemon (hv_kvp_daemon) has to be running. It reads the kvp_file when the host requests the guest to enumerate the KVP's. This reporter collates all events for a module (origin|name) in a single json string in the dictionary. For more information, see https://technet.microsoft.com/en-us/library/dn798287.aspx#Linux%20guests """ HV_KVP_EXCHANGE_MAX_VALUE_SIZE = 2048 HV_KVP_EXCHANGE_MAX_KEY_SIZE = 512 HV_KVP_RECORD_SIZE = (HV_KVP_EXCHANGE_MAX_KEY_SIZE + HV_KVP_EXCHANGE_MAX_VALUE_SIZE) EVENT_PREFIX = 'CLOUD_INIT' MSG_KEY = 'msg' RESULT_KEY = 'result' DESC_IDX_KEY = 'msg_i' JSON_SEPARATORS = (',', ':') KVP_POOL_FILE_GUEST = '/var/lib/hyperv/.kvp_pool_1' def __init__(self, kvp_file_path=KVP_POOL_FILE_GUEST, event_types=None): super(HyperVKvpReportingHandler, self).__init__() self._kvp_file_path = kvp_file_path self._event_types = event_types self.q = JQueue() self.kvp_file = None self.incarnation_no = self._get_incarnation_no() self.event_key_prefix = u"{0}|{1}".format(self.EVENT_PREFIX, self.incarnation_no) self._current_offset = 0 self.publish_thread = threading.Thread( target=self._publish_event_routine) self.publish_thread.daemon = True self.publish_thread.start() def _get_incarnation_no(self): """ use the time passed as the incarnation number. the incarnation number is the number which are used to distinguish the old data stored in kvp and the new data. """ uptime_str = util.uptime() try: return int(time.time() - float(uptime_str)) except ValueError: LOG.warning("uptime '%s' not in correct format.", uptime_str) return 0 def _iterate_kvps(self, offset): """iterate the kvp file from the current offset.""" try: with open(self._kvp_file_path, 'rb+') as f: self.kvp_file = f fcntl.flock(f, fcntl.LOCK_EX) f.seek(offset) record_data = f.read(self.HV_KVP_RECORD_SIZE) while len(record_data) == self.HV_KVP_RECORD_SIZE: self._current_offset += self.HV_KVP_RECORD_SIZE kvp_item = self._decode_kvp_item(record_data) yield kvp_item record_data = f.read(self.HV_KVP_RECORD_SIZE) fcntl.flock(f, fcntl.LOCK_UN) finally: self.kvp_file = None def _event_key(self, event): """ the event key format is: CLOUD_INIT||| """ return u"{0}|{1}|{2}".format(self.event_key_prefix, event.event_type, event.name) def _encode_kvp_item(self, key, value): data = (struct.pack("%ds%ds" % ( self.HV_KVP_EXCHANGE_MAX_KEY_SIZE, self.HV_KVP_EXCHANGE_MAX_VALUE_SIZE), key.encode('utf-8'), value.encode('utf-8'))) return data def _decode_kvp_item(self, record_data): record_data_len = len(record_data) if record_data_len != self.HV_KVP_RECORD_SIZE: raise ReportException( "record_data len not correct {0} {1}." .format(record_data_len, self.HV_KVP_RECORD_SIZE)) k = (record_data[0:self.HV_KVP_EXCHANGE_MAX_KEY_SIZE].decode('utf-8') .strip('\x00')) v = ( record_data[ self.HV_KVP_EXCHANGE_MAX_KEY_SIZE:self.HV_KVP_RECORD_SIZE ].decode('utf-8').strip('\x00')) return {'key': k, 'value': v} def _update_kvp_item(self, record_data): if self.kvp_file is None: raise ReportException( "kvp file '{0}' not opened." .format(self._kvp_file_path)) self.kvp_file.seek(-self.HV_KVP_RECORD_SIZE, 1) self.kvp_file.write(record_data) def _append_kvp_item(self, record_data): with open(self._kvp_file_path, 'rb+') as f: fcntl.flock(f, fcntl.LOCK_EX) # seek to end of the file f.seek(0, 2) f.write(record_data) f.flush() fcntl.flock(f, fcntl.LOCK_UN) self._current_offset = f.tell() def _break_down(self, key, meta_data, description): del meta_data[self.MSG_KEY] des_in_json = json.dumps(description) des_in_json = des_in_json[1:(len(des_in_json) - 1)] i = 0 result_array = [] message_place_holder = "\"" + self.MSG_KEY + "\":\"\"" while True: meta_data[self.DESC_IDX_KEY] = i meta_data[self.MSG_KEY] = '' data_without_desc = json.dumps(meta_data, separators=self.JSON_SEPARATORS) room_for_desc = ( self.HV_KVP_EXCHANGE_MAX_VALUE_SIZE - len(data_without_desc) - 8) value = data_without_desc.replace( message_place_holder, '"{key}":"{desc}"'.format( key=self.MSG_KEY, desc=des_in_json[:room_for_desc])) result_array.append(self._encode_kvp_item(key, value)) i += 1 des_in_json = des_in_json[room_for_desc:] if len(des_in_json) == 0: break return result_array def _encode_event(self, event): """ encode the event into kvp data bytes. if the event content reaches the maximum length of kvp value. then it would be cut to multiple slices. """ key = self._event_key(event) meta_data = { "name": event.name, "type": event.event_type, "ts": (datetime.utcfromtimestamp(event.timestamp) .isoformat() + 'Z'), } if hasattr(event, self.RESULT_KEY): meta_data[self.RESULT_KEY] = event.result meta_data[self.MSG_KEY] = event.description value = json.dumps(meta_data, separators=self.JSON_SEPARATORS) # if it reaches the maximum length of kvp value, # break it down to slices. # this should be very corner case. if len(value) > self.HV_KVP_EXCHANGE_MAX_VALUE_SIZE: return self._break_down(key, meta_data, event.description) else: data = self._encode_kvp_item(key, value) return [data] def _publish_event_routine(self): while True: try: event = self.q.get(block=True) need_append = True try: if not os.path.exists(self._kvp_file_path): LOG.warning( "skip writing events %s to %s. file not present.", event.as_string(), self._kvp_file_path) encoded_event = self._encode_event(event) # for each encoded_event for encoded_data in (encoded_event): for kvp in self._iterate_kvps(self._current_offset): match = ( re.match( r"^{0}\|(\d+)\|.+" .format(self.EVENT_PREFIX), kvp['key'] )) if match: match_groups = match.groups(0) if int(match_groups[0]) < self.incarnation_no: need_append = False self._update_kvp_item(encoded_data) continue if need_append: self._append_kvp_item(encoded_data) except IOError as e: LOG.warning( "failed posting event to kvp: %s e:%s", event.as_string(), e) finally: self.q.task_done() # when main process exits, q.get() will through EOFError # indicating we should exit this thread. except EOFError: return # since the saving to the kvp pool can be a time costing task # if the kvp pool already contains a chunk of data, # so defer it to another thread. def publish_event(self, event): if (not self._event_types or event.event_type in self._event_types): self.q.put(event) def flush(self): LOG.debug('HyperVReportingHandler flushing remaining events') self.q.join() available_handlers = DictRegistry() available_handlers.register_item('log', LogHandler) available_handlers.register_item('print', PrintHandler) available_handlers.register_item('webhook', WebHookHandler) available_handlers.register_item('hyperv', HyperVKvpReportingHandler) # vi: ts=4 expandtab