diff options
author | Joshua Harlow <harlowja@yahoo-inc.com> | 2012-06-11 17:12:49 -0700 |
---|---|---|
committer | Joshua Harlow <harlowja@yahoo-inc.com> | 2012-06-11 17:12:49 -0700 |
commit | 6a58c8ee7345c53dda47ae0845b0e8cae0043dd4 (patch) | |
tree | 993d8ed17786d0c1dd561d007a8a07205fddd616 | |
parent | a987baa675c19c8c2064c7eee786e93468a9c8ec (diff) | |
download | vyos-cloud-init-6a58c8ee7345c53dda47ae0845b0e8cae0043dd4.tar.gz vyos-cloud-init-6a58c8ee7345c53dda47ae0845b0e8cae0043dd4.zip |
This will now just contain the "public" view of the cloud objects, backed by whatever other set of objects.
-rw-r--r-- | cloudinit/cloud.py | 482 |
1 files changed, 11 insertions, 471 deletions
diff --git a/cloudinit/cloud.py b/cloudinit/cloud.py index 80d4f1ce..765e7d3a 100644 --- a/cloudinit/cloud.py +++ b/cloudinit/cloud.py @@ -20,158 +20,23 @@ # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. -from time import time - -import cPickle as pickle - -import contextlib -import copy -import os -import sys -import weakref - -from cloudinit.settings import (PER_INSTANCE, PER_ALWAYS) -from cloudinit.settings import (OLD_CLOUD_CONFIG, CLOUD_CONFIG) - -from cloudinit import (get_builtin_cfg, get_base_cfg) +from cloudinit import distros +from cloudinit import helpers from cloudinit import log as logging -from cloudinit import sources -from cloudinit import util -from cloudinit import handlers - -from cloudinit import user_data as ud -from cloudinit.user_data import boot_hook as bh_part -from cloudinit.user_data import cloud_config as cc_part -from cloudinit.user_data import processor as ud_proc -from cloudinit.user_data import shell_script as ss_part -from cloudinit.user_data import upstart_job as up_part LOG = logging.getLogger(__name__) -class CloudSemaphores(object): - def __init__(self, sem_path): - self.sem_path = sem_path - - # acquire lock on 'name' for given 'freq' and run function 'func' - # if 'clear_on_fail' is True and 'func' throws an exception - # then remove the lock (so it would run again) - def run_functor(self, name, freq, functor, args=None, clear_on_fail=False): - if not args: - args = [] - if self.has_run(name, freq): - LOG.debug("%s already ran %s", name, freq) - return False - with self.lock(name, freq, clear_on_fail) as lock: - if not lock: - raise RuntimeError("Failed to acquire lock on %s" % name) - else: - LOG.debug("Running %s with args %s using lock %s", func, args, lock) - func(*args) - return True - - @contextlib.contextmanager - def lock(self, name, freq, clear_on_fail=False): - try: - yield self._acquire(name, freq) - except: - if clear_on_fail: - self.clear(name, freq) - raise - - def clear(self, name, freq): - sem_file = self._getpath(name, freq) - try: - util.del_file(sem_file) - except IOError: - return False - return True - - def _acquire(self, name, freq): - if self.has_run(name, freq): - return None - # This is a race condition since nothing atomic is happening - # here, but this should be ok due to the nature of when - # and where cloud-init runs... (file writing is not a lock..) - sem_file = self._getpath(name, freq) - contents = "%s: %s\n" % (os.getpid(), time()) - try: - util.write_file(sem_file, contents) - except (IOError, OSError): - return None - return sem_file - - def has_run(self, name, freq): - if freq == PER_ALWAYS: - return False - sem_file = self._get_path(name, freq) - if os.path.exists(sem_file): - return True - return False - - def _get_path(self, name, freq): - sem_path = self.sem_path - if freq == PER_INSTANCE: - return os.path.join(sem_path, name) - return os.path.join(sem_path, "%s.%s" % (name, freq)) - - -class CloudPaths(object): - def __init__(self, sys_info): - self.cloud_dir = sys_info['cloud_dir'] - self.instance_link = os.path.join(self.cloud_dir, 'instance') - self.boot_finished = os.path.join(self.instance_link, "boot-finished") - self.upstart_conf_d = sys_info.get('upstart_dir') - self.template_dir = sys_info['templates_dir'] - self.seed_dir = os.path.join(self.cloud_dir, 'seed') - self.datasource = None - self.lookups = { - "handlers": "handlers", - "scripts": "scripts", - "sem": "sem", - "boothooks": "boothooks", - "userdata_raw": "user-data.txt", - "userdata": "user-data.txt.i", - "obj_pkl": "obj.pkl", - "cloud_config": "cloud-config.txt", - "data": "data", - } - - # get_ipath_cur: get the current instance path for an item - def get_ipath_cur(self, name=None): - ipath = os.path.join(self.cloud_dir, 'instance') - add_on = self.lookups.get(name) - if add_on: - ipath = os.path.join(ipath, add_on) - return ipath - - # get_cpath : get the "clouddir" (/var/lib/cloud/<name>) - # for a name in dirmap - def get_cpath(self, name=None): - cpath = self.var_dir - add_on = self.lookups.get(name) - if add_on: - cpath = os.path.join(cpath, add_on) - return cpath - - # get_ipath : get the instance path for a name in pathmap - # (/var/lib/cloud/instances/<instance>/<name>) - def get_ipath(self, name=None): - if not self.datasource: - raise RuntimeError("Unable to get instance path, datasource not available/set.") - iid = self.datasource.get_instance_id() - ipath = os.path.join(self.cloud_dir, 'instances', iid) - add_on = self.lookups.get(name) - if add_on: - ipath = os.path.join(ipath, add_on) - return ipath - +class Cloud(object): + def __init__(self, datasource, paths, cfg): + self.datasource = datasource + self.paths = paths + self.cfg = cfg + self.distro = distros.fetch(cfg, self) + self.runners = helpers.Runners(paths) -class CloudSimple(object): - def __init__(self, ci): - self.datasource = init.datasource - self.paths = init.paths - self.cfg = copy.deepcopy(ci.cfg) + def run(self, name, functor, args, freq=None, clear_on_fail=False): + return self.runners.run(name, functor, args, freq, clear_on_fail) def get_userdata(self): return self.datasource.get_userdata() @@ -199,328 +64,3 @@ class CloudSimple(object): def get_ipath(self, name=None): return self.paths.get_ipath(name) - - -class CloudInit(object): - def __init__(self, ds_deps=None): - self.datasource = None - if ds_deps: - self.ds_deps = ds_deps - else: - self.ds_deps = [sources.DEP_FILESYSTEM, sources.DEP_NETWORK] - self.cfg = self._read_cfg() - self.paths = CloudPaths(self.cfg['system_info']) - - def _read_cfg_old(self): - # support reading the old ConfigObj format file and merging - # it into the yaml dictionary - try: - from configobj import ConfigObj - except ImportError: - ConfigObj = None - if not ConfigObj: - return {} - old_cfg = ConfigObj(OLD_CLOUD_CONFIG) - return dict(old_cfg) - - def _initial_subdirs(self): - c_dir = self.paths.cloud_dir - initial_dirs = [ - os.path.join(c_dir, 'scripts'), - os.path.join(c_dir, 'scripts', 'per-instance'), - os.path.join(c_dir, 'scripts', 'per-once'), - os.path.join(c_dir, 'scripts', 'per-boot'), - os.path.join(c_dir, 'seed'), - os.path.join(c_dir, 'instances'), - os.path.join(c_dir, 'handlers'), - os.path.join(c_dir, 'sem'), - os.path.join(c_dir, 'data'), - ] - return initial_dirs - - def purge_cache(self, rmcur=True): - rmlist = [] - rmlist.append(self.paths.boot_finished) - if rmcur: - rmlist.append(self.paths.instance_link) - for f in rmlist: - util.unlink(f) - return len(rmlist) - - def init_fs(self): - util.ensure_dirs(self._initial_subdirs()) - log_file = util.get_cfg_option_str(self.cfg, 'def_log_file', None) - perms = util.get_cfg_option_str(self.cfg, 'syslog_fix_perms', None) - if log_file: - util.ensure_file(log_file) - if perms: - (u, g) = perms.split(':', 1) - if u == "-1" or u == "None": - u = None - if g == "-1" or g == "None": - g = None - util.chownbyname(log_file, u, g) - - def _read_cfg(self): - starting_config = get_builtin_cfg() - try: - conf = get_base_cfg(CLOUD_CONFIG, starting_config) - except Exception: - conf = starting_config - old_conf = self._read_cfg_old() - conf = util.mergedict(conf, old_conf) - return conf - - def _restore_from_cache(self): - pickled_fn = self.paths.get_ipath_cur('obj_pkl') - try: - # we try to restore from a current link and static path - # by using the instance link, if purge_cache was called - # the file wont exist - return pickle.loads(util.load_file(pickled_fn)) - except Exception as e: - LOG.debug("Failed loading pickled datasource from %s due to %s", pickled_fn, e) - return False - - def write_to_cache(self): - pickled_fn = self.paths.get_ipath_cur("obj_pkl") - try: - contents = pickle.dumps(self.datasource) - util.write_file(pickled_fn, contents, mode=0400) - except Exception as e: - LOG.debug("Failed pickling datasource to %s due to: %s", pickled_fn, e) - return False - - def _get_processor(self): - return ud_proc.UserDataProcessor(self.paths) - - def _get_datasources(self): - # Any config provided??? - pkg_list = self.cfg.get('datasource_pkg_list') or [] - # Add the defaults at the end - for n in [util.obj_name(sources), '']: - if n not in pkg_list: - pkg_list.append(n) - cfg_list = self.cfg.get('datasource_list') or [] - return (cfg_list, pkg_list) - - def get_data_source(self): - if self.datasource: - return True - ds = self._restore_from_cache() - if ds: - LOG.debug("Restored from cache datasource: %s" % ds) - else: - (cfg_list, pkg_list) = self._get_datasources() - ud_proc = self._get_processor() - (ds, dsname) = sources.find_source(self.cfg, - self.ds_deps, - cfg_list=cfg_list, - pkg_list=pkg_list, - ud_proc=ud_proc) - LOG.debug("Loaded datasource %s - %s", dsname, ds) - self.datasource = ds - # This allows the paths obj to have an ipath function that works - self.paths.datasource = ds - return True - - def set_cur_instance(self): - # Ensure we are hooked into the right symlink for the current instance - idir = self.paths.get_ipath() - util.del_file(self.paths.instance_link) - util.sym_link(idir, self.paths.instance_link) - - dlist = [] - for d in ["handlers", "scripts", "sem"]: - dlist.append(os.path.join(idir, d)) - util.ensure_dirs(dlist) - - # Write out information on what is being used for the current instance - # and what may have been used for a previous instance... - dp = self.paths.get_cpath('data') - ds = "%s: %s\n" % (self.datasource.__class__, self.datasource) - previous_ds = '' - ds_fn = os.path.join(idir, 'datasource') - try: - previous_ds = util.load_file(ds_fn).strip() - except IOError as e: - pass - if not previous_ds: - # TODO: ?? is this right - previous_ds = ds - util.write_file(ds_fn, ds) - util.write_file(os.path.join(dp, 'previous-datasource'), previous_ds) - iid = self.datasource.get_instance_id() - previous_iid = '' - p_iid_fn = os.path.join(dp, 'previous-instance-id') - try: - previous_iid = util.load_file(p_iid_fn).strip() - except IOError as e: - pass - if not previous_iid: - # TODO: ?? is this right - previous_iid = iid - util.write_file(p_iid_fn, "%s\n" % previous_iid) - - def update_cache(self): - self.write_to_cache() - self.store_userdata() - - def store_userdata(self): - raw_ud = "%s" % (self.datasource.get_userdata_raw()) - util.write_file(self.paths.get_ipath('userdata_raw'), raw_ud, 0600) - ud = "%s" % (self.datasource.get_userdata()) - util.write_file(self.paths.get_ipath('userdata'), ud, 0600) - - def consume_userdata(self, frequency=PER_INSTANCE): - cdir = self.paths.get_cpath("handlers") - idir = self.paths.get_ipath("handlers") - - # Add the path to the plugins dir to the top of our list for import - # instance dir should be read before cloud-dir - sys.path.insert(0, cdir) - sys.path.insert(0, idir) - - # Data will be a little proxy that modules can use - data = CloudSimple(self) - - # This keeps track of all the active handlers - handlers = CloudHandlers(self) - - # Add handlers in cdir - potential_handlers = utils.find_modules(cdir) - for (fname, modname) in potential_handlers.iteritems(): - try: - mod = parts.fixup_module(importer.import_module(modname)) - types = handlers.register(mod) - LOG.debug("Added handler for [%s] from %s", types, fname) - except: - LOG.exception("Failed to register handler from %s", fname) - - def_handlers = handlers.register_defaults() - if def_handlers: - LOG.debug("Registered default handlers for [%s]", def_handlers) - - # Init the handlers first - # Ensure userdata fetched before activation - called = [] - for (_mtype, mod) in handlers.iteritems(): - if mod in called: - continue - parts.call_begin(mod, data, frequency) - called.append(mod) - - # Walk the user data - part_data = { - 'handlers': handlers, - 'handlerdir': idir, - 'data': data, - 'frequency': frequency, - 'handlercount': 0, - } - ud.walk(data.get_userdata(), parts.walker_callback, data=part_data) - - # Give callbacks opportunity to finalize - called = [] - for (_mtype, mod) in handlers.iteritems(): - if mod in called: - continue - parts.call_end(mod, data, frequency) - called.append(mod) - - -class CloudHandlers(object): - - def __init__(self, paths): - self.paths = paths - self.registered = {} - - def __contains__(self, item): - return self.is_registered(item) - - def __getitem__(self, key): - return self._get_handler(key) - - def is_registered(self, content_type): - return content_type in self.registered - - def register(self, mod): - types = set() - for t in mod.list_types(): - self.registered[t] = handler - types.add(t) - return types - - def _get_handler(self, content_type): - return self.registered[content_type] - - def items(self): - return self.registered.items() - - def iteritems(self): - return self.registered.iteritems() - - def _get_default_handlers(self): - def_handlers = [] - if self.paths.get_ipath("cloud_config"): - def_handlers.append(cc_part.CloudConfigPartHandler(self.paths.get_ipath("cloud_config"))) - if self.paths.get_ipath_cur('scripts'): - def_handlers.append(ss_part.ShellScriptPartHandler(self.paths.get_ipath_cur('scripts'))) - if self.paths.get_ipath("boothooks"): - def_handlers.append(bh_part.BootHookPartHandler(self.paths.get_ipath("boothooks"))) - if self.paths.upstart_conf_d: - def_handlers.append(up_part.UpstartJobPartHandler(self.paths.upstart_conf_d)) - return def_handlers - - def register_defaults(self): - registered = set() - for h in self._get_default_handlers(): - for t in h.list_types(): - if not self.is_registered(t) - self.register_handler(t, h) - registered.add(t) - return registered - - -class CloudConfig(object): - def __init__(self, cfgfile, cloud): - self.cloud = cloud - self.cfg = self._get_config(cfgfile) - self.paths = cloud.paths - self.sems = CloudSemaphores(self.paths.get_ipath("sem")) - - def _get_config(self, cfgfile): - cfg = None - try: - cfg = util.read_conf(cfgfile) - except: - LOG.exception(("Failed loading of cloud config '%s'. " - "Continuing with empty config."), cfgfile) - if not cfg: - cfg = {} - - ds_cfg = None - try: - ds_cfg = self.cloud.datasource.get_config_obj() - except: - LOG.exception("Failed loading of datasource config.") - if not ds_cfg: - ds_cfg = {} - - cfg = util.mergedict(cfg, ds_cfg) - cloud_cfg = self.cloud.cfg or {} - return util.mergedict(cfg, cloud_cfg) - - def extract(self, name): - modname = handlers.form_module_name(name) - if not modname: - return None - return handlers.fixup_module(importer.import_module(modname)) - - def handle(self, name, mod, args, freq=None): - def_freq = mod.frequency - if not freq: - freq = def_freq - c_name = "config-%s" % (name) - real_args = [name, copy.deepcopy(self.cfg), CloudSimple(self.cloud), LOG, copy.deepcopy(args)] - return self.sems.run_functor(c_name, freq, mod.handle, real_args) |