diff options
author | Joshua Harlow <harlowja@yahoo-inc.com> | 2012-06-07 13:48:49 -0700 |
---|---|---|
committer | Joshua Harlow <harlowja@yahoo-inc.com> | 2012-06-07 13:48:49 -0700 |
commit | 9aa8060d463c8cdefa6f8b381584838794d4c509 (patch) | |
tree | fd93ef338fb668b6d2c38b43911f1b96c6b393f1 | |
parent | 7adde5f221ab3e8d79d3cf053b374c541a75fbf1 (diff) | |
download | vyos-cloud-init-9aa8060d463c8cdefa6f8b381584838794d4c509.tar.gz vyos-cloud-init-9aa8060d463c8cdefa6f8b381584838794d4c509.zip |
Add this file which will contain the cloud init and cloud config classes.
-rw-r--r-- | cloudinit/cloud.py | 368 |
1 files changed, 368 insertions, 0 deletions
diff --git a/cloudinit/cloud.py b/cloudinit/cloud.py new file mode 100644 index 00000000..eb71439b --- /dev/null +++ b/cloudinit/cloud.py @@ -0,0 +1,368 @@ +import os + +import cPickle as pickle + + +class CloudInit(object): + cfg = None + part_handlers = {} + old_conffile = '/etc/ec2-init/ec2-config.cfg' + ds_deps = [DataSource.DEP_FILESYSTEM, DataSource.DEP_NETWORK] + datasource = None + cloud_config_str = '' + datasource_name = '' + + builtin_handlers = [] + + def __init__(self, ds_deps=None, sysconfig=system_config): + self.builtin_handlers = [ + ['text/x-shellscript', self.handle_user_script, per_always], + ['text/cloud-config', self.handle_cloud_config, per_always], + ['text/upstart-job', self.handle_upstart_job, per_instance], + ['text/cloud-boothook', self.handle_cloud_boothook, per_always], + ] + + if ds_deps != None: + self.ds_deps = ds_deps + + self.sysconfig = sysconfig + + self.cfg = self.read_cfg() + + def read_cfg(self): + if self.cfg: + return(self.cfg) + + try: + conf = util.get_base_cfg(self.sysconfig, cfg_builtin, parsed_cfgs) + except Exception: + conf = get_builtin_cfg() + + # support reading the old ConfigObj format file and merging + # it into the yaml dictionary + try: + from configobj import ConfigObj + oldcfg = ConfigObj(self.old_conffile) + if oldcfg is None: + oldcfg = {} + conf = util.mergedict(conf, oldcfg) + except: + pass + + return(conf) + + def restore_from_cache(self): + try: + # we try to restore from a current link and static path + # by using the instance link, if purge_cache was called + # the file wont exist + cache = get_ipath_cur('obj_pkl') + f = open(cache, "rb") + data = cPickle.load(f) + f.close() + self.datasource = data + return True + except: + return False + + def write_to_cache(self): + cache = self.get_ipath("obj_pkl") + try: + os.makedirs(os.path.dirname(cache)) + except OSError as e: + if e.errno != errno.EEXIST: + return False + + try: + f = open(cache, "wb") + cPickle.dump(self.datasource, f) + f.close() + os.chmod(cache, 0400) + except: + raise + + def get_data_source(self): + if self.datasource is not None: + return True + + if self.restore_from_cache(): + log.debug("restored from cache type %s" % self.datasource) + return True + + cfglist = self.cfg['datasource_list'] + dslist = list_sources(cfglist, self.ds_deps) + dsnames = [f.__name__ for f in dslist] + + log.debug("searching for data source in %s" % dsnames) + for cls in dslist: + ds = cls.__name__ + try: + s = cls(sys_cfg=self.cfg) + if s.get_data(): + self.datasource = s + self.datasource_name = ds + log.debug("found data source %s" % ds) + return True + except Exception as e: + log.warn("get_data of %s raised %s" % (ds, e)) + util.logexc(log) + msg = "Did not find data source. searched classes: %s" % dsnames + log.debug(msg) + raise DataSourceNotFoundException(msg) + + def set_cur_instance(self): + try: + os.unlink(cur_instance_link) + except OSError as e: + if e.errno != errno.ENOENT: + raise + + iid = self.get_instance_id() + os.symlink("./instances/%s" % iid, cur_instance_link) + idir = self.get_ipath() + dlist = [] + for d in ["handlers", "scripts", "sem"]: + dlist.append("%s/%s" % (idir, d)) + + util.ensure_dirs(dlist) + + ds = "%s: %s\n" % (self.datasource.__class__, str(self.datasource)) + dp = self.get_cpath('data') + util.write_file("%s/%s" % (idir, 'datasource'), ds) + util.write_file("%s/%s" % (dp, 'previous-datasource'), ds) + util.write_file("%s/%s" % (dp, 'previous-instance-id'), "%s\n" % iid) + + def get_userdata(self): + return(self.datasource.get_userdata()) + + def get_userdata_raw(self): + return(self.datasource.get_userdata_raw()) + + def get_instance_id(self): + return(self.datasource.get_instance_id()) + + def update_cache(self): + self.write_to_cache() + self.store_userdata() + + def store_userdata(self): + util.write_file(self.get_ipath('userdata_raw'), + self.datasource.get_userdata_raw(), 0600) + util.write_file(self.get_ipath('userdata'), + self.datasource.get_userdata(), 0600) + + def sem_getpath(self, name, freq): + if freq == 'once-per-instance': + return("%s/%s" % (self.get_ipath("sem"), name)) + return("%s/%s.%s" % (get_cpath("sem"), name, freq)) + + def sem_has_run(self, name, freq): + if freq == per_always: + return False + semfile = self.sem_getpath(name, freq) + if os.path.exists(semfile): + return True + return False + + def sem_acquire(self, name, freq): + from time import time + semfile = self.sem_getpath(name, freq) + + try: + os.makedirs(os.path.dirname(semfile)) + except OSError as e: + if e.errno != errno.EEXIST: + raise e + + if os.path.exists(semfile) and freq != per_always: + return False + + # race condition + try: + f = open(semfile, "w") + f.write("%s\n" % str(time())) + f.close() + except: + return(False) + return(True) + + def sem_clear(self, name, freq): + semfile = self.sem_getpath(name, freq) + try: + os.unlink(semfile) + except OSError as e: + if e.errno != errno.ENOENT: + return False + + return True + + # acquire lock on 'name' for given 'freq' + # if that does not exist, then call 'func' with given 'args' + # if 'clear_on_fail' is True and func throws an exception + # then remove the lock (so it would run again) + def sem_and_run(self, semname, freq, func, args=None, clear_on_fail=False): + if args is None: + args = [] + if self.sem_has_run(semname, freq): + log.debug("%s already ran %s", semname, freq) + return False + try: + if not self.sem_acquire(semname, freq): + raise Exception("Failed to acquire lock on %s" % semname) + + func(*args) + except: + if clear_on_fail: + self.sem_clear(semname, freq) + raise + + return True + + # get_ipath : get the instance path for a name in pathmap + # (/var/lib/cloud/instances/<instance>/name)<name>) + def get_ipath(self, name=None): + return("%s/instances/%s%s" + % (varlibdir, self.get_instance_id(), pathmap[name])) + + def consume_userdata(self, frequency=per_instance): + self.get_userdata() + data = self + + cdir = get_cpath("handlers") + idir = self.get_ipath("handlers") + + # add the path to the plugins dir to the top of our list for import + # instance dir should be read before cloud-dir + sys.path.insert(0, cdir) + sys.path.insert(0, idir) + + part_handlers = {} + # add handlers in cdir + for fname in glob.glob("%s/*.py" % cdir): + if not os.path.isfile(fname): + continue + modname = os.path.basename(fname)[0:-3] + try: + mod = __import__(modname) + handler_register(mod, part_handlers, data, frequency) + log.debug("added handler for [%s] from %s" % (mod.list_types(), + fname)) + except: + log.warn("failed to initialize handler in %s" % fname) + util.logexc(log) + + # add the internal handers if their type hasn't been already claimed + for (btype, bhand, bfreq) in self.builtin_handlers: + if btype in part_handlers: + continue + handler_register(InternalPartHandler(bhand, [btype], bfreq), + part_handlers, data, frequency) + + # walk the data + pdata = {'handlers': part_handlers, 'handlerdir': idir, + 'data': data, 'frequency': frequency} + UserDataHandler.walk_userdata(self.get_userdata(), + partwalker_callback, data=pdata) + + # give callbacks opportunity to finalize + called = [] + for (_mtype, mod) in part_handlers.iteritems(): + if mod in called: + continue + handler_call_end(mod, data, frequency) + + def handle_user_script(self, _data, ctype, filename, payload, _frequency): + if ctype == "__end__": + return + if ctype == "__begin__": + # maybe delete existing things here + return + + filename = filename.replace(os.sep, '_') + scriptsdir = get_ipath_cur('scripts') + util.write_file("%s/%s" % + (scriptsdir, filename), util.dos2unix(payload), 0700) + + def handle_upstart_job(self, _data, ctype, filename, payload, frequency): + # upstart jobs are only written on the first boot + if frequency != per_instance: + return + + if ctype == "__end__" or ctype == "__begin__": + return + if not filename.endswith(".conf"): + filename = filename + ".conf" + + util.write_file("%s/%s" % ("/etc/init", filename), + util.dos2unix(payload), 0644) + + def handle_cloud_config(self, _data, ctype, filename, payload, _frequency): + if ctype == "__begin__": + self.cloud_config_str = "" + return + if ctype == "__end__": + cloud_config = self.get_ipath("cloud_config") + util.write_file(cloud_config, self.cloud_config_str, 0600) + + ## this could merge the cloud config with the system config + ## for now, not doing this as it seems somewhat circular + ## as CloudConfig does that also, merging it with this cfg + ## + # ccfg = yaml.load(self.cloud_config_str) + # if ccfg is None: ccfg = {} + # self.cfg = util.mergedict(ccfg, self.cfg) + + return + + self.cloud_config_str += "\n#%s\n%s" % (filename, payload) + + def handle_cloud_boothook(self, _data, ctype, filename, payload, + _frequency): + if ctype == "__end__": + return + if ctype == "__begin__": + return + + filename = filename.replace(os.sep, '_') + payload = util.dos2unix(payload) + prefix = "#cloud-boothook" + start = 0 + if payload.startswith(prefix): + start = len(prefix) + 1 + + boothooks_dir = self.get_ipath("boothooks") + filepath = "%s/%s" % (boothooks_dir, filename) + util.write_file(filepath, payload[start:], 0700) + try: + env = os.environ.copy() + env['INSTANCE_ID'] = self.datasource.get_instance_id() + subprocess.check_call([filepath], env=env) + except subprocess.CalledProcessError as e: + log.error("boothooks script %s returned %i" % + (filepath, e.returncode)) + except Exception as e: + log.error("boothooks unknown exception %s when running %s" % + (e, filepath)) + + def get_public_ssh_keys(self): + return(self.datasource.get_public_ssh_keys()) + + def get_locale(self): + return(self.datasource.get_locale()) + + def get_mirror(self): + return(self.datasource.get_local_mirror()) + + def get_hostname(self, fqdn=False): + return(self.datasource.get_hostname(fqdn=fqdn)) + + def device_name_to_device(self, name): + return(self.datasource.device_name_to_device(name)) + + # I really don't know if this should be here or not, but + # I needed it in cc_update_hostname, where that code had a valid 'cloud' + # reference, but did not have a cloudinit handle + # (ie, no cloudinit.get_cpath()) + def get_cpath(self, name=None): + return(get_cpath(name)) + |