X-Git-Url: http://xvm.mit.edu/gitweb/invirt/packages/invirt-base.git/blobdiff_plain/8bfea358c4a93a7c46efbefe0694036f38f390c4..refs/tags/sipb-xen-base/8.12:/files/usr/share/python-support/sipb-xen-base/invirt/config.py diff --git a/files/usr/share/python-support/sipb-xen-base/invirt/config.py b/files/usr/share/python-support/sipb-xen-base/invirt/config.py index fad48fe..4a27cea 100644 --- a/files/usr/share/python-support/sipb-xen-base/invirt/config.py +++ b/files/usr/share/python-support/sipb-xen-base/invirt/config.py @@ -1,13 +1,10 @@ -import json, yaml +import json from invirt.common import * -from os import error, makedirs -from os.path import dirname, getmtime +from os.path import getmtime default_src_path = '/etc/invirt/master.yaml' -default_cache_path = '/var/lib/invirt/invirt.json' - -try: default_loader = yaml.CSafeLoader -except: default_loader = yaml.SafeLoader +default_cache_path = '/var/lib/invirt/cache.json' +lock_file = '/var/lib/invirt/cache.lock' def load(src_path = default_src_path, cache_path = default_cache_path, @@ -18,25 +15,42 @@ def load(src_path = default_src_path, instead from the original YAML file at src_path and regenerate the cache. I assume I have the permissions to write to the cache directory. """ + # Namespace container for various state variables, so that they can be + # updated by closures. + ns = struct() + if force_refresh: - do_refresh = True + ns.do_refresh = True else: src_mtime = getmtime(src_path) - try: cache_mtime = getmtime(cache_path) - except OSError: do_refresh = True - else: do_refresh = src_mtime > cache_mtime + try: cache_mtime = getmtime(cache_path) + except OSError: ns.do_refresh = True + else: ns.do_refresh = src_mtime > cache_mtime - if not do_refresh: - # try reading from the cache first - try: cfg = wrap(file(cache_path), lambda f: json.read(f.read())) - except: do_refresh = True + if not ns.do_refresh: + # Try reading from the cache first. This must be transactionally + # isolated from concurrent writes to prevent reading an incomplete + # (changing) version of the data (but the transaction can share the + # lock with other concurrent reads). + @with_lock_file(lock_file, False) + def read_cache(): + try: ns.cfg = with_closing(file(cache_path)) ( + lambda f: json.read(f.read())) + except: ns.do_refresh = True - if do_refresh: - # reload the source and regenerate the cache - cfg = wrap(file(src_path), lambda f: yaml.load(f, default_loader)) - try: wrap(file(cache_path, 'w'), lambda f: f.write(json.write(cfg))) - except: pass # silent failure - return cfg + if ns.do_refresh: + # Atomically reload the source and regenerate the cache. The read and + # write must be a single transaction, or a stale version may be + # written. + @with_lock_file(lock_file) + def refresh_cache(): + import yaml + ns.cfg = with_closing(file(src_path)) ( + lambda f: yaml.load(f, yaml.CSafeLoader)) + try: with_closing(file(cache_path, 'w')) ( + lambda f: f.write(json.write(ns.cfg))) + except: pass # silent failure + return ns.cfg dicts = load() structs = dicts2struct(dicts)