9 SRC_PATH = '/etc/invirt/master.yaml'
10 SRC_DIRPATH = '/etc/invirt/conf.d'
11 CACHE_PATH = '/var/lib/invirt/cache.json'
12 LOCK_PATH = '/var/lib/invirt/cache.lock'
16 Splice dict-tree d2 into d1. Return d1.
18 d2 may be None for an empty dict-tree, because yaml.load produces that.
21 >>> d = {'a': {'b': 1}, 'c': 2}
22 >>> augment(d, {'a': {'d': 3}})
23 {'a': {'b', 1, 'd': 3}, 'c': 2}
25 {'a': {'b', 1, 'd': 3}, 'c': 2}
31 if k in d1 and isinstance(d1[k], dict):
37 def run_parts_list(dirname):
39 Reimplements Debian's run-parts --list.
41 One difference from run-parts's behavior: run-parts --list /foo/
42 will give output like /foo//bar, but run_parts_list('/foo/') gives
43 /foo/bar in deference to Python conventions.
45 Matches documented behavior of run-parts in debianutils v2.28.2, dated 2007.
49 lanana_re = re.compile('^[a-z0-9]+$')
50 lsb_re = re.compile('^_?([a-z0-9_.]+-)+[a-z0-9]+$')
51 deb_cron_re = re.compile('^[a-z0-9][a-z0-9-]*$')
52 for name in os.listdir(dirname):
53 if lanana_re.match(name) or lsb_re.match(name) or deb_cron_re.match(name):
54 yield os.path.join(dirname, name)
58 for name in run_parts_list(SRC_DIRPATH):
63 for filename in list_files():
64 with open(filename) as f:
65 augment(config, yaml.load(f, yaml.CSafeLoader))
69 return max(max(os.path.getmtime(filename) for filename in list_files()),
70 os.path.getmtime(SRC_DIRPATH))
72 def load(force_refresh=False):
74 Try loading the configuration from the faster-to-load JSON cache at
75 CACHE_PATH. If it doesn't exist or is outdated, load the configuration
76 instead from the original YAML file at SRC_PATH and regenerate the cache.
77 I assume I have the permissions to write to the cache directory.
80 # Namespace container for state variables, so that they can be updated by
87 src_mtime = get_src_mtime()
89 cache_mtime = os.path.getmtime(CACHE_PATH)
93 do_refresh = src_mtime + 1 >= cache_mtime
95 # We chose not to simply say
97 # do_refresh = src_mtime >= cache_time
99 # because between the getmtime(SRC_PATH) and the time the cache is
100 # rewritten, the master configuration may have been updated, so future
101 # checks here would find a cache with a newer mtime than the master
102 # (and thus treat the cache as containing the latest version of the
103 # master). The +1 means that for at least a full second following the
104 # update to the master, this function will refresh the cache, giving us
105 # 1 second to write the cache. Note that if it takes longer than 1
106 # second to write the cache, then this situation could still arise.
108 # The getmtime calls should logically be part of the same transaction
109 # as the rest of this function (cache read + conditional cache
110 # refresh), but to wrap everything in an flock would cause the
111 # following cache read to be less streamlined.
114 # Try reading from the cache first. This must be transactionally
115 # isolated from concurrent writes to prevent reading an incomplete
116 # (changing) version of the data (but the transaction can share the
117 # lock with other concurrent reads). This isolation is accomplished
118 # using an atomic filesystem rename in the refreshing stage.
120 with open(CACHE_PATH) as f:
121 ns.cfg = json.read(f.read())
126 # Atomically reload the source and regenerate the cache. The read and
127 # write must be a single transaction, or a stale version may be
128 # written (if another read/write of a more recent configuration
129 # is interleaved). The final atomic rename is to keep this
130 # transactionally isolated from the above cache read. If we fail to
131 # acquire the lock, just try to load the master configuration.
133 with invirt.common.open_locked(LOCK_PATH):
134 ns.cfg = load_master()
136 with open(CACHE_PATH + '.tmp', 'w') as f:
137 f.write(json.write(ns.cfg))
139 pass # silent failure
141 os.rename(CACHE_PATH + '.tmp', CACHE_PATH)
143 ns.cfg = load_master()
147 structs = dicts2struct(dicts, '')