1 | from __future__ import with_statement |
---|
2 | |
---|
3 | import json |
---|
4 | from invirt.common import * |
---|
5 | import os |
---|
6 | from os import rename |
---|
7 | from os.path import getmtime |
---|
8 | from contextlib import closing |
---|
9 | import yaml |
---|
10 | |
---|
11 | try: loader = yaml.CSafeLoader |
---|
12 | except: loader = yaml.SafeLoader |
---|
13 | |
---|
14 | src_path = '/etc/invirt/master.yaml' |
---|
15 | src_dirpath = '/etc/invirt/conf.d' |
---|
16 | cache_path = '/var/lib/invirt/cache.json' |
---|
17 | lock_path = '/var/lib/invirt/cache.lock' |
---|
18 | |
---|
19 | def augment(d1, d2): |
---|
20 | """Splice dict-tree d2 into d1. Return d1. |
---|
21 | |
---|
22 | Example: |
---|
23 | >>> d = {'a': {'b': 1}, 'c': 2} |
---|
24 | >>> augment(d, {'a': {'d': 3}}) |
---|
25 | {'a': {'b', 1, 'd': 3}, 'c': 2} |
---|
26 | >>> d |
---|
27 | {'a': {'b', 1, 'd': 3}, 'c': 2} |
---|
28 | """ |
---|
29 | for k in d2: |
---|
30 | if k in d1 and isinstance(d1[k], dict): |
---|
31 | augment(d1[k], d2[k]) |
---|
32 | else: |
---|
33 | d1[k] = d2[k] |
---|
34 | return d1 |
---|
35 | |
---|
36 | def list_files(): |
---|
37 | yield src_path |
---|
38 | for name in os.listdir(src_dirpath): |
---|
39 | yield os.path.join(src_dirpath, name) |
---|
40 | |
---|
41 | def load_master(): |
---|
42 | config = dict() |
---|
43 | for filename in list_files(): |
---|
44 | with closing(file(filename)) as f: |
---|
45 | augment(config, yaml.load(f, loader)) |
---|
46 | return config |
---|
47 | |
---|
48 | def get_src_mtime(): |
---|
49 | return max(max(getmtime(filename) for filename in list_files()), |
---|
50 | getmtime(src_dirpath)) |
---|
51 | |
---|
52 | def load(force_refresh = False): |
---|
53 | """ |
---|
54 | Try loading the configuration from the faster-to-load JSON cache at |
---|
55 | cache_path. If it doesn't exist or is outdated, load the configuration |
---|
56 | instead from the original YAML file at src_path and regenerate the cache. |
---|
57 | I assume I have the permissions to write to the cache directory. |
---|
58 | """ |
---|
59 | |
---|
60 | # Namespace container for state variables, so that they can be updated by |
---|
61 | # closures. |
---|
62 | ns = struct() |
---|
63 | |
---|
64 | if force_refresh: |
---|
65 | do_refresh = True |
---|
66 | else: |
---|
67 | src_mtime = get_src_mtime() |
---|
68 | try: cache_mtime = getmtime(cache_path) |
---|
69 | except OSError: do_refresh = True |
---|
70 | else: do_refresh = src_mtime + 1 >= cache_mtime |
---|
71 | |
---|
72 | # We chose not to simply say |
---|
73 | # |
---|
74 | # do_refresh = src_mtime >= cache_time |
---|
75 | # |
---|
76 | # because between the getmtime(src_path) and the time the cache is |
---|
77 | # rewritten, the master configuration may have been updated, so future |
---|
78 | # checks here would find a cache with a newer mtime than the master |
---|
79 | # (and thus treat the cache as containing the latest version of the |
---|
80 | # master). The +1 means that for at least a full second following the |
---|
81 | # update to the master, this function will refresh the cache, giving us |
---|
82 | # 1 second to write the cache. Note that if it takes longer than 1 |
---|
83 | # second to write the cache, then this situation could still arise. |
---|
84 | # |
---|
85 | # The getmtime calls should logically be part of the same transaction |
---|
86 | # as the rest of this function (cache read + conditional cache |
---|
87 | # refresh), but to wrap everything in an flock would cause the |
---|
88 | # following cache read to be less streamlined. |
---|
89 | |
---|
90 | if not do_refresh: |
---|
91 | # Try reading from the cache first. This must be transactionally |
---|
92 | # isolated from concurrent writes to prevent reading an incomplete |
---|
93 | # (changing) version of the data (but the transaction can share the |
---|
94 | # lock with other concurrent reads). This isolation is accomplished |
---|
95 | # using an atomic filesystem rename in the refreshing stage. |
---|
96 | try: |
---|
97 | with closing(file(cache_path)) as f: |
---|
98 | ns.cfg = json.read(f.read()) |
---|
99 | except: do_refresh = True |
---|
100 | |
---|
101 | if do_refresh: |
---|
102 | # Atomically reload the source and regenerate the cache. The read and |
---|
103 | # write must be a single transaction, or a stale version may be |
---|
104 | # written (if another read/write of a more recent configuration |
---|
105 | # is interleaved). The final atomic rename is to keep this |
---|
106 | # transactionally isolated from the above cache read. If we fail to |
---|
107 | # acquire the lock, just try to load the master configuration. |
---|
108 | try: |
---|
109 | with lock_file(lock_path): |
---|
110 | ns.cfg = load_master() |
---|
111 | try: |
---|
112 | with closing(file(cache_path + '.tmp', 'w')) as f: |
---|
113 | f.write(json.write(ns.cfg)) |
---|
114 | except: pass # silent failure |
---|
115 | else: rename(cache_path + '.tmp', cache_path) |
---|
116 | except IOError: |
---|
117 | ns.cfg = load_master() |
---|
118 | return ns.cfg |
---|
119 | |
---|
120 | dicts = load() |
---|
121 | structs = dicts2struct(dicts) |
---|
122 | |
---|
123 | # vim:et:sw=4:ts=4 |
---|