diff options
author | Aleksey Lim <alsroot@sugarlabs.org> | 2014-04-20 11:24:21 (GMT) |
---|---|---|
committer | Aleksey Lim <alsroot@sugarlabs.org> | 2014-04-20 11:24:21 (GMT) |
commit | 046073b04229021ec53833a353ffd069d0a5b561 (patch) | |
tree | 1930c720a4391daeaf3e8540b2b027f9cd1ab97f /sugar_network | |
parent | 71391e654f497234fac0a4602bba769820aa521c (diff) |
Pull node updates for checked-in resources
Diffstat (limited to 'sugar_network')
-rw-r--r-- | sugar_network/client/model.py | 12 | ||||
-rw-r--r-- | sugar_network/client/routes.py | 239 | ||||
-rw-r--r-- | sugar_network/db/blobs.py | 4 | ||||
-rw-r--r-- | sugar_network/db/directory.py | 23 | ||||
-rw-r--r-- | sugar_network/db/index.py | 19 | ||||
-rw-r--r-- | sugar_network/db/resource.py | 17 | ||||
-rw-r--r-- | sugar_network/db/routes.py | 125 | ||||
-rw-r--r-- | sugar_network/db/volume.py | 43 | ||||
-rw-r--r-- | sugar_network/model/__init__.py | 2 | ||||
-rw-r--r-- | sugar_network/model/context.py | 4 | ||||
-rw-r--r-- | sugar_network/model/routes.py | 15 | ||||
-rw-r--r-- | sugar_network/node/auth.py | 4 | ||||
-rw-r--r-- | sugar_network/node/model.py | 2 | ||||
-rw-r--r-- | sugar_network/node/routes.py | 9 | ||||
-rw-r--r-- | sugar_network/toolkit/__init__.py | 65 | ||||
-rw-r--r-- | sugar_network/toolkit/http.py | 1 | ||||
-rw-r--r-- | sugar_network/toolkit/parcel.py | 113 | ||||
-rw-r--r-- | sugar_network/toolkit/router.py | 70 |
18 files changed, 502 insertions, 265 deletions
diff --git a/sugar_network/client/model.py b/sugar_network/client/model.py index 70c8f46..fd85a4d 100644 --- a/sugar_network/client/model.py +++ b/sugar_network/client/model.py @@ -29,7 +29,7 @@ _logger = logging.getLogger('client.model') class Context(_Context): - @db.indexed_property(db.List, prefix='RP', default=[], + @db.indexed_property(db.List, prefix='P', default=[], acl=ACL.READ | ACL.LOCAL) def pins(self, value): return value + this.injector.pins(self.guid) @@ -37,7 +37,9 @@ class Context(_Context): class Volume(db.Volume): - def __init__(self, root): - db.Volume.__init__(self, root, [User, Context, Post, Report]) - for resource in ('user', 'context', 'post'): - self[resource].metadata['author'].acl |= ACL.LOCAL + def __init__(self, root, resources=None): + if resources is None: + resources = [User, Context, Post, Report] + db.Volume.__init__(self, root, resources) + for directory in self.values(): + directory.metadata['author'].acl |= ACL.LOCAL diff --git a/sugar_network/client/routes.py b/sugar_network/client/routes.py index f580789..f618df3 100644 --- a/sugar_network/client/routes.py +++ b/sugar_network/client/routes.py @@ -18,42 +18,46 @@ import logging from httplib import IncompleteRead from os.path import join -from sugar_network import db, client, node, toolkit, model -from sugar_network.client import journal -from sugar_network.toolkit.coroutine import this -from sugar_network.toolkit.router import Request, Router, File +from sugar_network import db, client, node, toolkit +from sugar_network.model import FrontRoutes +from sugar_network.client.journal import Routes as JournalRoutes +from sugar_network.toolkit.router import Request, Router, Response from sugar_network.toolkit.router import route, fallbackroute +from sugar_network.toolkit.coroutine import this from sugar_network.toolkit import netlink, zeroconf, coroutine, http, parcel from sugar_network.toolkit import ranges, lsb_release, enforce -# Flag file to recognize a directory as a synchronization directory +_SYNC_TIMEOUT = 30 _RECONNECT_TIMEOUT = 3 _RECONNECT_TIMEOUT_MAX = 60 * 15 _logger = logging.getLogger('client.routes') -class ClientRoutes(model.FrontRoutes, journal.Routes): +class ClientRoutes(FrontRoutes, JournalRoutes): def __init__(self, home_volume, creds, no_subscription=False): - model.FrontRoutes.__init__(self) - journal.Routes.__init__(self) + FrontRoutes.__init__(self) + JournalRoutes.__init__(self) this.localcast = this.broadcast self._local = _LocalRoutes(home_volume) + self._remote = None + self._remote_urls = [] self._creds = creds self._inline = coroutine.Event() self._inline_job = coroutine.Pool() - self._remote_urls = [] - self._node = None self._connect_jobs = coroutine.Pool() + self._sync_jobs = coroutine.Pool() self._no_subscription = no_subscription + self._pull_r = toolkit.Bin( + join(home_volume.root, 'var', 'pull'), [[1, None]]) self._push_r = toolkit.Bin( - join(home_volume.root, 'var', 'push'), - [[1, None]]) - self._push_job = coroutine.Pool() + join(home_volume.root, 'var', 'push'), [[1, None]]) + self._push_guids_map = toolkit.Bin( + join(home_volume.root, 'var', 'push-guids'), {}) def connect(self, api=None): if self._connect_jobs: @@ -64,11 +68,13 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): else: self._remote_urls.append(api) self._connect_jobs.spawn(self._wait_for_connectivity) + self._local.volume.populate() def close(self): self._connect_jobs.kill() self._got_offline() self._local.volume.close() + self._pull_r.commit() @fallbackroute('GET', ['hub']) def hub(self): @@ -99,7 +105,7 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): @fallbackroute('GET', ['packages']) def route_packages(self): - if self._inline.is_set(): + if self.inline(): return self.fallback() else: # Let caller know that we are in offline and @@ -113,7 +119,7 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): @route('GET', cmd='whoami', mime_type='application/json') def whoami(self): - if self._inline.is_set(): + if self.inline(): result = self.fallback() result['route'] = 'proxy' else: @@ -121,47 +127,6 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): result['guid'] = self._creds.login return result - @route('GET', [None], - arguments={'offset': int, 'limit': int, 'reply': ('guid',)}, - mime_type='application/json') - def find(self): - request = this.request - if not self._inline.is_set() or 'pins' in request: - return self._local.call(request, this.response) - - reply = request.setdefault('reply', ['guid']) - if 'pins' not in reply: - return self.fallback() - - if 'guid' not in reply: - # Otherwise there is no way to mixin `pins` - reply.append('guid') - result = self.fallback() - - directory = self._local.volume[request.resource] - for item in result['result']: - doc = directory[item['guid']] - if doc.exists: - item['pins'] += doc.repr('pins') - - return result - - @route('GET', [None, None], mime_type='application/json') - def get(self): - request = this.request - if self._local.volume[request.resource][request.guid].exists: - return self._local.call(request, this.response) - else: - return self.fallback() - - @route('GET', [None, None, None], mime_type='application/json') - def get_prop(self): - request = this.request - if self._local.volume[request.resource][request.guid].exists: - return self._local.call(request, this.response) - else: - return self.fallback() - @route('POST', ['report'], cmd='submit', mime_type='text/event-stream') def submit_report(self): props = this.request.content @@ -208,6 +173,62 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): def recycle(self): return this.injector.recycle() + @route('GET', [None], + arguments={'offset': int, 'limit': int, 'reply': ['guid']}, + mime_type='application/json') + def find(self, reply): + request = this.request + if not self.inline() or 'pins' in request: + return self._local.call(request, this.response) + if 'guid' not in reply: + # Otherwise no way to mixin `pins` or sync checkins + reply.append('guid') + if 'mtime' not in reply: + # To track updates for checked-in resources + reply.append('mtime') + result = self.fallback() + directory = self._local.volume[request.resource] + for item in result['result']: + checkin = directory[item['guid']] + if not checkin.exists: + continue + pins = item['pins'] = checkin.repr('pins') + if pins and item['mtime'] > checkin['mtime']: + pull = Request(method='GET', + path=[checkin.metadata.name, checkin.guid], cmd='diff') + self._sync_jobs.spawn(self._pull_checkin, pull, None, 'range') + return result + + @route('GET', [None, None], mime_type='application/json') + def get(self): + request = this.request + if self._local.volume[request.resource][request.guid].exists: + return self._local.call(request, this.response) + else: + return self.fallback() + + @route('GET', [None, None, None], mime_type='application/json') + def get_prop(self): + return self.get() + + @route('PUT', [None, None]) + def update(self): + if not self.inline(): + return self.fallback() + request = this.request + local = self._local.volume[request.resource][request.guid] + if not local.exists or not local.repr('pins'): + return self.fallback() + self._pull_checkin(request, None, 'pull') + + @route('PUT', [None, None, None]) + def update_prop(self): + self.update() + + @route('DELETE', [None, None]) + def delete(self): + self.update() + @fallbackroute() def fallback(self, request=None, response=None, **kwargs): if request is None: @@ -215,18 +236,18 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): if response is None: response = this.response - if not self._inline.is_set(): + if not self.inline(): return self._local.call(request, response) try: - reply = self._node.call(request, response) - if hasattr(reply, 'read'): + result = self._remote.call(request, response) + if hasattr(result, 'read'): if response.relocations: - return reply + return result else: - return _ResponseStream(reply, self._restart_online) + return _ResponseStream(result, self._restart_online) else: - return reply + return result except (http.ConnectionError, IncompleteRead): if response.relocations: raise @@ -234,28 +255,30 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): return self._local.call(request, response) def _got_online(self, url): - enforce(not self._inline.is_set()) - _logger.debug('Got online on %r', self._node) + enforce(not self.inline()) + _logger.debug('Got online on %r', self._remote) self._inline.set() self._local.volume.mute = True this.injector.api = url this.localcast({'event': 'inline', 'state': 'online'}) - self._push_job.spawn(self._push) + if not self._local.volume.empty: + self._sync_jobs.spawn_later(_SYNC_TIMEOUT, self._sync) def _got_offline(self): - if self._node is not None: - self._node.close() - if self._inline.is_set(): - _logger.debug('Got offline on %r', self._node) + if self._remote is not None: + self._remote.close() + self._remote = None + if self.inline(): + _logger.debug('Got offline on %r', self._remote) self._inline.clear() self._local.volume.mute = False this.injector.api = None this.localcast({'event': 'inline', 'state': 'offline'}) - self._push_job.kill() + self._sync_jobs.kill() def _restart_online(self): _logger.debug('Lost %r connection, try to reconnect in %s seconds', - self._node, _RECONNECT_TIMEOUT) + self._remote, _RECONNECT_TIMEOUT) self._remote_connect(_RECONNECT_TIMEOUT) def _discover_node(self): @@ -275,19 +298,19 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): def _remote_connect(self, timeout=0): def pull_events(): - for event in self._node.subscribe(): + for event in self._remote.subscribe(): if event.get('event') == 'release': this.injector.seqno = event['seqno'] this.broadcast(event) def handshake(url): _logger.debug('Connecting to %r node', url) - self._node = client.Connection(url, creds=self._creds) - status = self._node.get(cmd='status') + self._remote = client.Connection(url, creds=self._creds) + status = self._remote.get(cmd='status') seqno = status.get('seqno') if seqno and 'releases' in seqno: this.injector.seqno = seqno['releases'] - if self._inline.is_set(): + if self.inline(): _logger.info('Reconnected to %r node', url) else: self._got_online(url) @@ -322,36 +345,63 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): self._inline_job.spawn_later(timeout, connect) def _checkin_context(self, pin=None): - context = this.volume['context'][this.request.guid] - if not context.exists: + contexts = self._local.volume['context'] + local_context = contexts[this.request.guid] + if not local_context.exists: enforce(self.inline(), http.ServiceUnavailable, 'Not available in offline') - _logger.debug('Checkin %r context', context.guid) - clone = self.fallback( - method='GET', path=['context', context.guid], cmd='clone') - seqno, __ = this.volume.patch(next(parcel.decode(clone))) - if seqno: - ranges.exclude(self._push_r.value, seqno, seqno) - pins = context['pins'] + _logger.debug('Checkin %r context', local_context.guid) + pull = Request(method='GET', + path=['context', local_context.guid], cmd='diff') + self._pull_checkin(pull, None, 'range') + pins = local_context['pins'] if pin and pin not in pins: - this.volume['context'].update(context.guid, {'pins': pins + [pin]}) + contexts.update(local_context.guid, {'pins': pins + [pin]}) def _checkout_context(self, pin=None): - directory = this.volume['context'] - context = directory[this.request.guid] - if not context.exists: + contexts = self._local.volume['context'] + local_context = contexts[this.request.guid] + if not local_context.exists: return - pins = set(context.repr('pins')) + pins = set(local_context.repr('pins')) if pin: pins -= set([pin]) - if not self._inline.is_set() or pins: + if not self.inline() or pins: if pin: - directory.update(context.guid, {'pins': list(pins)}) + contexts.update(local_context.guid, {'pins': list(pins)}) else: - directory.delete(context.guid) + contexts.delete(local_context.guid) + + def _pull_checkin(self, request, response, header_key): + request.headers[header_key] = self._pull_r.value + patch = self.fallback(request, response) + __, committed = self._local.volume.patch(next(parcel.decode(patch)), + shift_seqno=False) + ranges.exclude(self._pull_r.value, committed) + + def _sync(self): + _logger.info('Start pulling updates') + + for directory in self._local.volume.values(): + if directory.empty: + continue + request = Request(method='GET', + path=[directory.metadata.name], cmd='diff') + response = Response() + while True: + request.headers['range'] = self._pull_r.value + r, guids = self.fallback(request, response) + if not r: + break + for guid in guids: + checkin = Request(method='GET', + path=[request.resource, guid], cmd='diff') + self._pull_checkin(checkin, response, 'range') + ranges.exclude(self._pull_r.value, r) + self._pull_r.commit() + this.localcast({'event': 'sync', 'state': 'pull'}) - def _push(self): - return + """ resource = None metadata = None @@ -396,6 +446,7 @@ class ClientRoutes(model.FrontRoutes, journal.Routes): request.content_type = 'application/json' request.content = props self.fallback(request) + """ class _LocalRoutes(db.Routes, Router): diff --git a/sugar_network/db/blobs.py b/sugar_network/db/blobs.py index 54fd78a..ce5bb1b 100644 --- a/sugar_network/db/blobs.py +++ b/sugar_network/db/blobs.py @@ -150,7 +150,7 @@ class Blobs(object): if exists(path): stat = os.stat(path) if seqno != int(stat.st_mtime): - _logger.debug('Found updated %r file', path) + _logger.debug('Found updated %r blob', path) seqno = self._seqno.next() meta = _read_meta(path) meta['x-seqno'] = str(seqno) @@ -169,7 +169,7 @@ class Blobs(object): elif not is_files or exists(path + _META_SUFFIX): continue else: - _logger.debug('Found new %r file', path) + _logger.debug('Found new %r blob', path) mime_type = mimetypes.guess_type(filename)[0] or \ 'application/octet-stream' if checkin_seqno is None: diff --git a/sugar_network/db/directory.py b/sugar_network/db/directory.py index ecda920..17ff27d 100644 --- a/sugar_network/db/directory.py +++ b/sugar_network/db/directory.py @@ -56,6 +56,10 @@ class Directory(object): self._open() + @property + def empty(self): + return True if self._index is None else (self._index.mtime == 0) + def wipe(self): self.close() _logger.debug('Wipe %r directory', self.metadata.name) @@ -182,21 +186,32 @@ class Directory(object): self._save_layout() self.commit() + def diff(self, r): + for start, end in r: + query = 'seqno:%s..' % start + if end: + query += str(end) + docs, __ = self.find(query=query, order_by='seqno') + for doc in docs: + yield doc + def patch(self, guid, patch, seqno=None): """Apply changes for documents.""" doc = self.resource(guid, self._storage.get(guid)) + merged = False for prop, meta in patch.items(): orig_meta = doc.meta(prop) if orig_meta and orig_meta['mtime'] >= meta['mtime']: continue - if doc.post_seqno is None: - if seqno is None: + if doc.post_seqno is None and seqno is not False: + if not seqno: seqno = self._seqno.next() doc.post_seqno = seqno doc.post(prop, **meta) + merged = True - if doc.post_seqno is not None and doc.exists: + if merged and doc.exists: # No need in after-merge event, further commit event # is enough to avoid increasing events flow self._index.store(guid, doc.posts, self._preindex) @@ -234,6 +249,8 @@ class Directory(object): if not doc.post_seqno and not doc.metadata[prop].acl & ACL.LOCAL: doc.post_seqno = self._seqno.next() doc.post(prop, changes[prop]) + if not doc.exists: + return None for prop in self.metadata.keys(): enforce(doc[prop] is not None, 'Empty %r property', prop) return doc diff --git a/sugar_network/db/index.py b/sugar_network/db/index.py index 89ea6e8..0270dd4 100644 --- a/sugar_network/db/index.py +++ b/sugar_network/db/index.py @@ -70,14 +70,13 @@ class IndexReader(object): @property def mtime(self): """UNIX seconds of the last `commit()` call.""" - return int(os.stat(self._mtime_path).st_mtime) + if exists(self._mtime_path): + return int(os.stat(self._mtime_path).st_mtime) + else: + return 0 def ensure_open(self): - if not exists(self._mtime_path): - with file(self._mtime_path, 'w'): - pass - # Outter code should understand the initial state - os.utime(self._mtime_path, (0, 0)) + pass def get_cached(self, guid): """Return cached document. @@ -337,6 +336,8 @@ class IndexWriter(IndexReader): if pre_cb is not None: properties = pre_cb(guid, properties, *args) + if properties is None: + return _logger.debug('Index %r object: %r', self.metadata.name, properties) @@ -419,7 +420,11 @@ class IndexWriter(IndexReader): self._db.flush() checkpoint = time.time() - os.utime(self._mtime_path, (checkpoint, checkpoint)) + if exists(self._mtime_path): + os.utime(self._mtime_path, (checkpoint, checkpoint)) + else: + with file(self._mtime_path, 'w'): + pass self._pending_updates = 0 _logger.debug('Commit to %r took %s seconds', diff --git a/sugar_network/db/resource.py b/sugar_network/db/resource.py index 9af5086..2c2e46b 100644 --- a/sugar_network/db/resource.py +++ b/sugar_network/db/resource.py @@ -55,7 +55,7 @@ class Resource(object): self._post_seqno = value self.post('seqno', value) - @indexed_property(Numeric, slot=1000, prefix='RS', acl=0) + @indexed_property(Numeric, slot=1000, prefix='RS', acl=0, default=0) def seqno(self, value): return value @@ -85,7 +85,8 @@ class Resource(object): def status(self, value): return value - @indexed_property(List, prefix='RP', default=[], acl=ACL.READ) + @indexed_property(List, prefix='RP', default=[], + acl=ACL.READ | ACL.LOCAL) def pins(self, value): return value @@ -93,6 +94,10 @@ class Resource(object): def exists(self): return self.record is not None and self.record.consistent + @property + def available(self): + return self.exists and self['state'] != 'deleted' + def created(self): ts = int(time.time()) self.posts['ctime'] = ts @@ -160,7 +165,7 @@ class Resource(object): if self.record is not None: return self.record.get(prop) - def diff(self, r): + def diff(self, r, out_r=None): patch = {} for name, prop in self.metadata.items(): if name == 'seqno' or prop.acl & (ACL.CALC | ACL.LOCAL): @@ -171,6 +176,8 @@ class Resource(object): seqno = meta.get('seqno') if not ranges.contains(r, seqno): continue + if out_r is not None: + ranges.include(out_r, seqno, seqno) value = meta.get('value') if isinstance(prop, Aggregated): value_ = {} @@ -178,6 +185,8 @@ class Resource(object): agg_seqno = agg.pop('seqno') if ranges.contains(r, agg_seqno): value_[key] = agg + if out_r is not None: + ranges.include(out_r, agg_seqno, agg_seqno) value = value_ patch[name] = {'mtime': meta['mtime'], 'value': value} return patch @@ -204,7 +213,7 @@ class Resource(object): if prop.on_set is not None: value = prop.on_set(self, value) seqno = None - if not prop.acl & ACL.LOCAL: + if self.post_seqno and not prop.acl & ACL.LOCAL: seqno = meta['seqno'] = self.post_seqno if seqno and isinstance(prop, Aggregated): for agg in value.values(): diff --git a/sugar_network/db/routes.py b/sugar_network/db/routes.py index c74a93e..a1bb75e 100644 --- a/sugar_network/db/routes.py +++ b/sugar_network/db/routes.py @@ -13,18 +13,22 @@ # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. +# pylint: disable-msg=W0611 + import re import logging from contextlib import contextmanager from sugar_network import toolkit from sugar_network.db.metadata import Aggregated -from sugar_network.toolkit.router import ACL, route, fallbackroute +from sugar_network.toolkit.router import ACL, File +from sugar_network.toolkit.router import route, postroute, fallbackroute from sugar_network.toolkit.coroutine import this -from sugar_network.toolkit import http, parcel, enforce +from sugar_network.toolkit import http, parcel, ranges, enforce _GUID_RE = re.compile('[a-zA-Z0-9_+-.]+$') +_GROUPED_DIFF_LIMIT = 1024 _logger = logging.getLogger('db.routes') @@ -35,6 +39,17 @@ class Routes(object): this.volume = self.volume = volume self._find_limit = find_limit + @postroute + def postroute(self, result, exception): + request = this.request + if not request.guid: + return result + pull = request.headers['pull'] + if pull is None: + return result + this.response.content_type = 'application/octet-stream' + return self._object_diff(pull) + @route('POST', [None], acl=ACL.AUTH, mime_type='application/json') def create(self): with self._post(ACL.CREATE) as doc: @@ -45,25 +60,6 @@ class Routes(object): self.volume[this.request.resource].create(doc.posts) return doc['guid'] - @route('GET', [None], - arguments={'offset': int, 'limit': int, 'reply': ('guid',)}, - mime_type='application/json') - def find(self, reply, limit): - self._preget() - request = this.request - if self._find_limit and limit > self._find_limit: - _logger.warning('The find limit is restricted to %s', - self._find_limit) - request['limit'] = self._find_limit - documents, total = self.volume[request.resource].find( - not_state='deleted', **request) - result = [self._postget(i, reply) for i in documents] - return {'total': total, 'result': result} - - @route('GET', [None, None], cmd='exists', mime_type='application/json') - def exists(self): - return self.volume[this.request.resource][this.request.guid].exists - @route('PUT', [None, None], acl=ACL.AUTH | ACL.AUTHOR) def update(self): with self._post(ACL.WRITE) as doc: @@ -88,11 +84,30 @@ class Routes(object): # to make master-slave synchronization possible directory = self.volume[this.request.resource] doc = directory[this.request.guid] - enforce(doc.exists, http.NotFound, 'Resource not found') + enforce(doc.available, http.NotFound, 'Resource not found') doc.posts['state'] = 'deleted' doc.updated() directory.update(doc.guid, doc.posts, 'delete') + @route('GET', [None], + arguments={'offset': int, 'limit': int, 'reply': ('guid',)}, + mime_type='application/json') + def find(self, reply, limit): + self._preget() + request = this.request + if self._find_limit and limit > self._find_limit: + _logger.warning('The find limit is restricted to %s', + self._find_limit) + request['limit'] = self._find_limit + documents, total = self.volume[request.resource].find( + not_state='deleted', **request) + result = [self._postget(i, reply) for i in documents] + return {'total': total, 'result': result} + + @route('GET', [None, None], cmd='exists', mime_type='application/json') + def exists(self): + return self.volume[this.request.resource][this.request.guid].available + @route('GET', [None, None], arguments={'reply': list}, mime_type='application/json') def get(self, reply): @@ -103,8 +118,7 @@ class Routes(object): reply.append(prop.name) self._preget() doc = self.volume[this.request.resource].get(this.request.guid) - enforce(doc.exists and doc['state'] != 'deleted', http.NotFound, - 'Resource not found') + enforce(doc.available, http.NotFound, 'Resource not found') return self._postget(doc, reply) @route('GET', [None, None, None], mime_type='application/json') @@ -166,15 +180,66 @@ class Routes(object): del authors[user] directory.update(request.guid, {'author': authors}) - @route('GET', [None, None], cmd='clone') - def clone(self): - clone = self.volume.clone(this.request.resource, this.request.guid) - return parcel.encode([('push', None, clone)]) + @route('GET', [None], cmd='diff', mime_type='application/json') + def grouped_diff(self, key): + if not key: + key = 'guid' + in_r = this.request.headers['range'] or [[1, None]] + out_r = [] + diff = set() + + for doc in self.volume[this.request.resource].diff(in_r): + diff.add(doc.guid) + if len(diff) > _GROUPED_DIFF_LIMIT: + break + ranges.include(out_r, doc['seqno'], doc['seqno']) + doc.diff(in_r, out_r) + + return out_r, list(diff) + + @route('GET', [None, None], cmd='diff') + def object_diff(self): + return self._object_diff(this.request.headers['range']) @fallbackroute('GET', ['blobs']) def blobs(self): return self.volume.blobs.get(this.request.guid) + def _object_diff(self, in_r): + request = this.request + doc = self.volume[request.resource][request.guid] + enforce(doc.exists, http.NotFound, 'Resource not found') + + out_r = [] + if in_r is None: + in_r = [[1, None]] + patch = doc.diff(in_r, out_r) + if not patch: + return parcel.encode([(None, None, [])], compresslevel=0) + + diff = [{'resource': request.resource}, + {'guid': request.guid, 'patch': patch}, + ] + + def add_blob(blob): + if not isinstance(blob, File): + return + seqno = int(blob.meta['x-seqno']) + ranges.include(out_r, seqno, seqno) + diff.append(blob) + + for prop, meta in patch.items(): + prop = doc.metadata[prop] + value = prop.reprcast(meta['value']) + if isinstance(prop, Aggregated): + for __, aggvalue in value: + add_blob(aggvalue) + else: + add_blob(value) + diff.append({'commit': out_r}) + + return parcel.encode([(None, None, diff)], compresslevel=0) + @contextmanager def _post(self, access): content = this.request.content @@ -197,7 +262,7 @@ class Routes(object): doc.posts[name] = prop.default else: doc = self.volume[this.request.resource][this.request.guid] - enforce(doc.exists, 'Resource not found') + enforce(doc.available, 'Resource not found') this.resource = doc def teardown(new, old): @@ -244,7 +309,7 @@ class Routes(object): def _useradd(self, authors, user, role): props = {} user_doc = self.volume['user'][user] - if user_doc.exists: + if user_doc.available: props['name'] = user_doc['name'] role |= ACL.INSYSTEM else: diff --git a/sugar_network/db/volume.py b/sugar_network/db/volume.py index 295fc02..382176c 100644 --- a/sugar_network/db/volume.py +++ b/sugar_network/db/volume.py @@ -19,7 +19,6 @@ from copy import deepcopy from os.path import exists, join, abspath from sugar_network import toolkit -from sugar_network.db.metadata import Blob from sugar_network.db.directory import Directory from sugar_network.db.index import IndexWriter from sugar_network.db.blobs import Blobs @@ -64,6 +63,13 @@ class Volume(dict): def root(self): return self._root + @property + def empty(self): + for directory in self.values(): + if not directory.empty: + return False + return True + def close(self): """Close operations with the server.""" _logger.info('Closing documents in %r', self._root) @@ -90,19 +96,13 @@ class Volume(dict): for resource, directory in self.items(): if one_way and directory.resource.one_way: continue - directory.commit() yield {'resource': resource} - for start, end in r: - query = 'seqno:%s..' % start - if end: - query += str(end) - docs, __ = directory.find(query=query, order_by='seqno') - for doc in docs: - patch = doc.diff(include) - if patch: - yield {'guid': doc.guid, 'patch': patch} - found = True - last_seqno = max(last_seqno, doc['seqno']) + for doc in directory.diff(r): + patch = doc.diff(include) + if patch: + yield {'guid': doc.guid, 'patch': patch} + found = True + last_seqno = max(last_seqno, doc['seqno']) if blobs: for blob in self.blobs.diff(include): seqno = int(blob.meta.pop('x-seqno')) @@ -124,27 +124,16 @@ class Volume(dict): ranges.exclude(r, None, last_seqno) yield {'commit': commit_r} - def clone(self, resource, guid): - doc = self[resource][guid] - patch = doc.diff([[1, None]]) - if not patch: - return - for name, prop in self[resource].metadata.items(): - if isinstance(prop, Blob) and name in patch: - yield self.blobs.get(patch[name]['value']) - yield {'resource': resource} - yield {'guid': guid, 'patch': patch} - - def patch(self, records): + def patch(self, records, shift_seqno=True): directory = None committed = [] - seqno = None + seqno = None if shift_seqno else False for record in records: if isinstance(record, File): if seqno is None: seqno = self.seqno.next() - self.blobs.patch(record, seqno) + self.blobs.patch(record, seqno or 0) continue resource = record.get('resource') if resource: diff --git a/sugar_network/model/__init__.py b/sugar_network/model/__init__.py index c6b3321..3f6aef1 100644 --- a/sugar_network/model/__init__.py +++ b/sugar_network/model/__init__.py @@ -199,7 +199,7 @@ def load_bundle(blob, context=None, initial=False, extra_deps=None): this.call(method='POST', path=['context'], content=context_meta, principal=principal) else: - enforce(doc.exists, http.NotFound, 'No context') + enforce(doc.available, http.NotFound, 'No context') enforce(context_type in doc['type'], http.BadRequest, 'Inappropriate bundle type') diff --git a/sugar_network/model/context.py b/sugar_network/model/context.py index 78df790..9153552 100644 --- a/sugar_network/model/context.py +++ b/sugar_network/model/context.py @@ -21,10 +21,6 @@ from sugar_network.toolkit import svg_to_png class Context(db.Resource): - @db.indexed_property(db.List, prefix='P', default=[]) - def pins(self, value): - return value - @db.indexed_property(db.List, prefix='T', subtype=db.Enum(model.CONTEXT_TYPES)) def type(self, value): diff --git a/sugar_network/model/routes.py b/sugar_network/model/routes.py index eda26dc..8012853 100644 --- a/sugar_network/model/routes.py +++ b/sugar_network/model/routes.py @@ -17,7 +17,7 @@ import logging from sugar_network.toolkit.router import route from sugar_network.toolkit.coroutine import this -from sugar_network.toolkit import coroutine +from sugar_network.toolkit import coroutine, http _logger = logging.getLogger('model.routes') @@ -30,9 +30,9 @@ class FrontRoutes(object): this.broadcast = self._broadcast this.localcast = self._broadcast - @route('GET', mime_type='text/html') + @route('GET') def hello(self): - return _HELLO_HTML + raise http.Redirect('http://wiki.sugarlabs.org/go/Sugar_Network/API') @route('OPTIONS') def options(self): @@ -86,7 +86,7 @@ class FrontRoutes(object): @route('GET', ['favicon.ico']) def favicon(self): - return this.volume.blobs.get('favicon.ico') + return this.volume.blobs.get('assets/favicon.ico') def _broadcast(self, event): _logger.debug('Broadcast event: %r', event) @@ -97,10 +97,3 @@ class FrontRoutes(object): coroutine.select([rfile.fileno()], [], []) finally: self._spooler.notify_all(rfile) - - -_HELLO_HTML = """\ -<h2>Welcome to Sugar Network API!</h2> -Visit the <a href="http://wiki.sugarlabs.org/go/Sugar_Network/API"> -Sugar Labs Wiki</a> to learn how it can be used. -""" diff --git a/sugar_network/node/auth.py b/sugar_network/node/auth.py index 27b334c..00054f5 100644 --- a/sugar_network/node/auth.py +++ b/sugar_network/node/auth.py @@ -17,7 +17,7 @@ import time import hashlib import logging from ConfigParser import ConfigParser -from os.path import join, dirname, exists, expanduser, abspath +from os.path import join, exists from sugar_network.toolkit.coroutine import this from sugar_network.toolkit import pylru, http, enforce @@ -89,7 +89,7 @@ class SugarAuth(object): signature = creds['signature'] nonce = int(creds['nonce']) user = this.volume['user'][login] - enforce(user.exists, Unauthorized, 'Principal does not exist') + enforce(user.available, Unauthorized, 'Principal does not exist') key = RSA.load_pub_key_bio(BIO.MemoryBuffer(str(user['pubkey']))) data = hashlib.sha1('%s:%s' % (login, nonce)).digest() enforce(key.verify(data, signature.decode('hex')), diff --git a/sugar_network/node/model.py b/sugar_network/node/model.py index b1cb401..144dab0 100644 --- a/sugar_network/node/model.py +++ b/sugar_network/node/model.py @@ -181,7 +181,7 @@ def solve(volume, top_context, command=None, lsb_id=None, lsb_release=None, if context in context_clauses: return context_clauses[context] context = volume['context'][context] - enforce(context.exists, http.NotFound, 'Context not found') + enforce(context.available, http.NotFound, 'Context not found') releases = context['releases'] clause = [] diff --git a/sugar_network/node/routes.py b/sugar_network/node/routes.py index 4457b2f..ac8a840 100644 --- a/sugar_network/node/routes.py +++ b/sugar_network/node/routes.py @@ -13,15 +13,16 @@ # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. +# pylint: disable-msg=W0611 + import logging from os.path import join from sugar_network import db from sugar_network.model import FrontRoutes, load_bundle from sugar_network.node import model -# pylint: disable-msg=W0611 -from sugar_network.toolkit.router import route, postroute, ACL, File -from sugar_network.toolkit.router import Request, fallbackroute, preroute +from sugar_network.toolkit.router import ACL, File +from sugar_network.toolkit.router import route, fallbackroute, preroute from sugar_network.toolkit.spec import parse_requires, parse_version from sugar_network.toolkit.bundle import Bundle from sugar_network.toolkit.coroutine import this @@ -83,7 +84,7 @@ class NodeRoutes(db.Routes, FrontRoutes): return {'guid': self.guid, 'seqno': { 'db': self.volume.seqno.value, - 'releases': self.volume.releases_seqno.value, + 'releases': self.volume.release_seqno.value, }, } diff --git a/sugar_network/toolkit/__init__.py b/sugar_network/toolkit/__init__.py index 675c25f..7585e29 100644 --- a/sugar_network/toolkit/__init__.py +++ b/sugar_network/toolkit/__init__.py @@ -464,6 +464,35 @@ class NamedTemporaryFile(object): return getattr(self._file, name) +class Variable(list): + + def __init__(self, default=None): + list.__init__(self, [default]) + + @property + def value(self): + return self[0] + + @value.setter + def value(self, value): + self[0] = value + + def __contains__(self, key): + return key in self[0] + + def __getitem__(self, key): + return self[0].get(key) + + def __setitem__(self, key, value): + self[0][key] = value + + def __delitem__(self, key): + del self[0][key] + + def __getattr__(self, name): + return getattr(self[0], name) + + class Bin(object): """Store variable in a file.""" @@ -471,10 +500,7 @@ class Bin(object): self._path = abspath(path) self.value = default_value - if exists(self._path): - with file(self._path) as f: - self.value = json.load(f) - else: + if not self.reset(): self.commit() @property @@ -491,6 +517,13 @@ class Bin(object): f.flush() os.fsync(f.fileno()) + def reset(self): + if not exists(self._path): + return False + with file(self._path) as f: + self.value = json.load(f) + return True + def __enter__(self): return self.value @@ -535,6 +568,30 @@ class Seqno(Bin): return self.value +class CaseInsensitiveDict(dict): + + def __contains__(self, key): + return dict.__contains__(self, key.lower()) + + def __getitem__(self, key): + return self.get(key.lower()) + + def __setitem__(self, key, value): + return self.set(key.lower(), value) + + def __delitem__(self, key): + self.remove(key.lower()) + + def get(self, key, default=None): + return dict.get(self, key, default) + + def set(self, key, value): + dict.__setitem__(self, key, value) + + def remove(self, key): + dict.__delitem__(self, key) + + class Pool(object): """Stack that keeps its iterators correct after changing content.""" diff --git a/sugar_network/toolkit/http.py b/sugar_network/toolkit/http.py index 0ebee86..0cbd535 100644 --- a/sugar_network/toolkit/http.py +++ b/sugar_network/toolkit/http.py @@ -267,6 +267,7 @@ class Connection(object): value = request.environ.get(env_key) if value is not None: headers[key] = value + headers.update(request.headers) path = request.path while True: diff --git a/sugar_network/toolkit/parcel.py b/sugar_network/toolkit/parcel.py index 9d583cd..edbbf02 100644 --- a/sugar_network/toolkit/parcel.py +++ b/sugar_network/toolkit/parcel.py @@ -46,7 +46,14 @@ _logger = logging.getLogger('parcel') def decode(stream, limit=None): _logger.debug('Decode %r stream limit=%r', stream, limit) - stream = _UnzipStream(stream, limit) + if limit is not None: + limit -= 2 + magic = stream.read(2) + enforce(len(magic) == 2, http.BadRequest, 'Malformed parcel') + if magic == '\037\213': + stream = _ZippedDecoder(stream, limit) + else: + stream = _Decoder(magic, stream, limit) header = stream.read_record() packet = _DecodeIterator(stream) @@ -63,7 +70,11 @@ def encode(packets, limit=None, header=None, compresslevel=None, _logger.debug('Encode %r packets limit=%r header=%r', packets, limit, header) - ostream = _ZipStream(compresslevel) + if compresslevel is 0: + ostream = _Encoder() + else: + ostream = _ZippedEncoder(compresslevel) + # In case of downloading blobs # (?) reuse current `this.http` this.http = http.Connection() @@ -242,16 +253,10 @@ class _DecodeIterator(object): pass -class _ZipStream(object): +class _Encoder(object): - def __init__(self, compresslevel=None): - if compresslevel is None: - compresslevel = DEFAULT_COMPRESSLEVEL - self._zipper = zlib.compressobj(compresslevel, - zlib.DEFLATED, -_ZLIB_WBITS, zlib.DEF_MEM_LEVEL, 0) + def __init__(self): self._offset = 0 - self._size = 0 - self._crc = zlib.crc32('') & 0xffffffffL def write_record(self, record, limit=None): chunk = json.dumps(record) + '\n' @@ -260,49 +265,58 @@ class _ZipStream(object): return self.write(chunk) def write(self, chunk): + chunk = self._encode(chunk) + if chunk: + self._offset += len(chunk) + return chunk + + def flush(self): + chunk = self._flush() + self._offset += len(chunk) + return chunk + + def _encode(self, chunk): + return chunk + + def _flush(self): + return '' + + +class _ZippedEncoder(_Encoder): + + def __init__(self, compresslevel=None): + _Encoder.__init__(self) + if compresslevel is None: + compresslevel = DEFAULT_COMPRESSLEVEL + self._zipper = zlib.compressobj(compresslevel, + zlib.DEFLATED, -_ZLIB_WBITS, zlib.DEF_MEM_LEVEL, 0) + self._size = 0 + self._crc = zlib.crc32('') & 0xffffffffL + + def _encode(self, chunk): self._size += len(chunk) self._crc = zlib.crc32(chunk, self._crc) & 0xffffffffL chunk = self._zipper.compress(chunk) - if self._offset == 0: chunk = '\037\213' + '\010' + chr(0) + \ struct.pack('<L', long(time.time())) + \ '\002' + '\377' + \ chunk self._offset = _ZLIB_WBITS_SIZE - if chunk: - self._offset += len(chunk) - return chunk - def flush(self): - chunk = self._zipper.flush() + \ + def _flush(self): + return self._zipper.flush() + \ struct.pack('<L', self._crc) + \ struct.pack('<L', self._size & 0xffffffffL) - self._offset += len(chunk) - return chunk -class _UnzipStream(object): +class _Decoder(object): - def __init__(self, stream, limit): + def __init__(self, prefix, stream, limit): + self._buffer = prefix self._stream = stream self._limit = limit - self._unzipper = zlib.decompressobj(-_ZLIB_WBITS) - self._crc = zlib.crc32('') & 0xffffffffL - self._size = 0 - self._buffer = '' - - if self._limit is not None: - self._limit -= 10 - magic = stream.read(2) - enforce(magic == '\037\213', http.BadRequest, - 'Not a gzipped file') - enforce(ord(stream.read(1)) == 8, http.BadRequest, - 'Unknown compression method') - enforce(ord(stream.read(1)) == 0, http.BadRequest, - 'Gzip flags should be empty') - stream.read(6) # Ignore the rest of header def read_record(self): while True: @@ -328,20 +342,41 @@ class _UnzipStream(object): if self._limit is not None: size = min(size, self._limit) chunk = self._stream.read(size) + if chunk and self._limit is not None: + self._limit -= len(chunk) + return self._decode(chunk) + + def _decode(self, chunk): + self._buffer += chunk + return bool(self._buffer) + +class _ZippedDecoder(_Decoder): + + def __init__(self, stream, limit): + _Decoder.__init__(self, '', stream, limit) + self._unzipper = zlib.decompressobj(-_ZLIB_WBITS) + self._crc = zlib.crc32('') & 0xffffffffL + self._size = 0 + + if self._limit is not None: + self._limit -= 8 + enforce(ord(stream.read(1)) == 8, http.BadRequest, + 'Unknown compression method') + enforce(ord(stream.read(1)) == 0, http.BadRequest, + 'Gzip flags should be empty') + stream.read(6) # Ignore the rest of header + + def _decode(self, chunk): if chunk: - if self._limit is not None: - self._limit -= len(chunk) self._add_to_buffer(self._unzipper.decompress(chunk)) return True - enforce(len(self._unzipper.unused_data) >= 8, http.BadRequest, 'Malformed gzipped file') crc = struct.unpack('<I', self._unzipper.unused_data[:4])[0] enforce(crc == self._crc, http.BadRequest, 'CRC check failed') size = struct.unpack('<I', self._unzipper.unused_data[4:8])[0] enforce(size == self._size, http.BadRequest, 'Incorrect length') - return self._add_to_buffer(self._unzipper.flush()) def _add_to_buffer(self, chunk): diff --git a/sugar_network/toolkit/router.py b/sugar_network/toolkit/router.py index e9e91fd..f4b23ce 100644 --- a/sugar_network/toolkit/router.py +++ b/sugar_network/toolkit/router.py @@ -140,6 +140,7 @@ class Request(dict): else: dict.__setitem__(self, key, value) self.environ = environ + self.headers = _RequestHeaders(self.environ) if method: self.environ['REQUEST_METHOD'] = method @@ -312,35 +313,15 @@ class Request(dict): (self.method, self.path, self.cmd, dict(self)) -class CaseInsensitiveDict(dict): - - def __contains__(self, key): - return dict.__contains__(self, key.lower()) - - def __getitem__(self, key): - return self.get(key.lower()) - - def __setitem__(self, key, value): - return self.set(key.lower(), value) - - def __delitem__(self, key): - self.remove(key.lower()) - - def get(self, key, default=None): - return dict.get(self, key, default) - - def set(self, key, value): - dict.__setitem__(self, key, value) - - def remove(self, key): - dict.__delitem__(self, key) - - -class Response(CaseInsensitiveDict): +class Response(toolkit.CaseInsensitiveDict): status = '200 OK' relocations = 0 + def __init__(self): + toolkit.CaseInsensitiveDict.__init__(self) + self.headers = _ResponseHeaders(self) + @property def content_length(self): return int(self.get('content-length') or '0') @@ -392,7 +373,7 @@ class File(str): pass def __new__(cls, path=None, digest=None, meta=None): - meta = CaseInsensitiveDict(meta or []) + meta = toolkit.CaseInsensitiveDict(meta or []) url = '' if meta: @@ -568,7 +549,7 @@ class Router(object): raise finally: for i in self._postroutes: - i(result, exception) + result = i(result, exception) return result @@ -915,4 +896,39 @@ class _Route(object): return '%s /%s (%s)' % (self.method, path, self.callback.__name__) +class _RequestHeaders(dict): + + def __init__(self, environ): + dict.__init__(self) + self._environ = environ + + def __contains__(self, key): + return 'HTTP_X_%s' % key.upper() in self._environ + + def __getitem__(self, key): + value = self._environ.get('HTTP_X_%s' % key.upper()) + if value is not None: + return json.loads(value) + + def __setitem__(self, key, value): + dict.__setitem__(self, 'x-%s' % key, json.dumps(value)) + + +class _ResponseHeaders(object): + + def __init__(self, headers): + self._headers = headers + + def __contains__(self, key): + return 'x-%s' % key.lower() in self._headers + + def __getitem__(self, key): + value = self._headers.get('x-%s' % key.lower()) + if value is not None: + return json.loads(value) + + def __setitem__(self, key, value): + self._headers.set('x-%s' % key.lower(), json.dumps(value)) + + File.AWAY = File(None) |