diff --git a/swh/web/browse/urls.py b/swh/web/browse/urls.py index 418fbec8..a4645807 100644 --- a/swh/web/browse/urls.py +++ b/swh/web/browse/urls.py @@ -1,44 +1,82 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information from django.conf.urls import url from django.shortcuts import redirect from swh.web.common.utils import reverse from swh.web.browse.views import ( - directory, content + directory, content, origin ) def default_browse_view(request): """Default django view used as an entry point - for the swh ui web application. + for the swh browse ui web application. The url that point to it is /browse/. - Currently, it points to the root directory view associated - to the last visit of the master branch for the linux kernel + Currently, it points to the origin view for the linux kernel source tree github mirror. Args: request: input django http request """ - linux_tree_sha1 = '3347b090b27c27082414070a9cbf08a7bb75cbc6' - linux_tree_url = reverse('browse-directory', - kwargs={'sha1_git': linux_tree_sha1}) - return redirect(linux_tree_url) + linux_origin_id = '2' + linux_origin_url = reverse('browse-origin', + kwargs={'origin_id': linux_origin_id}) + return redirect(linux_origin_url) urlpatterns = [ url(r'^$', default_browse_view), url(r'^directory/(?P[0-9a-f]+)/$', directory.directory_browse, name='browse-directory'), url(r'^directory/(?P[0-9a-f]+)/(?P.+)/$', directory.directory_browse, name='browse-directory'), + url(r'^content/(?P[0-9a-f]+)/$', content.content_display, name='browse-content'), url(r'^content/(?P[0-9a-f]+)/raw/$', - content.content_raw, name='browse-content-raw') + content.content_raw, name='browse-content-raw'), + + url(r'^origin/(?P[0-9]+)/$', origin.origin_browse, + name='browse-origin'), + url(r'^origin/(?P[a-z]+)/url/(?P.+)/$', + origin.origin_browse, name='browse-origin'), + + url(r'^origin/(?P[0-9]+)/directory/$', + directory.origin_directory_browse, + name='browse-origin-directory'), + url(r'^origin/(?P[0-9]+)/directory/(?P.+)/$', + directory.origin_directory_browse, + name='browse-origin-directory'), + url(r'^origin/(?P[0-9]+)/visit/(?P[0-9]+)/directory/$', # noqa + directory.origin_directory_browse, + name='browse-origin-directory'), + url(r'^origin/(?P[0-9]+)/visit/(?P[0-9]+)' + r'/directory/(?P.+)/$', + directory.origin_directory_browse, + name='browse-origin-directory'), + url(r'^origin/(?P[0-9]+)/ts/(?P[0-9]+)/directory/$', # noqa + directory.origin_directory_browse, + name='browse-origin-directory'), + url(r'^origin/(?P[0-9]+)/ts/(?P[0-9]+)' + r'/directory/(?P.+)/$', + directory.origin_directory_browse, + name='browse-origin-directory'), + + url(r'^origin/(?P[0-9]+)/content/(?P.+)/$', + content.origin_content_display, + name='browse-origin-content'), + url(r'^origin/(?P[0-9]+)/visit/(?P[0-9]+)' + r'/content/(?P.+)/$', + content.origin_content_display, + name='browse-origin-content'), + url(r'^origin/(?P[0-9]+)/ts/(?P[0-9]+)' + r'/content/(?P.+)/$', + content.origin_content_display, + name='browse-origin-content'), ] diff --git a/swh/web/browse/utils.py b/swh/web/browse/utils.py index a12a6992..3893c292 100644 --- a/swh/web/browse/utils.py +++ b/swh/web/browse/utils.py @@ -1,53 +1,183 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information +import dateutil import magic +import math + +from django.core.cache import cache + +from swh.web.common import service +from swh.web.common.exc import NotFoundExc def gen_path_info(path): """Function to generate path data navigation for use with a breadcrumb in the swh web ui. For instance, from a path /folder1/folder2/folder3, it returns the following list:: [{'name': 'folder1', 'path': 'folder1'}, {'name': 'folder2', 'path': 'folder1/folder2'}, {'name': 'folder3', 'path': 'folder1/folder2/folder3'}] Args: path: a filesystem path Returns: A list of path data for navigation as illustrated above. """ path_info = [] if path: sub_paths = path.strip('/').split('/') path_from_root = '' for p in sub_paths: path_from_root += '/' + p path_info.append({'name': p, 'path': path_from_root.strip('/')}) return path_info _mime_magic = magic.open(magic.MAGIC_MIME_TYPE) _mime_magic.load() def get_mimetype_for_content(content): """Function that returns the mime type associated to a content buffer using the magic module under the hood. Args: content (bytes): a content buffer Returns: The mime type (e.g. text/plain) associated to the provided content. """ return _mime_magic.buffer(content) + + +def get_origin_visits(origin_id): + """Function that returns the list of visits for a swh origin. + That list is put in cache in order to speedup the navigation + in the swh web browse ui. + + Args: + origin_id (int): the id of the swh origin to fetch visits from + + Returns: + A list of dict describing the origin visits:: + + [{'date': , + 'origin': , + 'status': <'full' | 'partial'>, + 'visit': + }, + ... + ] + + Raises: + NotFoundExc if the origin is not found + """ + cache_entry_id = 'origin_%s_visits' % origin_id + cache_entry = cache.get(cache_entry_id) + + if cache_entry: + return cache_entry + + origin_visits = [] + + per_page = service.MAX_LIMIT + last_visit = None + while 1: + visits = list(service.lookup_origin_visits(origin_id, + last_visit=last_visit, + per_page=per_page)) + origin_visits += visits + if len(visits) < per_page: + break + else: + if not last_visit: + last_visit = per_page + else: + last_visit += per_page + + cache.set(cache_entry_id, origin_visits) + + return origin_visits + + +def get_origin_visit_branches(origin_id, visit_id=None, visit_ts=None): + """Function that returns the list of branches associated to + a swh origin for a given visit. + The visit can be expressed by its id or its timestamp. + If no visit parameter is provided, it returns the list of branches + found for the latest visit. + That list is put in cache in order to speedup the navigation + in the swh web browse ui. + + Args: + origin_id (int): the id of the swh origin to fetch branches from + visit_id (int): the id of the origin visit + visit_ts (Unix timestamp): the timestamp of the origin visit + + Returns: + A list of dict describing the origin branches for the given visit:: + + [{'name': , + 'revision': , + 'directory': + }, + ... + ] + + Raises: + NotFoundExc if the origin or its visit are not found + """ + + if not visit_id and visit_ts: + visits = get_origin_visits(origin_id) + for visit in visits: + ts = dateutil.parser.parse(visit['date']).timestamp() + ts = str(math.floor(ts)) + if ts == visit_ts: + return get_origin_visit_branches(origin_id, visit['visit']) + raise NotFoundExc( + 'Visit with timestamp %s for origin with id %s not found!' % + (visit_ts, origin_id)) + + cache_entry_id = 'origin_%s_visit_%s_branches' % (origin_id, visit_id) + cache_entry = cache.get(cache_entry_id) + + if cache_entry: + return cache_entry + + origin_visit_data = service.lookup_origin_visit(origin_id, visit_id) + branches = [] + revision_ids = [] + occurrences = origin_visit_data['occurrences'] + for key in sorted(occurrences.keys()): + if occurrences[key]['target_type'] == 'revision': + branches.append({'name': key, + 'revision': occurrences[key]['target']}) + revision_ids.append(occurrences[key]['target']) + + revisions = service.lookup_revision_multiple(revision_ids) + + branches_to_remove = [] + + for idx, revision in enumerate(revisions): + if revision: + branches[idx]['directory'] = revision['directory'] + else: + branches_to_remove.append(branches[idx]) + + for b in branches_to_remove: + branches.remove(b) + + cache.set(cache_entry_id, branches) + + return branches diff --git a/swh/web/browse/views/content.py b/swh/web/browse/views/content.py index 9dd03048..89b0595a 100644 --- a/swh/web/browse/views/content.py +++ b/swh/web/browse/views/content.py @@ -1,159 +1,330 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import base64 -from django.http import ( - HttpResponse, HttpResponseBadRequest -) +from django.http import HttpResponse + from django.shortcuts import render from swh.web.common import service, highlightjs +from swh.web.common.utils import reverse +from swh.web.common.exc import NotFoundExc, handle_view_exception from swh.web.browse.utils import ( - gen_path_info, get_mimetype_for_content + gen_path_info, get_mimetype_for_content, + get_origin_visit_branches, get_origin_visits ) + _browsers_supported_image_mimes = set(['image/gif', 'image/png', 'image/jpeg', 'image/bmp', 'image/webp']) def _request_content(sha1_git): query_string = 'sha1_git:' + sha1_git content_raw = service.lookup_content_raw(query_string) mime_type = service.lookup_content_filetype(query_string) if mime_type: mime_type = mime_type['mimetype'] - return content_raw, mime_type + return content_raw['data'], mime_type + + +def _prepare_content_for_display(content_data, mime_type, path): + if not mime_type: + mime_type = get_mimetype_for_content(content_data) + + language = highlightjs.get_hljs_language_from_filename(path) + + if not language: + language = highlightjs.get_hljs_language_from_mime_type(mime_type) + + if not language: + language = 'nohighlight' + elif mime_type.startswith('application/'): + mime_type = mime_type.replace('application/', 'text/') + + if mime_type.startswith('image/'): + if mime_type in _browsers_supported_image_mimes: + content_data = base64.b64encode(content_data) + else: + content_data = None + + return {'content_data': content_data, + 'language': language, + 'mime_type': mime_type} def content_display(request, sha1_git): """Django view that produces an HTML display of a swh content identified by its sha1_git value. The url that points to it is /browse/content//[?path=] In the context of a navigation coming from a directory view, the path query parameter will be used and filled with the path of the file relative to a root directory. If the content to display is textual, it will be highlighted client-side if possible using highlight.js. In order for that operation to be performed, a programming language must first be associated to the content. The following procedure is used in order to find the language: 1) First try to find a language from the content's filename (provided as query parameter when navigating from a directory view). 2) If no language has been found from the filename, try to find one from the content's mime type. The mime type is retrieved from the content metadata stored in the swh archive or is computed server-side using Python magic module. When that view is called in the context of a navigation coming from a directory view, a breadcrumb will be displayed on top of the rendered content in order to easily navigate up to the associated root directory. Args: request: input django http request sha1_git: swh sha1_git identifier of the content to display Returns: The HTML rendering of the requested content. """ try: - content_raw, mime_type = _request_content(sha1_git) + content_data, mime_type = _request_content(sha1_git) except Exception as exc: - return HttpResponseBadRequest(str(exc)) + return handle_view_exception(exc) path = request.GET.get('path', None) - content_data = content_raw['data'] - - if not mime_type: - mime_type = get_mimetype_for_content(content_data) - language = highlightjs.get_hljs_language_from_filename(path) - - if not language: - language = highlightjs.get_hljs_language_from_mime_type(mime_type) - - if not language: - language = 'nohighlight' - elif mime_type.startswith('application/'): - mime_type = mime_type.replace('application/', 'text/') + content_display_data = _prepare_content_for_display(content_data, + mime_type, path) root_dir = None filename = None path_info = None + + breadcrumbs = [] + if path: split_path = path.split('/') root_dir = split_path[0] filename = split_path[-1] - path = path.replace(root_dir + '/', '').replace(filename, '') + path = path.replace(root_dir + '/', '') + path = path.replace(filename, '') path_info = gen_path_info(path) - - if mime_type.startswith('image/'): - if mime_type in _browsers_supported_image_mimes: - content_data = base64.b64encode(content_data) - else: - content_data = None - - return render(request, 'content.html', {'content': content_data, - 'mime_type': mime_type, - 'language': language, - 'content_sha1': sha1_git, - 'root_dir': root_dir, - 'filename': filename, - 'path_info': path_info}) + breadcrumbs.append({'name': root_dir[:7], + 'url': reverse('browse-directory', + kwargs={'sha1_git': root_dir})}) + for pi in path_info: + breadcrumbs.append({'name': pi['name'], + 'url': reverse('browse-directory', + kwargs={'sha1_git': root_dir, + 'path': pi['path']})}) + breadcrumbs.append({'name': filename, + 'url': None}) + + content_raw_url = reverse('browse-content-raw', + kwargs={'sha1_git': sha1_git}, + query_params={'filename': filename}) + + return render(request, 'content.html', + {'content_sha1_git': sha1_git, + 'content': content_display_data['content_data'], + 'content_raw_url': content_raw_url, + 'mime_type': content_display_data['mime_type'], + 'language': content_display_data['language'], + 'breadcrumbs': breadcrumbs, + 'branches': None, + 'branch': None}) def content_raw(request, sha1_git): """Django view that produces a raw display of a swh content identified by its sha1_git value. The url that points to it is /browse/content//raw/[?filename=] The behaviour of that view depends on the mime type of the requested content. If the mime type is from the text family, the view will return a response whose content type is 'text/plain' that will be rendered by the browser. Otherwise, the view will return a response whose content type is 'application/octet-stream' and the browser will then offer to download the file. In the context of a navigation coming from a directory view, the filename query parameter will be used in order to provide the real name of the file when one wants to save it locally. Args: request: input django http request sha1_git: swh sha1_git identifier of the requested content Returns: The raw bytes of the content. """ # noqa try: - content_raw, mime_type = _request_content(sha1_git) + content_data, mime_type = _request_content(sha1_git) except Exception as exc: - return HttpResponseBadRequest(str(exc)) + return handle_view_exception(exc) filename = request.GET.get('filename', None) if not filename: filename = sha1_git - content_data = content_raw['data'] mime_type = get_mimetype_for_content(content_data) if mime_type.startswith('text/'): response = HttpResponse(content_data, content_type="text/plain") response['Content-disposition'] = 'filename=%s' % filename else: response = HttpResponse(content_data, content_type='application/octet-stream') response['Content-disposition'] = 'attachment; filename=%s' % filename return response + + +def origin_content_display(request, origin_id, path, visit_id=None, ts=None): + """Django view that produces an HTML display of a swh content + associated to an origin for a given visit. + + The url scheme that points to it is: + + * /browse/origin//content//[?branch=] + + * /browse/origin//visit//content//[?branch=] + + * /browse/origin//ts//content//[?branch=] + + If the content to display is textual, it will be highlighted client-side + if possible using highlight.js. In order for that operation to be + performed, a programming language must first be associated to the content. + The following procedure is used in order to find the language: + + 1) First try to find a language from the content's filename + (provided as query parameter when navigating from a directory view). + + 2) If no language has been found from the filename, + try to find one from the content's mime type. + The mime type is retrieved from the content metadata stored + in the swh archive or is computed server-side using Python + magic module. + + The view displays a breadcrumb on top of the rendered + content in order to easily navigate up to the origin root directory. + + The view also enables to easily switch between the origin branches + through a dropdown menu. + + The origin branch (default to master) from which to retrieve the content + can also be specified by using the branch query parameter. + + Args: + request: input django http request + origin_id: id of a swh origin + path: path of the content relative to the origin root directory + visit_id: optionnal visit id parameter + (the last one will be used by default) + ts: optionnal visit timestamp parameter + (the last one will be used by default) + branch: optionnal query parameter that specifies the origin branch + from which to retrieve the content + + Returns: + The HTML rendering of the requested content associated to + the provided origin and visit. + + """ # noqa + try: + + if not visit_id and not ts: + origin_visits = get_origin_visits(origin_id) + return origin_content_display(request, origin_id, path, + origin_visits[-1]['visit']) + + if not visit_id and ts: + branches = get_origin_visit_branches(origin_id, visit_ts=ts) + kwargs = {'origin_id': origin_id, + 'ts': ts} + else: + branches = get_origin_visit_branches(origin_id, visit_id) + kwargs = {'origin_id': origin_id, + 'visit_id': visit_id} + + for b in branches: + bc_kwargs = dict(kwargs) + bc_kwargs['path'] = path + b['url'] = reverse('browse-origin-content', + kwargs=bc_kwargs, + query_params={'branch': b['name']}) + + branch = request.GET.get('branch', 'master') + filtered_branches = [b for b in branches if branch in b['name']] + + if len(filtered_branches) > 0: + root_sha1_git = filtered_branches[0]['directory'] + branch = filtered_branches[0]['name'] + else: + if visit_id: + raise NotFoundExc('Branch %s associated to visit with' + ' id %s for origin with id %s' + ' not found!' % + (branch, visit_id, origin_id)) + else: + raise NotFoundExc('Branch %s associated to visit with' + ' timestamp %s for origin with id %s' + ' not found!' % + (branch, ts, origin_id)) + + content_info = service.lookup_directory_with_path(root_sha1_git, path) + sha1_git = content_info['target'] + content_data, mime_type = _request_content(sha1_git) + + except Exception as exc: + return handle_view_exception(exc) + + content_display_data = _prepare_content_for_display(content_data, + mime_type, path) + + filename = None + path_info = None + + breadcrumbs = [] + + split_path = path.split('/') + filename = split_path[-1] + path = path.replace(filename, '') + path_info = gen_path_info(path) + breadcrumbs.append({'name': root_sha1_git[:7], + 'url': reverse('browse-origin-directory', + kwargs=kwargs, + query_params={'branch': branch})}) + for pi in path_info: + bc_kwargs = dict(kwargs) + bc_kwargs['path'] = pi['path'] + breadcrumbs.append({'name': pi['name'], + 'url': reverse('browse-origin-directory', + kwargs=bc_kwargs, + query_params={'branch': branch})}) + + breadcrumbs.append({'name': filename, + 'url': None}) + + content_raw_url = reverse('browse-content-raw', + kwargs={'sha1_git': sha1_git}, + query_params={'filename': filename}) + + return render(request, 'content.html', + {'content_sha1_git': sha1_git, + 'content': content_display_data['content_data'], + 'content_raw_url': content_raw_url, + 'mime_type': content_display_data['mime_type'], + 'language': content_display_data['language'], + 'breadcrumbs': breadcrumbs, + 'branches': branches, + 'branch': branch}) diff --git a/swh/web/browse/views/directory.py b/swh/web/browse/views/directory.py index a1285a78..75202595 100644 --- a/swh/web/browse/views/directory.py +++ b/swh/web/browse/views/directory.py @@ -1,59 +1,238 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information -from django.http import HttpResponseBadRequest + from django.shortcuts import render from swh.web.common import service -from swh.web.browse.utils import gen_path_info +from swh.web.common.utils import reverse +from swh.web.common.exc import NotFoundExc, handle_view_exception +from swh.web.browse.utils import ( + gen_path_info, get_origin_visit_branches, + get_origin_visits +) + + +def _get_directory_entries(sha1_git): + entries = list(service.lookup_directory(sha1_git)) + entries = sorted(entries, key=lambda e: e['name']) + dirs = [e for e in entries if e['type'] == 'dir'] + files = [e for e in entries if e['type'] == 'file'] + return dirs, files def directory_browse(request, sha1_git, path=None): """Django view for browsing the content of a swh directory identified by its sha1_git value. The url scheme that points to that view is the following: * /browse/directory// * /browse/directory/// The content of the directory is first sorted in lexicographical order and the sub-directories are displayed before the regular files. The view enables to navigate from the provided root directory to directories reachable from it in a recursive way. A breadcrumb located in the top part of the view allows to keep track of the paths navigated so far. Args: request: input django http request sha1_git: swh sha1_git identifer of the directory to browse path: optionnal path parameter used to navigate in directories reachable from the provided root one + Returns: + The HTML rendering for the content of the provided directory. """ root_sha1_git = sha1_git try: if path: dir_info = service.lookup_directory_with_path(sha1_git, path) sha1_git = dir_info['target'] - entries = list(service.lookup_directory(sha1_git)) + dirs, files = _get_directory_entries(sha1_git) except Exception as exc: - return HttpResponseBadRequest(str(exc)) + return handle_view_exception(exc) - entries = sorted(entries, key=lambda e: e['name']) - dirs = [e for e in entries if e['type'] == 'dir'] - files = [e for e in entries if e['type'] == 'file'] + path_info = gen_path_info(path) + + breadcrumbs = [] + breadcrumbs.append({'name': root_sha1_git[:7], + 'url': reverse('browse-directory', + kwargs={'sha1_git': root_sha1_git})}) + for pi in path_info: + breadcrumbs.append({'name': pi['name'], + 'url': reverse('browse-directory', + kwargs={'sha1_git': root_sha1_git, + 'path': pi['path']})}) + + path = '' if path is None else (path + '/') + + for d in dirs: + d['url'] = reverse('browse-directory', + kwargs={'sha1_git': root_sha1_git, + 'path': path + d['name']}) + + for f in files: + f['url'] = reverse('browse-content', + kwargs={'sha1_git': f['target']}, + query_params={'path': root_sha1_git + '/' + + path + f['name']}) + + return render(request, 'directory.html', + {'dir_sha1_git': sha1_git, + 'dirs': dirs, + 'files': files, + 'breadcrumbs': breadcrumbs, + 'branches': None, + 'branch': None}) + + +def origin_directory_browse(request, origin_id, visit_id=None, ts=None, + path=None): + """Django view for browsing the content of a swh directory associated + to an origin for a given visit. + + The url scheme that points to that view is the following: + + * /browse/origin//directory/[?branch=] + + * /browse/origin//directory//[?branch=] + + * /browse/origin//visit//directory/[?branch=] + + * /browse/origin//visit//directory//[?branch=] + + * /browse/origin//ts//directory/[?branch=] + + * /browse/origin//ts//directory//[?branch=] + + For the first two urls, the displayed directory will correspond to + the one associated to the latest swh visit. + + The content of the directory is first sorted in lexicographical order + and the sub-directories are displayed before the regular files. + + The view enables to navigate from the origin root directory to + directories reachable from it in a recursive way. + A breadcrumb located in the top part of the view allows + to keep track of the paths navigated so far. + + The view also enables to easily switch between the origin branches + through a dropdown menu. + + The origin branch (default to master) from which to retrieve the directory + content can also be specified by using the branch query parameter. + + Args: + request: input django http request + origin_id: a swh origin id + visit_id: optionnal visit id parameter + (the last one will be used by default) + ts: optionnal visit timestamp parameter + (the last one will be used by default) + path: optionnal path parameter used to navigate in directories + reachable from the origin root one + + Returns: + The HTML rendering for the content of the directory associated + to the provided origin and visit. + """ # noqa + try: + + if not visit_id and not ts: + origin_visits = get_origin_visits(origin_id) + return origin_directory_browse(request, origin_id, + origin_visits[-1]['visit'], + path=path) + + if not visit_id and ts: + branches = get_origin_visit_branches(origin_id, visit_ts=ts) + url_args = {'origin_id': origin_id, + 'ts': ts} + else: + branches = get_origin_visit_branches(origin_id, visit_id) + url_args = {'origin_id': origin_id, + 'visit_id': visit_id} + + branch = request.GET.get('branch', 'master') + filtered_branches = [b for b in branches if branch in b['name']] + + if len(filtered_branches) > 0: + root_sha1_git = filtered_branches[0]['directory'] + branch = filtered_branches[0]['name'] + else: + if visit_id: + raise NotFoundExc('Branch %s associated to visit with' + ' id %s for origin with id %s' + ' not found!' % + (branch, visit_id, origin_id)) + else: + raise NotFoundExc('Branch %s associated to visit with' + ' timestamp %s for origin with id %s' + ' not found!' % + (branch, ts, origin_id)) + + sha1_git = root_sha1_git + if path: + dir_info = service.lookup_directory_with_path(root_sha1_git, path) + sha1_git = dir_info['target'] + + dirs, files = _get_directory_entries(sha1_git) + + except Exception as exc: + return handle_view_exception(exc) + + for b in branches: + branch_url_args = dict(url_args) + if path: + b['path'] = path + branch_url_args['path'] = path + b['url'] = reverse('browse-origin-directory', + kwargs=branch_url_args, + query_params={'branch': b['name']}) path_info = gen_path_info(path) + breadcrumbs = [] + breadcrumbs.append({'name': root_sha1_git[:7], + 'url': reverse('browse-origin-directory', + kwargs=url_args, + query_params={'branch': branch})}) + for pi in path_info: + bc_url_args = dict(url_args) + bc_url_args['path'] = pi['path'] + breadcrumbs.append({'name': pi['name'], + 'url': reverse('browse-origin-directory', + kwargs=bc_url_args, + query_params={'branch': branch})}) + + path = '' if path is None else (path + '/') + + for d in dirs: + bc_url_args = dict(url_args) + bc_url_args['path'] = path + d['name'] + d['url'] = reverse('browse-origin-directory', + kwargs=bc_url_args, + query_params={'branch': branch}) + + for f in files: + bc_url_args = dict(url_args) + bc_url_args['path'] = path + f['name'] + f['url'] = reverse('browse-origin-content', + kwargs=bc_url_args, + query_params={'branch': branch}) + return render(request, 'directory.html', - {'root_dir': root_sha1_git, + {'dir_sha1_git': sha1_git, 'dirs': dirs, 'files': files, - 'path': '' if path is None else (path + '/'), - 'path_info': path_info}) + 'breadcrumbs': breadcrumbs, + 'branches': branches, + 'branch': branch}) diff --git a/swh/web/browse/views/origin.py b/swh/web/browse/views/origin.py new file mode 100644 index 00000000..e029087a --- /dev/null +++ b/swh/web/browse/views/origin.py @@ -0,0 +1,72 @@ +# Copyright (C) 2017 The Software Heritage developers +# See the AUTHORS file at the top-level directory of this distribution +# License: GNU General Public License version 3, or any later version +# See top-level LICENSE file for more information + +import dateutil + +from django.shortcuts import render + +from swh.web.common import service +from swh.web.common.utils import reverse +from swh.web.common.exc import handle_view_exception +from swh.web.browse.utils import get_origin_visits + + +def origin_browse(request, origin_id=None, origin_type=None, + origin_url=None): + """Django view that produces an HTML display of a swh origin identified + by its id or its url. + + The url scheme that points to it is the following: + + * /browse/origin// + * /browse/origin//url// + + The view displays the origin metadata and contains links + for browsing its directories and contents for each swh visit. + + Args: + request: input django http request + origin_id: a swh origin id + origin_type: type of origin (git, svn, ...) + origin_url: url of the origin (e.g. https://github.com//) + + Returns: + The HMTL rendering for the metadata of the provided origin. + """ + try: + if origin_id: + origin_request_params = { + 'id': origin_id, + } + else: + origin_request_params = { + 'type': origin_type, + 'url': origin_url + } + origin_info = service.lookup_origin(origin_request_params) + origin_id = origin_info['id'] + origin_visits = get_origin_visits(origin_id) + except Exception as exc: + return handle_view_exception(exc) + + origin_info['last swh visit browse url'] = \ + reverse('browse-origin-directory', + kwargs={'origin_id': origin_id}) + + origin_visits_data = [] + for visit in origin_visits: + visit_date = dateutil.parser.parse(visit['date']) + visit['date'] = visit_date.strftime('%d %B %Y, %H:%M UTC') + visit['browse_url'] = reverse('browse-origin-directory', + kwargs={'origin_id': origin_id, + 'visit_id': visit['visit']}) + origin_visits_data.append( + {'date': visit_date.timestamp()}) + + return render(request, 'origin.html', + {'origin': origin_info, + 'origin_visits_data': origin_visits_data, + 'visits': origin_visits, + 'browse_url_base': '/browse/origin/%s/' % origin_id}) diff --git a/swh/web/common/exc.py b/swh/web/common/exc.py index c92bbe35..21fe9203 100644 --- a/swh/web/common/exc.py +++ b/swh/web/common/exc.py @@ -1,34 +1,54 @@ # Copyright (C) 2015-2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU Affero General Public License version 3, or any later version # See top-level LICENSE file for more information +import traceback + +from django.http import ( + HttpResponseBadRequest, HttpResponseNotFound +) + +from swh.web.config import get_config + class BadInputExc(ValueError): """Wrong request to the api. Example: Asking a content with the wrong identifier format. """ pass class NotFoundExc(Exception): """Good request to the api but no result were found. Example: Asking a content with the right identifier format but that content does not exist. """ pass class ForbiddenExc(Exception): """Good request to the api, forbidden result to return due to enforce policy. Example: Asking for a raw content which exists but whose mimetype is not text. """ pass + + +def handle_view_exception(exc): + content_type = None + content = str(exc) + if get_config()['debug']: + content_type = 'text/plain' + content = traceback.format_exc() + if isinstance(exc, NotFoundExc): + return HttpResponseNotFound(content, content_type=content_type) + else: + return HttpResponseBadRequest(content, content_type=content_type) diff --git a/swh/web/common/service.py b/swh/web/common/service.py index b919ba90..7dd2cf0f 100644 --- a/swh/web/common/service.py +++ b/swh/web/common/service.py @@ -1,813 +1,824 @@ # Copyright (C) 2015-2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU Affero General Public License version 3, or any later version # See top-level LICENSE file for more information import os from collections import defaultdict from swh.model import hashutil from swh.web.common import converters from swh.web.common import query from swh.web.common.exc import NotFoundExc from swh.web import config storage = config.storage() MAX_LIMIT = 50 # Top limit the users can ask for def _first_element(l): """Returns the first element in the provided list or None if it is empty or None""" return next(iter(l or []), None) def lookup_multiple_hashes(hashes): """Lookup the passed hashes in a single DB connection, using batch processing. Args: An array of {filename: X, sha1: Y}, string X, hex sha1 string Y. Returns: The same array with elements updated with elem['found'] = true if the hash is present in storage, elem['found'] = false if not. """ hashlist = [hashutil.hash_to_bytes(elem['sha1']) for elem in hashes] content_missing = storage.content_missing_per_sha1(hashlist) missing = [hashutil.hash_to_hex(x) for x in content_missing] for x in hashes: x.update({'found': True}) for h in hashes: if h['sha1'] in missing: h['found'] = False return hashes def lookup_expression(expression, last_sha1, per_page): """Lookup expression in raw content. Args: expression (str): An expression to lookup through raw indexed content last_sha1 (str): Last sha1 seen per_page (int): Number of results per page Returns: List of ctags whose content match the expression """ limit = min(per_page, MAX_LIMIT) ctags = storage.content_ctags_search(expression, last_sha1=last_sha1, limit=limit) for ctag in ctags: ctag = converters.from_swh(ctag, hashess={'id'}) ctag['sha1'] = ctag['id'] ctag.pop('id') yield ctag def lookup_hash(q): """Checks if the storage contains a given content checksum Args: query string of the form Returns: Dict with key found containing the hash info if the hash is present, None if not. """ algo, hash = query.parse_hash(q) found = storage.content_find({algo: hash}) return {'found': found, 'algo': algo} def search_hash(q): """Checks if the storage contains a given content checksum Args: query string of the form Returns: Dict with key found to True or False, according to whether the checksum is present or not """ algo, hash = query.parse_hash(q) found = storage.content_find({algo: hash}) return {'found': found is not None} def lookup_content_provenance(q): """Return provenance information from a specified content. Args: q: query string of the form Yields: provenance information (dict) list if the content is found. """ algo, hash = query.parse_hash(q) provenances = storage.content_find_provenance({algo: hash}) if not provenances: return None return (converters.from_provenance(p) for p in provenances) def _lookup_content_sha1(q): """Given a possible input, query for the content's sha1. Args: q: query string of the form Returns: binary sha1 if found or None """ algo, hash = query.parse_hash(q) if algo != 'sha1': hashes = storage.content_find({algo: hash}) if not hashes: return None return hashes['sha1'] return hash def lookup_content_ctags(q): """Return ctags information from a specified content. Args: q: query string of the form Yields: ctags information (dict) list if the content is found. """ sha1 = _lookup_content_sha1(q) if not sha1: return None ctags = list(storage.content_ctags_get([sha1])) if not ctags: return None for ctag in ctags: yield converters.from_swh(ctag, hashess={'id'}) def lookup_content_filetype(q): """Return filetype information from a specified content. Args: q: query string of the form Yields: filetype information (dict) list if the content is found. """ sha1 = _lookup_content_sha1(q) if not sha1: return None filetype = _first_element(list(storage.content_mimetype_get([sha1]))) if not filetype: return None return converters.from_filetype(filetype) def lookup_content_language(q): """Return language information from a specified content. Args: q: query string of the form Yields: language information (dict) list if the content is found. """ sha1 = _lookup_content_sha1(q) if not sha1: return None lang = _first_element(list(storage.content_language_get([sha1]))) if not lang: return None return converters.from_swh(lang, hashess={'id'}) def lookup_content_license(q): """Return license information from a specified content. Args: q: query string of the form Yields: license information (dict) list if the content is found. """ sha1 = _lookup_content_sha1(q) if not sha1: return None lang = _first_element(storage.content_fossology_license_get([sha1])) if not lang: return None return converters.from_swh(lang, hashess={'id'}) def lookup_origin(origin): """Return information about the origin matching dict origin. Args: origin: origin's dict with keys either 'id' or ('type' AND 'url') Returns: origin information as dict. """ - return converters.from_origin(storage.origin_get(origin)) + origin_info = storage.origin_get(origin) + if not origin_info: + if 'id' in origin and origin['id']: + msg = 'Origin with id %s not found!' % origin['id'] + else: + msg = 'Origin of type %s and url %s not found!' % \ + (origin['type'], origin['url']) + raise NotFoundExc(msg) + return converters.from_origin(origin_info) def lookup_person(person_id): """Return information about the person with id person_id. Args: person_id as string Returns: person information as dict. """ person = _first_element(storage.person_get([person_id])) return converters.from_person(person) def lookup_directory(sha1_git): """Return information about the directory with id sha1_git. Args: sha1_git as string Returns: directory information as dict. """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( sha1_git, ['sha1'], # HACK: sha1_git really 'Only sha1_git is supported.') dir = _first_element(storage.directory_get([sha1_git_bin])) if not dir: - return None + raise NotFoundExc('Directory with sha1_git %s not found' % sha1_git) directory_entries = storage.directory_ls(sha1_git_bin) or [] return map(converters.from_directory_entry, directory_entries) def lookup_directory_with_path(directory_sha1_git, path_string): """Return directory information for entry with path path_string w.r.t. root directory pointed by directory_sha1_git Args: - directory_sha1_git: sha1_git corresponding to the directory to which we append paths to (hopefully) find the entry - the relative path to the entry starting from the directory pointed by directory_sha1_git Raises: NotFoundExc if the directory entry is not found """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( directory_sha1_git, ['sha1'], 'Only sha1_git is supported.') paths = path_string.strip(os.path.sep).split(os.path.sep) queried_dir = storage.directory_entry_get_by_path( sha1_git_bin, list(map(lambda p: p.encode('utf-8'), paths))) if not queried_dir: raise NotFoundExc(('Directory entry with path %s from %s not found') % (path_string, directory_sha1_git)) return converters.from_directory_entry(queried_dir) def lookup_release(release_sha1_git): """Return information about the release with sha1 release_sha1_git. Args: release_sha1_git: The release's sha1 as hexadecimal Returns: Release information as dict. Raises: ValueError if the identifier provided is not of sha1 nature. """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( release_sha1_git, ['sha1'], 'Only sha1_git is supported.') res = _first_element(storage.release_get([sha1_git_bin])) return converters.from_release(res) def lookup_revision(rev_sha1_git): """Return information about the revision with sha1 revision_sha1_git. Args: revision_sha1_git: The revision's sha1 as hexadecimal Returns: Revision information as dict. Raises: ValueError if the identifier provided is not of sha1 nature. """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( rev_sha1_git, ['sha1'], 'Only sha1_git is supported.') revision = _first_element(storage.revision_get([sha1_git_bin])) return converters.from_revision(revision) def lookup_revision_multiple(sha1_git_list): """Return information about the revision with sha1 revision_sha1_git. Args: revision_sha1_git: The revision's sha1 as hexadecimal Returns: Revision information as dict. Raises: ValueError if the identifier provided is not of sha1 nature. """ def to_sha1_bin(sha1_hex): _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( sha1_hex, ['sha1'], 'Only sha1_git is supported.') return sha1_git_bin sha1_bin_list = (to_sha1_bin(x) for x in sha1_git_list) revisions = storage.revision_get(sha1_bin_list) or [] return (converters.from_revision(x) for x in revisions) def lookup_revision_message(rev_sha1_git): """Return the raw message of the revision with sha1 revision_sha1_git. Args: revision_sha1_git: The revision's sha1 as hexadecimal Returns: Decoded revision message as dict {'message': } Raises: ValueError if the identifier provided is not of sha1 nature. NotFoundExc if the revision is not found, or if it has no message """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( rev_sha1_git, ['sha1'], 'Only sha1_git is supported.') revision = _first_element(storage.revision_get([sha1_git_bin])) if not revision: raise NotFoundExc('Revision with sha1_git %s not found.' % rev_sha1_git) if 'message' not in revision: raise NotFoundExc('No message for revision with sha1_git %s.' % rev_sha1_git) res = {'message': revision['message']} return res def lookup_revision_by(origin_id, branch_name="refs/heads/master", timestamp=None): """Lookup revisions by origin_id, branch_name and timestamp. If: - branch_name is not provided, lookup using 'refs/heads/master' as default. - ts is not provided, use the most recent Args: - origin_id: origin of the revision. - branch_name: revision's branch. - timestamp: revision's time frame. Yields: The revisions matching the criterions. """ res = _first_element(storage.revision_get_by(origin_id, branch_name, timestamp=timestamp, limit=1)) return converters.from_revision(res) def lookup_revision_log(rev_sha1_git, limit): """Return information about the revision with sha1 revision_sha1_git. Args: revision_sha1_git: The revision's sha1 as hexadecimal limit: the maximum number of revisions returned Returns: Revision information as dict. Raises: ValueError if the identifier provided is not of sha1 nature. """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( rev_sha1_git, ['sha1'], 'Only sha1_git is supported.') revision_entries = storage.revision_log([sha1_git_bin], limit) return map(converters.from_revision, revision_entries) def lookup_revision_log_by(origin_id, branch_name, timestamp, limit): """Return information about the revision with sha1 revision_sha1_git. Args: origin_id: origin of the revision branch_name: revision's branch timestamp: revision's time frame limit: the maximum number of revisions returned Returns: Revision information as dict. Raises: NotFoundExc if no revision corresponds to the criterion NotFoundExc if the corresponding revision has no log """ revision_entries = storage.revision_log_by(origin_id, branch_name, timestamp, limit=limit) if not revision_entries: return None return map(converters.from_revision, revision_entries) def lookup_revision_with_context_by(origin_id, branch_name, ts, sha1_git, limit=100): """Return information about revision sha1_git, limited to the sub-graph of all transitive parents of sha1_git_root. sha1_git_root being resolved through the lookup of a revision by origin_id, branch_name and ts. In other words, sha1_git is an ancestor of sha1_git_root. Args: - origin_id: origin of the revision. - branch_name: revision's branch. - timestamp: revision's time frame. - sha1_git: one of sha1_git_root's ancestors. - limit: limit the lookup to 100 revisions back. Returns: Pair of (root_revision, revision). Information on sha1_git if it is an ancestor of sha1_git_root including children leading to sha1_git_root Raises: - BadInputExc in case of unknown algo_hash or bad hash. - NotFoundExc if either revision is not found or if sha1_git is not an ancestor of sha1_git_root. """ rev_root = _first_element(storage.revision_get_by(origin_id, branch_name, timestamp=ts, limit=1)) if not rev_root: raise NotFoundExc('Revision with (origin_id: %s, branch_name: %s' ', ts: %s) not found.' % (origin_id, branch_name, ts)) return (converters.from_revision(rev_root), lookup_revision_with_context(rev_root, sha1_git, limit)) def lookup_revision_with_context(sha1_git_root, sha1_git, limit=100): """Return information about revision sha1_git, limited to the sub-graph of all transitive parents of sha1_git_root. In other words, sha1_git is an ancestor of sha1_git_root. Args: sha1_git_root: latest revision. The type is either a sha1 (as an hex string) or a non converted dict. sha1_git: one of sha1_git_root's ancestors limit: limit the lookup to 100 revisions back Returns: Information on sha1_git if it is an ancestor of sha1_git_root including children leading to sha1_git_root Raises: BadInputExc in case of unknown algo_hash or bad hash NotFoundExc if either revision is not found or if sha1_git is not an ancestor of sha1_git_root """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( sha1_git, ['sha1'], 'Only sha1_git is supported.') revision = _first_element(storage.revision_get([sha1_git_bin])) if not revision: raise NotFoundExc('Revision %s not found' % sha1_git) if isinstance(sha1_git_root, str): _, sha1_git_root_bin = query.parse_hash_with_algorithms_or_throws( sha1_git_root, ['sha1'], 'Only sha1_git is supported.') revision_root = _first_element(storage.revision_get([sha1_git_root_bin])) # noqa if not revision_root: raise NotFoundExc('Revision root %s not found' % sha1_git_root) else: sha1_git_root_bin = sha1_git_root['id'] revision_log = storage.revision_log([sha1_git_root_bin], limit) parents = {} children = defaultdict(list) for rev in revision_log: rev_id = rev['id'] parents[rev_id] = [] for parent_id in rev['parents']: parents[rev_id].append(parent_id) children[parent_id].append(rev_id) if revision['id'] not in parents: raise NotFoundExc('Revision %s is not an ancestor of %s' % (sha1_git, sha1_git_root)) revision['children'] = children[revision['id']] return converters.from_revision(revision) def lookup_directory_with_revision(sha1_git, dir_path=None, with_data=False): """Return information on directory pointed by revision with sha1_git. If dir_path is not provided, display top level directory. Otherwise, display the directory pointed by dir_path (if it exists). Args: sha1_git: revision's hash. dir_path: optional directory pointed to by that revision. with_data: boolean that indicates to retrieve the raw data if the path resolves to a content. Default to False (for the api) Returns: Information on the directory pointed to by that revision. Raises: BadInputExc in case of unknown algo_hash or bad hash. NotFoundExc either if the revision is not found or the path referenced does not exist. NotImplementedError in case of dir_path exists but do not reference a type 'dir' or 'file'. """ _, sha1_git_bin = query.parse_hash_with_algorithms_or_throws( sha1_git, ['sha1'], 'Only sha1_git is supported.') revision = _first_element(storage.revision_get([sha1_git_bin])) if not revision: raise NotFoundExc('Revision %s not found' % sha1_git) dir_sha1_git_bin = revision['directory'] if dir_path: paths = dir_path.strip(os.path.sep).split(os.path.sep) entity = storage.directory_entry_get_by_path( dir_sha1_git_bin, list(map(lambda p: p.encode('utf-8'), paths))) if not entity: raise NotFoundExc( "Directory or File '%s' pointed to by revision %s not found" % (dir_path, sha1_git)) else: entity = {'type': 'dir', 'target': dir_sha1_git_bin} if entity['type'] == 'dir': directory_entries = storage.directory_ls(entity['target']) or [] return {'type': 'dir', 'path': '.' if not dir_path else dir_path, 'revision': sha1_git, 'content': map(converters.from_directory_entry, directory_entries)} elif entity['type'] == 'file': # content content = storage.content_find({'sha1_git': entity['target']}) if with_data: c = _first_element(storage.content_get([content['sha1']])) content['data'] = c['data'] return {'type': 'file', 'path': '.' if not dir_path else dir_path, 'revision': sha1_git, 'content': converters.from_content(content)} else: raise NotImplementedError('Entity of type %s not implemented.' % entity['type']) def lookup_content(q): """Lookup the content designed by q. Args: q: The release's sha1 as hexadecimal """ algo, hash = query.parse_hash(q) c = storage.content_find({algo: hash}) return converters.from_content(c) def lookup_content_raw(q): """Lookup the content defined by q. Args: q: query string of the form Returns: dict with 'sha1' and 'data' keys. data representing its raw data decoded. """ algo, hash = query.parse_hash(q) c = storage.content_find({algo: hash}) if not c: return None content = _first_element(storage.content_get([c['sha1']])) return converters.from_content(content) def stat_counters(): """Return the stat counters for Software Heritage Returns: A dict mapping textual labels to integer values. """ return storage.stat_counters() def _lookup_origin_visits(origin_id, last_visit=None, limit=10): """Yields the origin origin_ids' visits. Args: origin_id (int): origin to list visits for last_visit (int): last visit to lookup from limit (int): Number of elements max to display Yields: Dictionaries of origin_visit for that origin """ limit = min(limit, MAX_LIMIT) yield from storage.origin_visit_get( origin_id, last_visit=last_visit, limit=limit) def lookup_origin_visits(origin_id, last_visit=None, per_page=10): """Yields the origin origin_ids' visits. Args: origin_id: origin to list visits for Yields: Dictionaries of origin_visit for that origin """ visits = _lookup_origin_visits(origin_id, last_visit=last_visit, limit=per_page) for visit in visits: yield converters.from_origin_visit(visit) def lookup_origin_visit(origin_id, visit_id): """Return information about visit visit_id with origin origin_id. Args: origin_id: origin concerned by the visit visit_id: the visit identifier to lookup Yields: The dict origin_visit concerned """ visit = storage.origin_visit_get_by(origin_id, visit_id) + if not visit: + raise NotFoundExc('Origin with id %s or its visit ' + 'with id %s not found!' % (origin_id, visit_id)) return converters.from_origin_visit(visit) def lookup_entity_by_uuid(uuid): """Return the entity's hierarchy from its uuid. Args: uuid: entity's identifier. Returns: List of hierarchy entities from the entity with uuid. """ uuid = query.parse_uuid4(uuid) for entity in storage.entity_get(uuid): entity = converters.from_swh(entity, convert={'last_seen', 'uuid'}, convert_fn=lambda x: str(x)) yield entity def lookup_revision_through(revision, limit=100): """Retrieve a revision from the criterion stored in revision dictionary. Args: revision: Dictionary of criterion to lookup the revision with. Here are the supported combination of possible values: - origin_id, branch_name, ts, sha1_git - origin_id, branch_name, ts - sha1_git_root, sha1_git - sha1_git Returns: None if the revision is not found or the actual revision. """ if 'origin_id' in revision and \ 'branch_name' in revision and \ 'ts' in revision and \ 'sha1_git' in revision: return lookup_revision_with_context_by(revision['origin_id'], revision['branch_name'], revision['ts'], revision['sha1_git'], limit) if 'origin_id' in revision and \ 'branch_name' in revision and \ 'ts' in revision: return lookup_revision_by(revision['origin_id'], revision['branch_name'], revision['ts']) if 'sha1_git_root' in revision and \ 'sha1_git' in revision: return lookup_revision_with_context(revision['sha1_git_root'], revision['sha1_git'], limit) if 'sha1_git' in revision: return lookup_revision(revision['sha1_git']) # this should not happen raise NotImplementedError('Should not happen!') def lookup_directory_through_revision(revision, path=None, limit=100, with_data=False): """Retrieve the directory information from the revision. Args: revision: dictionary of criterion representing a revision to lookup path: directory's path to lookup. limit: optional query parameter to limit the revisions log (default to 100). For now, note that this limit could impede the transitivity conclusion about sha1_git not being an ancestor of. with_data: indicate to retrieve the content's raw data if path resolves to a content. Returns: The directory pointing to by the revision criterions at path. """ rev = lookup_revision_through(revision, limit) if not rev: raise NotFoundExc('Revision with criterion %s not found!' % revision) return (rev['id'], lookup_directory_with_revision(rev['id'], path, with_data)) diff --git a/swh/web/common/swh_templatetags.py b/swh/web/common/swh_templatetags.py index 09ed6c09..970a9b96 100644 --- a/swh/web/common/swh_templatetags.py +++ b/swh/web/common/swh_templatetags.py @@ -1,89 +1,89 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import re from docutils.core import publish_parts from docutils.writers.html4css1 import Writer, HTMLTranslator from inspect import cleandoc from django import template from django.utils.safestring import mark_safe from pygments import highlight from pygments.lexers import JsonLexer from pygments.formatters import HtmlFormatter register = template.Library() class NoHeaderHTMLTranslator(HTMLTranslator): """ Docutils translator subclass to customize the generation of HTML from reST-formatted docstrings """ def __init__(self, document): super().__init__(document) self.body_prefix = [] self.body_suffix = [] def visit_bullet_list(self, node): self.context.append((self.compact_simple, self.compact_p)) self.compact_p = None self.compact_simple = self.is_compactable(node) self.body.append(self.starttag(node, 'ul', CLASS='docstring')) DOCSTRING_WRITER = Writer() DOCSTRING_WRITER.translator_class = NoHeaderHTMLTranslator @register.filter def safe_docstring_display(docstring): """ Utility function to htmlize reST-formatted documentation in browsable api. """ docstring = cleandoc(docstring) return publish_parts(docstring, writer=DOCSTRING_WRITER)['html_body'] @register.filter def urlize_api_links(text): """Utility function for decorating api links in browsable api. Args: text: whose content matching links should be transformed into contextual API or Browse html links. Returns The text transformed if any link is found. The text as is otherwise. """ - return re.sub(r'(/api/[^"<]*/|/browse/.*/)', + return re.sub(r'(/api/[^"<]*/|/browse/.*/|http.*$)', r'\1', text) @register.filter def urlize_header_links(text): """Utility function for decorating headers links in browsable api. Args text: Text whose content contains Link header value Returns: The text transformed with html link if any link is found. The text as is otherwise. """ return re.sub(r'<(/api/.*|/browse/.*)>', r'<\1>', text) @register.filter def highlight_json(text): return mark_safe(highlight(text, JsonLexer(), HtmlFormatter())) diff --git a/swh/web/settings/common.py b/swh/web/settings/common.py index 024185d8..5bc53e8b 100644 --- a/swh/web/settings/common.py +++ b/swh/web/settings/common.py @@ -1,192 +1,193 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information """ Django settings for swhweb project. Generated by 'django-admin startproject' using Django 1.11.3. For more information on this file, see https://docs.djangoproject.com/en/1.11/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.11/ref/settings/ """ import os from swh.web.config import get_config swh_web_config = get_config() # Build paths inside the project like this: os.path.join(BASE_DIR, ...) PROJECT_DIR = os.path.dirname(os.path.abspath(__file__)) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.11/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = swh_web_config['secret_key'] # SECURITY WARNING: don't run with debug turned on in production! DEBUG = swh_web_config['debug'] +DEBUG_PROPAGATE_EXCEPTIONS = swh_web_config['debug'] ALLOWED_HOSTS = ['127.0.0.1', 'localhost', 'testserver'] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'swh.web.api', 'swh.web.browse' ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'swh.web.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [os.path.join(PROJECT_DIR, "../templates")], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], 'libraries': { 'swh_templatetags': 'swh.web.common.swh_templatetags', }, }, }, ] TEMPLATE_DIRS = TEMPLATES[0]['DIRS'] WSGI_APPLICATION = 'swh.web.wsgi.application' # Database # https://docs.djangoproject.com/en/1.11/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(PROJECT_DIR, 'db.sqlite3'), } } # Password validation # https://docs.djangoproject.com/en/1.11/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', # noqa }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', # noqa }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', # noqa }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', # noqa }, ] # Internationalization # https://docs.djangoproject.com/en/1.11/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.11/howto/static-files/ STATIC_URL = '/static/' STATICFILES_DIRS = [ os.path.join(PROJECT_DIR, "../static") ] INTERNAL_IPS = ['127.0.0.1'] throttle_rates = {} throttling = swh_web_config['throttling'] for limiter_scope, limiter_conf in throttling['scopes'].items(): throttle_rates[limiter_scope] = limiter_conf['limiter_rate'] REST_FRAMEWORK = { 'DEFAULT_RENDERER_CLASSES': ( 'rest_framework.renderers.JSONRenderer', 'swh.web.api.renderers.YAMLRenderer', 'rest_framework.renderers.TemplateHTMLRenderer' ), 'DEFAULT_THROTTLE_CLASSES': ( 'swh.web.common.throttling.SwhWebRateThrottle', ), 'DEFAULT_THROTTLE_RATES': throttle_rates } LOGGING = { 'version': 1, 'disable_existing_loggers': False, 'filters': { 'require_debug_false': { '()': 'django.utils.log.RequireDebugFalse', }, 'require_debug_true': { '()': 'django.utils.log.RequireDebugTrue', }, }, 'handlers': { 'console': { 'level': 'DEBUG', 'filters': ['require_debug_true'], 'class': 'logging.StreamHandler', }, 'file': { 'level': 'DEBUG', 'filters': ['require_debug_false'], 'class': 'logging.FileHandler', 'filename': os.path.join(swh_web_config['log_dir'], 'swh-web.log'), }, }, 'loggers': { 'django': { 'handlers': ['console', 'file'], 'level': 'DEBUG', 'propagate': True, - }, + } }, } SILENCED_SYSTEM_CHECKS = ['1_7.W001', '1_8.W001'] diff --git a/swh/web/static/css/style.css b/swh/web/static/css/style.css index 33f24806..f1218512 100644 --- a/swh/web/static/css/style.css +++ b/swh/web/static/css/style.css @@ -1,328 +1,337 @@ /* version: 0.1 date: 21/09/15 author: swh email: swh website: softwareheritage.org version history: /style.css */ @import url(https://fonts.googleapis.com/css?family=Alegreya:400,400italic,700,700italic); @import url(https://fonts.googleapis.com/css?family=Alegreya+Sans:400,400italic,500,500italic,700,700italic,100,300,100italic,300italic); html { height: 100%; } body { font-family: 'Alegreya Sans', sans-serif; font-size: 1.7rem; line-height: 1.5; color: rgba(0, 0, 0, 0.55); padding-top: 80px; /* avoid fixed bootstrap navbar covers content */ padding-bottom: 120px; min-height: 100%; margin: 0; position: relative; } .heading { font-family: 'Alegreya', serif; } .shell, .text { font-size: 0.7em; } .logo img { max-height: 40px; } .logo .navbar-brand { padding: 5px; } .logo .sitename { padding: 15px 5px; } .jumbotron { padding: 0; background-color: rgba(0, 0, 0, 0); position: fixed; top: 0; width: 100%; } #swh-navbar-collapse { border-top-style: none; border-left-style: none; border-right-style: none; border-bottom: 5px solid; border-image: linear-gradient(to right, rgb(226, 0, 38) 0%, rgb(254, 205, 27) 100%) 1 1 1 1; width: 100%; padding: 5px; } .nav-horizontal { float: right; } h3[id], h4[id], a[id] { /* avoid in-page links covered by navbar */ padding-top: 80px; margin-top: -70px; } h1, h2, h3, h4 { margin: 0; color: #e20026; padding-bottom: 10px; } h1 { font-size: 1.8em; } h2 { font-size: 1.2em; } h3 { font-size: 1.1em; } a { color: rgba(0, 0, 0, 0.75); border-bottom-style: dotted; border-bottom-width: 1px; border-bottom-color: rgb(91, 94, 111); } a:hover { color: black; } ul.dropdown-menu a, .navbar-header a, ul.navbar-nav a { /* No decoration on links in dropdown menu */ border-bottom-style: none; color: #323232; font-weight: 700; } .navbar-header a:hover, ul.navbar-nav a:hover { color: #8f8f8f; } .sitename .first-word, .sitename .second-word { color: rgba(0, 0, 0, 0.75); font-weight: normal; font-size: 1.8rem; } .sitename .first-word { font-family: 'Alegreya Sans', sans-serif; } .sitename .second-word { font-family: 'Alegreya', serif; } ul.dropdown-menu > li, ul.dropdown-menu > li > ul > li { /* No decoration on bullet points in dropdown menu */ list-style-type: none; } .page { margin: 2em auto; width: 35em; border: 5px solid #ccc; padding: 0.8em; background: white; } .entries { list-style: none; margin: 0; padding: 0; } .entries li { margin: 0.8em 1.2em; } .entries li h2 { margin-left: -1em; } .add-entry { font-size: 0.9em; border-bottom: 1px solid #ccc; } .add-entry dl { font-weight: bold; } .metanav { text-align: right; font-size: 0.8em; padding: 0.3em; margin-bottom: 1em; background: #fafafa; } .flash { background: #cee5F5; padding: 0.5em; border: 1px solid #aacbe2; } .error { background: #f0d6d6; padding: 0.5em; } .file-found { color: #23BA49; } .file-notfound { color: #FF4747; } /* Bootstrap custom styling to correctly render multiple * form-controls in an input-group: * github.com/twbs/bootstrap/issues/12732 */ .input-group-field { display: table-cell; vertical-align: middle; border-radius:4px; min-width:1%; white-space: nowrap; } .input-group-field .form-control { border-radius: inherit !important; } .input-group-field:not(:first-child):not(:last-child) { border-radius:0; } .input-group-field:not(:first-child):not(:last-child) .form-control { border-left-width: 0; border-right-width: 0; } .input-group-field:last-child { border-top-left-radius:0; border-bottom-left-radius:0; } .input-group > span:not(:last-child) > button { border-radius: 0; } .multi-input-group > .input-group-btn { vertical-align: bottom; padding: 0; } .dataTables_filter { margin-top: 15px; } .dataTables_filter input { width: 70%; float: right; } tr.api-doc-route-upcoming > td, tr.api-doc-route-upcoming > td > a { font-size: 90%; } tr.api-doc-route-deprecated > td, tr.api-doc-route-deprecated > td > a { color: red; } #back-to-top { display: initial; position: fixed; bottom: 30px; right: 30px; z-index: 10; } #back-to-top a img { display: block; width: 32px; height: 32px; background-size: 32px 32px; text-indent: -999px; overflow: hidden; } .table > thead > tr > th { border-bottom: 1px solid #e20026; } .table > tbody > tr > td { border-style: none; } pre { background-color: hsl(47, 99%, 75%); } .dataTables_wrapper { position: static; } /* breadcrumbs */ .bread-crumbs{ display: inline-block; margin: 0 0 15px; overflow: hidden; color: rgba(0, 0, 0, 0.55); font-size: 1.2rem; } bread-crumbs ul { list-style-type: none; } .bread-crumbs li { float: left; margin-right: 10px; list-style-type: none; } .bread-crumbs a { color: rgba(0, 0, 0, 0.75); border-bottom-style: none; } .bread-crumbs a:hover { color: rgba(0, 0, 0, 0.85); text-decoration: underline; } .title-small .bread-crumbs{ margin: -30px 0 25px; } #footer { background-color: #262626; color: hsl(0, 0%, 100%); font-size: 1.2rem; text-align: center; padding-top: 20px; padding-bottom: 20px; position: absolute; bottom: 0; left: 0; right: 0; } #footer a, #footer a:visited { color: hsl(0, 0%, 100%); } #footer a:hover { text-decoration: underline; } .highlightjs pre { background-color: transparent; border-radius: 0px; border-color: transparent; } .hljs { background-color: transparent; white-space: pre; } - \ No newline at end of file + +.scrollable-menu { + max-height: 180px; + overflow-x: hidden; +} + +.browse-bread-crumbs { + font-size: inherit; +} + diff --git a/swh/web/static/js/calendar.js b/swh/web/static/js/calendar.js index 842d12c5..4429fba5 100644 --- a/swh/web/static/js/calendar.js +++ b/swh/web/static/js/calendar.js @@ -1,373 +1,365 @@ /** * Calendar: * A one-off object that makes an AJAX call to the API's visit stats * endpoint, then displays these statistics in a zoomable timeline-like * format. * Args: * browse_url: the relative URL for browsing a revision via the web ui, * accurate up to the origin * visit_url: the complete relative URL for getting the origin's visit * stats * origin_id: the origin being browsed * zoomw: the element that should contain the zoomable part of the calendar * staticw: the element that should contain the static part of the calendar * reset: the element that should reset the zoom level on click */ -var Calendar = function(browse_url, visit_url, origin_id, +var Calendar = function(browse_url, data, origin_id, zoomw, staticw, reset) { /** Constants **/ this.month_names = ['Jan', 'Feb', 'Mar', 'Apr', 'May', 'Jun', 'Jul', 'Aug', 'Sep', 'Oct', 'Nov', 'Dec']; /** Display **/ this.desiredPxWidth = 7; this.padding = 0.01; /** Object vars **/ this.origin_id = origin_id; this.zoomw = zoomw; this.staticw = staticw; /** Calendar data **/ this.cal_data = null; this.static = { group_factor: 3600 * 1000, group_data: null, plot_data: null }; this.zoom = { group_factor: 3600 * 1000, group_data: null, plot_data: null }; /** * Keep a reference to the Calendar object context. * Otherwise, 'this' changes to represent current function caller scope */ var self = this; - /** Start AJAX call **/ - $.ajax({ - type: 'GET', - url: visit_url, - success: function(data) { - self.calendar(data); - } - }); - /** * Group the plot's base data according to the grouping ratio and the * range required * Args: * groupFactor: the amount the data should be grouped by * range: * Returns: * A dictionary containing timestamps divided by the grouping ratio as * keys, a list of the corresponding complete timestamps as values */ this.dataGroupedRange = function(groupFactor, range) { var group_dict = {}; var start = range.xaxis.from; var end = range.xaxis.to; var range_data = self.cal_data.filter(function(item, index, arr) { return item >= start && item <= end; }); for (var date_idx in range_data) { var date = range_data[date_idx]; var floor = Math.floor(date / groupFactor); if (group_dict[floor] == undefined) group_dict[floor] = [date]; else group_dict[floor].push(date); } return group_dict; }; /** * Update the ratio that governs how the data is grouped based on changes * in the data range or the display size, and regroup the plot's data * according to this value. * * Args: * element: the element in which the plot is displayed * plotprops: the properties corresponding to that plot * range: the range of the data displayed */ this.updateGroupFactorAndData = function(element, plotprops, range) { var milli_length = range.xaxis.to - range.xaxis.from; var px_length = element.width(); plotprops.group_factor = Math.floor( self.desiredPxWidth * (milli_length / px_length)); plotprops.group_data = self.dataGroupedRange( plotprops.group_factor, range); }; /** Get plot data from the group data **/ this.getPlotData = function(grouped_data) { var plot_data = []; if (self.cal_data.length == 1) { plot_data = [[self.cal_data[0] - 3600*1000*24*30, 0], [self.cal_data[0], 1], [self.cal_data[0] + 3600*1000*24*30, 0]]; } else { $.each(grouped_data, function(key, value) { plot_data.push([value[0], value.length]); }); } return [{ label: 'Calendar', data: plot_data }]; }; this.plotZoom = function(zoom_options) { return $.plot(self.zoomw, self.zoom.plot_data, zoom_options); }; this.plotStatic = function(static_options) { return $.plot(self.staticw, self.static.plot_data, static_options); }; /** * Display a zoomable calendar with click-through links to revisions * of the same origin * * Args: * data: the data that the calendar should present, as a list of * POSIX second-since-epoch timestamps */ this.calendar = function(data) { // POSIX timestamps to JS timestamps self.cal_data = data.map(function(e) { return Math.floor(e['date'] * 1000); }); /** Bootstrap the group ratio **/ var cal_data_range = null; if (self.cal_data.length == 1) { var padding_qty = 3600*1000*24*30; cal_data_range = {xaxis: {from: self.cal_data[0] - padding_qty, to: self.cal_data[0] + padding_qty}}; } else cal_data_range = {xaxis: {from: self.cal_data[0], to: self.cal_data[self.cal_data.length -1] } }; self.updateGroupFactorAndData(self.zoomw, self.zoom, cal_data_range); self.updateGroupFactorAndData(self.staticw, self.static, cal_data_range); /** Bootstrap the plot data **/ self.zoom.plot_data = self.getPlotData(self.zoom.group_data); self.static.plot_data = self.getPlotData(self.zoom.group_data); /** * Return the flot-required function for displaying tooltips, according to * the group we want to display the tooltip for * Args: * group_options: the group we want to display the tooltip for (self.static * or self.zoom) */ function tooltip_fn(group_options) { return function (label, x_timestamp, y_hits, item) { var floor_index = Math.floor( item.datapoint[0] / group_options.group_factor); var tooltip_text = group_options.group_data[floor_index].map( function(elem) { var date = new Date(elem); - var year = (date.getYear() + 1900).toString(); - var month = self.month_names[date.getMonth()]; - var day = date.getDate(); - var hr = date.getHours(); - var min = date.getMinutes(); + var year = date.getUTCFullYear(); + var month = self.month_names[date.getUTCMonth()]; + var day = date.getUTCDate(); + var hr = date.getUTCHours(); + var min = date.getUTCMinutes(); if (min < 10) min = '0'+min; return [day, month, year + ',', hr+':'+min, 'UTC'].join(' '); } ); return tooltip_text.join('
'); }; } /** Plot options for both graph windows **/ var zoom_options = { legend: { show: false }, series: { clickable: true, bars: { show: true, lineWidth: 1, barWidth: self.zoom.group_factor } }, xaxis: { mode: 'time', minTickSize: [1, 'day'], // monthNames: self.month_names, position: 'top' }, yaxis: { show: false }, selection: { mode: 'x' }, grid: { clickable: true, hoverable: true }, tooltip: { show: true, content: tooltip_fn(self.zoom) } }; var overview_options = { legend: { show: false }, series: { clickable: true, bars: { show: true, lineWidth: 1, barWidth: self.static.group_factor }, shadowSize: 0 }, yaxis: { show: false }, xaxis: { mode: 'time', minTickSize: [1, 'day'] }, grid: { clickable: true, hoverable: true, color: '#999' }, selection: { mode: 'x' }, tooltip: { show: true, content: tooltip_fn(self.static) } }; function addPadding(options, range) { var len = range.xaxis.to - range.xaxis.from; return $.extend(true, {}, options, { xaxis: { min: range.xaxis.from - (self.padding * len), max: range.xaxis.to + (self.padding * len) } }); } /** draw the windows **/ var plot = self.plotZoom(addPadding(zoom_options, cal_data_range)); var overview = self.plotStatic( addPadding(overview_options, cal_data_range)); var current_ranges = $.extend(true, {}, cal_data_range); /** * Zoom to the mouse-selected range in the given window * * Args: * plotzone: the jQuery-selected element the zoomed plot should be * in (usually the same as the original 'zoom plot' element) * range: the data range as a dict {xaxis: {from:, to:}, * yaxis:{from:, to:}} */ function zoom(ranges) { current_ranges.xaxis.from = ranges.xaxis.from; current_ranges.xaxis.to = ranges.xaxis.to; self.updateGroupFactorAndData( self.zoomw, self.zoom, current_ranges); self.zoom.plot_data = self.getPlotData(self.zoom.group_data); var zoomedopts = $.extend(true, {}, zoom_options, { xaxis: { min: ranges.xaxis.from, max: ranges.xaxis.to }, series: { bars: {barWidth: self.zoom.group_factor} } }); return self.plotZoom(zoomedopts); } function resetZoomW(plot_options) { self.zoom.group_data = self.static.group_data; self.zoom.plot_data = self.static.plot_data; self.updateGroupFactorAndData(zoomw, self.zoom, cal_data_range); plot = self.plotZoom(addPadding(plot_options, cal_data_range)); } // now connect the two self.zoomw.bind('plotselected', function (event, ranges) { // clamp the zooming to prevent eternal zoom if (ranges.xaxis.to - ranges.xaxis.from < 0.00001) ranges.xaxis.to = ranges.xaxis.from + 0.00001; // do the zooming plot = zoom(ranges); // don't fire event on the overview to prevent eternal loop overview.setSelection(ranges, true); }); self.staticw.bind('plotselected', function (event, ranges) { plot.setSelection(ranges); }); function unbindClick() { self.zoomw.unbind('plotclick'); self.staticw.unbind('plotclick'); } function bindClick() { self.zoomw.bind('plotclick', redirect_to_revision); self.staticw.bind('plotclick', redirect_to_revision); } function redirect_to_revision(event, pos, item) { if (item) { var ts = Math.floor(item.datapoint[0] / 1000); // POSIX ts - var url = browse_url + 'ts/' + ts + '/'; + var url = browse_url + 'ts/' + ts + '/directory/'; window.location.href = url; } } reset.click(function(event) { plot.clearSelection(); overview.clearSelection(); current_ranges = $.extend(true, {}, cal_data_range); resetZoomW(zoom_options); }); $(window).resize(function(event) { /** Update zoom display **/ self.updateGroupFactorAndData(zoomw, self.zoom, current_ranges); self.zoom.plot_data = self.getPlotData(self.zoom.group_data); /** Update static display **/ self.updateGroupFactorAndData(staticw, self.static, cal_data_range); self.static.plot_data = self.getPlotData(self.static.group_data); /** Replot **/ plot = self.plotZoom( addPadding(zoom_options, current_ranges)); overview = self.plotStatic( addPadding(overview_options, cal_data_range)); }); bindClick(); }; + self.calendar(data); }; diff --git a/swh/web/templates/content-directory-top-navigation.html b/swh/web/templates/content-directory-top-navigation.html new file mode 100644 index 00000000..b261dbee --- /dev/null +++ b/swh/web/templates/content-directory-top-navigation.html @@ -0,0 +1,28 @@ + +
+ {% if branches %} + + {% endif %} + + +
+ \ No newline at end of file diff --git a/swh/web/templates/content.html b/swh/web/templates/content.html index e02add03..e2787d88 100644 --- a/swh/web/templates/content.html +++ b/swh/web/templates/content.html @@ -1,37 +1,29 @@ {% extends "layout.html" %} {% block title %}Content display{% endblock %} {% block content %} -{% if root_dir %} - -{% endif %} +
- Content mime type: {{ mime_type }} - Raw File + Content mime type: {{ mime_type }} , Content sha1 git: {{ content_sha1_git }} + Raw File
+ +{% include "content-directory-top-navigation.html" %} +
-{% if "text/" in mime_type or "inode/x-empty" == mime_type %} -
-
-      
-{{ content }}
-      
-    
-
+{% if "inode/x-empty" == mime_type %} +File is empty +{% elif "text/" in mime_type %} +
+
+    {{ content }}
+  
+
{% elif "image/" in mime_type and content %} {% else %} Content with mime type {{ mime_type }} can not be displayed {% endif %}
+ {% endblock %} diff --git a/swh/web/templates/directory.html b/swh/web/templates/directory.html index 0649414f..8e20e2ca 100644 --- a/swh/web/templates/directory.html +++ b/swh/web/templates/directory.html @@ -1,49 +1,41 @@ {% extends "layout.html" %} {% block title %}Directory browse{% endblock %} {% block content %} - +
+ Directory sha1 git: {{ dir_sha1_git }} +
+ +{% include "content-directory-top-navigation.html" %} +
- + {% for d in dirs %} - + {% endfor %} {% for f in files %} - + {% endfor %}
FileSha1Sha1 git
- {% with path|add:d.name as dir_path %} - {{d.name}} - {% endwith %} + {{ d.name }} {{d.target}}{{ d.target }}
- {% with root_dir|add:'/'|add:path|add:f.name as file_path %} - {{f.name}} - {% endwith %} + {{ f.name }} {{f.target}}{{ f.target }}
{% endblock %} diff --git a/swh/web/templates/origin.html b/swh/web/templates/origin.html index eaaeac4b..04f029d4 100644 --- a/swh/web/templates/origin.html +++ b/swh/web/templates/origin.html @@ -1,51 +1,71 @@ {% extends "layout.html" %} -{% block title %}Origin{% endblock %} -{% block content %} - -{% if message is not none %} -{{ message }} -{% endif %} +{% load static %} +{% load swh_templatetags %} -{% if origin is not none %} +{% block title %}Origin information{% endblock %} +{% block content %} - + + + +

Origin information

+ +{% for key, val in origin.items|dictsort:0 %} + + + + +{% endfor %} +
{{ key }}{{ val | safe | urlize_api_links | safe }}

Origin visit history

+

Calendar

- +
- -

Origin information

-
Details on origin {{ origin['id'] }}: - {% for key in ['type', 'lister', 'project', 'url'] %} - {% if origin[key] is not none %} -
-
{{ key }}
-
{{ origin[key] }}
-
- {% endif %} - {% endfor %} - {% if 'decoding_failures' in content %} -
-
(some decoding errors)
-
- {% endif %} +

Visit list

+
+ + + + + + + + + + + {% for v in visits %} + + + + + + + {% endfor %} + +
Visit idVisit dateVisit statusBrowse revision url
{{ v.visit }}{{ v.date }}{{ v.status }}{{ v.browse_url }}
- -{% endif %} + {% endblock %} diff --git a/swh/web/tests/browse/test_utils.py b/swh/web/tests/browse/test_utils.py index a243d2d8..7f1c0516 100644 --- a/swh/web/tests/browse/test_utils.py +++ b/swh/web/tests/browse/test_utils.py @@ -1,34 +1,103 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import unittest +from unittest.mock import patch from nose.tools import istest + from swh.web.browse import utils -class SwhUiUtilsTestCase(unittest.TestCase): +class SwhBrowseUtilsTestCase(unittest.TestCase): @istest def gen_path_info(self): input_path = '/home/user/swh-environment/swh-web/' expected_result = [ {'name': 'home', 'path': 'home'}, {'name': 'user', 'path': 'home/user'}, {'name': 'swh-environment', 'path': 'home/user/swh-environment'}, {'name': 'swh-web', 'path': 'home/user/swh-environment/swh-web'} ] path_info = utils.gen_path_info(input_path) self.assertEquals(path_info, expected_result) input_path = 'home/user/swh-environment/swh-web' path_info = utils.gen_path_info(input_path) self.assertEquals(path_info, expected_result) @istest def get_mimetype_for_content(self): text = b'Hello world!' self.assertEqual(utils.get_mimetype_for_content(text), 'text/plain') + + @patch('swh.web.browse.utils.service') + @istest + def get_origin_visits(self, mock_service): + mock_service.MAX_LIMIT = 2 + + def _lookup_origin_visits(*args, **kwargs): + if kwargs['last_visit'] is None: + return [{'visit': 1}, {'visit': 2}] + else: + return [{'visit': 3}] + + mock_service.lookup_origin_visits.side_effect = _lookup_origin_visits + + origin_visits = utils.get_origin_visits(1) + + self.assertEqual(len(origin_visits), 3) + + @patch('swh.web.browse.utils.service') + @istest + def test_get_origin_visit_branches(self, mock_service): + + mock_service.lookup_origin_visit.return_value = \ + {'date': '2015-08-04T22:26:14.804009+00:00', + 'metadata': {}, + 'occurrences': { + 'refs/heads/master': { + 'target': '9fbd21adbac36be869514e82e2e98505dc47219c', + 'target_type': 'revision', + 'target_url': '/api/1/revision/9fbd21adbac36be869514e82e2e98505dc47219c/' # noqa + }, + 'refs/tags/0.10.0': { + 'target': '6072557b6c10cd9a21145781e26ad1f978ed14b9', + 'target_type': 'revision', + 'target_url': '/api/1/revision/6072557b6c10cd9a21145781e26ad1f978ed14b9/' # noqa + }, + 'refs/tags/0.10.1': { + 'target': 'ecc003b43433e5b46511157598e4857a761007bf', + 'target_type': 'revision', + 'target_url': '/api/1/revision/ecc003b43433e5b46511157598e4857a761007bf/' # noqa + } + }, + 'origin': 1, + 'origin_url': '/api/1/origin/1/', + 'status': 'full', + 'visit': 1} + + mock_service.lookup_revision_multiple.return_value = \ + [{'directory': '828da2b80e41aa958b2c98526f4a1d2cc7d298b7'}, + {'directory': '2df4cd84ecc65b50b1d5318d3727e02a39b8a4cf'}, + {'directory': '28ba64f97ef709e54838ae482c2da2619a74a0bd'}] + + expected_result = [ + {'name': 'refs/heads/master', + 'revision': '9fbd21adbac36be869514e82e2e98505dc47219c', + 'directory': '828da2b80e41aa958b2c98526f4a1d2cc7d298b7'}, + {'name': 'refs/tags/0.10.0', + 'revision': '6072557b6c10cd9a21145781e26ad1f978ed14b9', + 'directory': '2df4cd84ecc65b50b1d5318d3727e02a39b8a4cf'}, + {'name': 'refs/tags/0.10.1', + 'revision': 'ecc003b43433e5b46511157598e4857a761007bf', + 'directory': '28ba64f97ef709e54838ae482c2da2619a74a0bd'} + ] + + origin_visit_branches = utils.get_origin_visit_branches(1, 1) + + self.assertEqual(origin_visit_branches, expected_result) diff --git a/swh/web/tests/browse/views/data/content_test_data.py b/swh/web/tests/browse/views/data/content_test_data.py index 26932496..d79a6470 100644 --- a/swh/web/tests/browse/views/data/content_test_data.py +++ b/swh/web/tests/browse/views/data/content_test_data.py @@ -1,78 +1,240 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information # flake8: noqa import os stub_content_root_dir = '08e8329257dad3a3ef7adea48aa6e576cd82de5b' stub_content_text_file = \ """ /* This file is part of the KDE project * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public License * along with this library; see the file COPYING.LIB. If not, write to * the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifndef KATE_SESSION_TEST_H #define KATE_SESSION_TEST_H #include class KateSessionTest : public QObject { Q_OBJECT private Q_SLOTS: void init(); void cleanup(); void initTestCase(); void cleanupTestCase(); void create(); void createAnonymous(); void createAnonymousFrom(); void createFrom(); void documents(); void setFile(); void setName(); void timestamp(); private: class QTemporaryFile *m_tmpfile; }; #endif """ stub_content_text_sha1 = '5ecd9f37b7a2d2e9980d201acd6286116f2ba1f1' -stub_content_text_path = stub_content_root_dir + '/kate/autotests/session_test.h' +stub_content_text_path = 'kate/autotests/session_test.h' + +stub_content_text_path_with_root_dir = stub_content_root_dir + '/' + stub_content_text_path stub_content_text_data = {'data': str.encode(stub_content_text_file), 'sha1': stub_content_text_sha1} stub_content_bin_filename = 'swh-logo.png' png_file_path = os.path.dirname(__file__) + '/' + stub_content_bin_filename stub_content_bin_sha1 = '02328b91cfad800e1d2808cfb379511b79679ebc' with open(png_file_path, 'rb') as png_file: stub_content_bin_data = {'data': png_file.read(), 'sha1': stub_content_bin_sha1} + +stub_content_origin_id = 10357753 + +stub_content_origin_visit_id = 10 + +stub_content_origin_visit_ts = 1494032350 + +stub_content_origin_branch = 'refs/heads/master' + +stub_content_origin_visits = [ + {'date': '2015-09-26T09:30:52.373449+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 1}, + {'date': '2016-03-10T05:36:11.118989+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 2}, + {'date': '2016-03-24T07:39:29.727793+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 3}, + {'date': '2016-03-31T22:55:31.402863+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 4}, + {'date': '2016-05-26T06:25:54.879676+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 5}, + {'date': '2016-06-07T17:16:33.964164+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 6}, + {'date': '2016-07-27T01:38:20.345358+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 7}, + {'date': '2016-08-13T04:46:45.987508+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 8}, + {'date': '2016-08-16T23:24:13.214496+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 9}, + {'date': '2016-08-17T18:10:39.841005+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 10}, + {'date': '2016-08-30T17:28:02.476486+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 11}, + {'date': '2016-09-08T09:32:37.152054+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 12}, + {'date': '2016-09-15T09:47:37.758093+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 13}, + {'date': '2016-12-04T06:14:02.688518+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 14}, + {'date': '2017-02-16T08:45:57.719974+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'partial', + 'visit': 15}, + {'date': '2017-05-06T00:59:10.495727+00:00', + 'metadata': {}, + 'origin': 10357753, + 'status': 'full', + 'visit': 16} +] + +stub_content_origin_branches = [ + {'directory': '08e8329257dad3a3ef7adea48aa6e576cd82de5b', + 'name': 'HEAD', + 'revision': '11f15b0789344427ddf17b8d75f38577c4395ce0'}, + {'directory': '2371baf0411e3adf12d65daf86c3b135633dd5e4', + 'name': 'refs/heads/Applications/14.12', + 'revision': '5b27ad32f8c8da9b6fc898186d59079488fb74c9'}, + {'directory': '5d024d33a218eeb164936301a2f89231d1f0854a', + 'name': 'refs/heads/Applications/15.04', + 'revision': '4f1e29120795ac643044991e91f24d02c9980202'}, + {'directory': 'f33984df50ec29dbbc86295adb81ebb831e3b86d', + 'name': 'refs/heads/Applications/15.08', + 'revision': '52722e588f46a32b480b5f304ba21480fc8234b1'}, + {'directory': 'e706b836cf32929a48b6f92c07766f237f9d068f', + 'name': 'refs/heads/Applications/15.12', + 'revision': '38c4e42c4a653453fc668c704bb8995ae31b5baf'}, + {'directory': 'ebf8ae783b44df5c827bfa46227e5dbe98f25eb4', + 'name': 'refs/heads/Applications/16.04', + 'revision': 'd0fce3b880ab37a551d75ec940137e0f46bf2143'}, + {'directory': '68ea0543fa80cc512d969fc2294d391a904e04fa', + 'name': 'refs/heads/Applications/16.08', + 'revision': '0b05000bfdde06aec2dc6528411ec24c9e20e672'}, + {'directory': 'b9481c652d57b2e0e36c63f2bf795bc6ffa0b6a1', + 'name': 'refs/heads/Applications/16.12', + 'revision': '2a52ca09fce28e29f5afd0ba4622635679036837'}, + {'directory': '415ea4716870c59feabde3210da6f60bcf897479', + 'name': 'refs/heads/Applications/17.04', + 'revision': 'c7ba6cef1ebfdb743e4f3f53f51f44917981524a'}, + {'directory': 'a9d27a5cd354f2f1e50304ef72818141231f7876', + 'name': 'refs/heads/KDE/4.10', + 'revision': 'e0bc3d8ab537d06c817c459f0be7c7f21d670b6e'}, + {'directory': 'a273331f42e6998099ac98934f33431eb244b222', + 'name': 'refs/heads/KDE/4.11', + 'revision': 'e9db108b584aabe88eff1969f408146b0b9eac32'}, + {'directory': '00be5902593157c55a9888b9e5c17c3b416d1f89', + 'name': 'refs/heads/KDE/4.12', + 'revision': 'c2a1c24f28613342985aa40573fb922370900a3a'}, + {'directory': 'f25aa509cc4ad99478a71407850575d267ae4106', + 'name': 'refs/heads/KDE/4.13', + 'revision': 'b739b7a67882408b4378d901c38b2c88108f1312'}, + {'directory': 'e39f1a6967c33635c9e0c3ee627fbd987612417b', + 'name': 'refs/heads/KDE/4.14', + 'revision': 'dd6530d110b165dfeed8dc1b20b8cfab0e4bd25b'}, + {'directory': '5ac8842a402fe3136be5e2ddd31cb24232152994', + 'name': 'refs/heads/KDE/4.7', + 'revision': '776b581f5f724b1179f2fe013c2da835bb0d5cfc'}, + {'directory': '1e6d88a64ecfa70a6883efd977bfd6248344b108', + 'name': 'refs/heads/KDE/4.8', + 'revision': 'fe3723f6ab789ecf21864e198c91092d10a5289b'}, + {'directory': 'aa03927b4d5738c67646509b4b5d55faef03f024', + 'name': 'refs/heads/KDE/4.9', + 'revision': '69121e434e25f8f4c8ee92a1771a8e87913b3559'}, + {'directory': '57844f10b9ade482ece88ae07a406570e5c0b35d', + 'name': 'refs/heads/goinnn-kate-plugins', + 'revision': 'f51e7810338fe5648319a88712d0ce560cc5f847'}, + {'directory': 'eed636cada058599df292eb59180896cd8aeceac', + 'name': 'refs/heads/kfunk/fix-katecompletionmodel', + 'revision': 'dbf7cae67c5db0737fcf37235000b867cd839f3e'}, + {'directory': '08e8329257dad3a3ef7adea48aa6e576cd82de5b', + 'name': 'refs/heads/master', + 'revision': '11f15b0789344427ddf17b8d75f38577c4395ce0'}, + {'directory': '7b5bdcb46cfaa25229af6b038190b004a26397ff', + 'name': 'refs/heads/plasma/sreich/declarative-kate-applet', + 'revision': '51ab3ea145abd3219c3fae06ff99fa911a6a8993'}, + {'directory': 'e39f1a6967c33635c9e0c3ee627fbd987612417b', + 'name': 'refs/pull/1/head', + 'revision': 'dd6530d110b165dfeed8dc1b20b8cfab0e4bd25b'} +] diff --git a/swh/web/tests/browse/views/data/origin_directory_test_data.py b/swh/web/tests/browse/views/data/origin_directory_test_data.py new file mode 100644 index 00000000..9355a0ae --- /dev/null +++ b/swh/web/tests/browse/views/data/origin_directory_test_data.py @@ -0,0 +1,473 @@ +# Copyright (C) 2017 The Software Heritage developers +# See the AUTHORS file at the top-level directory of this distribution +# License: GNU General Public License version 3, or any later version +# See top-level LICENSE file for more information + +# flake8: noqa + +stub_origin_id = 7416001 +stub_visit_id = 10 +stub_visit_ts = 1493909263 + +stub_origin_visits = [ +{'date': '2015-08-05T18:55:20.899865+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 1}, + {'date': '2016-03-06T12:16:26.240919+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 2}, + {'date': '2016-03-21T11:40:10.329221+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 3}, + {'date': '2016-03-29T08:05:17.602649+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 4}, + {'date': '2016-07-26T20:11:03.827577+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 5}, + {'date': '2016-08-13T04:10:22.142897+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 6}, + {'date': '2016-08-16T22:57:46.201737+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 7}, + {'date': '2016-08-17T17:58:43.346437+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 8}, + {'date': '2016-08-29T23:29:09.445945+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 9}, + {'date': '2016-09-07T13:49:15.096109+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 10}, + {'date': '2016-09-14T15:01:09.017257+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 11}, + {'date': '2016-09-23T12:29:15.921727+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 12}, + {'date': '2017-02-16T07:44:23.302439+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'partial', + 'visit': 13}, + {'date': '2017-05-04T14:47:43.228455+00:00', + 'metadata': {}, + 'origin': 7416001, + 'status': 'full', + 'visit': 14} +] + +stub_origin_branches = [ + {'directory': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'name': 'HEAD', + 'revision': '7bc08e1aa0b08cb23e18715a32aa38517ad34672'}, + {'directory': 'c47a824f95109ca7cafdd1c3206332a0d10df55d', + 'name': 'refs/heads/0.10', + 'revision': 'f944553c77254732c4ce22c0add32aa1f641959d'}, + {'directory': '45e31184ebb7699cd74175145c7eb11cce3f085e', + 'name': 'refs/heads/0.11', + 'revision': '0a29109a6e4579926ebc9b03a6301c61861cce62'}, + {'directory': '42346b33e2d16019490c273ff586ee88817327b3', + 'name': 'refs/heads/0.8', + 'revision': 'e42701dc6f9b035bfbb5d0fffded905d8b456db4'}, + {'directory': '828c7e9385523f852f8d4dac3cb241e319a9ce61', + 'name': 'refs/heads/0.9', + 'revision': '6c3f51e6d9491a2463ad099a2ca49255ec83ff00'}, + {'directory': '2c50e78d63bdc4441c8d2691f5729b04f0ab3ecd', + 'name': 'refs/heads/1.0', + 'revision': 'fb7958d172e1ef6fb77f23bf56818ad24e896e5c'}, + {'directory': '31a3355c4d0a464aa311c5fa11c7f8b20aede6b4', + 'name': 'refs/heads/IgnorePluginHotfix', + 'revision': 'fdc922a2fa007e71b7ec07252012ffab9a178d4a'}, + {'directory': 'e566db1fc65cb61b3799c6e0f0ad06b2406f095f', + 'name': 'refs/heads/beta', + 'revision': '40428853da5d9ce6a8751e13b5e54145337b6a7e'}, + {'directory': '0f3cf363d4184a2cc36fdbebe1657d40fad6a591', + 'name': 'refs/heads/break/add-timeout', + 'revision': '14c43e0cc54af67344f8304708ccdd7b8ab91110'}, + {'directory': '8dc14b0625b2280d630aeae9dd424054c37db820', + 'name': 'refs/heads/bugfix/child-records-cache', + 'revision': '101850c5a90dbc51f1a95813fe1fece42bc1b9dc'}, + {'directory': '08c703fdf5cdb4f9321a1e1ad616eeeb78c96670', + 'name': 'refs/heads/bugfix/disable-module-in-harmony', + 'revision': 'bebe688e6cd0cfe35b967069ddfe1ca42a83f99b'}, + {'directory': '2378f4c463a54f057e42b2e489cd641645f82517', + 'name': 'refs/heads/bugfix/dll-extension', + 'revision': '38264a092e130b83bfe521e6e35a3b3e7600f183'}, + {'directory': 'def1489e00e12ed6ed2515c782da4ed29bf7a2dc', + 'name': 'refs/heads/bugfix/order', + 'revision': '089356faa3d27c1ef4d5c9eee1c235bc8ea505ea'}, + {'directory': 'fe1f7ebe60b771b6494be316fc01c305e0ba5c3c', + 'name': 'refs/heads/bugfix/status', + 'revision': '206ef6d079515a57fbca113cbdcb53cec3def4d6'}, + {'directory': '4cb66b94d3f2438fc095271f7184ca643ce1cdc4', + 'name': 'refs/heads/ci/bench', + 'revision': '21af9e96908e62616eaf04337dfc1b69630765d9'}, + {'directory': 'c1aff7c49f9a38edb1d2e037cc401a9f9e2366ff', + 'name': 'refs/heads/dep/dependency-update', + 'revision': '0bb4b14118ce8ce745222c89c75c166f219a6193'}, + {'directory': 'b4db55e82f9d7b7a9ebc8ce6a2b511d86efb3012', + 'name': 'refs/heads/feature/emoji_presets_to_options', + 'revision': 'a6ebd61bb967d66e1573562f58b2d36893cf338d'}, + {'directory': '676474786d3e6334f1a71bf49adb1e53b0681d34', + 'name': 'refs/heads/feature/fix_errors_only_bug', + 'revision': '5dd1d0cb130032fe176e70b770c38e7df3d804d2'}, + {'directory': '8c8576394828f016d7981e10199ba7a4dee58cf8', + 'name': 'refs/heads/feature/fix_watch_test_cases_timeout_for_caching_harmony', + 'revision': 'd507632afaa4ec2257601e58e17b455ab2491a58'}, + {'directory': 'b1968bb5bbbd604ef52b10668b04869f5790f7c2', + 'name': 'refs/heads/feature/jsonp-to-push', + 'revision': 'cc26df3dd65062d19fca6268adb866dbf9ec0d8e'}, + {'directory': '2a13a1b4e098680cfc4a3a9a103873a0fce232c4', + 'name': 'refs/heads/feature/postinstall-autolink', + 'revision': '7d4e0b72491819e51e84a8747ebcb07bebcf0f6a'}, + {'directory': '27226447d3a8e6de4dec3878eac3617f922b358c', + 'name': 'refs/heads/feature/travis_yarn_command', + 'revision': 'd370967ac8874214023aa69a22c120d4a0cbdef9'}, + {'directory': '4bd9839b5ade32398ff62a7aced9a32cf89981d0', + 'name': 'refs/heads/feature/webpack_analytics', + 'revision': 'a3461fdc6232bef3e4df3bab5079ab5323df1216'}, + {'directory': '1d427c810ab43e1abfdda73a09af41ea4e314a9f', + 'name': 'refs/heads/feature/webpack_options_apply_coverage_increase', + 'revision': 'b92f3fd3ab21959fcdfe492fb927c70014f16def'}, + {'directory': 'fe6b88dd54f3c1043d32e8af812655ea9b8ba000', + 'name': 'refs/heads/fix/node_modules_mangle', + 'revision': '26f79f5b30ba51afae6219ba53bf52be138c2fd3'}, + {'directory': 'b01142c405b9dba0d73dd8232a310ec469c89379', + 'name': 'refs/heads/inline', + 'revision': '427798d4f0e6b72ca3ecb4bfe849b3a17de08729'}, + {'directory': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'name': 'refs/heads/master', + 'revision': '7bc08e1aa0b08cb23e18715a32aa38517ad34672'}, + {'directory': '63fe66a1ecbb35be33f617f96080a78463f89948', + 'name': 'refs/heads/perf/chunks-set', + 'revision': 'f6f22ef0a733931e217050ad9c3b95265978a556'}, + {'directory': '6b813dd7a42991507ef11bc347d53a28340f563c', + 'name': 'refs/heads/perf/modules-set', + 'revision': 'df44599731b4b1d185b20ccd36e9d37d817ad8c2'}, + {'directory': '29cf488a00126c0e45c78b92f97cdbc733f4d07e', + 'name': 'refs/heads/test/circleci', + 'revision': '0f91f949e2d41ef5cb92493bcc6c1fa7578ac27d'}, + {'directory': 'ea3a49842790b4eb8c3805a3318d8ade15cea6c8', + 'name': 'refs/heads/test/dependency-upgrade', + 'revision': 'c6265492ec886405256e23b9fe4de6dfd1e39c49'}, + {'directory': '4de09be1628ed81def03f78d8d832c93efdf0af4', + 'name': 'refs/heads/test/move-entry', + 'revision': 'a244879a07e04e6b5951520ca3cd80c3ef160f8e'}, + {'directory': '0fa77c20564364056319570bd46607111c97cb42', + 'name': 'refs/heads/webpack-1', + 'revision': 'd4878a55be7a4b7ac4e5db1ae5eef89e15811072'} +] + +stub_origin_master_branch = 'refs/heads/master' + +stub_origin_root_directory_sha1 = 'ae59ceecf46367e8e4ad800e231fc76adc3afffb' + +stub_origin_root_directory_entries = [ + {'checksums': {'sha1': '1a17dd2c8245559b43a90aa7c084572e917effff', + 'sha1_git': '012966bd94e648f23b53e71a3f9918e28abc5d81', + 'sha256': 'd65ab1f8cdb323e2b568a8e99814b1b986a38beed85a380981b383c0feb93525'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 394, + 'name': '.editorconfig', + 'perms': 33188, + 'status': 'visible', + 'target': '012966bd94e648f23b53e71a3f9918e28abc5d81', + 'type': 'file'}, + {'checksums': {'sha1': '2e727ec452dc592ae6038d3e09cd35d83d7ea265', + 'sha1_git': '291a4e25598633cd7c286ad8d6cbe9eee5a6291a', + 'sha256': 'd5951c8b796288e0dae1da50575d1b8619462a8df2272cd250146872a1fe804a'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 1839, + 'name': '.eslintrc.js', + 'perms': 33188, + 'status': 'visible', + 'target': '291a4e25598633cd7c286ad8d6cbe9eee5a6291a', + 'type': 'file'}, + {'checksums': {'sha1': '5c59880c0576b2789ec126b61b09fad7a982763b', + 'sha1_git': 'ac579eb7bc04ba44fe84f3c8d1082573e9f4f514', + 'sha256': '8a59a61ff6c0f568a8f76bab434baf3318c80a75ef6fb1b6eb861a0c97518de0'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 67, + 'name': '.gitattributes', + 'perms': 33188, + 'status': 'visible', + 'target': 'ac579eb7bc04ba44fe84f3c8d1082573e9f4f514', + 'type': 'file'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': '.github', + 'perms': 16384, + 'target': '93bdcf98e9c05307b39a9d9e00e48cda6dbd036c', + 'type': 'dir'}, + {'checksums': {'sha1': '7e1008eee2a373f0db7746d0416856aec6b95c22', + 'sha1_git': '84bc35a3abab38bdf87a8f32cc82ce9c136d331e', + 'sha256': '7de369f1d26bc34c7b6329de78973db07e341320eace6a8704a65d4c5bf5993f'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 167, + 'name': '.gitignore', + 'perms': 33188, + 'status': 'visible', + 'target': '84bc35a3abab38bdf87a8f32cc82ce9c136d331e', + 'type': 'file'}, + {'checksums': {'sha1': '06d96508b7d343ff42868f9b6406864517935da7', + 'sha1_git': '79b049846744a2da3eb1c4ac3b01543f2bdca44a', + 'sha256': '697733061d96dd2e061df04dcd86392bb792e2dbe5725a6cb14a436d3c8b76f1'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 706, + 'name': '.jsbeautifyrc', + 'perms': 33188, + 'status': 'visible', + 'target': '79b049846744a2da3eb1c4ac3b01543f2bdca44a', + 'type': 'file'}, + {'checksums': {'sha1': '8041a4a66f46e615c99a850700850a8bd1079dce', + 'sha1_git': '90e4f1ef5beb167891b2e029da6eb9b14ab17add', + 'sha256': '3d6a76a57351b9e3acc5843ff2127dc2cf70c023133312143f86ee74ba9ef6d3'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 1059, + 'name': '.travis.yml', + 'perms': 33188, + 'status': 'visible', + 'target': '90e4f1ef5beb167891b2e029da6eb9b14ab17add', + 'type': 'file'}, + {'checksums': {'sha1': 'cd52973e43c6f4294e8cdfd3106df602b9993f20', + 'sha1_git': 'e5279ebcecd87445648d003c36e6abfebed0ed73', + 'sha256': '130672b16dff61b1541b6d26c2e568ac11830a31d04faace1583d3ad4a38720e'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 2058, + 'name': 'CONTRIBUTING.md', + 'perms': 33188, + 'status': 'visible', + 'target': 'e5279ebcecd87445648d003c36e6abfebed0ed73', + 'type': 'file'}, + {'checksums': {'sha1': '3bebb9ba92e45dd02a0512e144f6a46b14a9b8ab', + 'sha1_git': '8c11fc7289b75463fe07534fcc8224e333feb7ff', + 'sha256': '9068a8782d2fb4c6e432cfa25334efa56f722822180570802bf86e71b6003b1e'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 1071, + 'name': 'LICENSE', + 'perms': 33188, + 'status': 'visible', + 'target': '8c11fc7289b75463fe07534fcc8224e333feb7ff', + 'type': 'file'}, + {'checksums': {'sha1': '6892825420196e84c7104a7ff71ec75db20a1fca', + 'sha1_git': '8f96a0a6d3bfe7183765938483585f3981151553', + 'sha256': 'b0170cfc28f56ca718b43ab086ca5428f853268687c8c033b4fbf028c66d663e'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 46700, + 'name': 'README.md', + 'perms': 33188, + 'status': 'visible', + 'target': '8f96a0a6d3bfe7183765938483585f3981151553', + 'type': 'file'}, + {'checksums': {'sha1': '9bc4902b282f9f1c9f8f885a6947f3bf0f6e6e5f', + 'sha1_git': 'dd6912c8fc97eff255d64da84cfd9837ebf0a05a', + 'sha256': 'e06dbc101195ec7ea0b9aa236be4bdc03784a01f64d6e11846ce3a3f6e1080c6'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 590, + 'name': 'appveyor.yml', + 'perms': 33188, + 'status': 'visible', + 'target': 'dd6912c8fc97eff255d64da84cfd9837ebf0a05a', + 'type': 'file'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'benchmark', + 'perms': 16384, + 'target': '6bd2996b76e051982aa86499a2b485594e607fe3', + 'type': 'dir'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'bin', + 'perms': 16384, + 'target': '681da97ea1ce9a2bd29e3e72781d80e8b961cd51', + 'type': 'dir'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'buildin', + 'perms': 16384, + 'target': '35cfb25d1b3a4063bf04a43f9cbb7e1e87703708', + 'type': 'dir'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'ci', + 'perms': 16384, + 'target': 'efccd3ce0a0304c8cbcffcfdfcafcf1e598819b8', + 'type': 'dir'}, + {'checksums': {'sha1': '9eb3d0e3711f68f82d29785e64ebff2c0d7cec7a', + 'sha1_git': '1ecf877e445bcf865ef53cfcecadda7e9691aace', + 'sha256': '2007e0883c2784bb82584a10d53a0f0c36286dd913741bfd5e4d22b812db529c'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 529, + 'name': 'circle.yml', + 'perms': 33188, + 'status': 'visible', + 'target': '1ecf877e445bcf865ef53cfcecadda7e9691aace', + 'type': 'file'}, + {'checksums': {'sha1': '63209428718e101492c3bb91509f1b4e319b0d7d', + 'sha1_git': 'b3fa4e6abe22977e6267e9969a593e790bf2cd36', + 'sha256': '5d14c8d70215f46a9722d29c7ebff8cc9bd24509650d7ee601fd461e52a52f7f'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 254, + 'name': 'codecov.yml', + 'perms': 33188, + 'status': 'visible', + 'target': 'b3fa4e6abe22977e6267e9969a593e790bf2cd36', + 'type': 'file'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'examples', + 'perms': 16384, + 'target': '7e3ac01795317fbc36a031a9117e7963d6c7da90', + 'type': 'dir'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'hot', + 'perms': 16384, + 'target': 'a5eea6ca952fba9f7ae4177627ed5e22754df9f5', + 'type': 'dir'}, + {'checksums': {'sha1': '92d9367db4ba049f698f5bf78b6946b8e2d91345', + 'sha1_git': 'eaa9cc4a247b01d6a9c0adc91997fefe6a62be1f', + 'sha256': 'd4b42fa0651cf3d99dea0ca5bd6ba64cc21e80be7d9ea05b2b4423ef8f16ec36'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 19, + 'name': 'input.js', + 'perms': 33188, + 'status': 'visible', + 'target': 'eaa9cc4a247b01d6a9c0adc91997fefe6a62be1f', + 'type': 'file'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'lib', + 'perms': 16384, + 'target': '187d40104aa21475d8af88ccd77fc582cf6ac7a6', + 'type': 'dir'}, + {'checksums': {'sha1': 'f17ffa2dc14262292e2275efa3730a96fe060c44', + 'sha1_git': 'd55b7110929cbba3d94da01494a272b39878ac0f', + 'sha256': '012d4446ef8ab6656251b1b7f8e0217a5666ec04ad952e8a617b70946de17166'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 9132, + 'name': 'open-bot.yaml', + 'perms': 33188, + 'status': 'visible', + 'target': 'd55b7110929cbba3d94da01494a272b39878ac0f', + 'type': 'file'}, + {'checksums': {'sha1': '3a6638e72fcc2499f1a4c9b46d4d00d239bbe1c8', + 'sha1_git': '6d1aa82c90ecd184d136151eb81d240e1fea723e', + 'sha256': '00faf7dde1eb0742f3ca567af4dbcd8c01a38cf30d8faa7f0208f46dbc6b5201'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 4034, + 'name': 'package.json', + 'perms': 33188, + 'status': 'visible', + 'target': '6d1aa82c90ecd184d136151eb81d240e1fea723e', + 'type': 'file'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'schemas', + 'perms': 16384, + 'target': 'f1f89c389f73c29e7a5d1a0ce5f9e0f166857815', + 'type': 'dir'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'test', + 'perms': 16384, + 'target': '318c279189d186a1e06653fc5c78c539878c4d7d', + 'type': 'dir'}, + {'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': None, + 'name': 'web_modules', + 'perms': 16384, + 'target': '93a5cc8e492d0b0323386814a72536381019ef7b', + 'type': 'dir'}, + {'checksums': {'sha1': '8047389fcc8e286ceed5536c677c2e803032cf84', + 'sha1_git': 'eb8509f70158c231a3fd864aecf2649590bbedf3', + 'sha256': '8cbe1ce94349ac3bc6cbcc952efd45d838c6b4524af8a773b18e1ebe8b4f936b'}, + 'dir_id': 'ae59ceecf46367e8e4ad800e231fc76adc3afffb', + 'length': 141192, + 'name': 'yarn.lock', + 'perms': 33188, + 'status': 'visible', + 'target': 'eb8509f70158c231a3fd864aecf2649590bbedf3', + 'type': 'file'} +] + +stub_origin_sub_directory_path = 'lib/webworker' + +stub_origin_sub_directory_entries = [ + {'checksums': {'sha1': '7bf366cd9f4a9835c73aafb70e44f640bab7ad16', + 'sha1_git': '870252b7a175ee5ec2edfe2c22b2d56aa04bece4', + 'sha256': 'e0af438932627dd9d53b36bfe69c3dbad6dc4d4569f6cdb29d606c9df2b128fa'}, + 'dir_id': '02b626051e0935ecd28f50337f452db76803f980', + 'length': 921, + 'name': 'WebWorkerChunkTemplatePlugin.js', + 'perms': 33188, + 'status': 'visible', + 'target': '870252b7a175ee5ec2edfe2c22b2d56aa04bece4', + 'type': 'file'}, + {'checksums': {'sha1': 'e2862b2787702bd3eb856f73627d5d8df5a8b550', + 'sha1_git': 'b3e90d26a68ad9da0a7cc97a262db585fa4c73ba', + 'sha256': '1c254e76248ff5ec7e2185cdb1cfd2e0338087244d2d617a868c346317b7646b'}, + 'dir_id': '02b626051e0935ecd28f50337f452db76803f980', + 'length': 1039, + 'name': 'WebWorkerHotUpdateChunkTemplatePlugin.js', + 'perms': 33188, + 'status': 'visible', + 'target': 'b3e90d26a68ad9da0a7cc97a262db585fa4c73ba', + 'type': 'file'}, + {'checksums': {'sha1': 'a1e04061d3e50bb8c024b07e9464da7392f37bf1', + 'sha1_git': '1e503e028fdd5322c9f7d8ec50f54006cacf334e', + 'sha256': '72dea06510d1a4435346f8dca20d8898a394c52c7382a97bd73d1840e31f90b3'}, + 'dir_id': '02b626051e0935ecd28f50337f452db76803f980', + 'length': 1888, + 'name': 'WebWorkerMainTemplate.runtime.js', + 'perms': 33188, + 'status': 'visible', + 'target': '1e503e028fdd5322c9f7d8ec50f54006cacf334e', + 'type': 'file'}, + {'checksums': {'sha1': 'b95c16e90784cf7025352839133b482149526da0', + 'sha1_git': '46c9fe382d606ce19e556deeae6a23af47a8027d', + 'sha256': 'c78c7ca9ee0aa341f843a431ef27c75c386607be3037d44ff530bfe3218edb3c'}, + 'dir_id': '02b626051e0935ecd28f50337f452db76803f980', + 'length': 4051, + 'name': 'WebWorkerMainTemplatePlugin.js', + 'perms': 33188, + 'status': 'visible', + 'target': '46c9fe382d606ce19e556deeae6a23af47a8027d', + 'type': 'file'}, + {'checksums': {'sha1': 'ec9df36b1e8dd689d84dbeeeb9f45fe9f9d96605', + 'sha1_git': 'd850018bb0d2ad41dd0ae9e5c887dff8a23601e9', + 'sha256': 'f995f6a13511955244850c2344c6cef09c10ab24c49f8448544e2b34aa69d03c'}, + 'dir_id': '02b626051e0935ecd28f50337f452db76803f980', + 'length': 763, + 'name': 'WebWorkerTemplatePlugin.js', + 'perms': 33188, + 'status': 'visible', + 'target': 'd850018bb0d2ad41dd0ae9e5c887dff8a23601e9', + 'type': 'file'} +] \ No newline at end of file diff --git a/swh/web/tests/browse/views/data/origin_test_data.py b/swh/web/tests/browse/views/data/origin_test_data.py new file mode 100644 index 00000000..60584633 --- /dev/null +++ b/swh/web/tests/browse/views/data/origin_test_data.py @@ -0,0 +1,125 @@ +# Copyright (C) 2017 The Software Heritage developers +# See the AUTHORS file at the top-level directory of this distribution +# License: GNU General Public License version 3, or any later version +# See top-level LICENSE file for more information + +# flake8: noqa + +origin_info_test_data = { + 'id': 2, + 'type': 'git', + 'url': 'https://github.com/torvalds/linux' +} + +origin_visits_test_data = [ + {'date': '2015-07-09T21:09:24+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 1}, + {'date': '2016-02-23T18:05:23.312045+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 2}, + {'date': '2016-03-28T01:35:06.554111+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 3}, + {'date': '2016-06-18T01:22:24.808485+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 4}, + {'date': '2016-08-14T12:10:00.536702+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 5}, + {'date': '2016-08-17T09:16:22.052065+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 6}, + {'date': '2016-08-29T18:55:54.153721+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 7}, + {'date': '2016-09-07T08:44:47.861875+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 8}, + {'date': '2016-09-14T10:36:21.505296+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 9}, + {'date': '2016-09-23T10:14:02.169862+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 10}, + {'date': '2017-02-16T07:53:39.467657+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'partial', + 'visit': 11}, + {'date': '2017-05-04T19:40:09.336451+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 12}, + {'date': '2017-09-07T18:43:13.021746+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 13}, + {'date': '2017-09-09T05:14:33.466107+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 14}, + {'date': '2017-09-09T17:18:54.307789+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 15}, + {'date': '2017-09-10T05:29:01.462971+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 16}, + {'date': '2017-09-10T17:35:20.158515+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 17}, + {'date': '2017-09-11T05:49:58.300518+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 18}, + {'date': '2017-09-11T18:00:15.037345+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 19}, + {'date': '2017-09-12T06:06:34.703343+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 20}, + {'date': '2017-09-12T18:12:35.344511+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 21}, + {'date': '2017-09-13T06:26:36.580675+00:00', + 'metadata': {}, + 'origin': 2, + 'status': 'full', + 'visit': 22} +] diff --git a/swh/web/tests/browse/views/test_content.py b/swh/web/tests/browse/views/test_content.py index 7a5c6172..4c68533a 100644 --- a/swh/web/tests/browse/views/test_content.py +++ b/swh/web/tests/browse/views/test_content.py @@ -1,190 +1,316 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import base64 from unittest.mock import patch -from nose.tools import istest +from nose.tools import istest, nottest from django.test import TestCase from django.utils.html import escape from swh.web.common.utils import reverse from swh.web.browse.utils import ( gen_path_info ) from .data.content_test_data import ( + stub_content_root_dir, stub_content_text_data, stub_content_text_sha1, + stub_content_text_path_with_root_dir, stub_content_text_path, stub_content_bin_data, - stub_content_bin_sha1, stub_content_bin_filename + stub_content_bin_sha1, stub_content_bin_filename, + stub_content_origin_id, stub_content_origin_visit_id, + stub_content_origin_visit_ts, stub_content_origin_branch, + stub_content_origin_visits, stub_content_origin_branches ) -class SwhUiContentViewTest(TestCase): +class SwhBrowseContentViewTest(TestCase): @patch('swh.web.browse.views.content.service') @istest def content_view_text(self, mock_service): mock_service.lookup_content_raw.return_value =\ stub_content_text_data mock_service.lookup_content_filetype.return_value =\ None url = reverse('browse-content', kwargs={'sha1_git': stub_content_text_sha1}) url_raw = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_text_sha1}) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertTemplateUsed('content.html') self.assertContains(resp, '') self.assertContains(resp, escape(stub_content_text_data['data'])) self.assertContains(resp, url_raw) @patch('swh.web.browse.views.content.service') @istest def content_view_image(self, mock_service): mock_service.lookup_content_raw.return_value =\ stub_content_bin_data mime_type = 'image/png' mock_service.lookup_content_filetype.return_value =\ {'mimetype': mime_type} url = reverse('browse-content', kwargs={'sha1_git': stub_content_bin_sha1}) url_raw = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_bin_sha1}) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertTemplateUsed('content.html') pngEncoded = base64.b64encode(stub_content_bin_data['data']) \ .decode('utf-8') self.assertContains(resp, '' % (mime_type, pngEncoded)) self.assertContains(resp, url_raw) @patch('swh.web.browse.views.content.service') @istest def content_view_with_path(self, mock_service): mock_service.lookup_content_raw.return_value =\ stub_content_text_data mock_service.lookup_content_filetype.return_value =\ {'mimetype': 'text/x-c++'} url = reverse('browse-content', kwargs={'sha1_git': stub_content_text_sha1}, - query_params={'path': stub_content_text_path}) + query_params={'path': stub_content_text_path_with_root_dir}) # noqa resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertTemplateUsed('content.html') self.assertContains(resp, '') self.assertContains(resp, escape(stub_content_text_data['data'])) - split_path = stub_content_text_path.split('/') + split_path = stub_content_text_path_with_root_dir.split('/') root_dir_sha1 = split_path[0] filename = split_path[-1] - path = stub_content_text_path.replace(root_dir_sha1 + '/', '') \ - .replace(filename, '') + path = stub_content_text_path_with_root_dir \ + .replace(root_dir_sha1 + '/', '') \ + .replace(filename, '') path_info = gen_path_info(path) root_dir_url = reverse('browse-directory', kwargs={'sha1_git': root_dir_sha1}) self.assertContains(resp, '
  • ', count=len(path_info)+1) self.assertContains(resp, '' + root_dir_sha1[:7] + '') for p in path_info: dir_url = reverse('browse-directory', kwargs={'sha1_git': root_dir_sha1, 'path': p['path']}) self.assertContains(resp, '' + p['name'] + '') self.assertContains(resp, '
  • ' + filename + '
  • ') url_raw = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_text_sha1}, query_params={'filename': filename}) self.assertContains(resp, url_raw) @patch('swh.web.browse.views.content.service') @istest def content_raw_text(self, mock_service): mock_service.lookup_content_raw.return_value =\ stub_content_text_data url = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_text_sha1}) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertEqual(resp['Content-Type'], 'text/plain') self.assertEqual(resp['Content-disposition'], 'filename=%s' % stub_content_text_sha1) self.assertEqual(resp.content, stub_content_text_data['data']) - filename = stub_content_text_path.split('/')[-1] + filename = stub_content_text_path_with_root_dir.split('/')[-1] url = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_text_sha1}, query_params={'filename': filename}) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertEqual(resp['Content-Type'], 'text/plain') self.assertEqual(resp['Content-disposition'], 'filename=%s' % filename) self.assertEqual(resp.content, stub_content_text_data['data']) @patch('swh.web.browse.views.content.service') @istest def content_raw_bin(self, mock_service): mock_service.lookup_content_raw.return_value =\ stub_content_bin_data url = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_bin_sha1}) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertEqual(resp['Content-Type'], 'application/octet-stream') self.assertEqual(resp['Content-disposition'], 'attachment; filename=%s' % stub_content_bin_sha1) self.assertEqual(resp.content, stub_content_bin_data['data']) url = reverse('browse-content-raw', kwargs={'sha1_git': stub_content_bin_sha1}, query_params={'filename': stub_content_bin_filename}) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertEqual(resp['Content-Type'], 'application/octet-stream') self.assertEqual(resp['Content-disposition'], 'attachment; filename=%s' % stub_content_bin_filename) self.assertEqual(resp.content, stub_content_bin_data['data']) + + @nottest + def origin_content_view_test(self, origin_id, origin_visits, + origin_branches, origin_branch, + root_dir_sha1, content_sha1, content_path, + content_data, content_language, + visit_id=None, ts=None): + + url_args = {'origin_id': origin_id, + 'path': content_path} + + if not visit_id: + visit_id = origin_visits[-1]['visit'] + + if ts: + url_args['ts'] = ts + else: + url_args['visit_id'] = visit_id + + url = reverse('browse-origin-content', + kwargs=url_args) + + resp = self.client.get(url) + self.assertEquals(resp.status_code, 200) + self.assertTemplateUsed('content.html') + + self.assertContains(resp, '' % content_language) + self.assertContains(resp, escape(content_data)) + + split_path = content_path.split('/') + + filename = split_path[-1] + path = content_path.replace(filename, '')[:-1] + + path_info = gen_path_info(path) + + del url_args['path'] + + root_dir_url = reverse('browse-origin-directory', + kwargs=url_args, + query_params={'branch': origin_branch}) + + self.assertContains(resp, '
  • ', + count=len(path_info)+1) + + self.assertContains(resp, '%s' % + (root_dir_url, root_dir_sha1[:7])) + + for p in path_info: + url_args['path'] = p['path'] + dir_url = reverse('browse-origin-directory', + kwargs=url_args, + query_params={'branch': origin_branch}) + self.assertContains(resp, '%s' % + (dir_url, p['name'])) + + self.assertContains(resp, '
  • %s
  • ' % filename) + + url_raw = reverse('browse-content-raw', + kwargs={'sha1_git': content_sha1}, + query_params={'filename': filename}) + self.assertContains(resp, url_raw) + + self.assertContains(resp, '
  • ', + count=len(origin_branches)) + + url_args['path'] = content_path + + for branch in origin_branches: + root_dir_branch_url = \ + reverse('browse-origin-content', + kwargs=url_args, + query_params={'branch': branch['name']}) + + self.assertContains(resp, '%s' % + (root_dir_branch_url, branch['name'])) + + @patch('swh.web.browse.views.content.get_origin_visits') + @patch('swh.web.browse.views.content.get_origin_visit_branches') + @patch('swh.web.browse.views.content.service') + @istest + def origin_content_view(self, mock_service, + mock_get_origin_visit_branches, + mock_get_origin_visits): + + mock_get_origin_visits.return_value = stub_content_origin_visits + mock_get_origin_visit_branches.return_value = stub_content_origin_branches # noqa + mock_service.lookup_directory_with_path.return_value = \ + {'target': stub_content_text_sha1} + mock_service.lookup_content_raw.return_value = stub_content_text_data + mock_service.lookup_content_filetype.return_value = {'mimetype': 'text/x-c++'} # noqa + + self.origin_content_view_test(stub_content_origin_id, + stub_content_origin_visits, + stub_content_origin_branches, + stub_content_origin_branch, + stub_content_root_dir, + stub_content_text_sha1, + stub_content_text_path, + stub_content_text_data['data'], 'cpp') + + self.origin_content_view_test(stub_content_origin_id, + stub_content_origin_visits, + stub_content_origin_branches, + stub_content_origin_branch, + stub_content_root_dir, + stub_content_text_sha1, + stub_content_text_path, + stub_content_text_data['data'], 'cpp', + visit_id=stub_content_origin_visit_id) + + self.origin_content_view_test(stub_content_origin_id, + stub_content_origin_visits, + stub_content_origin_branches, + stub_content_origin_branch, + stub_content_root_dir, + stub_content_text_sha1, + stub_content_text_path, + stub_content_text_data['data'], 'cpp', + ts=stub_content_origin_visit_ts) diff --git a/swh/web/tests/browse/views/test_directory.py b/swh/web/tests/browse/views/test_directory.py index b88b45ab..3f05c0d2 100644 --- a/swh/web/tests/browse/views/test_directory.py +++ b/swh/web/tests/browse/views/test_directory.py @@ -1,112 +1,262 @@ # Copyright (C) 2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information from unittest.mock import patch -from nose.tools import istest +from nose.tools import istest, nottest from django.test import TestCase from swh.web.common.utils import reverse from swh.web.browse.utils import gen_path_info from .data.directory_test_data import ( stub_root_directory_sha1, stub_root_directory_data, stub_sub_directory_path, stub_sub_directory_data ) +from .data.origin_directory_test_data import ( + stub_origin_id, stub_visit_id, + stub_origin_visits, stub_origin_branches, + stub_origin_root_directory_entries, stub_origin_master_branch, + stub_origin_root_directory_sha1, stub_origin_sub_directory_path, + stub_origin_sub_directory_entries, stub_visit_ts +) -class SwhUiDirectoryViewTest(TestCase): - @patch('swh.web.browse.views.directory.service') - @istest - def root_directory_view(self, mock_service): - mock_service.lookup_directory.return_value = \ - stub_root_directory_data +class SwhBrowseDirectoryViewTest(TestCase): - dirs = [e for e in stub_root_directory_data if e['type'] == 'dir'] - files = [e for e in stub_root_directory_data if e['type'] == 'file'] + @nottest + def directory_view(self, root_directory_sha1, directory_entries, + path=None): + dirs = [e for e in directory_entries if e['type'] == 'dir'] + files = [e for e in directory_entries if e['type'] == 'file'] + + url_args = {'sha1_git': root_directory_sha1} + if path: + url_args['path'] = path url = reverse('browse-directory', - kwargs={'sha1_git': stub_root_directory_sha1}) + kwargs=url_args) + + root_dir_url = reverse('browse-directory', + kwargs={'sha1_git': root_directory_sha1}) + resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertTemplateUsed('directory.html') + self.assertContains(resp, '' + + root_directory_sha1[:7] + '') self.assertContains(resp, '', count=len(dirs)) self.assertContains(resp, '', count=len(files)) for d in dirs: + dir_path = d['name'] + if path: + dir_path = "%s/%s" % (path, d['name']) dir_url = reverse('browse-directory', - kwargs={'sha1_git': stub_root_directory_sha1, - 'path': d['name']}) + kwargs={'sha1_git': root_directory_sha1, + 'path': dir_path}) self.assertContains(resp, dir_url) for f in files: - file_path = stub_root_directory_sha1 + '/' + f['name'] + file_path = "%s/%s" % (root_directory_sha1, f['name']) + if path: + file_path = "%s/%s/%s" % (root_directory_sha1, path, f['name']) file_url = reverse('browse-content', kwargs={'sha1_git': f['target']}, query_params={'path': file_path}) self.assertContains(resp, file_url) + path_info = gen_path_info(path) + self.assertContains(resp, '
  • ', - count=1) - self.assertContains(resp, '' + - stub_root_directory_sha1[:7] + '') + count=len(path_info)+1) + self.assertContains(resp, '%s' % + (root_dir_url, root_directory_sha1[:7])) + + for p in path_info: + dir_url = reverse('browse-directory', + kwargs={'sha1_git': root_directory_sha1, + 'path': p['path']}) + self.assertContains(resp, '%s' % + (dir_url, p['name'])) + + @patch('swh.web.browse.views.directory.service') + @istest + def root_directory_view(self, mock_service): + mock_service.lookup_directory.return_value = \ + stub_root_directory_data + + self.directory_view(stub_root_directory_sha1, stub_root_directory_data) @patch('swh.web.browse.views.directory.service') @istest def sub_directory_view(self, mock_service): mock_service.lookup_directory.return_value = \ stub_sub_directory_data - dirs = [e for e in stub_sub_directory_data if e['type'] == 'dir'] - files = [e for e in stub_sub_directory_data if e['type'] == 'file'] + self.directory_view(stub_root_directory_sha1, stub_sub_directory_data, + stub_sub_directory_path) - url = reverse('browse-directory', - kwargs={'sha1_git': stub_root_directory_sha1, - 'path': stub_sub_directory_path}) + @nottest + def origin_directory_view(self, origin_id, origin_visits, + origin_branches, origin_branch, + root_directory_sha1, directory_entries, + visit_id=None, ts=None, path=None): - root_dir_url = reverse('browse-directory', - kwargs={'sha1_git': stub_root_directory_sha1}) + dirs = [e for e in directory_entries + if e['type'] == 'dir'] + files = [e for e in directory_entries + if e['type'] == 'file'] + + if not visit_id: + visit_id = origin_visits[-1]['visit'] + + url_args = {'origin_id': origin_id} + + if ts: + url_args['ts'] = ts + else: + url_args['visit_id'] = visit_id + + if path: + url_args['path'] = path + + url = reverse('browse-origin-directory', + kwargs=url_args) resp = self.client.get(url) self.assertEquals(resp.status_code, 200) self.assertTemplateUsed('directory.html') - self.assertContains(resp, '' + - stub_root_directory_sha1[:7] + '') self.assertContains(resp, '', count=len(dirs)) self.assertContains(resp, '', count=len(files)) for d in dirs: - dir_url = reverse('browse-directory', - kwargs={'sha1_git': stub_root_directory_sha1, - 'path': d['name']}) + dir_path = d['name'] + if path: + dir_path = "%s/%s" % (path, d['name']) + dir_url_args = dict(url_args) + dir_url_args['path'] = dir_path + dir_url = reverse('browse-origin-directory', + kwargs=dir_url_args, + query_params={'branch': origin_branch}) # noqa self.assertContains(resp, dir_url) for f in files: - file_path = stub_root_directory_sha1 + '/' + \ - stub_sub_directory_path + '/' + f['name'] - file_url = reverse('browse-content', - kwargs={'sha1_git': f['target']}, - query_params={'path': file_path}) + file_path = f['name'] + if path: + file_path = "%s/%s" % (path, f['name']) + file_url_args = dict(url_args) + file_url_args['path'] = file_path + file_url = reverse('browse-origin-content', + kwargs=file_url_args, + query_params={'branch': origin_branch}) # noqa self.assertContains(resp, file_url) - path_info = gen_path_info(stub_sub_directory_path) + if 'path' in url_args: + del url_args['path'] - self.assertContains(resp, '
  • ', - count=len(path_info)+1) - self.assertContains(resp, '' + - stub_root_directory_sha1[:7] + '') + root_dir_branch_url = \ + reverse('browse-origin-directory', + kwargs=url_args, + query_params={'branch': origin_branch}) - for p in path_info: - dir_url = reverse('browse-directory', - kwargs={'sha1_git': stub_root_directory_sha1, - 'path': p['path']}) - self.assertContains(resp, '' + - p['name'] + '') + nb_bc_paths = 1 + if path: + nb_bc_paths = len(path.split('/')) + 1 + + self.assertContains(resp, '
  • ', count=nb_bc_paths) + self.assertContains(resp, '%s' % + (root_dir_branch_url, + root_directory_sha1[:7])) + + self.assertContains(resp, '
  • ', + count=len(origin_branches)) + + if path: + url_args['path'] = path + + for branch in origin_branches: + root_dir_branch_url = \ + reverse('browse-origin-directory', + kwargs=url_args, + query_params={'branch': branch['name']}) + + self.assertContains(resp, '%s' % + (root_dir_branch_url, branch['name'])) + + @patch('swh.web.browse.views.directory.get_origin_visits') + @patch('swh.web.browse.views.directory.get_origin_visit_branches') + @patch('swh.web.browse.views.directory.service') + @istest + def origin_root_directory_view(self, mock_service, + mock_get_origin_visit_branches, + mock_get_origin_visits): + + mock_get_origin_visits.return_value = stub_origin_visits + mock_get_origin_visit_branches.return_value = stub_origin_branches + mock_service.lookup_directory.return_value = \ + stub_origin_root_directory_entries + + self.origin_directory_view(stub_origin_id, stub_origin_visits, + stub_origin_branches, + stub_origin_master_branch, + stub_origin_root_directory_sha1, + stub_origin_root_directory_entries) + + self.origin_directory_view(stub_origin_id, stub_origin_visits, + stub_origin_branches, + stub_origin_master_branch, + stub_origin_root_directory_sha1, + stub_origin_root_directory_entries, + visit_id=stub_visit_id) + + self.origin_directory_view(stub_origin_id, stub_origin_visits, + stub_origin_branches, + stub_origin_master_branch, + stub_origin_root_directory_sha1, + stub_origin_root_directory_entries, + ts=stub_visit_ts) + + @patch('swh.web.browse.views.directory.get_origin_visits') + @patch('swh.web.browse.views.directory.get_origin_visit_branches') + @patch('swh.web.browse.views.directory.service') + @istest + def origin_sub_directory_view(self, mock_service, + mock_get_origin_visit_branches, + mock_get_origin_visits): + + mock_get_origin_visits.return_value = stub_origin_visits + mock_get_origin_visit_branches.return_value = stub_origin_branches + mock_service.lookup_directory.return_value = \ + stub_origin_sub_directory_entries + + self.origin_directory_view(stub_origin_id, stub_origin_visits, + stub_origin_branches, + stub_origin_master_branch, + stub_origin_root_directory_sha1, + stub_origin_sub_directory_entries, + path=stub_origin_sub_directory_path) + + self.origin_directory_view(stub_origin_id, stub_origin_visits, + stub_origin_branches, + stub_origin_master_branch, + stub_origin_root_directory_sha1, + stub_origin_sub_directory_entries, + visit_id=stub_visit_id, + path=stub_origin_sub_directory_path) + + self.origin_directory_view(stub_origin_id, stub_origin_visits, + stub_origin_branches, + stub_origin_master_branch, + stub_origin_root_directory_sha1, + stub_origin_sub_directory_entries, + ts=stub_visit_ts, + path=stub_origin_sub_directory_path) diff --git a/swh/web/tests/browse/views/test_origin.py b/swh/web/tests/browse/views/test_origin.py new file mode 100644 index 00000000..efd10297 --- /dev/null +++ b/swh/web/tests/browse/views/test_origin.py @@ -0,0 +1,48 @@ +# Copyright (C) 2017 The Software Heritage developers +# See the AUTHORS file at the top-level directory of this distribution +# License: GNU General Public License version 3, or any later version +# See top-level LICENSE file for more information + +from unittest.mock import patch +from nose.tools import istest + +from django.test import TestCase + +from swh.web.common.utils import reverse + +from .data.origin_test_data import ( + origin_info_test_data, + origin_visits_test_data +) + + +class SwhBrowseOriginViewTest(TestCase): + + @patch('swh.web.browse.views.origin.get_origin_visits') + @patch('swh.web.browse.views.origin.service') + @istest + def test_origin_browse(self, mock_service, mock_get_origin_visits): + mock_service.lookup_origin.return_value = origin_info_test_data + mock_get_origin_visits.return_value = origin_visits_test_data + + url = reverse('browse-origin', + kwargs={'origin_id': origin_info_test_data['id']}) + resp = self.client.get(url) + + self.assertEquals(resp.status_code, 200) + self.assertTemplateUsed('origin.html') + self.assertContains(resp, '%s' % origin_info_test_data['id']) + self.assertContains(resp, '%s' % origin_info_test_data['type']) # noqa + self.assertContains(resp, '%s' % + (origin_info_test_data['url'], + origin_info_test_data['url'])) + + self.assertContains(resp, '', + count=len(origin_visits_test_data)) + + for visit in origin_visits_test_data: + browse_url = reverse('browse-origin-directory', + kwargs={'origin_id': visit['origin'], + 'visit_id': visit['visit']}) + self.assertContains(resp, '%s' % + (browse_url, browse_url)) diff --git a/swh/web/tests/common/test_service.py b/swh/web/tests/common/test_service.py index 5bbcb18d..fe6fef86 100644 --- a/swh/web/tests/common/test_service.py +++ b/swh/web/tests/common/test_service.py @@ -1,2071 +1,2072 @@ # Copyright (C) 2015-2017 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU Affero General Public License version 3, or any later version # See top-level LICENSE file for more information import datetime import unittest from nose.tools import istest from unittest.mock import MagicMock, patch, call from swh.model.hashutil import hash_to_bytes, hash_to_hex from swh.web.common import service from swh.web.common.exc import BadInputExc, NotFoundExc class ServiceTestCase(unittest.TestCase): def setUp(self): self.BLAKE2S256_SAMPLE = ('685395c5dc57cada459364f0946d3dd45b' 'ad5fcbabc1048edb44380f1d31d0aa') self.BLAKE2S256_SAMPLE_BIN = hash_to_bytes(self.BLAKE2S256_SAMPLE) self.SHA1_SAMPLE = '40e71b8614fcd89ccd17ca2b1d9e66c5b00a6d03' self.SHA1_SAMPLE_BIN = hash_to_bytes(self.SHA1_SAMPLE) self.SHA256_SAMPLE = ('8abb0aa566452620ecce816eecdef4792d77a' '293ad8ea82a4d5ecb4d36f7e560') self.SHA256_SAMPLE_BIN = hash_to_bytes(self.SHA256_SAMPLE) self.SHA1GIT_SAMPLE = '25d1a2e8f32937b0f498a5ca87f823d8df013c01' self.SHA1GIT_SAMPLE_BIN = hash_to_bytes(self.SHA1GIT_SAMPLE) self.DIRECTORY_ID = '7834ef7e7c357ce2af928115c6c6a42b7e2a44e6' self.DIRECTORY_ID_BIN = hash_to_bytes(self.DIRECTORY_ID) self.AUTHOR_ID_BIN = { 'name': b'author', 'email': b'author@company.org', } self.AUTHOR_ID = { 'name': 'author', 'email': 'author@company.org', } self.COMMITTER_ID_BIN = { 'name': b'committer', 'email': b'committer@corp.org', } self.COMMITTER_ID = { 'name': 'committer', 'email': 'committer@corp.org', } self.SAMPLE_DATE_RAW = { 'timestamp': datetime.datetime( 2000, 1, 17, 11, 23, 54, tzinfo=datetime.timezone.utc, ).timestamp(), 'offset': 0, 'negative_utc': False, } self.SAMPLE_DATE = '2000-01-17T11:23:54+00:00' self.SAMPLE_MESSAGE_BIN = b'elegant fix for bug 31415957' self.SAMPLE_MESSAGE = 'elegant fix for bug 31415957' self.SAMPLE_REVISION = { 'id': self.SHA1_SAMPLE, 'directory': self.DIRECTORY_ID, 'author': self.AUTHOR_ID, 'committer': self.COMMITTER_ID, 'message': self.SAMPLE_MESSAGE, 'date': self.SAMPLE_DATE, 'committer_date': self.SAMPLE_DATE, 'synthetic': False, 'type': 'git', 'parents': [], 'metadata': {}, 'merge': False } self.SAMPLE_REVISION_RAW = { 'id': self.SHA1_SAMPLE_BIN, 'directory': self.DIRECTORY_ID_BIN, 'author': self.AUTHOR_ID_BIN, 'committer': self.COMMITTER_ID_BIN, 'message': self.SAMPLE_MESSAGE_BIN, 'date': self.SAMPLE_DATE_RAW, 'committer_date': self.SAMPLE_DATE_RAW, 'synthetic': False, 'type': 'git', 'parents': [], 'metadata': [], } self.SAMPLE_CONTENT = { 'checksums': { 'blake2s256': self.BLAKE2S256_SAMPLE, 'sha1': self.SHA1_SAMPLE, 'sha256': self.SHA256_SAMPLE, 'sha1_git': self.SHA1GIT_SAMPLE, }, 'length': 190, 'status': 'absent' } self.SAMPLE_CONTENT_RAW = { 'blake2s256': self.BLAKE2S256_SAMPLE_BIN, 'sha1': self.SHA1_SAMPLE_BIN, 'sha256': self.SHA256_SAMPLE_BIN, 'sha1_git': self.SHA1GIT_SAMPLE_BIN, 'length': 190, 'status': 'hidden' } self.date_origin_visit1 = datetime.datetime( 2015, 1, 1, 22, 0, 0, tzinfo=datetime.timezone.utc) self.origin_visit1 = { 'date': self.date_origin_visit1, 'origin': 1, 'visit': 1 } @patch('swh.web.common.service.storage') @istest def lookup_multiple_hashes_ball_missing(self, mock_storage): # given mock_storage.content_missing_per_sha1 = MagicMock(return_value=[]) # when actual_lookup = service.lookup_multiple_hashes( [{'filename': 'a', 'sha1': '456caf10e9535160d90e874b45aa426de762f19f'}, {'filename': 'b', 'sha1': '745bab676c8f3cec8016e0c39ea61cf57e518865'}]) # then self.assertEquals(actual_lookup, [ {'filename': 'a', 'sha1': '456caf10e9535160d90e874b45aa426de762f19f', 'found': True}, {'filename': 'b', 'sha1': '745bab676c8f3cec8016e0c39ea61cf57e518865', 'found': True} ]) @patch('swh.web.common.service.storage') @istest def lookup_multiple_hashes_some_missing(self, mock_storage): # given mock_storage.content_missing_per_sha1 = MagicMock(return_value=[ hash_to_bytes('456caf10e9535160d90e874b45aa426de762f19f') ]) # when actual_lookup = service.lookup_multiple_hashes( [{'filename': 'a', 'sha1': '456caf10e9535160d90e874b45aa426de762f19f'}, {'filename': 'b', 'sha1': '745bab676c8f3cec8016e0c39ea61cf57e518865'}]) # then self.assertEquals(actual_lookup, [ {'filename': 'a', 'sha1': '456caf10e9535160d90e874b45aa426de762f19f', 'found': False}, {'filename': 'b', 'sha1': '745bab676c8f3cec8016e0c39ea61cf57e518865', 'found': True} ]) @patch('swh.web.common.service.storage') @istest def lookup_hash_does_not_exist(self, mock_storage): # given mock_storage.content_find = MagicMock(return_value=None) # when actual_lookup = service.lookup_hash( 'sha1_git:123caf10e9535160d90e874b45aa426de762f19f') # then self.assertEquals({'found': None, 'algo': 'sha1_git'}, actual_lookup) # check the function has been called with parameters mock_storage.content_find.assert_called_with( {'sha1_git': hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')}) @patch('swh.web.common.service.storage') @istest def lookup_hash_exist(self, mock_storage): # given stub_content = { 'sha1': hash_to_bytes( '456caf10e9535160d90e874b45aa426de762f19f') } mock_storage.content_find = MagicMock(return_value=stub_content) # when actual_lookup = service.lookup_hash( 'sha1:456caf10e9535160d90e874b45aa426de762f19f') # then self.assertEquals({'found': stub_content, 'algo': 'sha1'}, actual_lookup) mock_storage.content_find.assert_called_with( {'sha1': hash_to_bytes('456caf10e9535160d90e874b45aa426de762f19f')} ) @patch('swh.web.common.service.storage') @istest def search_hash_does_not_exist(self, mock_storage): # given mock_storage.content_find = MagicMock(return_value=None) # when actual_lookup = service.search_hash( 'sha1_git:123caf10e9535160d90e874b45aa426de762f19f') # then self.assertEquals({'found': False}, actual_lookup) # check the function has been called with parameters mock_storage.content_find.assert_called_with( {'sha1_git': hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')}) @patch('swh.web.common.service.storage') @istest def search_hash_exist(self, mock_storage): # given stub_content = { 'sha1': hash_to_bytes( '456caf10e9535160d90e874b45aa426de762f19f') } mock_storage.content_find = MagicMock(return_value=stub_content) # when actual_lookup = service.search_hash( 'sha1:456caf10e9535160d90e874b45aa426de762f19f') # then self.assertEquals({'found': True}, actual_lookup) mock_storage.content_find.assert_called_with( {'sha1': hash_to_bytes('456caf10e9535160d90e874b45aa426de762f19f')}, ) @patch('swh.web.common.service.storage') @istest def lookup_content_ctags(self, mock_storage): # given mock_storage.content_ctags_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'line': 100, 'name': 'hello', 'kind': 'function', 'tool_name': 'ctags', 'tool_version': 'some-version', }]) expected_ctags = [{ 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'line': 100, 'name': 'hello', 'kind': 'function', 'tool_name': 'ctags', 'tool_version': 'some-version', }] # when actual_ctags = list(service.lookup_content_ctags( 'sha1:123caf10e9535160d90e874b45aa426de762f19f')) # then self.assertEqual(actual_ctags, expected_ctags) mock_storage.content_ctags_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_content_ctags_no_hash(self, mock_storage): # given mock_storage.content_find.return_value = None mock_storage.content_ctags_get = MagicMock( return_value=None) # when actual_ctags = list(service.lookup_content_ctags( 'sha1_git:123caf10e9535160d90e874b45aa426de762f19f')) # then self.assertEqual(actual_ctags, []) mock_storage.content_find.assert_called_once_with( {'sha1_git': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f')}) @patch('swh.web.common.service.storage') @istest def lookup_content_filetype(self, mock_storage): # given mock_storage.content_mimetype_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'mimetype': b'text/x-c++', 'encoding': b'us-ascii', }]) expected_filetype = { 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'mimetype': 'text/x-c++', 'encoding': 'us-ascii', } # when actual_filetype = service.lookup_content_filetype( 'sha1:123caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(actual_filetype, expected_filetype) mock_storage.content_mimetype_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_content_filetype_2(self, mock_storage): # given mock_storage.content_find = MagicMock( return_value={ 'sha1': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f') } ) mock_storage.content_mimetype_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'mimetype': b'text/x-python', 'encoding': b'us-ascii', }] ) expected_filetype = { 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'mimetype': 'text/x-python', 'encoding': 'us-ascii', } # when actual_filetype = service.lookup_content_filetype( 'sha1_git:456caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(actual_filetype, expected_filetype) mock_storage.content_find( 'sha1_git', hash_to_bytes( '456caf10e9535160d90e874b45aa426de762f19f') ) mock_storage.content_mimetype_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_content_language(self, mock_storage): # given mock_storage.content_language_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'lang': 'python', }]) expected_language = { 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'lang': 'python', } # when actual_language = service.lookup_content_language( 'sha1:123caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(actual_language, expected_language) mock_storage.content_language_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_content_language_2(self, mock_storage): # given mock_storage.content_find = MagicMock( return_value={ 'sha1': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f') } ) mock_storage.content_language_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'lang': 'haskell', }] ) expected_language = { 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'lang': 'haskell', } # when actual_language = service.lookup_content_language( 'sha1_git:456caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(actual_language, expected_language) mock_storage.content_find( 'sha1_git', hash_to_bytes( '456caf10e9535160d90e874b45aa426de762f19f') ) mock_storage.content_language_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_expression(self, mock_storage): # given mock_storage.content_ctags_search = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'name': 'foobar', 'kind': 'variable', 'lang': 'C', 'line': 10 }]) expected_ctags = [{ 'sha1': '123caf10e9535160d90e874b45aa426de762f19f', 'name': 'foobar', 'kind': 'variable', 'lang': 'C', 'line': 10 }] # when actual_ctags = list(service.lookup_expression( 'foobar', last_sha1='hash', per_page=10)) # then self.assertEqual(actual_ctags, expected_ctags) mock_storage.content_ctags_search.assert_called_with( 'foobar', last_sha1='hash', limit=10) @patch('swh.web.common.service.storage') @istest def lookup_expression_no_result(self, mock_storage): # given mock_storage.content_ctags_search = MagicMock( return_value=[]) expected_ctags = [] # when actual_ctags = list(service.lookup_expression( 'barfoo', last_sha1='hash', per_page=10)) # then self.assertEqual(actual_ctags, expected_ctags) mock_storage.content_ctags_search.assert_called_with( 'barfoo', last_sha1='hash', limit=10) @patch('swh.web.common.service.storage') @istest def lookup_content_license(self, mock_storage): # given mock_storage.content_fossology_license_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'lang': 'python', }]) expected_license = { 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'lang': 'python', } # when actual_license = service.lookup_content_license( 'sha1:123caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(actual_license, expected_license) mock_storage.content_fossology_license_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_content_license_2(self, mock_storage): # given mock_storage.content_find = MagicMock( return_value={ 'sha1': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f') } ) mock_storage.content_fossology_license_get = MagicMock( return_value=[{ 'id': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'lang': 'haskell', }] ) expected_license = { 'id': '123caf10e9535160d90e874b45aa426de762f19f', 'lang': 'haskell', } # when actual_license = service.lookup_content_license( 'sha1_git:456caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(actual_license, expected_license) mock_storage.content_find( 'sha1_git', hash_to_bytes( '456caf10e9535160d90e874b45aa426de762f19f') ) mock_storage.content_fossology_license_get.assert_called_with( [hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')]) @patch('swh.web.common.service.storage') @istest def lookup_content_provenance(self, mock_storage): # given mock_storage.content_find_provenance = MagicMock( return_value=(p for p in [{ 'content': hash_to_bytes( '123caf10e9535160d90e874b45aa426de762f19f'), 'revision': hash_to_bytes( '456caf10e9535160d90e874b45aa426de762f19f'), 'origin': 100, 'visit': 1, 'path': b'octavio-3.4.0/octave.html/doc_002dS_005fISREG.html' }])) expected_provenances = [{ 'content': '123caf10e9535160d90e874b45aa426de762f19f', 'revision': '456caf10e9535160d90e874b45aa426de762f19f', 'origin': 100, 'visit': 1, 'path': 'octavio-3.4.0/octave.html/doc_002dS_005fISREG.html' }] # when actual_provenances = service.lookup_content_provenance( 'sha1_git:123caf10e9535160d90e874b45aa426de762f19f') # then self.assertEqual(list(actual_provenances), expected_provenances) mock_storage.content_find_provenance.assert_called_with( {'sha1_git': hash_to_bytes('123caf10e9535160d90e874b45aa426de762f19f')}) @patch('swh.web.common.service.storage') @istest def lookup_content_provenance_not_found(self, mock_storage): # given mock_storage.content_find_provenance = MagicMock(return_value=None) # when actual_provenances = service.lookup_content_provenance( 'sha1_git:456caf10e9535160d90e874b45aa426de762f19f') # then self.assertIsNone(actual_provenances) mock_storage.content_find_provenance.assert_called_with( {'sha1_git': hash_to_bytes('456caf10e9535160d90e874b45aa426de762f19f')}) @patch('swh.web.common.service.storage') @istest def stat_counters(self, mock_storage): # given input_stats = { "content": 1770830, "directory": 211683, "directory_entry_dir": 209167, "directory_entry_file": 1807094, "directory_entry_rev": 0, "entity": 0, "entity_history": 0, "occurrence": 0, "occurrence_history": 19600, "origin": 1096, "person": 0, "release": 8584, "revision": 7792, "revision_history": 0, "skipped_content": 0 } mock_storage.stat_counters = MagicMock(return_value=input_stats) # when actual_stats = service.stat_counters() # then expected_stats = input_stats self.assertEqual(actual_stats, expected_stats) mock_storage.stat_counters.assert_called_with() @patch('swh.web.common.service._lookup_origin_visits') @istest def lookup_origin_visits(self, mock_lookup_visits): # given date_origin_visit2 = datetime.datetime( 2013, 7, 1, 20, 0, 0, tzinfo=datetime.timezone.utc) date_origin_visit3 = datetime.datetime( 2015, 1, 1, 21, 0, 0, tzinfo=datetime.timezone.utc) stub_result = [self.origin_visit1, { 'date': date_origin_visit2, 'origin': 1, 'visit': 2, 'target': hash_to_bytes( '65a55bbdf3629f916219feb3dcc7393ded1bc8db'), 'branch': b'master', 'target_type': 'release', 'metadata': None, }, { 'date': date_origin_visit3, 'origin': 1, 'visit': 3 }] mock_lookup_visits.return_value = stub_result # when expected_origin_visits = [{ 'date': self.origin_visit1['date'].isoformat(), 'origin': self.origin_visit1['origin'], 'visit': self.origin_visit1['visit'] }, { 'date': date_origin_visit2.isoformat(), 'origin': 1, 'visit': 2, 'target': '65a55bbdf3629f916219feb3dcc7393ded1bc8db', 'branch': 'master', 'target_type': 'release', 'metadata': {}, }, { 'date': date_origin_visit3.isoformat(), 'origin': 1, 'visit': 3 }] actual_origin_visits = service.lookup_origin_visits(6) # then self.assertEqual(list(actual_origin_visits), expected_origin_visits) mock_lookup_visits.assert_called_once_with( 6, last_visit=None, limit=10) @patch('swh.web.common.service.storage') @istest def lookup_origin_visit(self, mock_storage): # given stub_result = self.origin_visit1 mock_storage.origin_visit_get_by.return_value = stub_result expected_origin_visit = { 'date': self.origin_visit1['date'].isoformat(), 'origin': self.origin_visit1['origin'], 'visit': self.origin_visit1['visit'] } # when actual_origin_visit = service.lookup_origin_visit(1, 1) # then self.assertEqual(actual_origin_visit, expected_origin_visit) mock_storage.origin_visit_get_by.assert_called_once_with(1, 1) @patch('swh.web.common.service.storage') @istest def lookup_origin(self, mock_storage): # given mock_storage.origin_get = MagicMock(return_value={ 'id': 'origin-id', 'lister': 'uuid-lister', 'project': 'uuid-project', 'url': 'ftp://some/url/to/origin', 'type': 'ftp'}) # when actual_origin = service.lookup_origin({'id': 'origin-id'}) # then self.assertEqual(actual_origin, {'id': 'origin-id', 'lister': 'uuid-lister', 'project': 'uuid-project', 'url': 'ftp://some/url/to/origin', 'type': 'ftp'}) mock_storage.origin_get.assert_called_with({'id': 'origin-id'}) @patch('swh.web.common.service.storage') @istest def lookup_release_ko_id_checksum_not_ok_because_not_a_sha1(self, mock_storage): # given mock_storage.release_get = MagicMock() with self.assertRaises(BadInputExc) as cm: # when service.lookup_release('not-a-sha1') self.assertIn('invalid checksum', cm.exception.args[0]) mock_storage.release_get.called = False @patch('swh.web.common.service.storage') @istest def lookup_release_ko_id_checksum_ok_but_not_a_sha1(self, mock_storage): # given mock_storage.release_get = MagicMock() # when with self.assertRaises(BadInputExc) as cm: service.lookup_release( '13c1d34d138ec13b5ebad226dc2528dc7506c956e4646f62d4daf5' '1aea892abe') self.assertIn('sha1_git supported', cm.exception.args[0]) mock_storage.release_get.called = False @patch('swh.web.common.service.storage') @istest def lookup_directory_with_path_not_found(self, mock_storage): # given mock_storage.lookup_directory_with_path = MagicMock(return_value=None) sha1_git = '65a55bbdf3629f916219feb3dcc7393ded1bc8db' # when actual_directory = mock_storage.lookup_directory_with_path( sha1_git, 'some/path/here') self.assertIsNone(actual_directory) @patch('swh.web.common.service.storage') @istest def lookup_directory_with_path_found(self, mock_storage): # given sha1_git = '65a55bbdf3629f916219feb3dcc7393ded1bc8db' entry = {'id': 'dir-id', 'type': 'dir', 'name': 'some/path/foo'} mock_storage.lookup_directory_with_path = MagicMock(return_value=entry) # when actual_directory = mock_storage.lookup_directory_with_path( sha1_git, 'some/path/here') self.assertEqual(entry, actual_directory) @patch('swh.web.common.service.storage') @istest def lookup_release(self, mock_storage): # given mock_storage.release_get = MagicMock(return_value=[{ 'id': hash_to_bytes('65a55bbdf3629f916219feb3dcc7393ded1bc8db'), 'target': None, 'date': { 'timestamp': datetime.datetime( 2015, 1, 1, 22, 0, 0, tzinfo=datetime.timezone.utc).timestamp(), 'offset': 0, 'negative_utc': True, }, 'name': b'v0.0.1', 'message': b'synthetic release', 'synthetic': True, }]) # when actual_release = service.lookup_release( '65a55bbdf3629f916219feb3dcc7393ded1bc8db') # then self.assertEqual(actual_release, { 'id': '65a55bbdf3629f916219feb3dcc7393ded1bc8db', 'target': None, 'date': '2015-01-01T22:00:00-00:00', 'name': 'v0.0.1', 'message': 'synthetic release', 'synthetic': True, }) mock_storage.release_get.assert_called_with( [hash_to_bytes('65a55bbdf3629f916219feb3dcc7393ded1bc8db')]) @istest def lookup_revision_with_context_ko_not_a_sha1_1(self): # given sha1_git = '13c1d34d138ec13b5ebad226dc2528dc7506c956e4646f62d4' \ 'daf51aea892abe' sha1_git_root = '65a55bbdf3629f916219feb3dcc7393ded1bc8db' # when with self.assertRaises(BadInputExc) as cm: service.lookup_revision_with_context(sha1_git_root, sha1_git) self.assertIn('Only sha1_git is supported', cm.exception.args[0]) @istest def lookup_revision_with_context_ko_not_a_sha1_2(self): # given sha1_git_root = '65a55bbdf3629f916219feb3dcc7393ded1bc8db' sha1_git = '13c1d34d138ec13b5ebad226dc2528dc7506c956e4646f6' \ '2d4daf51aea892abe' # when with self.assertRaises(BadInputExc) as cm: service.lookup_revision_with_context(sha1_git_root, sha1_git) self.assertIn('Only sha1_git is supported', cm.exception.args[0]) @patch('swh.web.common.service.storage') @istest def lookup_revision_with_context_ko_sha1_git_does_not_exist( self, mock_storage): # given sha1_git_root = '65a55bbdf3629f916219feb3dcc7393ded1bc8db' sha1_git = '777777bdf3629f916219feb3dcc7393ded1bc8db' sha1_git_bin = hash_to_bytes(sha1_git) mock_storage.revision_get.return_value = None # when with self.assertRaises(NotFoundExc) as cm: service.lookup_revision_with_context(sha1_git_root, sha1_git) self.assertIn('Revision 777777bdf3629f916219feb3dcc7393ded1bc8db' ' not found', cm.exception.args[0]) mock_storage.revision_get.assert_called_once_with( [sha1_git_bin]) @patch('swh.web.common.service.storage') @istest def lookup_revision_with_context_ko_root_sha1_git_does_not_exist( self, mock_storage): # given sha1_git_root = '65a55bbdf3629f916219feb3dcc7393ded1bc8db' sha1_git = '777777bdf3629f916219feb3dcc7393ded1bc8db' sha1_git_root_bin = hash_to_bytes(sha1_git_root) sha1_git_bin = hash_to_bytes(sha1_git) mock_storage.revision_get.side_effect = ['foo', None] # when with self.assertRaises(NotFoundExc) as cm: service.lookup_revision_with_context(sha1_git_root, sha1_git) self.assertIn('Revision 65a55bbdf3629f916219feb3dcc7393ded1bc8db' ' not found', cm.exception.args[0]) mock_storage.revision_get.assert_has_calls([call([sha1_git_bin]), call([sha1_git_root_bin])]) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_revision_with_context(self, mock_query, mock_storage): # given sha1_git_root = '666' sha1_git = '883' sha1_git_root_bin = b'666' sha1_git_bin = b'883' sha1_git_root_dict = { 'id': sha1_git_root_bin, 'parents': [b'999'], } sha1_git_dict = { 'id': sha1_git_bin, 'parents': [], 'directory': b'278', } stub_revisions = [ sha1_git_root_dict, { 'id': b'999', 'parents': [b'777', b'883', b'888'], }, { 'id': b'777', 'parents': [b'883'], }, sha1_git_dict, { 'id': b'888', 'parents': [b'889'], }, { 'id': b'889', 'parents': [], }, ] # inputs ok mock_query.parse_hash_with_algorithms_or_throws.side_effect = [ ('sha1', sha1_git_bin), ('sha1', sha1_git_root_bin) ] # lookup revision first 883, then 666 (both exists) mock_storage.revision_get.return_value = [ sha1_git_dict, sha1_git_root_dict ] mock_storage.revision_log = MagicMock( return_value=stub_revisions) # when actual_revision = service.lookup_revision_with_context( sha1_git_root, sha1_git) # then self.assertEquals(actual_revision, { 'id': hash_to_hex(sha1_git_bin), 'parents': [], 'children': [hash_to_hex(b'999'), hash_to_hex(b'777')], 'directory': hash_to_hex(b'278'), 'merge': False }) mock_query.parse_hash_with_algorithms_or_throws.assert_has_calls( [call(sha1_git, ['sha1'], 'Only sha1_git is supported.'), call(sha1_git_root, ['sha1'], 'Only sha1_git is supported.')]) mock_storage.revision_log.assert_called_with( [sha1_git_root_bin], 100) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_revision_with_context_sha1_git_root_already_retrieved_as_dict( self, mock_query, mock_storage): # given sha1_git = '883' sha1_git_root_bin = b'666' sha1_git_bin = b'883' sha1_git_root_dict = { 'id': sha1_git_root_bin, 'parents': [b'999'], } sha1_git_dict = { 'id': sha1_git_bin, 'parents': [], 'directory': b'278', } stub_revisions = [ sha1_git_root_dict, { 'id': b'999', 'parents': [b'777', b'883', b'888'], }, { 'id': b'777', 'parents': [b'883'], }, sha1_git_dict, { 'id': b'888', 'parents': [b'889'], }, { 'id': b'889', 'parents': [], }, ] # inputs ok mock_query.parse_hash_with_algorithms_or_throws.return_value = ( 'sha1', sha1_git_bin) # lookup only on sha1 mock_storage.revision_get.return_value = [sha1_git_dict] mock_storage.revision_log.return_value = stub_revisions # when actual_revision = service.lookup_revision_with_context( {'id': sha1_git_root_bin}, sha1_git) # then self.assertEquals(actual_revision, { 'id': hash_to_hex(sha1_git_bin), 'parents': [], 'children': [hash_to_hex(b'999'), hash_to_hex(b'777')], 'directory': hash_to_hex(b'278'), 'merge': False }) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with( # noqa sha1_git, ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([sha1_git_bin]) mock_storage.revision_log.assert_called_with( [sha1_git_root_bin], 100) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_ko_revision_not_found(self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') mock_storage.revision_get.return_value = None # when with self.assertRaises(NotFoundExc) as cm: service.lookup_directory_with_revision('123') self.assertIn('Revision 123 not found', cm.exception.args[0]) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_ko_revision_with_path_to_nowhere( self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') dir_id = b'dir-id-as-sha1' mock_storage.revision_get.return_value = [{ 'directory': dir_id, }] mock_storage.directory_entry_get_by_path.return_value = None # when with self.assertRaises(NotFoundExc) as cm: service.lookup_directory_with_revision( '123', 'path/to/something/unknown') self.assertIn("Directory/File 'path/to/something/unknown' " + "pointed to by revision 123 not found", cm.exception.args[0]) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) mock_storage.directory_entry_get_by_path.assert_called_once_with( b'dir-id-as-sha1', [b'path', b'to', b'something', b'unknown']) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_ko_type_not_implemented( self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') dir_id = b'dir-id-as-sha1' mock_storage.revision_get.return_value = [{ 'directory': dir_id, }] mock_storage.directory_entry_get_by_path.return_value = { 'type': 'rev', 'name': b'some/path/to/rev', 'target': b'456' } stub_content = { 'id': b'12', 'type': 'file' } mock_storage.content_get.return_value = stub_content # when with self.assertRaises(NotImplementedError) as cm: service.lookup_directory_with_revision( '123', 'some/path/to/rev') self.assertIn("Entity of type 'rev' not implemented.", cm.exception.args[0]) # then mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) mock_storage.directory_entry_get_by_path.assert_called_once_with( b'dir-id-as-sha1', [b'some', b'path', b'to', b'rev']) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_revision_without_path(self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') dir_id = b'dir-id-as-sha1' mock_storage.revision_get.return_value = [{ 'directory': dir_id, }] stub_dir_entries = [{ 'id': b'123', 'type': 'dir' }, { 'id': b'456', 'type': 'file' }] mock_storage.directory_ls.return_value = stub_dir_entries # when actual_directory_entries = service.lookup_directory_with_revision( '123') self.assertEqual(actual_directory_entries['type'], 'dir') self.assertEqual(list(actual_directory_entries['content']), stub_dir_entries) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) mock_storage.directory_ls.assert_called_once_with(dir_id) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_revision_with_path_to_dir(self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') dir_id = b'dir-id-as-sha1' mock_storage.revision_get.return_value = [{ 'directory': dir_id, }] stub_dir_entries = [{ 'id': b'12', 'type': 'dir' }, { 'id': b'34', 'type': 'file' }] mock_storage.directory_entry_get_by_path.return_value = { 'type': 'dir', 'name': b'some/path', 'target': b'456' } mock_storage.directory_ls.return_value = stub_dir_entries # when actual_directory_entries = service.lookup_directory_with_revision( '123', 'some/path') self.assertEqual(actual_directory_entries['type'], 'dir') self.assertEqual(actual_directory_entries['revision'], '123') self.assertEqual(actual_directory_entries['path'], 'some/path') self.assertEqual(list(actual_directory_entries['content']), stub_dir_entries) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) mock_storage.directory_entry_get_by_path.assert_called_once_with( dir_id, [b'some', b'path']) mock_storage.directory_ls.assert_called_once_with(b'456') @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_revision_with_path_to_file_without_data( self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') dir_id = b'dir-id-as-sha1' mock_storage.revision_get.return_value = [{ 'directory': dir_id, }] mock_storage.directory_entry_get_by_path.return_value = { 'type': 'file', 'name': b'some/path/to/file', 'target': b'789' } stub_content = { 'status': 'visible', } mock_storage.content_find.return_value = stub_content # when actual_content = service.lookup_directory_with_revision( '123', 'some/path/to/file') # then self.assertEqual(actual_content, {'type': 'file', 'revision': '123', 'path': 'some/path/to/file', 'content': stub_content}) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) mock_storage.directory_entry_get_by_path.assert_called_once_with( b'dir-id-as-sha1', [b'some', b'path', b'to', b'file']) mock_storage.content_find.assert_called_once_with({'sha1_git': b'789'}) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_with_revision_revision_with_path_to_file_with_data( self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ('sha1', b'123') dir_id = b'dir-id-as-sha1' mock_storage.revision_get.return_value = [{ 'directory': dir_id, }] mock_storage.directory_entry_get_by_path.return_value = { 'type': 'file', 'name': b'some/path/to/file', 'target': b'789' } stub_content = { 'status': 'visible', 'sha1': b'content-sha1' } mock_storage.content_find.return_value = stub_content mock_storage.content_get.return_value = [{ 'sha1': b'content-sha1', 'data': b'some raw data' }] expected_content = { 'status': 'visible', 'checksums': { 'sha1': hash_to_hex(b'content-sha1'), }, 'data': b'some raw data' } # when actual_content = service.lookup_directory_with_revision( '123', 'some/path/to/file', with_data=True) # then self.assertEqual(actual_content, {'type': 'file', 'revision': '123', 'path': 'some/path/to/file', 'content': expected_content}) mock_query.parse_hash_with_algorithms_or_throws.assert_called_once_with ('123', ['sha1'], 'Only sha1_git is supported.') mock_storage.revision_get.assert_called_once_with([b'123']) mock_storage.directory_entry_get_by_path.assert_called_once_with( b'dir-id-as-sha1', [b'some', b'path', b'to', b'file']) mock_storage.content_find.assert_called_once_with({'sha1_git': b'789'}) mock_storage.content_get.assert_called_once_with([b'content-sha1']) @patch('swh.web.common.service.storage') @istest def lookup_revision(self, mock_storage): # given mock_storage.revision_get = MagicMock( return_value=[self.SAMPLE_REVISION_RAW]) # when actual_revision = service.lookup_revision( self.SHA1_SAMPLE) # then self.assertEqual(actual_revision, self.SAMPLE_REVISION) mock_storage.revision_get.assert_called_with( [self.SHA1_SAMPLE_BIN]) @patch('swh.web.common.service.storage') @istest def lookup_revision_invalid_msg(self, mock_storage): # given stub_rev = self.SAMPLE_REVISION_RAW stub_rev['message'] = b'elegant fix for bug \xff' expected_revision = self.SAMPLE_REVISION expected_revision['message'] = None expected_revision['message_decoding_failed'] = True mock_storage.revision_get = MagicMock(return_value=[stub_rev]) # when actual_revision = service.lookup_revision( self.SHA1_SAMPLE) # then self.assertEqual(actual_revision, expected_revision) mock_storage.revision_get.assert_called_with( [self.SHA1_SAMPLE_BIN]) @patch('swh.web.common.service.storage') @istest def lookup_revision_msg_ok(self, mock_storage): # given mock_storage.revision_get.return_value = [self.SAMPLE_REVISION_RAW] # when rv = service.lookup_revision_message( self.SHA1_SAMPLE) # then self.assertEquals(rv, {'message': self.SAMPLE_MESSAGE_BIN}) mock_storage.revision_get.assert_called_with( [self.SHA1_SAMPLE_BIN]) @patch('swh.web.common.service.storage') @istest def lookup_revision_msg_absent(self, mock_storage): # given stub_revision = self.SAMPLE_REVISION_RAW del stub_revision['message'] mock_storage.revision_get.return_value = stub_revision # when with self.assertRaises(NotFoundExc) as cm: service.lookup_revision_message( self.SHA1_SAMPLE) # then mock_storage.revision_get.assert_called_with( self.SHA1_SAMPLE_BIN) self.assertEqual(cm.exception.args[0], 'No message for revision ' 'with sha1_git ' '18d8be353ed3480476f032475e7c233eff7371d5.') @patch('swh.web.common.service.storage') @istest def lookup_revision_msg_norev(self, mock_storage): # given mock_storage.revision_get.return_value = None # when with self.assertRaises(NotFoundExc) as cm: service.lookup_revision_message( self.SHA1_SAMPLE) # then mock_storage.revision_get.assert_called_with( self.SHA1_SAMPLE_BIN) self.assertEqual(cm.exception.args[0], 'Revision with sha1_git ' '18d8be353ed3480476f032475e7c233eff7371d5 ' 'not found.') @patch('swh.web.common.service.storage') @istest def lookup_revision_multiple(self, mock_storage): # given sha1 = self.SHA1_SAMPLE sha1_other = 'adc83b19e793491b1c6ea0fd8b46cd9f32e592fc' stub_revisions = [ self.SAMPLE_REVISION_RAW, { 'id': hash_to_bytes(sha1_other), 'directory': 'abcdbe353ed3480476f032475e7c233eff7371d5', 'author': { 'name': b'name', 'email': b'name@surname.org', }, 'committer': { 'name': b'name', 'email': b'name@surname.org', }, 'message': b'ugly fix for bug 42', 'date': { 'timestamp': datetime.datetime( 2000, 1, 12, 5, 23, 54, tzinfo=datetime.timezone.utc).timestamp(), 'offset': 0, 'negative_utc': False }, 'date_offset': 0, 'committer_date': { 'timestamp': datetime.datetime( 2000, 1, 12, 5, 23, 54, tzinfo=datetime.timezone.utc).timestamp(), 'offset': 0, 'negative_utc': False }, 'committer_date_offset': 0, 'synthetic': False, 'type': 'git', 'parents': [], 'metadata': [], } ] mock_storage.revision_get.return_value = stub_revisions # when actual_revisions = service.lookup_revision_multiple( [sha1, sha1_other]) # then self.assertEqual(list(actual_revisions), [ self.SAMPLE_REVISION, { 'id': sha1_other, 'directory': 'abcdbe353ed3480476f032475e7c233eff7371d5', 'author': { 'name': 'name', 'email': 'name@surname.org', }, 'committer': { 'name': 'name', 'email': 'name@surname.org', }, 'message': 'ugly fix for bug 42', 'date': '2000-01-12T05:23:54+00:00', 'date_offset': 0, 'committer_date': '2000-01-12T05:23:54+00:00', 'committer_date_offset': 0, 'synthetic': False, 'type': 'git', 'parents': [], 'metadata': {}, 'merge': False } ]) self.assertEqual( list(mock_storage.revision_get.call_args[0][0]), [hash_to_bytes(sha1), hash_to_bytes(sha1_other)]) @patch('swh.web.common.service.storage') @istest def lookup_revision_multiple_none_found(self, mock_storage): # given sha1_bin = self.SHA1_SAMPLE sha1_other = 'adc83b19e793491b1c6ea0fd8b46cd9f32e592fc' mock_storage.revision_get.return_value = [] # then actual_revisions = service.lookup_revision_multiple( [sha1_bin, sha1_other]) self.assertEqual(list(actual_revisions), []) self.assertEqual( list(mock_storage.revision_get.call_args[0][0]), [hash_to_bytes(self.SHA1_SAMPLE), hash_to_bytes(sha1_other)]) @patch('swh.web.common.service.storage') @istest def lookup_revision_log(self, mock_storage): # given stub_revision_log = [self.SAMPLE_REVISION_RAW] mock_storage.revision_log = MagicMock(return_value=stub_revision_log) # when actual_revision = service.lookup_revision_log( 'abcdbe353ed3480476f032475e7c233eff7371d5', limit=25) # then self.assertEqual(list(actual_revision), [self.SAMPLE_REVISION]) mock_storage.revision_log.assert_called_with( [hash_to_bytes('abcdbe353ed3480476f032475e7c233eff7371d5')], 25) @patch('swh.web.common.service.storage') @istest def lookup_revision_log_by(self, mock_storage): # given stub_revision_log = [self.SAMPLE_REVISION_RAW] mock_storage.revision_log_by = MagicMock( return_value=stub_revision_log) # when actual_log = service.lookup_revision_log_by( 1, 'refs/heads/master', None, limit=100) # then self.assertEqual(list(actual_log), [self.SAMPLE_REVISION]) mock_storage.revision_log_by.assert_called_with( 1, 'refs/heads/master', None, limit=100) @patch('swh.web.common.service.storage') @istest def lookup_revision_log_by_nolog(self, mock_storage): # given mock_storage.revision_log_by = MagicMock(return_value=None) # when res = service.lookup_revision_log_by( 1, 'refs/heads/master', None, limit=100) # then self.assertEquals(res, None) mock_storage.revision_log_by.assert_called_with( 1, 'refs/heads/master', None, limit=100) @patch('swh.web.common.service.storage') @istest def lookup_content_raw_not_found(self, mock_storage): # given mock_storage.content_find = MagicMock(return_value=None) # when actual_content = service.lookup_content_raw( 'sha1:' + self.SHA1_SAMPLE) # then self.assertIsNone(actual_content) mock_storage.content_find.assert_called_with( {'sha1': hash_to_bytes(self.SHA1_SAMPLE)}) @patch('swh.web.common.service.storage') @istest def lookup_content_raw(self, mock_storage): # given mock_storage.content_find = MagicMock(return_value={ 'sha1': self.SHA1_SAMPLE, }) mock_storage.content_get = MagicMock(return_value=[{ 'data': b'binary data'}]) # when actual_content = service.lookup_content_raw( 'sha256:%s' % self.SHA256_SAMPLE) # then self.assertEquals(actual_content, {'data': b'binary data'}) mock_storage.content_find.assert_called_once_with( {'sha256': self.SHA256_SAMPLE_BIN}) mock_storage.content_get.assert_called_once_with( [self.SHA1_SAMPLE]) @patch('swh.web.common.service.storage') @istest def lookup_content_not_found(self, mock_storage): # given mock_storage.content_find = MagicMock(return_value=None) # when actual_content = service.lookup_content( 'sha1:%s' % self.SHA1_SAMPLE) # then self.assertIsNone(actual_content) mock_storage.content_find.assert_called_with( {'sha1': self.SHA1_SAMPLE_BIN}) @patch('swh.web.common.service.storage') @istest def lookup_content_with_sha1(self, mock_storage): # given mock_storage.content_find = MagicMock( return_value=self.SAMPLE_CONTENT_RAW) # when actual_content = service.lookup_content( 'sha1:%s' % self.SHA1_SAMPLE) # then self.assertEqual(actual_content, self.SAMPLE_CONTENT) mock_storage.content_find.assert_called_with( {'sha1': hash_to_bytes(self.SHA1_SAMPLE)}) @patch('swh.web.common.service.storage') @istest def lookup_content_with_sha256(self, mock_storage): # given stub_content = self.SAMPLE_CONTENT_RAW stub_content['status'] = 'visible' expected_content = self.SAMPLE_CONTENT expected_content['status'] = 'visible' mock_storage.content_find = MagicMock( return_value=stub_content) # when actual_content = service.lookup_content( 'sha256:%s' % self.SHA256_SAMPLE) # then self.assertEqual(actual_content, expected_content) mock_storage.content_find.assert_called_with( {'sha256': self.SHA256_SAMPLE_BIN}) @patch('swh.web.common.service.storage') @istest def lookup_person(self, mock_storage): # given mock_storage.person_get = MagicMock(return_value=[{ 'id': 'person_id', 'name': b'some_name', 'email': b'some-email', }]) # when actual_person = service.lookup_person('person_id') # then self.assertEqual(actual_person, { 'id': 'person_id', 'name': 'some_name', 'email': 'some-email', }) mock_storage.person_get.assert_called_with(['person_id']) @patch('swh.web.common.service.storage') @istest def lookup_directory_bad_checksum(self, mock_storage): # given mock_storage.directory_ls = MagicMock() # when with self.assertRaises(BadInputExc): service.lookup_directory('directory_id') # then mock_storage.directory_ls.called = False @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory_not_found(self, mock_query, mock_storage): # given mock_query.parse_hash_with_algorithms_or_throws.return_value = ( 'sha1', 'directory-id-bin') mock_storage.directory_get.return_value = None # when - actual_dir = service.lookup_directory('directory_id') + with self.assertRaises(NotFoundExc) as cm: + service.lookup_directory('directory_id') + self.assertIn('Directory with sha1_git directory_id not found', + cm.exception.args[0]) # then - self.assertIsNone(actual_dir) - mock_query.parse_hash_with_algorithms_or_throws.assert_called_with( 'directory_id', ['sha1'], 'Only sha1_git is supported.') mock_storage.directory_get.assert_called_with(['directory-id-bin']) mock_storage.directory_ls.called = False @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_directory(self, mock_query, mock_storage): mock_query.parse_hash_with_algorithms_or_throws.return_value = ( 'sha1', 'directory-sha1-bin') # something that exists is all that matters here mock_storage.directory_get.return_value = {'id': b'directory-sha1-bin'} # given stub_dir_entries = [{ 'sha1': self.SHA1_SAMPLE_BIN, 'sha256': self.SHA256_SAMPLE_BIN, 'sha1_git': self.SHA1GIT_SAMPLE_BIN, 'blake2s256': self.BLAKE2S256_SAMPLE_BIN, 'target': hash_to_bytes( '40e71b8614fcd89ccd17ca2b1d9e66c5b00a6d03'), 'dir_id': self.DIRECTORY_ID_BIN, 'name': b'bob', 'type': 10, }] expected_dir_entries = [{ 'checksums': { 'sha1': self.SHA1_SAMPLE, 'sha256': self.SHA256_SAMPLE, 'sha1_git': self.SHA1GIT_SAMPLE, 'blake2s256': self.BLAKE2S256_SAMPLE }, 'target': '40e71b8614fcd89ccd17ca2b1d9e66c5b00a6d03', 'dir_id': self.DIRECTORY_ID, 'name': 'bob', 'type': 10, }] mock_storage.directory_ls.return_value = stub_dir_entries # when actual_directory_ls = list(service.lookup_directory( 'directory-sha1')) # then self.assertEqual(actual_directory_ls, expected_dir_entries) mock_query.parse_hash_with_algorithms_or_throws.assert_called_with( 'directory-sha1', ['sha1'], 'Only sha1_git is supported.') mock_storage.directory_ls.assert_called_with( 'directory-sha1-bin') @patch('swh.web.common.service.storage') @istest def lookup_revision_by_nothing_found(self, mock_storage): # given mock_storage.revision_get_by.return_value = None # when actual_revisions = service.lookup_revision_by(1) # then self.assertIsNone(actual_revisions) mock_storage.revision_get_by.assert_called_with(1, 'refs/heads/master', limit=1, timestamp=None) @patch('swh.web.common.service.storage') @istest def lookup_revision_by(self, mock_storage): # given stub_rev = self.SAMPLE_REVISION_RAW expected_rev = self.SAMPLE_REVISION mock_storage.revision_get_by.return_value = [stub_rev] # when actual_revision = service.lookup_revision_by(10, 'master2', 'some-ts') # then self.assertEquals(actual_revision, expected_rev) mock_storage.revision_get_by.assert_called_with(10, 'master2', limit=1, timestamp='some-ts') @patch('swh.web.common.service.storage') @istest def lookup_revision_by_nomerge(self, mock_storage): # given stub_rev = self.SAMPLE_REVISION_RAW stub_rev['parents'] = [ hash_to_bytes('adc83b19e793491b1c6ea0fd8b46cd9f32e592fc')] expected_rev = self.SAMPLE_REVISION expected_rev['parents'] = ['adc83b19e793491b1c6ea0fd8b46cd9f32e592fc'] mock_storage.revision_get_by.return_value = [stub_rev] # when actual_revision = service.lookup_revision_by(10, 'master2', 'some-ts') # then self.assertEquals(actual_revision, expected_rev) mock_storage.revision_get_by.assert_called_with(10, 'master2', limit=1, timestamp='some-ts') @patch('swh.web.common.service.storage') @istest def lookup_revision_by_merge(self, mock_storage): # given stub_rev = self.SAMPLE_REVISION_RAW stub_rev['parents'] = [ hash_to_bytes('adc83b19e793491b1c6ea0fd8b46cd9f32e592fc'), hash_to_bytes('ffff3b19e793491b1c6db0fd8b46cd9f32e592fc') ] expected_rev = self.SAMPLE_REVISION expected_rev['parents'] = [ 'adc83b19e793491b1c6ea0fd8b46cd9f32e592fc', 'ffff3b19e793491b1c6db0fd8b46cd9f32e592fc' ] expected_rev['merge'] = True mock_storage.revision_get_by.return_value = [stub_rev] # when actual_revision = service.lookup_revision_by(10, 'master2', 'some-ts') # then self.assertEquals(actual_revision, expected_rev) mock_storage.revision_get_by.assert_called_with(10, 'master2', limit=1, timestamp='some-ts') @patch('swh.web.common.service.storage') @istest def lookup_revision_with_context_by_ko(self, mock_storage): # given mock_storage.revision_get_by.return_value = None # when with self.assertRaises(NotFoundExc) as cm: origin_id = 1 branch_name = 'master3' ts = None service.lookup_revision_with_context_by(origin_id, branch_name, ts, 'sha1') # then self.assertIn( 'Revision with (origin_id: %s, branch_name: %s' ', ts: %s) not found.' % (origin_id, branch_name, ts), cm.exception.args[0]) mock_storage.revision_get_by.assert_called_once_with( origin_id, branch_name, ts) @patch('swh.web.common.service.lookup_revision_with_context') @patch('swh.web.common.service.storage') @istest def lookup_revision_with_context_by(self, mock_storage, mock_lookup_revision_with_context): # given stub_root_rev = {'id': 'root-rev-id'} mock_storage.revision_get_by.return_value = [{'id': 'root-rev-id'}] stub_rev = {'id': 'rev-found'} mock_lookup_revision_with_context.return_value = stub_rev # when origin_id = 1 branch_name = 'master3' ts = None sha1_git = 'sha1' actual_root_rev, actual_rev = service.lookup_revision_with_context_by( origin_id, branch_name, ts, sha1_git) # then self.assertEquals(actual_root_rev, stub_root_rev) self.assertEquals(actual_rev, stub_rev) mock_storage.revision_get_by.assert_called_once_with( origin_id, branch_name, limit=1, timestamp=ts) mock_lookup_revision_with_context.assert_called_once_with( stub_root_rev, sha1_git, 100) @patch('swh.web.common.service.storage') @patch('swh.web.common.service.query') @istest def lookup_entity_by_uuid(self, mock_query, mock_storage): # given uuid_test = 'correct-uuid' mock_query.parse_uuid4.return_value = uuid_test stub_entities = [{'uuid': uuid_test}] mock_storage.entity_get.return_value = stub_entities # when actual_entities = list(service.lookup_entity_by_uuid(uuid_test)) # then self.assertEquals(actual_entities, stub_entities) mock_query.parse_uuid4.assert_called_once_with(uuid_test) mock_storage.entity_get.assert_called_once_with(uuid_test) @istest def lookup_revision_through_ko_not_implemented(self): # then with self.assertRaises(NotImplementedError): service.lookup_revision_through({ 'something-unknown': 10, }) @patch('swh.web.common.service.lookup_revision_with_context_by') @istest def lookup_revision_through_with_context_by(self, mock_lookup): # given stub_rev = {'id': 'rev'} mock_lookup.return_value = stub_rev # when actual_revision = service.lookup_revision_through({ 'origin_id': 1, 'branch_name': 'master', 'ts': None, 'sha1_git': 'sha1-git' }, limit=1000) # then self.assertEquals(actual_revision, stub_rev) mock_lookup.assert_called_once_with( 1, 'master', None, 'sha1-git', 1000) @patch('swh.web.common.service.lookup_revision_by') @istest def lookup_revision_through_with_revision_by(self, mock_lookup): # given stub_rev = {'id': 'rev'} mock_lookup.return_value = stub_rev # when actual_revision = service.lookup_revision_through({ 'origin_id': 2, 'branch_name': 'master2', 'ts': 'some-ts', }, limit=10) # then self.assertEquals(actual_revision, stub_rev) mock_lookup.assert_called_once_with( 2, 'master2', 'some-ts') @patch('swh.web.common.service.lookup_revision_with_context') @istest def lookup_revision_through_with_context(self, mock_lookup): # given stub_rev = {'id': 'rev'} mock_lookup.return_value = stub_rev # when actual_revision = service.lookup_revision_through({ 'sha1_git_root': 'some-sha1-root', 'sha1_git': 'some-sha1', }) # then self.assertEquals(actual_revision, stub_rev) mock_lookup.assert_called_once_with( 'some-sha1-root', 'some-sha1', 100) @patch('swh.web.common.service.lookup_revision') @istest def lookup_revision_through_with_revision(self, mock_lookup): # given stub_rev = {'id': 'rev'} mock_lookup.return_value = stub_rev # when actual_revision = service.lookup_revision_through({ 'sha1_git': 'some-sha1', }) # then self.assertEquals(actual_revision, stub_rev) mock_lookup.assert_called_once_with( 'some-sha1') @patch('swh.web.common.service.lookup_revision_through') @istest def lookup_directory_through_revision_ko_not_found( self, mock_lookup_rev): # given mock_lookup_rev.return_value = None # when with self.assertRaises(NotFoundExc): service.lookup_directory_through_revision( {'id': 'rev'}, 'some/path', 100) mock_lookup_rev.assert_called_once_with({'id': 'rev'}, 100) @patch('swh.web.common.service.lookup_revision_through') @patch('swh.web.common.service.lookup_directory_with_revision') @istest def lookup_directory_through_revision_ok_with_data( self, mock_lookup_dir, mock_lookup_rev): # given mock_lookup_rev.return_value = {'id': 'rev-id'} mock_lookup_dir.return_value = {'type': 'dir', 'content': []} # when rev_id, dir_result = service.lookup_directory_through_revision( {'id': 'rev'}, 'some/path', 100) # then self.assertEquals(rev_id, 'rev-id') self.assertEquals(dir_result, {'type': 'dir', 'content': []}) mock_lookup_rev.assert_called_once_with({'id': 'rev'}, 100) mock_lookup_dir.assert_called_once_with('rev-id', 'some/path', False) @patch('swh.web.common.service.lookup_revision_through') @patch('swh.web.common.service.lookup_directory_with_revision') @istest def lookup_directory_through_revision_ok_with_content( self, mock_lookup_dir, mock_lookup_rev): # given mock_lookup_rev.return_value = {'id': 'rev-id'} stub_result = {'type': 'file', 'revision': 'rev-id', 'content': {'data': b'blah', 'sha1': 'sha1'}} mock_lookup_dir.return_value = stub_result # when rev_id, dir_result = service.lookup_directory_through_revision( {'id': 'rev'}, 'some/path', 10, with_data=True) # then self.assertEquals(rev_id, 'rev-id') self.assertEquals(dir_result, stub_result) mock_lookup_rev.assert_called_once_with({'id': 'rev'}, 10) mock_lookup_dir.assert_called_once_with('rev-id', 'some/path', True)