diff --git a/README.md b/README.md index f5c29d1..b6ee69e 100644 --- a/README.md +++ b/README.md @@ -1,233 +1,237 @@ swh-lister ========== This component from the Software Heritage stack aims to produce listings of software origins and their urls hosted on various public developer platforms or package managers. As these operations are quite similar, it provides a set of Python modules abstracting common software origins listing behaviors. It also provides several lister implementations, contained in the following Python modules: - `swh.lister.bitbucket` - `swh.lister.debian` - `swh.lister.github` - `swh.lister.gitlab` - `swh.lister.gnu` - `swh.lister.pypi` - `swh.lister.npm` - `swh.lister.phabricator` - `swh.lister.cran` - `swh.lister.cgit` Dependencies ------------ All required dependencies can be found in the `requirements*.txt` files located at the root of the repository. Local deployment ---------------- ## lister configuration Each lister implemented so far by Software Heritage (`github`, `gitlab`, `debian`, `pypi`, `npm`) must be configured by following the instructions below (please note that you have to replace `` by one of the lister name introduced above). ### Preparation steps 1. `mkdir ~/.config/swh/ ~/.cache/swh/lister//` 2. create configuration file `~/.config/swh/lister_.yml` 3. Bootstrap the db instance schema ```lang=bash $ createdb lister- $ python3 -m swh.lister.cli --db-url postgres:///lister- ``` Note: This bootstraps a minimum data set needed for the lister to run. ### Configuration file sample Minimalistic configuration shared by all listers to add in file `~/.config/swh/lister_.yml`: ```lang=yml storage: cls: 'remote' args: url: 'http://localhost:5002/' scheduler: cls: 'remote' args: url: 'http://localhost:5008/' lister: cls: 'local' args: # see http://docs.sqlalchemy.org/en/latest/core/engines.html#database-urls db: 'postgresql:///lister-' credentials: [] cache_responses: True cache_dir: /home/user/.cache/swh/lister// ``` Note: This expects storage (5002) and scheduler (5008) services to run locally ## lister-github Once configured, you can execute a GitHub lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.github.tasks import range_github_lister logging.basicConfig(level=logging.DEBUG) range_github_lister(364, 365) ... ``` ## lister-gitlab Once configured, you can execute a GitLab lister using the instructions detailed in the `python3` scripts below: ```lang=python import logging from swh.lister.gitlab.tasks import range_gitlab_lister logging.basicConfig(level=logging.DEBUG) range_gitlab_lister(1, 2, { 'instance': 'debian', 'api_baseurl': 'https://salsa.debian.org/api/v4', 'sort': 'asc', 'per_page': 20 }) ``` ```lang=python import logging from swh.lister.gitlab.tasks import full_gitlab_relister logging.basicConfig(level=logging.DEBUG) full_gitlab_relister({ 'instance': '0xacab', 'api_baseurl': 'https://0xacab.org/api/v4', 'sort': 'asc', 'per_page': 20 }) ``` ```lang=python import logging from swh.lister.gitlab.tasks import incremental_gitlab_lister logging.basicConfig(level=logging.DEBUG) incremental_gitlab_lister({ 'instance': 'freedesktop.org', 'api_baseurl': 'https://gitlab.freedesktop.org/api/v4', 'sort': 'asc', 'per_page': 20 }) ``` ## lister-debian Once configured, you can execute a Debian lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.debian.tasks import debian_lister logging.basicConfig(level=logging.DEBUG) debian_lister('Debian') ``` ## lister-pypi Once configured, you can execute a PyPI lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.pypi.tasks import pypi_lister logging.basicConfig(level=logging.DEBUG) pypi_lister() ``` ## lister-npm Once configured, you can execute a npm lister using the following instructions in a `python3` REPL: ```lang=python import logging from swh.lister.npm.tasks import npm_lister logging.basicConfig(level=logging.DEBUG) npm_lister() ``` ## lister-phabricator Once configured, you can execute a Phabricator lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.phabricator.tasks import incremental_phabricator_lister logging.basicConfig(level=logging.DEBUG) incremental_phabricator_lister(forge_url='https://forge.softwareheritage.org', api_token='XXXX') ``` ## lister-gnu Once configured, you can execute a PyPI lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.gnu.tasks import gnu_lister logging.basicConfig(level=logging.DEBUG) gnu_lister() ``` ## lister-cran Once configured, you can execute a CRAN lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.cran.tasks import cran_lister logging.basicConfig(level=logging.DEBUG) cran_lister() ``` ## lister-cgit Once configured, you can execute a cgit lister using the following instructions in a `python3` script: ```lang=python import logging from swh.lister.cgit.tasks import cgit_lister logging.basicConfig(level=logging.DEBUG) -cgit_lister(base_url='http://git.savannah.gnu.org/cgit/') +# simple cgit instance +cgit_lister(url='https://git.kernel.org/') +# cgit instance whose listed repositories differ from the base url +cgit_lister(url='https://cgit.kde.org/', + url_prefix='https://anongit.kde.org/') ``` Licensing --------- This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. See top-level LICENSE file for the full text of the GNU General Public License along with this program. diff --git a/swh/lister/cgit/lister.py b/swh/lister/cgit/lister.py index 4f5db9b..3897adb 100644 --- a/swh/lister/cgit/lister.py +++ b/swh/lister/cgit/lister.py @@ -1,276 +1,251 @@ # Copyright (C) 2019 the Software Heritage developers # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import random +import logging from bs4 import BeautifulSoup -from collections import defaultdict import requests from urllib.parse import urlparse from .models import CGitModel from swh.lister.core.simple_lister import SimpleLister from swh.lister.core.lister_transports import ListerOnePageApiTransport class CGitLister(ListerOnePageApiTransport, SimpleLister): MODEL = CGitModel LISTER_NAME = 'cgit' PAGE = None + url_prefix_present = True - def __init__(self, base_url, instance=None, override_config=None): + def __init__(self, url, instance=None, url_prefix=None, + override_config=None): + """Inits Class with PAGE url and origin url prefix. - self.PAGE = base_url + Args: + url (str): URL of the CGit instance. + instance (str): Name of cgit instance. + url_prefix (str): Prefix of the origin_url. Origin link of the + repos of some special instances do not match + the url of the repository page, they have origin + url in the format /. + + """ + self.PAGE = url + if url_prefix is None: + self.url_prefix = url + self.url_prefix_present = False + else: + self.url_prefix = url_prefix + + if not self.url_prefix.endswith('/'): + self.url_prefix += '/' url = urlparse(self.PAGE) self.url_netloc = find_netloc(url) if not instance: instance = url.hostname self.instance = instance + ListerOnePageApiTransport .__init__(self) SimpleLister.__init__(self, override_config=override_config) def list_packages(self, response): """List the actual cgit instance origins from the response. - Find the repos in all the pages by parsing over the HTML of - the `base_url`. Find the details for all the repos and return - them in the format of list of dictionaries. + Find repositories metadata by parsing the html page (response's raw + content). If there are links in the html page, retrieve those + repositories metadata from those pages as well. Return the + repositories as list of dictionaries. + + Args: + response (Response): http api request response. + + Returns: + List of repository origin urls (as dict) included in the response. """ repos_details = [] - repos = get_repo_list(response) - soup = make_repo_soup(response) - pages = self.get_page(soup) - if len(pages) > 1: - repos.extend(self.get_all_pages(pages)) - for repo in repos: + for repo in self.yield_repo_from_responses(response): repo_name = repo.a.text - repo_url = self.get_url(repo) - origin_url = find_origin_url(repo_url) + origin_url = self.find_origin_url(repo, repo_name) try: time = repo.span['title'] except Exception: time = None if origin_url is not None: repos_details.append({ - 'name': repo_name, - 'time': time, - 'origin_url': origin_url, - }) + 'name': repo_name, + 'time': time, + 'origin_url': origin_url, + }) random.shuffle(repos_details) return repos_details - def get_page(self, soup): - """Find URL of all pages + def yield_repo_from_responses(self, response): + """Yield repositories from all pages of the cgit instance. + + Finds the number of pages present and yields the list of + repositories present. + + Args: + response (Response): server response. + + Yields: + List of beautifulsoup object of repository rows. + + """ + html = response.text + yield from get_repo_list(html) + pages = self.get_pages(make_soup(html)) + if len(pages) > 1: + yield from self.get_repos_from_pages(pages[1:]) - Finds URL of all the pages that are present by parsing over the HTML of + def find_origin_url(self, repo, repo_name): + """Finds the origin url for a repository + + Args: + repo (Beautifulsoup): Beautifulsoup object of the repository + row present in base url. + repo_name (str): Repository name. + + Returns: + string: origin url. + + """ + if self.url_prefix_present: + return self.url_prefix + repo_name + + return self.get_url(repo) + + def get_pages(self, url_soup): + """Find URL of all pages. + + Finds URL of pages that are present by parsing over the HTML of pagination present at the end of the page. Args: - soup (Beautifulsoup): a beautifulsoup object of base URL + url_soup (Beautifulsoup): a beautifulsoup object of base URL Returns: - list: URL of all the pages present for a cgit instance + list: URL of pages present for a cgit instance """ - pages = soup.find('div', {"class": "content"}).find_all('li') + pages = url_soup.find('div', {"class": "content"}).find_all('li') if not pages: return [self.PAGE] return [self.get_url(page) for page in pages] - def get_all_pages(self, pages): - """Find repos from all the pages + def get_repos_from_pages(self, pages): + """Find repos from all pages. - Make the request for all the pages (except the first) present for a - particular cgit instance and finds the repos that are available - for each and every page. + Request the available repos from the pages. This yields + the available repositories found as beautiful object representation. Args: - pages ([str]): list of urls of all the pages present for a - particular cgit instance + pages ([str]): list of urls of all pages present for a + particular cgit instance. - Returns: - List of beautifulsoup object of all the repositories (url) row - present in all the pages(except first). + Yields: + List of beautifulsoup object of repository (url) rows + present in pages(except first). """ - all_repos = [] - for page in pages[1:]: + for page in pages: response = requests.get(page) - repos = get_repo_list(response) - all_repos.extend(repos) + if not response.ok: + logging.warning('Failed to retrieve repositories from page %s', + page) + continue - return all_repos + yield from get_repo_list(response.text) def get_url(self, repo): """Finds url of a repo page. Finds the url of a repo page by parsing over the html of the row of that repo present in the base url. Args: repo (Beautifulsoup): a beautifulsoup object of the repository row present in base url. Returns: string: The url of a repo. """ suffix = repo.a['href'] return self.url_netloc + suffix def get_model_from_repo(self, repo): """Transform from repository representation to model. """ return { 'uid': self.PAGE + repo['name'], 'name': repo['name'], 'full_name': repo['name'], 'html_url': repo['origin_url'], 'origin_url': repo['origin_url'], 'origin_type': 'git', 'time_updated': repo['time'], 'instance': self.instance, } def transport_response_simplified(self, repos_details): """Transform response to list for model manipulation. """ return [self.get_model_from_repo(repo) for repo in repos_details] def find_netloc(url): - """Finds the network location from then base_url + """Finds the network location from then url. - All the url in the repo are relative to the network location part of base - url, so we need to compute it to reconstruct all the urls. + URL in the repo are relative to the network location part of base + URL, so we need to compute it to reconstruct URLs. Args: - url (urllib): urllib object of base_url + url (urllib): urllib object of url. Returns: string: Scheme and Network location part in the base URL. Example: - For base_url = https://git.kernel.org/pub/scm/ + For url = https://git.kernel.org/pub/scm/ >>> find_netloc(url) 'https://git.kernel.org' """ return '%s://%s' % (url.scheme, url.netloc) def get_repo_list(response): - """Find all the rows with repo for a particualar page on the base url - - Finds all the repos on page and retuens a list of all the repos. Each - element of the list is a beautifulsoup object representing a repo. + """Find repositories (as beautifulsoup object) available within the server + response. Args: response (Response): server response Returns: - List of all the repos on a page. + List all repositories as beautifulsoup object within the response. """ - repo_soup = make_repo_soup(response) + repo_soup = make_soup(response) return repo_soup \ .find('div', {"class": "content"}).find_all("tr", {"class": ""}) -def make_repo_soup(response): - """Makes BeautifulSoup object of the response - - """ - return BeautifulSoup(response.text, features="html.parser") - - -def find_origin_url(repo_url): - """Finds origin url for a repo. - - Finds the origin url for a particular repo by parsing over the page of - that repo. - - Args: - repo_url: URL of the repo. - - Returns: - string: Origin url for the repo. - - Examples: - - >>> find_origin_url( - 'http://git.savannah.gnu.org/cgit/fbvbconv-py.git/') - 'https://git.savannah.gnu.org/git/fbvbconv-py.git' - - """ - - response = requests.get(repo_url) - repo_soup = make_repo_soup(response) - - origin_urls = find_all_origin_url(repo_soup) - return priority_origin_url(origin_urls) - - -def find_all_origin_url(soup): - """Finds all possible origin url for a repo. - - Finds all the origin url for a particular repo by parsing over the html of - repo page. - - Args: - soup: a beautifulsoup object repo representation. - - Returns: - dictionary: All possible origin urls for a repository (dict with - key 'protocol', value the associated url). - - Examples: - If soup is beautifulsoup object of the html code at - http://git.savannah.gnu.org/cgit/fbvbconv-py.git/ - - >>> print(find_all_origin_url(soup)) - { 'https': 'https://git.savannah.gnu.org/git/fbvbconv-py.git', - 'ssh': 'ssh://git.savannah.gnu.org/srv/git/fbvbconv-py.git', - 'git': 'git://git.savannah.gnu.org/fbvbconv-py.git'} - """ - origin_urls = defaultdict(dict) - found_clone_word = False - - for i in soup.find_all('tr'): - if found_clone_word: - link = i.text - protocol = link[:link.find(':')] - origin_urls[protocol] = link - if i.text == 'Clone': - found_clone_word = True - - return origin_urls - - -def priority_origin_url(origin_url): - """Finds the highest priority link for a particular repo. - - Priority order is https>http>git>ssh. - - Args: - origin_urls (Dict): All possible origin urls for a repository - (key 'protocol', value the associated url) - - Returns: - Url (str) with the highest priority. +def make_soup(response): + """Instantiates a beautiful soup object from the response object. """ - for protocol in ['https', 'http', 'git', 'ssh']: - if protocol in origin_url: - return origin_url[protocol] + return BeautifulSoup(response, features="html.parser") diff --git a/swh/lister/cgit/tasks.py b/swh/lister/cgit/tasks.py index af1ab63..31148dd 100644 --- a/swh/lister/cgit/tasks.py +++ b/swh/lister/cgit/tasks.py @@ -1,23 +1,25 @@ # Copyright (C) 2019 the Software Heritage developers # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information from swh.scheduler.celery_backend.config import app from .lister import CGitLister -def new_lister(base_url='https://git.savannah.gnu.org/cgit/', - instance='savannah-gnu', **kw): - return CGitLister(base_url=base_url, instance=instance, **kw) +def new_lister(url='https://git.kernel.org/', + url_prefix=None, + instance='kernal', **kw): + return CGitLister(url=url, instance=instance, url_prefix=url_prefix, + **kw) @app.task(name=__name__ + '.CGitListerTask') def cgit_lister(**lister_args): lister = new_lister(**lister_args) lister.run() @app.task(name=__name__ + '.ping') def ping(): return 'OK' diff --git a/swh/lister/cgit/tests/api_response.html b/swh/lister/cgit/tests/api_response.html deleted file mode 100644 index 1d34a4c..0000000 --- a/swh/lister/cgit/tests/api_response.html +++ /dev/null @@ -1,47 +0,0 @@ - - - -fbvbconv-py.git - Unnamed repository; edit this file 'description' to name the repository. - - - - - - - - - - -
- - - - -
-summaryrefslogtreecommitdiff
- - - -
-
-
- - - - - - - -
BranchCommit messageAuthorAge
masterinitial importJohannes Stezenbach2 years
 
 
AgeCommit messageAuthorFilesLines
2017-06-02initial importHEADmasterJohannes Stezenbach3-0/+889
 
Clone
git://git.savannah.gnu.org/fbvbconv-py.git
https://git.savannah.gnu.org/git/fbvbconv-py.git
ssh://git.savannah.gnu.org/srv/git/fbvbconv-py.git
- -
- - diff --git a/swh/lister/cgit/tests/repo_list.txt b/swh/lister/cgit/tests/repo_list.txt new file mode 100644 index 0000000..ea15c96 --- /dev/null +++ b/swh/lister/cgit/tests/repo_list.txt @@ -0,0 +1,15 @@ +openembedded-coreOpenEmbedded Core layerOpenEmbedded5 hourssummarylogtree +openembedded-core-contribOpenEmbedded Core user contribution treesOpenEmbedded5 hourssummarylogtree +meta-openembeddedCollection of OpenEmbedded layersOpenEmbedded21 hourssummarylogtree +meta-openembedded-contribOpenEmbedded layers collection contribution treesOpenEmbedded21 hourssummarylogtree +bitbakeBitbake Development treeOpenEmbedded7 dayssummarylogtree +bitbake-contribBitbake user contribution treesOpenEmbedded8 dayssummarylogtree +meta-handheldHandheld device meta layerOpenEmbedded9 monthssummarylogtree +meta-opieOPIE meta layerOpenEmbedded3 yearssummarylogtree +openembeddedClassic OpenEmbedded Development TreeOpenEmbedded4 yearssummarylogtree +openembedded-web-frontpagesOpenEmbedded Website Source CodeOpenEmbedded5 yearssummarylogtree +openembedded-adminOE Admin toolsOpenEmbedded6 yearssummarylogtree +meta-microMicro distribution meta layerOpenEmbedded7 yearssummarylogtree +eclipsetoolsEclipse tools for OpenEmbeddedOpenEmbedded8 yearssummarylogtree +oetestTest utilities for OpenEmbeddedOpenEmbedded10 yearssummarylogtree +oebuildstatsOE Build StatsOpenEmbeddedsummarylogtree diff --git a/swh/lister/cgit/tests/response.html b/swh/lister/cgit/tests/response.html new file mode 100644 index 0000000..cd95ccb --- /dev/null +++ b/swh/lister/cgit/tests/response.html @@ -0,0 +1,41 @@ + + + +OpenEmbedded Git Repository Browser + + + + + + + + + diff --git a/swh/lister/cgit/tests/test_lister.py b/swh/lister/cgit/tests/test_lister.py index e3c3610..049893e 100644 --- a/swh/lister/cgit/tests/test_lister.py +++ b/swh/lister/cgit/tests/test_lister.py @@ -1,50 +1,27 @@ # Copyright (C) 2019 the Software Heritage developers # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information -from bs4 import BeautifulSoup from urllib.parse import urlparse -from swh.lister.cgit.lister import priority_origin_url, find_all_origin_url -from swh.lister.cgit.lister import find_netloc +from swh.lister.cgit.lister import find_netloc, get_repo_list -def test_find_all_origin_url(): - f = open('swh/lister/cgit/tests/api_response.html') - soup = BeautifulSoup(f.read(), features="html.parser") - expected_output = {'https': 'https://git.savannah.gnu.org/git/' - 'fbvbconv-py.git', - 'ssh': 'ssh://git.savannah.gnu.org/srv/git/' - 'fbvbconv-py.git', - 'git': 'git://git.savannah.gnu.org/fbvbconv-py.git'} - - output = find_all_origin_url(soup) - - for protocol, url in expected_output.items(): - assert url == output[protocol] - - -def test_priority_origin_url(): - first_input = {'https': 'https://kernel.googlesource.com/pub/scm/docs/' - 'man-pages/man-pages.git', - 'git': 'git://git.kernel.org/pub/scm/docs/man-pages/' - 'man-pages.git'} - second_input = {'git': 'git://git.savannah.gnu.org/perl-pesel.git', - 'ssh': 'ssh://git.savannah.gnu.org/srv/git/perl-pesel.git'} - third_input = {} - - assert (priority_origin_url(first_input) == - 'https://kernel.googlesource.com/pub/scm/docs/man-pages/' - 'man-pages.git') - assert (priority_origin_url(second_input) == - 'git://git.savannah.gnu.org/perl-pesel.git') - assert priority_origin_url(third_input) is None +def test_get_repo_list(): + f = open('swh/lister/cgit/tests/response.html') + repos = get_repo_list(f.read()) + f = open('swh/lister/cgit/tests/repo_list.txt') + expected_repos = f.readlines() + expected_repos = list(map((lambda repo: repo[:-1]), expected_repos)) + assert len(repos) == len(expected_repos) + for i in range(len(repos)): + assert str(repos[i]) == expected_repos[i] def test_find_netloc(): first_url = urlparse('http://git.savannah.gnu.org/cgit/') second_url = urlparse('https://cgit.kde.org/') assert find_netloc(first_url) == 'http://git.savannah.gnu.org' assert find_netloc(second_url) == 'https://cgit.kde.org' diff --git a/swh/lister/cgit/tests/test_tasks.py b/swh/lister/cgit/tests/test_tasks.py index b8faabc..4a36a05 100644 --- a/swh/lister/cgit/tests/test_tasks.py +++ b/swh/lister/cgit/tests/test_tasks.py @@ -1,29 +1,53 @@ from unittest.mock import patch def test_ping(swh_app, celery_session_worker): res = swh_app.send_task( 'swh.lister.cgit.tasks.ping') assert res res.wait() assert res.successful() assert res.result == 'OK' @patch('swh.lister.cgit.tasks.CGitLister') -def test_lister(lister, swh_app, celery_session_worker): +def test_lister_no_url_prefix(lister, swh_app, celery_session_worker): # setup the mocked CGitLister lister.return_value = lister lister.run.return_value = None res = swh_app.send_task( - 'swh.lister.cgit.tasks.CGitListerTask') + 'swh.lister.cgit.tasks.CGitListerTask', + kwargs=dict(url='https://git.kernel.org/', instance='kernel')) assert res res.wait() assert res.successful() lister.assert_called_once_with( - base_url='https://git.savannah.gnu.org/cgit/', - instance='savannah-gnu') + url='https://git.kernel.org/', + url_prefix=None, + instance='kernel') + lister.db_last_index.assert_not_called() + lister.run.assert_called_once_with() + + +@patch('swh.lister.cgit.tasks.CGitLister') +def test_lister_with_url_prefix(lister, swh_app, celery_session_worker): + # setup the mocked CGitLister + lister.return_value = lister + lister.run.return_value = None + + res = swh_app.send_task( + 'swh.lister.cgit.tasks.CGitListerTask', + kwargs=dict(url='https://cgit.kde.org/', + url_prefix='https://anongit.kde.org/', instance='kde')) + assert res + res.wait() + assert res.successful() + + lister.assert_called_once_with( + url='https://cgit.kde.org/', + url_prefix='https://anongit.kde.org/', + instance='kde') lister.db_last_index.assert_not_called() lister.run.assert_called_once_with() diff --git a/swh/lister/cli.py b/swh/lister/cli.py index bf5c439..3a6f38f 100644 --- a/swh/lister/cli.py +++ b/swh/lister/cli.py @@ -1,157 +1,158 @@ # Copyright (C) 2018 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import logging import click from swh.core.cli import CONTEXT_SETTINGS logger = logging.getLogger(__name__) SUPPORTED_LISTERS = ['github', 'gitlab', 'bitbucket', 'debian', 'pypi', 'npm', 'phabricator', 'gnu', 'cran', 'cgit'] @click.group(name='lister', context_settings=CONTEXT_SETTINGS) @click.pass_context def lister(ctx): '''Software Heritage Lister tools.''' pass @lister.command(name='db-init', context_settings=CONTEXT_SETTINGS) @click.option( '--db-url', '-d', default='postgres:///lister-gitlab.com', help='SQLAlchemy DB URL; see ' '') # noqa @click.argument('listers', required=1, nargs=-1, type=click.Choice(SUPPORTED_LISTERS + ['all'])) @click.option('--drop-tables', '-D', is_flag=True, default=False, help='Drop tables before creating the database schema') @click.pass_context def cli(ctx, db_url, listers, drop_tables): """Initialize the database model for given listers. """ override_conf = { 'lister': { 'cls': 'local', 'args': {'db': db_url} } } if 'all' in listers: listers = SUPPORTED_LISTERS for lister in listers: logger.info('Initializing lister %s', lister) insert_minimum_data = None if lister == 'github': from .github.models import IndexingModelBase as ModelBase from .github.lister import GitHubLister _lister = GitHubLister( api_baseurl='https://api.github.com', override_config=override_conf) elif lister == 'bitbucket': from .bitbucket.models import IndexingModelBase as ModelBase from .bitbucket.lister import BitBucketLister _lister = BitBucketLister( api_baseurl='https://api.bitbucket.org/2.0', override_config=override_conf) elif lister == 'gitlab': from .gitlab.models import ModelBase from .gitlab.lister import GitLabLister _lister = GitLabLister( api_baseurl='https://gitlab.com/api/v4/', override_config=override_conf) elif lister == 'debian': from .debian.lister import DebianLister ModelBase = DebianLister.MODEL # noqa _lister = DebianLister(override_config=override_conf) def insert_minimum_data(lister): from swh.storage.schemata.distribution import ( Distribution, Area) d = Distribution( name='Debian', type='deb', mirror_uri='http://deb.debian.org/debian/') lister.db_session.add(d) areas = [] for distribution_name in ['stretch']: for area_name in ['main', 'contrib', 'non-free']: areas.append(Area( name='%s/%s' % (distribution_name, area_name), distribution=d, )) lister.db_session.add_all(areas) lister.db_session.commit() elif lister == 'pypi': from .pypi.models import ModelBase from .pypi.lister import PyPILister _lister = PyPILister(override_config=override_conf) elif lister == 'npm': from .npm.models import IndexingModelBase as ModelBase from .npm.models import NpmVisitModel from .npm.lister import NpmLister _lister = NpmLister(override_config=override_conf) if drop_tables: NpmVisitModel.metadata.drop_all(_lister.db_engine) NpmVisitModel.metadata.create_all(_lister.db_engine) elif lister == 'phabricator': from .phabricator.models import IndexingModelBase as ModelBase from .phabricator.lister import PhabricatorLister _lister = PhabricatorLister( forge_url='https://forge.softwareheritage.org', api_token='', override_config=override_conf) elif lister == 'gnu': from .gnu.models import ModelBase from .gnu.lister import GNULister _lister = GNULister(override_config=override_conf) elif lister == 'cran': from .cran.models import ModelBase from .cran.lister import CRANLister _lister = CRANLister(override_config=override_conf) elif lister == 'cgit': from .cgit.models import ModelBase from .cgit.lister import CGitLister _lister = CGitLister( - base_url='http://git.savannah.gnu.org/cgit/', + url='http://git.savannah.gnu.org/cgit/', + url_prefix='http://git.savannah.gnu.org/git/', override_config=override_conf) else: raise ValueError( 'Invalid lister %s: only supported listers are %s' % (lister, SUPPORTED_LISTERS)) if drop_tables: logger.info('Dropping tables for %s', lister) ModelBase.metadata.drop_all(_lister.db_engine) logger.info('Creating tables for %s', lister) ModelBase.metadata.create_all(_lister.db_engine) if insert_minimum_data: logger.info('Inserting minimal data for %s', lister) try: insert_minimum_data(_lister) except Exception: logger.warning( 'Failed to insert minimum data in %s', lister) if __name__ == '__main__': cli()