Page MenuHomeSoftware Heritage

loader.py
No OneTemporary

loader.py

# Copyright (C) 2015 The Software Heritage developers
# See the AUTHORS file at the top-level directory of this distribution
# License: GNU General Public License version 3, or any later version
# See top-level LICENSE file for more information
import logging
import traceback
import uuid
import psycopg2
import pygit2
from pygit2 import Oid, GIT_OBJ_BLOB, GIT_OBJ_TREE, GIT_OBJ_COMMIT, GIT_OBJ_TAG
import requests
from retrying import retry
from swh.core import config
from . import converters
from .utils import get_objects_per_object_type
def send_in_packets(source_list, formatter, sender, packet_size,
packet_size_bytes=None, *args, **kwargs):
"""Send objects from `source_list`, passed through `formatter` (with
extra args *args, **kwargs), using the `sender`, in packets of
`packet_size` objects (and of max `packet_size_bytes`).
"""
formatted_objects = []
count = 0
if not packet_size_bytes:
packet_size_bytes = 0
for obj in source_list:
formatted_object = formatter(obj, *args, **kwargs)
if formatted_object:
formatted_objects.append(formatted_object)
else:
continue
if packet_size_bytes:
count += formatted_object['length']
if len(formatted_objects) >= packet_size or count > packet_size_bytes:
sender(formatted_objects)
formatted_objects = []
count = 0
if formatted_objects:
sender(formatted_objects)
def retry_loading(error):
"""Retry policy when we catch a recoverable error"""
exception_classes = [
# raised when two parallel insertions insert the same data.
psycopg2.IntegrityError,
# raised when uWSGI restarts and hungs up on the worker.
requests.exceptions.ConnectionError,
]
if not any(isinstance(error, exc) for exc in exception_classes):
return False
logger = logging.getLogger('swh.loader.git.BulkLoader')
error_name = error.__module__ + '.' + error.__class__.__name__
logger.warning('Retry loading a batch', exc_info=False, extra={
'swh_type': 'storage_retry',
'swh_exception_type': error_name,
'swh_exception': traceback.format_exception(
error.__class__,
error,
error.__traceback__,
),
})
return True
class BulkLoader(config.SWHConfig):
"""A bulk loader for a git repository"""
DEFAULT_CONFIG = {
'storage_class': ('str', 'remote_storage'),
'storage_args': ('list[str]', ['http://localhost:5000/']),
'send_contents': ('bool', True),
'send_directories': ('bool', True),
'send_revisions': ('bool', True),
'send_releases': ('bool', True),
'send_occurrences': ('bool', True),
'content_packet_size': ('int', 10000),
'content_packet_size_bytes': ('int', 1024 * 1024 * 1024),
'directory_packet_size': ('int', 25000),
'revision_packet_size': ('int', 100000),
'release_packet_size': ('int', 100000),
'occurrence_packet_size': ('int', 100000),
}
def __init__(self, config):
self.config = config
if self.config['storage_class'] == 'remote_storage':
from swh.storage.api.client import RemoteStorage as Storage
else:
from swh.storage import Storage
self.storage = Storage(*self.config['storage_args'])
self.log = logging.getLogger('swh.loader.git.BulkLoader')
@retry(retry_on_exception=retry_loading, stop_max_attempt_number=3)
def send_contents(self, content_list):
"""Actually send properly formatted contents to the database"""
num_contents = len(content_list)
log_id = str(uuid.uuid4())
self.log.debug("Sending %d contents" % num_contents,
extra={
'swh_type': 'storage_send_start',
'swh_content_type': 'content',
'swh_num': num_contents,
'swh_id': log_id,
})
self.storage.content_add(content_list)
self.log.debug("Done sending %d contents" % num_contents,
extra={
'swh_type': 'storage_send_end',
'swh_content_type': 'content',
'swh_num': num_contents,
'swh_id': log_id,
})
@retry(retry_on_exception=retry_loading, stop_max_attempt_number=3)
def send_directories(self, directory_list):
"""Actually send properly formatted directories to the database"""
num_directories = len(directory_list)
log_id = str(uuid.uuid4())
self.log.debug("Sending %d directories" % num_directories,
extra={
'swh_type': 'storage_send_start',
'swh_content_type': 'directory',
'swh_num': num_directories,
'swh_id': log_id,
})
self.storage.directory_add(directory_list)
self.log.debug("Done sending %d directories" % num_directories,
extra={
'swh_type': 'storage_send_end',
'swh_content_type': 'directory',
'swh_num': num_directories,
'swh_id': log_id,
})
@retry(retry_on_exception=retry_loading, stop_max_attempt_number=3)
def send_revisions(self, revision_list):
"""Actually send properly formatted revisions to the database"""
num_revisions = len(revision_list)
log_id = str(uuid.uuid4())
self.log.debug("Sending %d revisions" % num_revisions,
extra={
'swh_type': 'storage_send_start',
'swh_content_type': 'revision',
'swh_num': num_revisions,
'swh_id': log_id,
})
self.storage.revision_add(revision_list)
self.log.debug("Done sending %d revisions" % num_revisions,
extra={
'swh_type': 'storage_send_end',
'swh_content_type': 'revision',
'swh_num': num_revisions,
'swh_id': log_id,
})
@retry(retry_on_exception=retry_loading, stop_max_attempt_number=3)
def send_releases(self, release_list):
"""Actually send properly formatted releases to the database"""
num_releases = len(release_list)
log_id = str(uuid.uuid4())
self.log.debug("Sending %d releases" % num_releases,
extra={
'swh_type': 'storage_send_start',
'swh_content_type': 'release',
'swh_num': num_releases,
'swh_id': log_id,
})
self.storage.release_add(release_list)
self.log.debug("Done sending %d releases" % num_releases,
extra={
'swh_type': 'storage_send_end',
'swh_content_type': 'release',
'swh_num': num_releases,
'swh_id': log_id,
})
@retry(retry_on_exception=retry_loading, stop_max_attempt_number=3)
def send_occurrences(self, occurrence_list):
"""Actually send properly formatted occurrences to the database"""
num_occurrences = len(occurrence_list)
log_id = str(uuid.uuid4())
self.log.debug("Sending %d occurrences" % num_occurrences,
extra={
'swh_type': 'storage_send_start',
'swh_content_type': 'occurrence',
'swh_num': num_occurrences,
'swh_id': log_id,
})
self.storage.occurrence_add(occurrence_list)
self.log.debug("Done sending %d occurrences" % num_occurrences,
extra={
'swh_type': 'storage_send_end',
'swh_content_type': 'occurrence',
'swh_num': num_occurrences,
'swh_id': log_id,
})
def get_or_create_origin(self, origin_url):
origin = converters.origin_url_to_origin(origin_url)
origin['id'] = self.storage.origin_add_one(origin)
return origin
def repo_origin(self, repo, origin_url):
log_id = str(uuid.uuid4())
self.log.debug('Creating origin for %s' % origin_url,
extra={
'swh_type': 'storage_send_start',
'swh_content_type': 'origin',
'swh_num': 1,
'swh_id': log_id
})
origin = self.get_or_create_origin(origin_url)
self.log.debug('Done creating origin for %s' % origin_url,
extra={
'swh_type': 'storage_send_end',
'swh_content_type': 'origin',
'swh_num': 1,
'swh_id': log_id
})
return origin
def bulk_send_blobs(self, repo, blobs, origin_id):
"""Format blobs as swh contents and send them to the database"""
packet_size = self.config['content_packet_size']
packet_size_bytes = self.config['content_packet_size_bytes']
max_content_size = self.config['content_size_limit']
send_in_packets(blobs, converters.blob_to_content,
self.send_contents, packet_size, repo=repo,
packet_size_bytes=packet_size_bytes,
log=self.log, max_content_size=max_content_size,
origin_id=origin_id)
def bulk_send_trees(self, repo, trees):
"""Format trees as swh directories and send them to the database"""
packet_size = self.config['directory_packet_size']
send_in_packets(trees, converters.tree_to_directory,
self.send_directories, packet_size, repo=repo,
log=self.log)
def bulk_send_commits(self, repo, commits):
"""Format commits as swh revisions and send them to the database"""
packet_size = self.config['revision_packet_size']
send_in_packets(commits, converters.commit_to_revision,
self.send_revisions, packet_size, repo=repo,
log=self.log)
def bulk_send_annotated_tags(self, repo, tags):
"""Format annotated tags (pygit2.Tag objects) as swh releases and send
them to the database
"""
packet_size = self.config['release_packet_size']
send_in_packets(tags, converters.annotated_tag_to_release,
self.send_releases, packet_size, repo=repo,
log=self.log)
def bulk_send_refs(self, repo, refs):
"""Format git references as swh occurrences and send them to the
database
"""
packet_size = self.config['occurrence_packet_size']
send_in_packets(refs, converters.ref_to_occurrence,
self.send_occurrences, packet_size)
def list_repo_refs(self, repo, origin_id, authority_id, validity):
"""List all the refs from the given repository.
Args:
- repo (pygit2.Repository): the repository to list
- origin_id (int): the id of the origin from which the repo is
taken
- validity (datetime.datetime): the validity date for the
repository's refs
- authority_id (int): the id of the authority on `validity`.
Returns:
A list of dicts with keys:
- branch (str): name of the ref
- revision (sha1_git): revision pointed at by the ref
- origin (int)
- validity (datetime.DateTime)
- authority (int)
Compatible with occurrence_add.
"""
log_id = str(uuid.uuid4())
refs = []
ref_names = repo.listall_references()
for ref_name in ref_names:
ref = repo.lookup_reference(ref_name)
target = ref.target
if not isinstance(target, Oid):
self.log.debug("Peeling symbolic ref %s pointing at %s" % (
ref_name, ref.target), extra={
'swh_type': 'git_sym_ref_peel',
'swh_name': ref_name,
'swh_target': str(ref.target),
'swh_id': log_id,
})
target_obj = ref.peel()
else:
target_obj = repo[target]
if target_obj.type == GIT_OBJ_TAG:
self.log.debug("Peeling ref %s pointing at tag %s" % (
ref_name, target_obj.name), extra={
'swh_type': 'git_ref_peel',
'swh_name': ref_name,
'swh_target': str(target_obj.name),
'swh_id': log_id,
})
target_obj = ref.peel()
if not target_obj.type == GIT_OBJ_COMMIT:
self.log.info("Skipping ref %s pointing to %s %s" % (
ref_name, target_obj.__class__.__name__,
target_obj.id.hex), extra={
'swh_type': 'git_ref_skip',
'swh_name': ref_name,
'swh_target': str(target_obj),
'swh_id': log_id,
})
refs.append({
'branch': ref_name,
'revision': target_obj.id.raw,
'origin': origin_id,
'validity': validity,
'authority': authority_id,
})
return refs
def list_repo_objs(self, repo):
"""List all the objects from repo.
Args:
- repo (pygit2.Repository): the repository to list
Returns:
a dict containing lists of `Oid`s with keys for each object type:
- GIT_OBJ_BLOB
- GIT_OBJ_TREE
- GIT_OBJ_COMMIT
- GIT_OBJ_TAG
"""
log_id = str(uuid.uuid4())
self.log.info("Started listing %s" % repo.path, extra={
'swh_type': 'git_list_objs_start',
'swh_repo': repo.path,
'swh_id': log_id,
})
objects = get_objects_per_object_type(repo)
self.log.info("Done listing the objects in %s: %d contents, "
"%d directories, %d revisions, %d releases" % (
repo.path,
len(objects[GIT_OBJ_BLOB]),
len(objects[GIT_OBJ_TREE]),
len(objects[GIT_OBJ_COMMIT]),
len(objects[GIT_OBJ_TAG]),
), extra={
'swh_type': 'git_list_objs_end',
'swh_repo': repo.path,
'swh_num_blobs': len(objects[GIT_OBJ_BLOB]),
'swh_num_trees': len(objects[GIT_OBJ_TREE]),
'swh_num_commits': len(objects[GIT_OBJ_COMMIT]),
'swh_num_tags': len(objects[GIT_OBJ_TAG]),
'swh_id': log_id,
})
return objects
def open_repo(self, repo_path):
return pygit2.Repository(repo_path)
def load_repo(self, repo, objects, refs, origin_id):
if self.config['send_contents']:
self.bulk_send_blobs(repo, objects[GIT_OBJ_BLOB], origin_id)
else:
self.log.info('Not sending contents')
if self.config['send_directories']:
self.bulk_send_trees(repo, objects[GIT_OBJ_TREE])
else:
self.log.info('Not sending directories')
if self.config['send_revisions']:
self.bulk_send_commits(repo, objects[GIT_OBJ_COMMIT])
else:
self.log.info('Not sending revisions')
if self.config['send_releases']:
self.bulk_send_annotated_tags(repo, objects[GIT_OBJ_TAG])
else:
self.log.info('Not sending releases')
if self.config['send_occurrences']:
self.bulk_send_refs(repo, refs)
else:
self.log.info('Not sending occurrences')
def process(self, repo_path, origin_url, authority_id, validity):
# Open repository
repo = self.open_repo(repo_path)
# Add origin to storage if needed, use the one from config if not
origin = self.repo_origin(repo, origin_url)
# Parse all the refs from our repo
refs = self.list_repo_refs(repo, origin['id'], authority_id,
validity)
if not refs:
self.log.info('Skipping empty repository %s' % repo_path, extra={
'swh_type': 'git_repo_list_refs',
'swh_repo': repo_path,
'swh_num_refs': 0,
})
return
else:
self.log.info('Listed %d refs for repo %s' % (
len(refs), repo_path), extra={
'swh_type': 'git_repo_list_refs',
'swh_repo': repo_path,
'swh_num_refs': len(refs),
})
# We want to load the repository, walk all the objects
objects = self.list_repo_objs(repo)
# Finally, load the repository
self.load_repo(repo, objects, refs, origin['id'])

File Metadata

Mime Type
text/x-python
Expires
Jun 4 2025, 7:05 PM (10 w, 1 d ago)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
3278175

Event Timeline