diff --git a/swh/provenance/tests/test_provenance_db.py b/swh/provenance/tests/test_provenance_db.py index 8e4d933..d7ef760 100644 --- a/swh/provenance/tests/test_provenance_db.py +++ b/swh/provenance/tests/test_provenance_db.py @@ -1,292 +1,367 @@ # Copyright (C) 2021 The Software Heritage developers # See the AUTHORS file at the top-level directory of this distribution # License: GNU General Public License version 3, or any later version # See top-level LICENSE file for more information import datetime import pytest from swh.model.tests.swh_model_data import TEST_OBJECTS from swh.provenance.model import RevisionEntry from swh.provenance.origin import OriginEntry from swh.provenance.provenance import origin_add, revision_add from swh.provenance.storage.archive import ArchiveStorage from swh.provenance.tests.conftest import synthetic_result def ts2dt(ts: dict) -> datetime.datetime: timestamp = datetime.datetime.fromtimestamp( ts["timestamp"]["seconds"], datetime.timezone(datetime.timedelta(minutes=ts["offset"])), ) return timestamp.replace(microsecond=ts["timestamp"]["microseconds"]) def test_provenance_origin_add(provenance, swh_storage_with_objects): """Test the ProvenanceDB.origin_add() method""" for origin in TEST_OBJECTS["origin"]: entry = OriginEntry(url=origin.url, revisions=[]) origin_add(ArchiveStorage(swh_storage_with_objects), provenance, entry) # TODO: check some facts here def test_provenance_add_revision(provenance, storage_and_CMDBTS, archive): storage, data = storage_and_CMDBTS for i in range(2): # do it twice, there should be no change in results for revision in data["revision"]: entry = RevisionEntry( id=revision["id"], date=ts2dt(revision["date"]), root=revision["directory"], ) revision_add(provenance, archive, [entry]) # there should be as many entries in 'revision' as revisions from the # test dataset provenance.cursor.execute("SELECT count(*) FROM revision") assert provenance.cursor.fetchone()[0] == len(data["revision"]) # there should be no 'location' for the empty path provenance.cursor.execute("SELECT count(*) FROM location WHERE path=''") assert provenance.cursor.fetchone()[0] == 0 # there should be 32 'location' for non-empty path provenance.cursor.execute("SELECT count(*) FROM location WHERE path!=''") assert provenance.cursor.fetchone()[0] == 32 # there should be as many entries in 'revision' as revisions from the # test dataset provenance.cursor.execute("SELECT count(*) FROM revision") assert provenance.cursor.fetchone()[0] == len(data["revision"]) # 7 directories provenance.cursor.execute("SELECT count(*) FROM directory") assert provenance.cursor.fetchone()[0] == 7 # 12 D-R entries provenance.cursor.execute("SELECT count(*) FROM directory_in_rev") assert provenance.cursor.fetchone()[0] == 12 provenance.cursor.execute("SELECT count(*) FROM content") assert provenance.cursor.fetchone()[0] == len(data["content"]) provenance.cursor.execute("SELECT count(*) FROM content_in_dir") assert provenance.cursor.fetchone()[0] == 16 provenance.cursor.execute("SELECT count(*) FROM content_early_in_rev") assert provenance.cursor.fetchone()[0] == 13 def test_provenance_content_find_first(provenance, storage_and_CMDBTS, archive): storage, data = storage_and_CMDBTS for revision in data["revision"]: entry = RevisionEntry( id=revision["id"], date=ts2dt(revision["date"]), root=revision["directory"], ) revision_add(provenance, archive, [entry]) first_expected_content = [ { "content": "43f3c871310a8e524004e91f033e7fb3b0bc8475", "rev": "35ccb8dd1b53d2d8a5c1375eb513ef2beaa79ae5", "date": 1609757158, "path": "README.md", }, { "content": "6dc7e44ead5c0e300fe94448c3e046dfe33ad4d1", "rev": "9e36e095b79e36a3da104ce272989b39cd68aefd", "date": 1610644094, "path": "Red/Blue/Green/a", }, { "content": "9f6e04be05297905f1275d3f4e0bb0583458b2e8", "rev": "bfbfcc72ae7fc35d6941386c36280512e6b38440", "date": 1610644097, "path": "Red/Blue/Green/b", }, { "content": "a28fa70e725ebda781e772795ca080cd737b823c", "rev": "0a31c9d509783abfd08f9fdfcd3acae20f17dfd0", "date": 1610644099, "path": "Red/Blue/c", }, { "content": "c0229d305adf3edf49f031269a70e3e87665fe88", "rev": "1d1fcf1816a8a2a77f9b1f342ba11d0fe9fd7f17", "date": 1610644105, "path": "Purple/d", }, { "content": "94ba40161084e8b80943accd9d24e1f9dd47189b", "rev": "55d4dc9471de6144f935daf3c38878155ca274d5", "date": 1610644113, "path": ("Dark/Brown/Purple/f", "Dark/Brown/Purple/g", "Dark/h"), # XXX }, { "content": "5e8f9ceaee9dafae2e3210e254fdf170295f8b5b", "rev": "a8939755d0be76cfea136e9e5ebce9bc51c49fef", "date": 1610644116, "path": "Dark/h", }, { "content": "bbd54b961764094b13f10cef733e3725d0a834c3", "rev": "ca1774a07b6e02c1caa7ae678924efa9259ee7c6", "date": 1610644118, "path": "Paris/i", }, { "content": "7ce4fe9a22f589fa1656a752ea371b0ebc2106b1", "rev": "611fe71d75b6ea151b06e3845c09777acc783d82", "date": 1610644120, "path": "Paris/j", }, { "content": "cb79b39935c9392fa5193d9f84a6c35dc9c22c75", "rev": "4c5551b4969eb2160824494d40b8e1f6187fc01e", "date": 1610644122, "path": "Paris/k", }, ] for expected in first_expected_content: contentid = bytes.fromhex(expected["content"]) (blob, rev, date, path) = provenance.content_find_first(contentid) if isinstance(expected["path"], tuple): assert bytes(path).decode() in expected["path"] else: assert bytes(path).decode() == expected["path"] assert bytes(blob) == contentid assert bytes(rev).hex() == expected["rev"] assert int(date.timestamp()) == expected["date"] +def test_provenance_content_find_all(provenance, storage_and_CMDBTS, archive): + storage, data = storage_and_CMDBTS + for revision in data["revision"]: + entry = RevisionEntry( + id=revision["id"], + date=ts2dt(revision["date"]), + root=revision["directory"], + ) + revision_add(provenance, archive, [entry]) + + expected_content = { + # fmt: off + '43f3c871310a8e524004e91f033e7fb3b0bc8475': [ + ('35ccb8dd1b53d2d8a5c1375eb513ef2beaa79ae5', 1609757158.0, b'README.md')], + '6dc7e44ead5c0e300fe94448c3e046dfe33ad4d1': [ + ('9e36e095b79e36a3da104ce272989b39cd68aefd', 1610644094.0, b'Red/Blue/Green/a'), # noqa: E501 + ('bfbfcc72ae7fc35d6941386c36280512e6b38440', 1610644097.0, b'Red/Blue/Green/a'), # noqa: E501 + ('0a31c9d509783abfd08f9fdfcd3acae20f17dfd0', 1610644099.0, b'Red/Blue/Green/a'), # noqa: E501 + ('ca6ec564c69efd2e5c70fb05486fd3f794765a04', 1610644101.0, b'Red/Green/a'), + ('ca6ec564c69efd2e5c70fb05486fd3f794765a04', 1610644101.0, b'Red/a'), + ('fc6e10b7d41b1d56a94091134e3683ce91e80d91', 1610644103.0, b'Red/Blue/Green/a'), # noqa: E501 + ('ba00e89d47dc820bb32c783af7123ffc6e58b56d', 1610644111.0, b'Dark/a')], + '9f6e04be05297905f1275d3f4e0bb0583458b2e8': [ + ('bfbfcc72ae7fc35d6941386c36280512e6b38440', 1610644097.0, b'Red/Blue/Green/b'), # noqa: E501 + ('0a31c9d509783abfd08f9fdfcd3acae20f17dfd0', 1610644099.0, b'Red/Blue/Green/b'), # noqa: E501 + ('ca6ec564c69efd2e5c70fb05486fd3f794765a04', 1610644101.0, b'Red/Green/b'), + ('fc6e10b7d41b1d56a94091134e3683ce91e80d91', 1610644103.0, b'Red/Blue/Green/b')], # noqa: E501 + 'a28fa70e725ebda781e772795ca080cd737b823c': [ + ('0a31c9d509783abfd08f9fdfcd3acae20f17dfd0', 1610644099.0, b'Red/Blue/c'), + ('fc6e10b7d41b1d56a94091134e3683ce91e80d91', 1610644103.0, b'Red/Blue/c')], + 'c0229d305adf3edf49f031269a70e3e87665fe88': [ + ('1d1fcf1816a8a2a77f9b1f342ba11d0fe9fd7f17', 1610644105.0, b'Purple/d'), + ('9a71f967ae1a125be9b6569cc4eccec0aecabb7c', 1610644107.0, b'Purple/Brown/Purple/d'), # noqa: E501 + ('4fde4ea4494a630030a4bda99d03961d9add00c7', 1610644109.0, b'Dark/Brown/Purple/d'), # noqa: E501 + ('4fde4ea4494a630030a4bda99d03961d9add00c7', 1610644109.0, b'Dark/d'), + ('ba00e89d47dc820bb32c783af7123ffc6e58b56d', 1610644111.0, b'Dark/Brown/Purple/d'), # noqa: E501 + ('ba00e89d47dc820bb32c783af7123ffc6e58b56d', 1610644111.0, b'Dark/Brown/Purple/e')], # noqa: E501 + '94ba40161084e8b80943accd9d24e1f9dd47189b': [ + ('55d4dc9471de6144f935daf3c38878155ca274d5', 1610644113.0, b'Dark/Brown/Purple/f'), # noqa: E501 + ('55d4dc9471de6144f935daf3c38878155ca274d5', 1610644113.0, b'Dark/Brown/Purple/g'), # noqa: E501 + ('55d4dc9471de6144f935daf3c38878155ca274d5', 1610644113.0, b'Dark/f'), + ('a8939755d0be76cfea136e9e5ebce9bc51c49fef', 1610644116.0, b'Dark/Brown/Purple/f'), # noqa: E501 + ('a8939755d0be76cfea136e9e5ebce9bc51c49fef', 1610644116.0, b'Dark/Brown/Purple/g'), # noqa: E501 + ('ca1774a07b6e02c1caa7ae678924efa9259ee7c6', 1610644118.0, b'Paris/Brown/Purple/f'), # noqa: E501 + ('ca1774a07b6e02c1caa7ae678924efa9259ee7c6', 1610644118.0, b'Paris/Brown/Purple/g'), # noqa: E501 + ('611fe71d75b6ea151b06e3845c09777acc783d82', 1610644120.0, b'Paris/Berlin/Purple/f'), # noqa: E501 + ('611fe71d75b6ea151b06e3845c09777acc783d82', 1610644120.0, b'Paris/Berlin/Purple/g'), # noqa: E501 + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/Berlin/Purple/f'), # noqa: E501 + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/Berlin/Purple/g'), # noqa: E501 + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/Munich/Purple/f'), # noqa: E501 + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/Munich/Purple/g'), # noqa: E501 + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/Purple/f'), # noqa: E501 + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/Purple/g')], # noqa: E501 + '5e8f9ceaee9dafae2e3210e254fdf170295f8b5b': [ + ('a8939755d0be76cfea136e9e5ebce9bc51c49fef', 1610644116.0, b'Dark/h')], + 'bbd54b961764094b13f10cef733e3725d0a834c3': [ + ('ca1774a07b6e02c1caa7ae678924efa9259ee7c6', 1610644118.0, b'Paris/i')], + '7ce4fe9a22f589fa1656a752ea371b0ebc2106b1': [ + ('611fe71d75b6ea151b06e3845c09777acc783d82', 1610644120.0, b'Paris/j')], + 'cb79b39935c9392fa5193d9f84a6c35dc9c22c75': [ + ('4c5551b4969eb2160824494d40b8e1f6187fc01e', 1610644122.0, b'Paris/k')], + # fmt: on + } + + for content, results in expected_content.items(): + contentid = bytes.fromhex(content) + occurrences = [ + (blob.hex(), rev.hex(), date.timestamp(), path) + for blob, rev, date, path in provenance.content_find_all(contentid) + ] + expected = [(content, *result) for result in results] + assert len(occurrences) == len(expected) + assert set(occurrences) == set(expected) + + def sha1s(cur, table): """return the 'sha1' column from the DB 'table' (as hex) 'cur' is a cursor to the provenance index DB. """ cur.execute(f"SELECT sha1 FROM {table}") return set(sha1.hex() for (sha1,) in cur.fetchall()) def locations(cur): """return the 'path' column from the DB location table 'cur' is a cursor to the provenance index DB. """ cur.execute("SELECT encode(location.path::bytea, 'escape') FROM location") return set(x for (x,) in cur.fetchall()) def relations(cur, src, dst): """return the triplets ('sha1', 'sha1', 'path') from the DB for the relation between 'src' table and 'dst' table (i.e. for C-R, C-D and D-R relations). 'cur' is a cursor to the provenance index DB. """ relation = { ("content", "revision"): "content_early_in_rev", ("content", "directory"): "content_in_dir", ("directory", "revision"): "directory_in_rev", }[(src, dst)] srccol = {"content": "blob", "directory": "dir"}[src] dstcol = {"directory": "dir", "revision": "rev"}[dst] cur.execute( f"SELECT encode(src.sha1::bytea, 'hex')," f" encode(dst.sha1::bytea, 'hex')," f" encode(location.path::bytea, 'escape') " f"FROM {relation} as rel, " f" {src} as src, {dst} as dst, location " f"WHERE rel.{srccol}=src.id AND rel.{dstcol}=dst.id AND rel.loc=location.id" ) return set(cur.fetchall()) @pytest.mark.parametrize( "syntheticfile, args", ( ("synthetic_lower_1.txt", {"lower": True, "mindepth": 1}), ("synthetic_upper_1.txt", {"lower": False, "mindepth": 1}), ("synthetic_lower_2.txt", {"lower": True, "mindepth": 2}), ("synthetic_upper_2.txt", {"lower": False, "mindepth": 2}), ), ) def test_provenance_heuristics( provenance, storage_and_CMDBTS, archive, syntheticfile, args ): storage, data = storage_and_CMDBTS revisions = {rev["id"]: rev for rev in data["revision"]} rows = { "content": set(), "content_in_dir": set(), "content_early_in_rev": set(), "directory": set(), "directory_in_rev": set(), "location": set(), "revision": set(), } for synth_rev in synthetic_result(syntheticfile): revision = revisions[synth_rev["sha1"]] entry = RevisionEntry( id=revision["id"], date=ts2dt(revision["date"]), root=revision["directory"], ) revision_add(provenance, archive, [entry], **args) # each "entry" in the synth file is one new revision rows["revision"].add(synth_rev["sha1"].hex()) assert rows["revision"] == sha1s(provenance.cursor, "revision"), synth_rev[ "msg" ] # this revision might have added new content objects rows["content"] |= set(x["dst"].hex() for x in synth_rev["R_C"]) rows["content"] |= set(x["dst"].hex() for x in synth_rev["D_C"]) assert rows["content"] == sha1s(provenance.cursor, "content"), synth_rev["msg"] # check for R-C (direct) entries rows["content_early_in_rev"] |= set( (x["dst"].hex(), x["src"].hex(), x["path"]) for x in synth_rev["R_C"] ) assert rows["content_early_in_rev"] == relations( provenance.cursor, "content", "revision" ), synth_rev["msg"] # check directories rows["directory"] |= set(x["dst"].hex() for x in synth_rev["R_D"]) assert rows["directory"] == sha1s(provenance.cursor, "directory"), synth_rev[ "msg" ] # check for R-D entries rows["directory_in_rev"] |= set( (x["dst"].hex(), x["src"].hex(), x["path"]) for x in synth_rev["R_D"] ) assert rows["directory_in_rev"] == relations( provenance.cursor, "directory", "revision" ), synth_rev["msg"] # check for D-C entries rows["content_in_dir"] |= set( (x["dst"].hex(), x["src"].hex(), x["path"]) for x in synth_rev["D_C"] ) assert rows["content_in_dir"] == relations( provenance.cursor, "content", "directory" ), synth_rev["msg"] # check for location entries rows["location"] |= set(x["path"] for x in synth_rev["R_C"]) rows["location"] |= set(x["path"] for x in synth_rev["D_C"]) rows["location"] |= set(x["path"] for x in synth_rev["R_D"]) assert rows["location"] == locations(provenance.cursor), synth_rev["msg"]