Sep 2 2020
Sep 1 2020
May 28 2020
Oct 18 2018
Oct 5 2018
Oct 1 2018
Sep 6 2018
Aug 27 2018
Then again, i'll check the pypi api's documentation. Hopefully, it's explained somewhere ;)
Aug 23 2018
So, having one branch in the snapshot per distribution format (tar/zip/etc.) is a nice and clean way of handling this.
Aug 22 2018
The Debian loader doesn't create release objects. Our data model doesn't allow to attach arbitrary structured metadata to release objects (as Git doesn't either), so we've shortcut this level of indirection.
Aug 21 2018
There remains 3 actions to do for the current implementation to be complete:
Aug 2 2018
As far as I can tell from those examples, the metadata that PyPI gives you are the most recent ones, probably the ones extracted from the most recent version, so it would be incorrect to associate them to other releases.
For (1), I think what we currently do for Debian packages is as you said, i.e., snapshot -> release -> revision -> tarball root dir. Maybe you can check for comparison (or @olasd can chime in?). We should do the same here.
The basic loader will be the tarball loader, yes. In addition to that there are two aspects to be defined:
- the stack of objects to be added to the DAG
- the metadata to extract
Aug 1 2018
capable of extracting upstream metadata that are meaningful (and specific to) PyPI.
Jul 25 2018
Jul 24 2018
They have multiple apis:
- basic json one [1] which permits to request information on a per project basis (no listing) [1] (~> foresee the use of this one for the loader)
- xmlrpc deprecated one [2] (this one lists ~> that would be for the lister use)
- html page (listing all packages)
- rss feed (update events)
Jul 10 2018
Sep 14 2017
Sep 7 2017
Jul 3 2017
May 15 2017
Feb 14 2017
Feb 8 2017
Feb 7 2017
Jan 24 2017
Aug 23 2016
Aug 19 2016
Aug 16 2016
We have no guarantee that the internal object ids are monotonic: concurrent transactions can make object_ids of objects go backwards.