Leveraging azure infrastructure, trigger the blake2s256 update on the existing contents.
This means:
- [ ] provisioning azure vms (sizing/cost to determine)
- [ ] deploying the swh.indexer.rehash module (needs task puppetization)
- [ ] Send all contents to the swh_indexer_rehash queue (table archiver.content or swh.content to determine).
- [ ] Cost/speed projection as soon as some threshold contents have been updated (~1M sounds reasonable given our size)
Note:
In regards to the storage stack to use, we can:
- either use the azure's objstorage (copy is complete now). This will be the starting point.
- or use uffizi's objstorage (or banco) as the azure's in-transit's cost is null if the cost projection is too high.