- Queries
- All Stories
- Search
- Advanced Search
- Transactions
- Transaction Logs
Advanced Search
Jan 8 2023
Dec 22 2022
We have a new one that went unnoticed until 19 days ago: b'superduper/super/sub/bye.txt' is not a valid directory entry name.
Dec 16 2022
Nov 22 2022
To force kafka compaction to run I've done the following:
Oct 31 2022
Possibly relevant for the Azure storage: https://learn.microsoft.com/en-us/rest/api/storageservices/find-blobs-by-tags
Oct 25 2022
This is now done, the objects are fixed in the production DB and kafka.
@vlorentz I'm running the following adaptation to your script:
Oh yeah, I was thinking of just removing the entire project, but your solution also works.
Do you actually want to keep these objects? This would be inconsistent with the fixed loader behavior that would just reject those objects, and not load the repository at all.
I tried to add a workaround in the backfiller, but it is incredibly hard to do properly, especially as entries as disordered, so raw_manifest needs to be fixed in two different ways.
Oct 24 2022
Oct 19 2022
Oct 18 2022
Aug 23 2022
Jul 19 2022
Jul 1 2022
In T2309#87779, @douardda wrote:do you have in mind to make the actual hash used as primary key in an objstorage a configuration of said storage instance? e.g. create a pathslicer or s3 objstorage using sha256 is just a matter of configuration of the objstorage?
do you have in mind to make the actual hash used as primary key in an objstorage a configuration of said storage instance? e.g. create a pathslicer or s3 objstorage using sha256 is just a matter of configuration of the objstorage?
Jun 27 2022
Jun 21 2022
May 1 2022
Apr 29 2022
Apr 8 2022
Apr 5 2022
Mar 30 2022
Mar 25 2022
Mar 23 2022
Jan 25 2022
@vsellier already did this to benchmark cassandra. it's indeed necessary to see how the backends behave with real loader and vault workloads. (less so for the objstorage, since the workloads should be much more uniform)
I'mt not exactly sure why I thought that would be necessary for benchmarking. In any case... it's not ;-)
The documentation is at:
It's documented in the winery test environment and was actually able to use the instructions successfully (after a few fixes...). It does work an this can be closed as resolved.
Added a wiki page to be a more accessible version of the benchmark process than the README in the sources.
Jan 22 2022
Dec 16 2021
Dec 14 2021
In D6834#177663, @vlorentz wrote:Ah wait I got it, docker/services/swh-winery/entrypoint.sh launches winery itself, not an actual objstorage backend; you're only reusing the scafholding.
Ah wait I got it, docker/services/swh-winery/entrypoint.sh launches winery itself, not an actual objstorage backend; you're only reusing the scafholding.
I don't understand why. Currently, we have: client --> objstorage pathslicing backend (port 5003) --> disk.
What you want to do is: client --> objstorage proxy (port 5003) --> objstorage winery backend (port 5012) --> winery --> ceph, right?
(objstorage winery backend is what is launched by docker/services/swh-winery/entrypoint.sh in this diff)
In D6834#177606, @vlorentz wrote:Instead of defining a new service, could you provide an alternative docker-compose config file? This way, it can be used as to switch all services to use it, just by adding a CLI parameter. eg. we do this to replace the postgres storage backend with cassandra: https://docs.softwareheritage.org/devel/getting-started/using-docker.html#cassandra
Instead of defining a new service, could you provide an alternative docker-compose config file? This way, it can be used as to switch all services to use it, just by adding a CLI parameter. eg. we do this to replace the postgres storage backend with cassandra: https://docs.softwareheritage.org/devel/getting-started/using-docker.html#cassandra
It depends on https://forge.softwareheritage.org/D6796 and will fail until it is merged. It can be tested from sources with an override like this:
Dec 13 2021
In practical terms, the two winery objstorage database servers and Ceph itself will be hosted at CEA, while the main ingestion storage / graph storage / ... will remain in Rocquencourt (separated sites, with fairly high bandwidth networking between them).
Dec 12 2021
@olasd I split the debian packaging in its own task at T3797 so that this task can be closed. I'll let you revert this if you think it is not appropriate. My rationale is that it would be easier to figure out what's left to be done with this one other task. Rather than coming back to this rather overloaded ticket. But it's just a matter of personal taste :-)
The documentation now shows as expected. The previous problems in rendering it were probably because the package was not published.