migration done.
- Queries
- All Stories
- Search
- Advanced Search
- Transactions
- Transaction Logs
Advanced Search
Jan 28 2022
Jan 24 2022
Jan 21 2022
Jan 20 2022
belvedere migration status: first index on directory created, ongoing index creation for revision, and then release.
staging, prod: storage deployed and service restarted (it's not dependent on the sql migration to be complete).
- staging db already migrated during the deployment of T3861.
- production db migration ongoing
Jan 19 2022
Jan 18 2022
Jan 13 2022
Jan 12 2022
Jan 11 2022
Jan 7 2022
Jan 6 2022
Dec 22 2021
Dec 7 2021
Dec 2 2021
We don't care anymore, this will be handled by T3753.
The slide of the restrospective of the experiment are available at : https://hedgedoc.softwareheritage.org/VOP9qh1MTqm4DjPQfFgNbQ
It was not easy to know if it's a lot of call or long running calls because it's regular sample and we don't have this granularity.
Dec 1 2021
Nov 26 2021
We decided to store manifests instead. T3594#74385
We decided to store manifests instead. T3594#74385
We decided to store manifests instead. T3594#74385
Copy of an email I sent today:
Nov 15 2021
Nov 8 2021
Oct 27 2021
Oct 26 2021
Oct 22 2021
I came across a rather small repository [1] which i believe raise the same issue.
So it may help to keep its reference to ease the testing of the improvment discussed here.
Feel free to dismiss if not that useful.
Oct 21 2021
Stopping here the investigations as the limit is now the hardware running cassandra (HDD). Great improvements were made compared to the first tries
Changing the cache size doesn't seem very effective for the particular workload of the loaders:
Exception some contextual differences due probably due to the different origin loaded, the performances are quite similar.
The hit ratios between the different configurations are also very close
Oct 20 2021
Closing here the tests of the workers.
This is the impact of configuration the add batch insert size. Tested values were 500, 1000 and 2000.
A full repair of a cassandra node was in progress due to a datalost (disk erased by someone else's experiment after the disk reservation expiration)
Oct 15 2021
Oct 14 2021
What "directory_entries_insert_algo" did you use for this?
Some flame graphs of storage was performed during the ingestion with 50 workers in //
Oct 12 2021
Some runs with the fix:
It globally improves the stability of the benchmark by reducing the timeouts.
Oct 11 2021
Oct 6 2021
The loader were finally stabilized. It was due to a wrong celery configuration.
Changing the pool configuration from solo to prefork solved the problem even if the concurrency is kept to one.
Solo looked indicated in environment like the POC but for obvious reasons, it was not working as expected:
Oct 1 2021
intermediary status:
- the bench lab is easily deployable on g5k on several workers to distribute the load [1]
- it's working well when the load is not so high. When the number of worker is increased, it seems the workers have some issues to talk with rabbitmq:
[loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,449: INFO/MainProcess] missed heartbeat from celery@loaders-77cdd444df-p9ds5 [loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,449: INFO/MainProcess] missed heartbeat from celery@loaders-77cdd444df-n6pvm [loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,449: INFO/MainProcess] missed heartbeat from celery@loaders-77cdd444df-mrcjj [loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,449: INFO/MainProcess] missed heartbeat from celery@loaders-77cdd444df-7bn4s [loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,449: INFO/MainProcess] missed heartbeat from celery@loaders-77cdd444df-lg2bd
and also an unexplained time drift:
[loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,447: WARNING/MainProcess] Substantial drift from celery@loaders-77cdd444df-lxjpl may mean clocks are out of sync. Current drift is [loaders-77cdd444df-flcv9 loaders] 356 seconds. [orig: 2021-09-30 23:46:55.447181 recv: 2021-09-30 23:40:59.633444] [loaders-77cdd444df-flcv9 loaders] [loaders-77cdd444df-flcv9 loaders] [2021-09-30 23:46:55,447: WARNING/MainProcess] Substantial drift from celery@loaders-77cdd444df-jd6v9 may mean clocks are out of sync. Current drift is [loaders-77cdd444df-flcv9 loaders] 355 seconds. [orig: 2021-09-30 23:46:55.447552 recv: 2021-09-30 23:41:00.723983] [loaders-77cdd444df-flcv9 loaders]