complete other tools section
- Queries
- All Stories
- Search
- Advanced Search
- Transactions
- Transaction Logs
Advanced Search
Sep 20 2021
- fix a couple of staging urls
- add the production urls
Do you reference only the staging part?
fix typos
Sep 17 2021
Sep 16 2021
the content of this comment was moved to the task description
changing the status to resolved as the main issues are solved.
Other tests with more parallel workers will be launched, if other problems will be detected, they will be tracked in new dedicated tickets.
Sep 15 2021
Test of a the new D6269 patch:
2 flame graphs of the previous directory_ls:
- one-by-one
first run (cache cold):
This is the results of the different runs:
The disk was received Monday and replaced Thuesday by Christophe from the DSI.
The raid card automatically launch the raid rebuild. Everything is ok now.
root@beaubourg:~# megacli -PDList -aALL ...
Sep 13 2021
The new datacenter is active since a couple of week.
It allowed to test:
- how to declare a new dc and bootstrap it
- how the data is replicated between the DC
- how to perform inter/intra DC repairs
- how to add nodes on a DC on bootstrap it
- how to remove a datacenter
The quote is done and validated, we will launch the command when the new matinfo deal will be available
Sep 10 2021
A replacement disk will be sent by DELL. It should be delivered the 2021-09-13 if everything is ok.
The DSI is notified of the delivery
rebase
rebase
- fix the wrong box url
- Explain the move to nfs v4
- Explain the puppet directories hack
final tests with the last version, everything looks good with almost the same performances a better ingestion rate in batch:
4 nodes before (batch only):
Sep 9 2021
ensure it works with stretch and versions >= bullseye
Thanks for the last fix, it looks better with a smaller batch size:
5 nodes:
The ingestion is ~7500 ops/s in batch compared to ~6500 before
Sep 8 2021
metadata searches are now done in Elasticsearch since the deployment of T3433
Everything is deployed and look functional.
According to the documentation of the cassandra concurrent api[1], it seems the concurrency can by specified as an argument of the execute_concurrent_with_args method. The default is 100, but it could be interesting to check with higher or lower values
These are more results with different number of replayers.
Each line represents a server with 20 directory replayers, the renages are for one-by-one, concurrent, batch
- 1 node
- 2 nodes
Sep 7 2021
Sep 6 2021
Sep 3 2021
Some feedback, I had to delay the benchmarks because the servers was almost full and the cluster needed to be expanded to 7 nodes. The cluster is in stabilization phase (rebuild/repair of the new node and cleanup of the old one)
When it will be done, I will be able to finalize the tests Hopefully at the beginning of the next week
production deployment:
- disable puppet
- stop and disable the journal clients and the search backend
- update the swh-search configuration to change the index name to origin-v0.11
root@search1:/etc/softwareheritage/search# diff -U3 /tmp/server.yml server.yml --- /tmp/server.yml 2021-09-03 14:06:07.896137122 +0000 +++ server.yml 2021-09-03 14:05:47.072081879 +0000 @@ -10,7 +10,7 @@ port: 9200 indexes: origin: - index: origin-production + index: origin-v0.11 read_alias: origin-read write_alias: origin-write
- update the journal-clients to use a group id swh.search.journal_client.[indexed|object]-v0.11
root@search1:/etc/softwareheritage/search# diff -U3 /tmp/journal_client_objects.yml journal_client_objects.yml --- /tmp/journal_client_objects.yml 2021-09-03 14:06:52.660255797 +0000 +++ journal_client_objects.yml 2021-09-03 14:07:10.684303568 +0000 @@ -8,7 +8,7 @@ - kafka2.internal.softwareheritage.org - kafka3.internal.softwareheritage.org - kafka4.internal.softwareheritage.org - group_id: swh.search.journal_client + group_id: swh.search.journal_client-v0.11 prefix: swh.journal.objects object_types: - origin root@search1:/etc/softwareheritage/search# diff -U3 /tmp/journal_client_indexed.yml journal_client_indexed.yml --- /tmp/journal_client_indexed.yml 2021-09-03 14:06:52.660255797 +0000 +++ journal_client_indexed.yml 2021-09-03 14:07:25.760343512 +0000 @@ -8,7 +8,7 @@ - kafka2.internal.softwareheritage.org - kafka3.internal.softwareheritage.org - kafka4.internal.softwareheritage.org - group_id: swh.search.journal_client.indexed + group_id: swh.search.journal_client.indexed-v0.11 prefix: swh.journal.indexed object_types: - origin_intrinsic_metadata
- perform a system upgrade
root@search1:/etc/softwareheritage/search# apt dist-upgrade -V ... The following NEW packages will be installed: python3-tree-sitter (0.19.0-1+swh1~bpo10+1) The following packages will be upgraded: libnss-systemd (247.3-3~bpo10+1 => 247.3-6~bpo10+1) libpam-systemd (247.3-3~bpo10+1 => 247.3-6~bpo10+1) libsystemd0 (247.3-3~bpo10+1 => 247.3-6~bpo10+1) libudev1 (247.3-3~bpo10+1 => 247.3-6~bpo10+1) python3-swh.core (0.14.3-1~swh1~bpo10+1 => 0.14.5-1~swh1~bpo10+1) python3-swh.model (2.6.1-1~swh1~bpo10+1 => 2.8.0-1~swh1~bpo10+1) python3-swh.scheduler (0.15.0-1~swh1~bpo10+1 => 0.18.0-1~swh1~bpo10+1) python3-swh.search (0.9.0-1~swh1~bpo10+1 => 0.11.4-2~swh1~bpo10+1) python3-swh.storage (0.30.1-1~swh1~bpo10+1 => 0.36.0-1~swh1~bpo10+1) systemd (247.3-3~bpo10+1 => 247.3-6~bpo10+1) systemd-sysv (247.3-3~bpo10+1 => 247.3-6~bpo10+1) systemd-timesyncd (247.3-3~bpo10+1 => 247.3-6~bpo10+1) udev (247.3-3~bpo10+1 => 247.3-6~bpo10+1) 13 upgraded, 1 newly installed, 0 to remove and 0 not upgraded. ...
There is no need to reboot
- enable and restart the swh-search backend
- check the new index creation
root@search1:/etc/softwareheritage/search# curl ${ES_SERVER}/_cat/indices\?v health status index uuid pri rep docs.count docs.deleted store.size pri.store.size green open origin-v0.11 XOUR_jKcTtWKjlPk_8EAlA 90 1 0 0 34.3kb 18.2kb green open origin-v0.9.0 TH9xlECuS4CcJTDw0Fqieg 90 1 175001478 36494554 293gb 146.9gb green open origin-production hZfuv0lVRImjOjO_rYgDzg 90 1 176722078 56232582 311gb 155.1gb
- update the write index alias
root@search1:~/T3433# ./update-write-alias.sh {"acknowledged":true}{"acknowledged":true}root@search1:~/T3433# root@search1:~/T3433# curl ${ES_SERVER}/_cat/aliases\?v alias index filter routing.index routing.search is_write_index origin-write origin-v0.11 - - - - origin-read-v0.9.0 origin-v0.9.0 - - - - origin-v0.9.0-read origin-v0.9.0 - - - - origin-v0.9.0-write origin-v0.9.0 - - - - origin-write-v0.9.0 origin-v0.9.0 - - - - origin-read origin-production - - - -
All the v0.9.0 stuff will be cleared once the migration to the v0.11 done
- restart the journal clients
root@search1:~# systemctl enable swh-search-journal-client@objects Created symlink /etc/systemd/system/multi-user.target.wants/swh-search-journal-client@objects.service → /etc/systemd/system/swh-search-journal-client@.service. root@search1:~# systemctl enable swh-search-journal-client@indexed Created symlink /etc/systemd/system/multi-user.target.wants/swh-search-journal-client@indexed.service → /etc/systemd/system/swh-search-journal-client@.service. root@search1:~# systemctl start swh-search-journal-client@objects root@search1:~# systemctl start swh-search-journal-client@indexed