Date
Attendees
Notes from the previous meeting
Discussion items
Discussed | Item | Notes |
---|---|---|
Project news |
Colin Slater community workshop (not the All Hands) at SLAC this June. The non-person format. | |
USDF | Summary of news from this week's meeting: Joint Data Facilities Meeting - 2024-03-18:
| |
Current status of Qserv and Qserv builds | The current production release is Status of the ongoing "garbage collection" for the temporary message tables:
Status of Qserv cluster at IDF:
The relevant section of the database server's log: 2024-03-18 20:43:13 3 [Warning] IP address '10.137.1.1' could not be resolved: Name or service not known 2024-03-18 20:43:13 3 [Warning] Aborted connection 3 to db: 'unconnected' user: 'unauthenticated' host: '10.137.1.1' (This connection closed normally without authentication) 2024-03-18 20:43:14 0x7f42e81e5700 InnoDB: Assertion failure in file /home/buildbot/buildbot/build/mariadb-10.6.8/storage/innobase/btr/btr0cur.cc line 324 InnoDB: Failing assertion: btr_page_get_prev(get_block->page.frame) == block->page.id().page_no() InnoDB: We intentionally generate a memory trap. InnoDB: Submit a detailed bug report to https://jira.mariadb.org/ InnoDB: If you get repeated assertion failures or crashes, even InnoDB: immediately after the mariadbd startup, there may be InnoDB: corruption in the InnoDB tablespace. Please refer to InnoDB: https://mariadb.com/kb/en/library/innodb-recovery-modes/ InnoDB: about forcing recovery. 240318 20:43:14 [ERROR] mysqld got signal 6 ; This could be because you hit a bug. It is also possible that this binary or one of the libraries it was linked against is corrupt, improperly built, or misconfigured. This error can also be caused by malfunctioning hardware. ... I will keep an eye on the problem. Fritz Mueller suggested Igor Gaponenko run the integrity test on the Replication system's database tomorrow during the Thursday outage at IDF (3:00 pm). UPDATE 2024-03-21: Igor Gaponenko : the database analyzer didn't report any problems with the Replication database on -int or -prod : % kubectl exec -it qserv-repl-db-0 -- mysqlcheck --analyze -uroot -pxxxxxx -h127.0.0.1 --protocol=tcp -P3306 qservReplica qservReplica.QMetadata OK qservReplica.config_database OK qservReplica.config_database_family OK qservReplica.config_database_table OK qservReplica.config_database_table_schema OK qservReplica.config_worker OK qservReplica.config_worker_ext OK qservReplica.controller OK qservReplica.controller_log OK qservReplica.controller_log_ext OK qservReplica.database_ingest OK qservReplica.job OK qservReplica.job_ext OK qservReplica.replica OK qservReplica.replica_file OK qservReplica.request OK qservReplica.request_ext OK qservReplica.stats_table_rows OK qservReplica.transaction OK qservReplica.transaction_contrib OK qservReplica.transaction_contrib_ext OK qservReplica.transaction_contrib_retry OK qservReplica.transaction_contrib_warn OK qservReplica.transaction_log OK | |
Qserv query analysis and query processing performance | Context: IDF A bunch of queries were sent to the "medium" queue From the relevant discussion in the team channel this week, it sounds like we may (possibly) have a suboptimal implementation of the following query class: SELECT DISTINCT <column> FROM <database>.<partitioned-table> LIMIT <N> Do we clearly understand what's going on here, and is this a problem or a "problem"? Shall we further discuss this?
| |
Merging qserv-operator into qserv source tree and changing container builds |
| |
Addressing an issue with the "dark" queries |
| |
HTTP-based Qserv frontend | Igor Gaponenko: ongoing work on expanding the integration test
Igor Gaponenko: no progress (has not started working on) on | |
New dispatch (new Qserv) | John Gates looking at: |