1. 30 Jul, 2021 4 commits
  2. 29 Jul, 2021 7 commits
    • Marko Mäkelä's avatar
      MDEV-19445/MDEV-16678 fixup: Acquire proper MDL on innodb_ft_aux_table · 0aa2bc7a
      Marko Mäkelä authored
      In commit 2647fd10 (MDEV-19445)
      we fixed a race condition around the INFORMATION_SCHEMA tables
      that access the table identified by the global variable
      innodb_ft_aux_table. Thanks to MDEV-16678 we could fix it
      even better by using MDL instead of the InnoDB dict_sys latches.
      0aa2bc7a
    • Marko Mäkelä's avatar
      MDEV-21175 follow-up: Remove redundant locking; rely on MDL · e305493b
      Marko Mäkelä authored
      Before entering DML or DDL execution in the storage engine, the SQL layer
      will have acquired metadata lock (MDL) on the current table name as well
      as the names of FOREIGN KEY (grand)child tables (that is,
      tables whose REFERENCES clauses point to the current table).
      The MDL prevents any metadata changes to these tables, such as
      RENAME, TRUNCATE, DROP, ALTER.
      
      While the MDL on the current table prevents dict_table_t::foreign_set
      from being modified, it does not prevent the table metadata that the
      stored pointers are pointing to from being modified.
      
      The MDL on the child tables will prevent both dict_table_t::referenced_set
      as well as the pointed child table metadata from being modified.
      
      wsrep_row_upd_index_is_foreign(): Do not unnecessarily acquire the
      data dictionary latch if Galera replication is not enabled.
      
      ha_innobase::can_switch_engines(): Rely on MDL. We are not dereferencing
      any pointers stored in the sets.
      
      row_mysql_freeze_data_dictionary(), row_mysql_unfreeze_data_dictionary():
      Remove.
      
      row_update_for_mysql(): Call init_fts_doc_id_for_ref() only once.
      
      In ALTER TABLE...IMPORT TABLESPACE and FLUSH TABLES...FOR EXPORT
      the SQL layer is protecting the current table with MDL. We do not
      need InnoDB latches.
      e305493b
    • Marko Mäkelä's avatar
      MDEV-23484 Rollback unnecessarily acquires dict_sys.latch · 86a14289
      Marko Mäkelä authored
      row_undo(): Remove the unnecessary acquisition and release of
      dict_sys.latch. This was supposed to prevent the table from being
      dropped while the undo log record is being rolled back. But, thanks to
      trx_resurrect_table_locks() that was introduced in
      mysql/mysql-server@935ba09d52c1908bde273ad1940b5ab919d9763d
      and commit c291ddfd
      as well as commit 1bd681c8
      (MDEV-25506 part 3) tables will be protected against dropping
      due to table locks.
      
      This reverts commit 0049d5b5
      (which had reverted a previous attempt of fixing this) and addresses
      an earlier race condition with the following:
      
      prepare_inplace_alter_table_dict(): If recovered transactions hold
      locks on the table while we are executing online ADD INDEX, acquire
      a table lock so that the rollback of those recovered transactions will
      not interfere with the ADD INDEX.
      86a14289
    • Marko Mäkelä's avatar
      Cleanup: Remove pars_stored_procedure_call() · 15363a4f
      Marko Mäkelä authored
      The InnoDB internal SQL parser never supported this syntax.
      15363a4f
    • Marko Mäkelä's avatar
      MDEV-25506 fixup: Correct a bogus comment · 92046ba2
      Marko Mäkelä authored
      Since commit 1bd681c8 tablespaces are
      dropped after the table has been dropped. No dict_sys latch protection
      is relevant during the execution of fil_delete_tablespace().
      92046ba2
    • Marko Mäkelä's avatar
      Cleanup: Remove a workaround · 72764b2f
      Marko Mäkelä authored
      Since commit 1bd681c8
      dict_stats_exec_sql() is no longer playing dirty tricks.
      We can remove the workaround in lock_release().
      72764b2f
    • Marko Mäkelä's avatar
      MDEV-26274 fts_lock_table() attempts to release unreserved dict_sys.mutex · e8ba6426
      Marko Mäkelä authored
      fts_lock_table(): Fix a regression that was introduced in
      commit da65cb4d (MDEV-25936).
      e8ba6426
  3. 26 Jul, 2021 2 commits
  4. 23 Jul, 2021 1 commit
  5. 22 Jul, 2021 11 commits
    • Marko Mäkelä's avatar
      Update libmariadb · c35ac548
      Marko Mäkelä authored
      c35ac548
    • Marko Mäkelä's avatar
      Merge 10.5 into 10.6 · 7783161f
      Marko Mäkelä authored
      7783161f
    • Daniel Black's avatar
      mtr: aix test fix innodb.temporary_table · 8cb2027b
      Daniel Black authored
      AIX error message didn't include the filename like:
      
      2021-07-22  9:43:04 0 [ERROR] mariadbd: Can't create/write to file '' (Errcode: 20 "Not a directory")
      2021-07-22  9:43:04 0 [ERROR] mariadbd: Can't create/write to file '' (Errcode: 20 "Not a directory")
      
      Reviewer: Marko
      8cb2027b
    • Marko Mäkelä's avatar
      MDEV-26110 fixup: GCC 4.8.5 -Wtype-limits · efd90937
      Marko Mäkelä authored
      row_ins_index_entry_set_vals(): Remove an assertion that trivially
      holds because the 16-bit dict_col_t::len cannot represent the value
      UNIV_PAGE_SIZE_MAX.
      efd90937
    • Marko Mäkelä's avatar
      MDEV-26210: Disable mariabackup.log_page_corruption · 5f4314f0
      Marko Mäkelä authored
      MDEV-26193 exposed a race condition in the test
      mariabackup.log_page_corruption by no longer waking up purge tasks
      on every transaction commit. (Note that there never was a guarantee
      that the purge tasks would actually run as quickly as the test expected
      it to happen; that would depend on the CPU load.)
      5f4314f0
    • Marko Mäkelä's avatar
      MDEV-26193: Wake up purge less often · a4dc9265
      Marko Mäkelä authored
      Starting with commit 6e12ebd4
      (MDEV-25062), srv_wake_purge_thread_if_not_active() became
      more expensive operation, especially on NUMA systems, because
      instead of reading an atomic global variable trx_sys.rseg_history_len
      we are traversing up to 128 cache lines in trx_sys.history_exists().
      
      trx_t::commit_cleanup(): Do not wake up purge at all.
      We will wake up purge about once per second in srv_master_callback().
      
      srv_master_do_active_tasks(), srv_master_do_idle_tasks():
      Move some duplicated code to srv_master_callback().
      
      srv_master_callback(): Invoke purge_coordinator_timer_callback()
      to ensure that purge will be periodically woken up, even if the
      latest execution of trx_t::commit_cleanup() allowed the purge view
      to advance but did not wake up purge.
      Do not call log_free_check(), because every thread that is going
      to generate redo log is supposed to call that function anyway,
      before acquiring any page latches. Additional calls to the function
      every few seconds should not make any difference.
      
      srv_shutdown_threads(): Ensure that srv_shutdown_state can be at most
      SRV_SHUTDOWN_INITIATED in srv_master_callback(), by first invoking
      srv_master_timer.reset() before changing srv_shutdown_state.
      (Note: We first terminate the srv_master_callback and only then
      terminate the purge tasks. Thus, the purge subsystem should exist
      when srv_master_callback() invokes purge_coordinator_timer_callback()
      if it was initiated in the first place.
      a4dc9265
    • Marko Mäkelä's avatar
      Merge 10.5 into 10.6 · 641f0939
      Marko Mäkelä authored
      641f0939
    • Marko Mäkelä's avatar
      MDEV-26110: Do not rely on alignment on static allocation · 82d59945
      Marko Mäkelä authored
      It is implementation-defined whether alignment requirements
      that are larger than std::max_align_t (typically 8 or 16 bytes)
      will be honored by the compiler and linker.
      
      It turns out that on IBM AIX, both alignas() and MY_ALIGNED()
      only guarantees alignment up to 16 bytes.
      
      For some data structures, specifying alignment to the CPU
      cache line size (typically 64 or 128 bytes) is a mere performance
      optimization, and we do not really care whether the requested
      alignment is guaranteed.
      
      But, for the correct operation of direct I/O, we do require that
      the buffers be aligned at a block size boundary.
      
      field_ref_zero: Define as a pointer, not an array.
      For innochecksum, we can make this point to unaligned memory;
      for anything else, we will allocate an aligned buffer from the heap.
      This buffer will be used for overwriting freed data pages when
      innodb_immediate_scrub_data_uncompressed=ON. And exactly that code
      hit an assertion failure on AIX, in the test innodb.innodb_scrub.
      
      log_sys.checkpoint_buf: Define as a pointer to aligned memory
      that is allocated from heap.
      
      log_t::file::write_header_durable(): Reuse log_sys.checkpoint_buf
      instead of trying to allocate an aligned buffer from the stack.
      82d59945
    • Marko Mäkelä's avatar
      MDEV-26200 buf_pool.flush_list corrupted by buffer pool resizing or ROW_FORMAT=COMPRESSED · bf435a3f
      Marko Mäkelä authored
      The lazy deletion of clean blocks from buf_pool.flush_list that was
      introduced in commit 6441bc61 (MDEV-25113)
      introduced a race condition around the function
      buf_flush_relocate_on_flush_list().
      
      The test innodb_zip.wl5522_debug_zip as well as the buffer pool
      resizing tests would occasionally fail in debug builds due to
      buf_pool.flush_list.count disagreeing with the actual length of the
      doubly-linked list.
      
      The safe procedure for relocating a block in buf_pool.flush_list should be
      as follows, now that we implement lazy deletion from buf_pool.flush_list:
      
      1. Acquire buf_pool.mutex.
      2. Acquire the exclusive buf_pool.page_hash.latch.
      3. Acquire buf_pool.flush_list_mutex.
      4. Copy the block descriptor.
      5. Invoke buf_flush_relocate_on_flush_list().
      6. Release buf_pool.flush_list_mutex.
      
      buf_flush_relocate_on_flush_list(): Assert that
      buf_pool.flush_list_mutex is being held. Invoke
      buf_page_t::oldest_modification() only once, using
      std::memory_order_relaxed, now that the mutex protects us.
      
      buf_LRU_free_page(), buf_LRU_block_remove_hashed(): Avoid
      an unlock-lock cycle on hash_lock. (We must not acquire hash_lock
      while already holding buf_pool.flush_list_mutex, because that
      could lead to a deadlock due to latching order violation.)
      bf435a3f
    • nia's avatar
      Fix building crc32_arm64 on NetBSD/aarch64 · 316a8ceb
      nia authored
      pmull_supported is not necessarily defined before crc32c_aarch64
      Signed-off-by: default avatarNia Alarie <nia@NetBSD.org>
      316a8ceb
    • Daniel Black's avatar
      mtr: aix - stack-trace is optional · 1519013f
      Daniel Black authored
      1519013f
  6. 21 Jul, 2021 2 commits
  7. 20 Jul, 2021 7 commits
  8. 19 Jul, 2021 4 commits
  9. 16 Jul, 2021 2 commits