/***************************************************************************** Copyright (c) 1994, 2018, Oracle and/or its affiliates. All Rights Reserved. This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License, version 2.0, as published by the Free Software Foundation. This program is also distributed with certain software (including but not limited to OpenSSL) that is licensed under separate terms, as designated in a particular file or component or in included license documentation. The authors of MySQL hereby grant you an additional permission to link the program and your derivative works with the separately licensed software that they have included with MySQL. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License, version 2.0, for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA *****************************************************************************/ /** @file ha/ha0ha.cc The hash table with external chains Created 8/22/1994 Heikki Tuuri *************************************************************************/ #include "ha0ha.h" #include #ifdef UNIV_DEBUG #include "buf0buf.h" #endif /* UNIV_DEBUG */ #include "btr0sea.h" #include "page0page.h" #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG /** Maximum number of records in a page */ static const ulint MAX_N_POINTERS = UNIV_PAGE_SIZE_MAX / REC_N_NEW_EXTRA_BYTES; #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ /** Creates a hash table with at least n array cells. The actual number of cells is chosen to be a prime number slightly bigger than n. @return own: created table */ hash_table_t *ib_create(ulint n, /*!< in: number of array cells */ latch_id_t id, /*!< in: latch ID */ ulint n_sync_obj, /*!< in: number of mutexes to protect the hash table: must be a power of 2, or 0 */ ulint type) /*!< in: type of datastructure for which MEM_HEAP_FOR_PAGE_HASH */ { hash_table_t *table; ut_a(type == MEM_HEAP_FOR_BTR_SEARCH || type == MEM_HEAP_FOR_PAGE_HASH); ut_ad(ut_is_2pow(n_sync_obj)); table = hash_create(n); /* Creating MEM_HEAP_BTR_SEARCH type heaps can potentially fail, but in practise it never should in this case, hence the asserts. */ if (n_sync_obj == 0) { table->heap = mem_heap_create_typed( ut_min(static_cast(4096), MEM_MAX_ALLOC_IN_BUF / 2 - MEM_BLOCK_HEADER_SIZE - MEM_SPACE_NEEDED(0)), type); ut_a(table->heap); return (table); } if (type == MEM_HEAP_FOR_PAGE_HASH) { /* We create a hash table protected by rw_locks for buf_pool->page_hash. */ hash_create_sync_obj(table, HASH_TABLE_SYNC_RW_LOCK, id, n_sync_obj); } else { hash_create_sync_obj(table, HASH_TABLE_SYNC_MUTEX, id, n_sync_obj); } table->heaps = static_cast(ut_malloc_nokey(n_sync_obj * sizeof(void *))); for (ulint i = 0; i < n_sync_obj; i++) { table->heaps[i] = mem_heap_create_typed( ut_min(static_cast(4096), MEM_MAX_ALLOC_IN_BUF / 2 - MEM_BLOCK_HEADER_SIZE - MEM_SPACE_NEEDED(0)), type); ut_a(table->heaps[i]); } return (table); } /** Recreate a hash table with at least n array cells. The actual number of cells is chosen to be a prime number slightly bigger than n. The new cells are all cleared. The heaps are recreated. The sync objects are reused. @param[in,out] table hash table to be resuzed (to be freed later) @param[in] n number of array cells @return resized new table */ hash_table_t *ib_recreate(hash_table_t *table, ulint n) { /* This function is for only page_hash for now */ ut_ad(table->type == HASH_TABLE_SYNC_RW_LOCK); ut_ad(table->n_sync_obj > 0); hash_table_t *new_table = hash_create(n); new_table->type = table->type; new_table->n_sync_obj = table->n_sync_obj; new_table->sync_obj = table->sync_obj; for (ulint i = 0; i < table->n_sync_obj; i++) { mem_heap_free(table->heaps[i]); } ut_free(table->heaps); new_table->heaps = static_cast( ut_malloc_nokey(new_table->n_sync_obj * sizeof(void *))); for (ulint i = 0; i < new_table->n_sync_obj; i++) { new_table->heaps[i] = mem_heap_create_typed( ut_min(static_cast(4096), MEM_MAX_ALLOC_IN_BUF / 2 - MEM_BLOCK_HEADER_SIZE - MEM_SPACE_NEEDED(0)), MEM_HEAP_FOR_PAGE_HASH); ut_a(new_table->heaps[i]); } return (new_table); } /** Empties a hash table and frees the memory heaps. */ void ha_clear(hash_table_t *table) /*!< in, own: hash table */ { ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); ut_ad(!table->adaptive || btr_search_own_all(RW_LOCK_X)); for (ulint i = 0; i < table->n_sync_obj; i++) { mem_heap_free(table->heaps[i]); } ut_free(table->heaps); switch (table->type) { case HASH_TABLE_SYNC_MUTEX: for (ulint i = 0; i < table->n_sync_obj; ++i) { mutex_destroy(&table->sync_obj.mutexes[i]); } ut_free(table->sync_obj.mutexes); table->sync_obj.mutexes = NULL; break; case HASH_TABLE_SYNC_RW_LOCK: for (ulint i = 0; i < table->n_sync_obj; ++i) { rw_lock_free(&table->sync_obj.rw_locks[i]); } ut_free(table->sync_obj.rw_locks); table->sync_obj.rw_locks = NULL; break; case HASH_TABLE_SYNC_NONE: /* do nothing */ break; } table->n_sync_obj = 0; table->type = HASH_TABLE_SYNC_NONE; /* Clear the hash table. */ ulint n = hash_get_n_cells(table); for (ulint i = 0; i < n; i++) { hash_get_nth_cell(table, i)->node = NULL; } } /** Inserts an entry into a hash table. If an entry with the same fold number is found, its node is updated to point to the new data, and no new node is inserted. If btr_search_enabled is set to FALSE, we will only allow updating existing nodes, but no new node is allowed to be added. @return true if succeed, false if no more memory could be allocated */ ibool ha_insert_for_fold_func( hash_table_t *table, /*!< in: hash table */ ulint fold, /*!< in: folded value of data; if a node with the same fold value already exists, it is updated to point to the same data, and no new node is created! */ #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG buf_block_t *block, /*!< in: buffer block containing the data */ #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ const rec_t *data) /*!< in: data, must not be NULL */ { hash_cell_t *cell; ha_node_t *node; ha_node_t *prev_node; ulint hash; ut_ad(data); ut_ad(table); ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG ut_a(block->frame == page_align(data)); #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ hash_assert_can_modify(table, fold); ut_ad(btr_search_enabled); hash = hash_calc_hash(fold, table); cell = hash_get_nth_cell(table, hash); prev_node = static_cast(cell->node); while (prev_node != NULL) { if (prev_node->fold == fold) { #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG if (table->adaptive) { buf_block_t *prev_block = prev_node->block; ut_a(prev_block->frame == page_align(prev_node->data)); ut_a(os_atomic_decrement_ulint(&prev_block->n_pointers, 1) < MAX_N_POINTERS); ut_a(os_atomic_increment_ulint(&block->n_pointers, 1) < MAX_N_POINTERS); } prev_node->block = block; #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ prev_node->data = data; return (TRUE); } prev_node = prev_node->next; } /* We have to allocate a new chain node */ node = static_cast( mem_heap_alloc(hash_get_heap(table, fold), sizeof(ha_node_t))); if (node == NULL) { /* It was a btr search type memory heap and at the moment no more memory could be allocated: return */ ut_ad(hash_get_heap(table, fold)->type & MEM_HEAP_BTR_SEARCH); return (FALSE); } ha_node_set_data(node, block, data); #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG if (table->adaptive) { ut_a(os_atomic_increment_ulint(&block->n_pointers, 1) < MAX_N_POINTERS); } #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ node->fold = fold; node->next = NULL; prev_node = static_cast(cell->node); if (prev_node == NULL) { cell->node = node; return (TRUE); } while (prev_node->next != NULL) { prev_node = prev_node->next; } prev_node->next = node; return (TRUE); } #ifdef UNIV_DEBUG /** Verify if latch corresponding to the hash table is x-latched @param[in] table hash table */ static void ha_btr_search_latch_x_locked(const hash_table_t *table) { ulint i; for (i = 0; i < btr_ahi_parts; ++i) { if (btr_search_sys->hash_tables[i] == table) { break; } } ut_ad(i < btr_ahi_parts); ut_ad(rw_lock_own(btr_search_latches[i], RW_LOCK_X)); } #endif /* UNIV_DEBUG */ /** Deletes a hash node. */ void ha_delete_hash_node(hash_table_t *table, /*!< in: hash table */ ha_node_t *del_node) /*!< in: node to be deleted */ { ut_ad(table); ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); ut_d(ha_btr_search_latch_x_locked(table)); ut_ad(btr_search_enabled); #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG if (table->adaptive) { ut_a(del_node->block->frame = page_align(del_node->data)); ut_a(os_atomic_decrement_ulint(&del_node->block->n_pointers, 1) < MAX_N_POINTERS); } #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ HASH_DELETE_AND_COMPACT(ha_node_t, next, table, del_node); } /** Looks for an element when we know the pointer to the data, and updates the pointer to data, if found. @return true if found */ ibool ha_search_and_update_if_found_func( hash_table_t *table, /*!< in/out: hash table */ ulint fold, /*!< in: folded value of the searched data */ const rec_t *data, /*!< in: pointer to the data */ #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG buf_block_t *new_block, /*!< in: block containing new_data */ #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ const rec_t *new_data) /*!< in: new pointer to the data */ { ha_node_t *node; ut_ad(table); ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); hash_assert_can_modify(table, fold); #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG ut_a(new_block->frame == page_align(new_data)); #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ ut_d(ha_btr_search_latch_x_locked(table)); if (!btr_search_enabled) { return (FALSE); } node = ha_search_with_data(table, fold, data); if (node) { #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG if (table->adaptive) { ut_a(os_atomic_decrement_ulint(&node->block->n_pointers, 1) < MAX_N_POINTERS); ut_a(os_atomic_increment_ulint(&new_block->n_pointers, 1) < MAX_N_POINTERS); } node->block = new_block; #endif /* UNIV_AHI_DEBUG || UNIV_DEBUG */ node->data = new_data; return (TRUE); } return (FALSE); } /** Removes from the chain determined by fold all nodes whose data pointer points to the page given. */ void ha_remove_all_nodes_to_page(hash_table_t *table, /*!< in: hash table */ ulint fold, /*!< in: fold value */ const page_t *page) /*!< in: buffer page */ { ha_node_t *node; ut_ad(table); ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); hash_assert_can_modify(table, fold); ut_ad(btr_search_enabled); node = ha_chain_get_first(table, fold); while (node) { if (page_align(ha_node_get_data(node)) == page) { /* Remove the hash node */ ha_delete_hash_node(table, node); /* Start again from the first node in the chain because the deletion may compact the heap of nodes and move other nodes! */ node = ha_chain_get_first(table, fold); } else { node = ha_chain_get_next(node); } } #ifdef UNIV_DEBUG /* Check that all nodes really got deleted */ node = ha_chain_get_first(table, fold); while (node) { ut_a(page_align(ha_node_get_data(node)) != page); node = ha_chain_get_next(node); } #endif /* UNIV_DEBUG */ } #if defined UNIV_AHI_DEBUG || defined UNIV_DEBUG /** Validates a given range of the cells in hash table. @return true if ok */ ibool ha_validate(hash_table_t *table, /*!< in: hash table */ ulint start_index, /*!< in: start index */ ulint end_index) /*!< in: end index */ { ibool ok = TRUE; ulint i; ut_ad(table); ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); ut_a(start_index <= end_index); ut_a(start_index < hash_get_n_cells(table)); ut_a(end_index < hash_get_n_cells(table)); for (i = start_index; i <= end_index; i++) { ha_node_t *node; hash_cell_t *cell; cell = hash_get_nth_cell(table, i); for (node = static_cast(cell->node); node != 0; node = node->next) { if (hash_calc_hash(node->fold, table) != i) { ib::error(ER_IB_MSG_522) << "Hash table node fold value " << node->fold << " does not match the" " cell number " << i << "."; ok = FALSE; } } } return (ok); } #endif /* defined UNIV_AHI_DEBUG || defined UNIV_DEBUG */ /** Prints info of a hash table. */ void ha_print_info(FILE *file, /*!< in: file where to print */ hash_table_t *table) /*!< in: hash table */ { #ifdef UNIV_DEBUG /* Some of the code here is disabled for performance reasons in production builds, see http://bugs.mysql.com/36941 */ #define PRINT_USED_CELLS #endif /* UNIV_DEBUG */ #ifdef PRINT_USED_CELLS hash_cell_t *cell; ulint cells = 0; ulint i; #endif /* PRINT_USED_CELLS */ ulint n_bufs; ut_ad(table); ut_ad(table->magic_n == HASH_TABLE_MAGIC_N); #ifdef PRINT_USED_CELLS for (i = 0; i < hash_get_n_cells(table); i++) { cell = hash_get_nth_cell(table, i); if (cell->node) { cells++; } } #endif /* PRINT_USED_CELLS */ fprintf(file, "Hash table size %lu", (ulong)hash_get_n_cells(table)); #ifdef PRINT_USED_CELLS fprintf(file, ", used cells %lu", (ulong)cells); #endif /* PRINT_USED_CELLS */ if (table->heaps == NULL && table->heap != NULL) { /* This calculation is intended for the adaptive hash index: how many buffer frames we have reserved? */ n_bufs = UT_LIST_GET_LEN(table->heap->base) - 1; if (table->heap->free_block) { n_bufs++; } fprintf(file, ", node heap has %lu buffer(s)\n", (ulong)n_bufs); } }