polardbxengine/storage/ndb/plugin/ha_ndbcluster.h

768 lines
30 KiB
C++

/*
Copyright (c) 2000, 2019, Oracle and/or its affiliates. All rights reserved.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License, version 2.0,
as published by the Free Software Foundation.
This program is also distributed with certain software (including
but not limited to OpenSSL) that is licensed under separate terms,
as designated in a particular file or component or in included license
documentation. The authors of MySQL hereby grant you an additional
permission to link the program and your derivative works with the
separately licensed software that they have included with MySQL.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License, version 2.0, for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
*/
#ifndef SQL_HA_NDBCLUSTER_INCLUDED
#define SQL_HA_NDBCLUSTER_INCLUDED
/*
This file defines the NDB Cluster handler: the interface between MySQL and
NDB Cluster
*/
#include "sql/partitioning/partition_handler.h"
#include "sql/sql_base.h"
#include "sql/table.h"
#include "storage/ndb/include/kernel/ndb_limits.h"
#include "storage/ndb/include/ndbapi/NdbApi.hpp"
#include "storage/ndb/include/ndbapi/ndbapi_limits.h"
#include "storage/ndb/plugin/ha_ndbcluster_cond.h"
#include "storage/ndb/plugin/ndb_conflict.h"
#include "storage/ndb/plugin/ndb_table_map.h"
#define NDB_HIDDEN_PRIMARY_KEY_LENGTH 8
class Ndb; // Forward declaration
class NdbOperation; // Forward declaration
class NdbTransaction; // Forward declaration
class NdbRecAttr; // Forward declaration
class NdbScanOperation;
class NdbIndexScanOperation;
class NdbBlob;
class NdbIndexStat;
class NdbEventOperation;
class ha_ndbcluster_cond;
class NdbQuery;
class NdbQueryOperation;
class NdbQueryOperationTypeWrapper;
class NdbQueryParamValue;
class ndb_pushed_join;
enum NDB_INDEX_TYPE {
UNDEFINED_INDEX = 0,
PRIMARY_KEY_INDEX = 1,
PRIMARY_KEY_ORDERED_INDEX = 2,
UNIQUE_INDEX = 3,
UNIQUE_ORDERED_INDEX = 4,
ORDERED_INDEX = 5
};
struct NDB_INDEX_DATA {
NDB_INDEX_TYPE type;
enum { UNDEFINED = 0, ACTIVE = 1, TO_BE_DROPPED = 2 } status;
const NdbDictionary::Index *index;
const NdbDictionary::Index *unique_index;
unsigned char *unique_index_attrid_map;
bool null_in_unique_index;
/*
In mysqld, keys and rows are stored differently (using KEY_PART_INFO for
keys and Field for rows).
So we need to use different NdbRecord for an index for passing values
from a key and from a row.
*/
NdbRecord *ndb_record_key;
NdbRecord *ndb_unique_record_key;
NdbRecord *ndb_unique_record_row;
};
// Wrapper class for list to hold NDBFKs
class Ndb_fk_list : public List<NdbDictionary::ForeignKey> {
public:
~Ndb_fk_list() { delete_elements(); }
};
#include "storage/ndb/plugin/ndb_ndbapi_util.h"
#include "storage/ndb/plugin/ndb_share.h"
struct Ndb_local_table_statistics {
int no_uncommitted_rows_count;
ulong last_count;
ha_rows records;
};
#include "storage/ndb/plugin/ndb_thd_ndb.h"
struct st_ndb_status {
st_ndb_status() { memset(this, 0, sizeof(struct st_ndb_status)); }
long cluster_node_id;
const char *connected_host;
long connected_port;
long number_of_data_nodes;
long number_of_ready_data_nodes;
long connect_count;
long execute_count;
long trans_hint_count;
long scan_count;
long pruned_scan_count;
long schema_locks_count;
long sorted_scan_count;
long pushed_queries_defined;
long pushed_queries_dropped;
long pushed_queries_executed;
long pushed_reads;
long long last_commit_epoch_server;
long long last_commit_epoch_session;
long long api_client_stats[Ndb::NumClientStatistics];
const char *system_name;
};
int ndbcluster_commit(handlerton *, THD *thd, bool all);
class ha_ndbcluster : public handler, public Partition_handler {
friend class ndb_pushed_builder_ctx;
public:
ha_ndbcluster(handlerton *hton, TABLE_SHARE *table);
~ha_ndbcluster() override;
std::string explain_extra() const override;
int open(const char *name, int mode, uint test_if_locked,
const dd::Table *table_def) override;
private:
void local_close(THD *thd, bool release_metadata);
public:
int close(void) override;
int optimize(THD *thd, HA_CHECK_OPT *) override;
private:
int analyze_index();
public:
int analyze(THD *thd, HA_CHECK_OPT *) override;
int write_row(uchar *buf) override;
int update_row(const uchar *old_data, uchar *new_data) override;
int delete_row(const uchar *buf) override;
int index_init(uint index, bool sorted) override;
int index_end() override;
int index_read(uchar *buf, const uchar *key, uint key_len,
enum ha_rkey_function find_flag) override;
int index_next(uchar *buf) override;
int index_prev(uchar *buf) override;
int index_first(uchar *buf) override;
int index_last(uchar *buf) override;
int index_next_same(uchar *buf, const uchar *key, uint keylen) override;
int index_read_last(uchar *buf, const uchar *key, uint key_len) override;
int rnd_init(bool scan) override;
int rnd_end() override;
int rnd_next(uchar *buf) override;
int rnd_pos(uchar *buf, uchar *pos) override;
void position(const uchar *record) override;
int cmp_ref(const uchar *ref1, const uchar *ref2) const override;
private:
int read_range_first_to_buf(const key_range *start_key,
const key_range *end_key, bool eq_range,
bool sorted, uchar *buf);
public:
int read_range_first(const key_range *start_key, const key_range *end_key,
bool eq_range, bool sorted) override;
int read_range_next() override;
/**
* Multi Range Read interface
*/
int multi_range_read_init(RANGE_SEQ_IF *seq, void *seq_init_param,
uint n_ranges, uint mode,
HANDLER_BUFFER *buf) override;
int multi_range_read_next(char **range_info) override;
ha_rows multi_range_read_info_const(uint keyno, RANGE_SEQ_IF *seq,
void *seq_init_param, uint n_ranges,
uint *bufsz, uint *flags,
Cost_estimate *cost) override;
ha_rows multi_range_read_info(uint keyno, uint n_ranges, uint keys,
uint *bufsz, uint *flags,
Cost_estimate *cost) override;
void append_create_info(String *packet) override;
/* Get partition row type
@param[in] table partition table
@param[in] part_id Id of partition for which row type to be retrieved
@return Partition row type. */
enum row_type get_partition_row_type(const dd::Table *table_def,
uint part_id) override;
private:
bool choose_mrr_impl(uint keyno, uint n_ranges, ha_rows n_rows, uint *bufsz,
uint *flags, Cost_estimate *);
private:
uint first_running_range;
uint first_range_in_batch;
uint first_unstarted_range;
int multi_range_start_retrievals(uint first_range);
public:
bool get_error_message(int error, String *buf) override;
int records(ha_rows *num_rows) override;
ha_rows estimate_rows_upper_bound() override { return HA_POS_ERROR; }
int info(uint) override;
uint32 calculate_key_hash_value(Field **field_array) override;
bool start_read_removal(void) override;
ha_rows end_read_removal(void) override;
int extra(enum ha_extra_function operation) override;
int reset() override;
int external_lock(THD *thd, int lock_type) override;
void unlock_row() override;
int start_stmt(THD *thd, thr_lock_type) override;
void update_create_info(HA_CREATE_INFO *create_info) override;
private:
void update_comment_info(THD *thd, HA_CREATE_INFO *create_info,
const NdbDictionary::Table *tab);
public:
void print_error(int error, myf errflag) override;
const char *table_type() const override;
ulonglong table_flags(void) const override;
ulong index_flags(uint idx, uint part, bool all_parts) const override;
bool primary_key_is_clustered() const override;
uint max_supported_keys() const override;
uint max_supported_key_parts() const override;
uint max_supported_key_length() const override;
uint max_supported_key_part_length(
HA_CREATE_INFO *create_info) const override;
private:
int get_child_or_parent_fk_list(List<FOREIGN_KEY_INFO> *f_key_list,
bool is_child, bool is_parent);
public:
int get_foreign_key_list(THD *thd,
List<FOREIGN_KEY_INFO> *f_key_list) override;
int get_parent_foreign_key_list(THD *thd,
List<FOREIGN_KEY_INFO> *f_key_list) override;
uint referenced_by_foreign_key() override;
char *get_foreign_key_create_info() override;
void free_foreign_key_create_info(char *str) override;
int rename_table(const char *from, const char *to,
const dd::Table *from_table_def,
dd::Table *to_table_def) override;
int delete_table(const char *name, const dd::Table *table_def) override;
bool upgrade_table(THD *thd, const char *, const char *table_name,
dd::Table *dd_table) override;
row_type get_real_row_type(const HA_CREATE_INFO *create_info) const override {
DBUG_TRACE;
// ROW_RORMAT=FIXED -> using FIXED
if (create_info->row_type == ROW_TYPE_FIXED) return ROW_TYPE_FIXED;
// All other values uses DYNAMIC
return ROW_TYPE_DYNAMIC;
}
int create(const char *name, TABLE *form, HA_CREATE_INFO *info,
dd::Table *table_def) override;
int truncate(dd::Table *table_def) override;
bool is_ignorable_error(int error) override {
if (handler::is_ignorable_error(error) ||
error == HA_ERR_NO_PARTITION_FOUND)
return true;
return false;
}
THR_LOCK_DATA **store_lock(THD *thd, THR_LOCK_DATA **to,
enum thr_lock_type lock_type) override;
bool low_byte_first() const override;
enum ha_key_alg get_default_index_algorithm() const override {
/* NDB uses hash indexes only when explicitly requested. */
return HA_KEY_ALG_BTREE;
}
bool is_index_algorithm_supported(enum ha_key_alg key_alg) const override {
return key_alg == HA_KEY_ALG_BTREE || key_alg == HA_KEY_ALG_HASH;
}
double scan_time() override;
ha_rows records_in_range(uint inx, key_range *min_key,
key_range *max_key) override;
void start_bulk_insert(ha_rows rows) override;
int end_bulk_insert() override;
bool start_bulk_update() override;
int bulk_update_row(const uchar *old_data, uchar *new_data,
uint *dup_key_found) override;
int exec_bulk_update(uint *dup_key_found) override;
void end_bulk_update() override;
private:
int ndb_update_row(const uchar *old_data, uchar *new_data,
int is_bulk_update);
public:
static void set_dbname(const char *pathname, char *dbname);
static void set_tabname(const char *pathname, char *tabname);
/*
static member function as it needs to access private
NdbTransaction methods
*/
static void release_completed_operations(NdbTransaction *);
/*
Condition pushdown
*/
/*
Push condition down to the table handler.
SYNOPSIS
cond_push()
cond Condition to be pushed. The condition tree must not be
modified by the by the caller.
other_tbls_ok Are other tables allowed to be referred
from the condition terms pushed down.
RETURN
The 'remainder' condition that caller must use to filter out records.
NULL means the handler will not return rows that do not match the
passed condition.
NOTES
The table handler filters out rows using (pushed_cond1 AND pushed_cond2
AND ... AND pushed_condN)
or less restrictive condition, depending on handler's capabilities.
handler->reset() call discard any pushed conditions.
Calls to rnd_init/rnd_end, index_init/index_end etc do not affect
any condition being pushed.
The current implementation supports arbitrary AND/OR nested conditions
with comparisons between columns and constants (including constant
expressions and function calls) and the following comparison operators:
=, !=, >, >=, <, <=, like, "not like", "is null", and "is not null".
Negated conditions are supported by NOT which generate NAND/NOR groups.
*/
const Item *cond_push(const Item *cond, bool other_tbls_ok) override;
public:
/**
* Generate the ScanFilters code for the condition(s) previously
* accepted for cond_push'ing.
* If code generation failed, the handler will evaluate the
* condition for every row returned from NDB.
*/
void generate_scan_filter(NdbInterpretedCode *code,
NdbScanOperation::ScanOptions *options);
/**
* Generate a ScanFilter using both the pushed condition AND
* add equality predicates matching the 'key' supplied as
* arguments.
* @return 1 if generation of the key part failed.
*/
int generate_scan_filter_with_key(NdbInterpretedCode *code,
NdbScanOperation::ScanOptions *options,
const KEY *key_info,
const key_range *start_key,
const key_range *end_key);
private:
bool maybe_pushable_join(const char *&reason) const;
public:
int assign_pushed_join(const ndb_pushed_join *pushed_join);
uint number_of_pushed_joins() const override;
const TABLE *member_of_pushed_join() const override;
const TABLE *parent_of_pushed_join() const override;
int index_read_pushed(uchar *buf, const uchar *key,
key_part_map keypart_map) override;
int index_next_pushed(uchar *buf) override;
/*
* Internal to ha_ndbcluster, used by C functions
*/
int ndb_err(NdbTransaction *);
enum_alter_inplace_result check_if_supported_inplace_alter(
TABLE *altered_table, Alter_inplace_info *ha_alter_info) override;
private:
bool parse_comment_changes(NdbDictionary::Table *new_tab,
const NdbDictionary::Table *old_tab,
HA_CREATE_INFO *create_info, THD *thd,
bool &max_rows_changed) const;
public:
bool prepare_inplace_alter_table(TABLE *altered_table,
Alter_inplace_info *ha_alter_info,
const dd::Table *old_table_def,
dd::Table *new_table_def) override;
bool inplace_alter_table(TABLE *altered_table,
Alter_inplace_info *ha_alter_info,
const dd::Table *old_table_def,
dd::Table *new_table_def) override;
bool commit_inplace_alter_table(TABLE *altered_table,
Alter_inplace_info *ha_alter_info,
bool commit, const dd::Table *old_table_def,
dd::Table *new_table_def) override;
void notify_table_changed(Alter_inplace_info *alter_info) override;
private:
void prepare_inplace__drop_index(uint key_num);
int inplace__final_drop_index(TABLE *table_arg);
enum_alter_inplace_result supported_inplace_field_change(Alter_inplace_info *,
Field *, Field *,
bool, bool) const;
bool table_storage_changed(HA_CREATE_INFO *) const;
bool column_has_index(TABLE *, uint, uint, uint) const;
enum_alter_inplace_result supported_inplace_ndb_column_change(
uint, TABLE *, Alter_inplace_info *, bool, bool) const;
enum_alter_inplace_result supported_inplace_column_change(
THD *, TABLE *, uint, Field *, Alter_inplace_info *) const;
enum_alter_inplace_result check_inplace_alter_supported(
TABLE *altered_table, Alter_inplace_info *ha_alter_info);
void check_implicit_column_format_change(
TABLE *altered_table, Alter_inplace_info *ha_alter_info) const;
bool abort_inplace_alter_table(TABLE *altered_table,
Alter_inplace_info *ha_alter_info);
int prepare_conflict_detection(
enum_conflicting_op_type op_type, const NdbRecord *key_rec,
const NdbRecord *data_rec, const uchar *old_data, const uchar *new_data,
const MY_BITMAP *write_set, NdbTransaction *trans,
NdbInterpretedCode *code, NdbOperation::OperationOptions *options,
bool &conflict_handled, bool &avoid_ndbapi_write);
void setup_key_ref_for_ndb_record(const NdbRecord **key_rec,
const uchar **key_row, const uchar *record,
bool use_active_index);
void check_read_before_write_removal();
int prepare_inplace__add_index(THD *thd, KEY *key_info,
uint num_of_keys) const;
int create_index_in_NDB(THD *thd, const char *name, KEY *key_info,
const NdbDictionary::Table *ndbtab,
bool unique) const;
int create_index(THD *thd, const char *name, KEY *key_info,
NDB_INDEX_TYPE idx_type,
const NdbDictionary::Table *ndbtab) const;
// Index list management
int create_indexes(THD *thd, TABLE *tab,
const NdbDictionary::Table *ndbtab) const;
int open_indexes(Ndb *ndb, TABLE *tab);
void release_indexes(NdbDictionary::Dictionary *dict, int invalidate);
void inplace__renumber_indexes(uint dropped_index_num);
int inplace__drop_indexes(Ndb *ndb, TABLE *tab);
int add_index_handle(NdbDictionary::Dictionary *dict, KEY *key_info,
const char *key_name, uint index_no);
int add_table_ndb_record(NdbDictionary::Dictionary *dict);
int add_hidden_pk_ndb_record(NdbDictionary::Dictionary *dict);
int add_index_ndb_record(NdbDictionary::Dictionary *dict, KEY *key_info,
uint index_no);
int get_fk_data(THD *thd, Ndb *ndb);
void release_fk_data();
int create_fks(THD *thd, Ndb *ndb);
int copy_fk_for_offline_alter(THD *thd, Ndb *, const char *tabname);
int inplace__drop_fks(THD *, Ndb *, NdbDictionary::Dictionary *,
const NdbDictionary::Table *);
static int get_fk_data_for_truncate(NdbDictionary::Dictionary *,
const NdbDictionary::Table *,
Ndb_fk_list &);
static int recreate_fk_for_truncate(THD *, Ndb *, const char *,
Ndb_fk_list &);
bool has_fk_dependency(THD *, const NdbDictionary::Column *) const;
int check_default_values(const NdbDictionary::Table *ndbtab);
int get_metadata(THD *thd, const dd::Table *table_def);
void release_metadata(THD *thd, Ndb *ndb);
NDB_INDEX_TYPE get_index_type(uint idx_no) const;
NDB_INDEX_TYPE get_index_type_from_table(uint index_no) const;
NDB_INDEX_TYPE get_index_type_from_key(uint index_no, KEY *key_info,
bool primary) const;
bool has_null_in_unique_index(uint idx_no) const;
bool check_index_fields_not_null(KEY *key_info) const;
bool check_if_pushable(int type, // NdbQueryOperationDef::Type,
uint idx = MAX_KEY) const;
bool check_is_pushed() const;
int create_pushed_join(const NdbQueryParamValue *keyFieldParams = NULL,
uint paramCnt = 0);
int ndb_pk_update_row(THD *thd, const uchar *old_data, uchar *new_data);
int pk_read(const uchar *key, uchar *buf, uint32 *part_id);
int ordered_index_scan(const key_range *start_key, const key_range *end_key,
bool sorted, bool descending, uchar *buf,
part_id_range *part_spec);
int unique_index_read(const uchar *key, uchar *buf);
int full_table_scan(const KEY *key_info, const key_range *start_key,
const key_range *end_key, uchar *buf);
int flush_bulk_insert(bool allow_batch = false);
int ndb_write_row(uchar *record, bool primary_key_update,
bool batched_update);
bool start_bulk_delete() override;
int end_bulk_delete() override;
int ndb_delete_row(const uchar *record, bool primary_key_update);
int ndb_optimize_table(THD *thd, uint delay) const;
bool check_all_operations_for_error(NdbTransaction *trans,
const NdbOperation *first,
const NdbOperation *last, uint errcode);
enum NDB_WRITE_OP { NDB_INSERT = 0, NDB_UPDATE = 1, NDB_PK_UPDATE = 2 };
int peek_indexed_rows(const uchar *record, NDB_WRITE_OP write_op);
int scan_handle_lock_tuple(NdbScanOperation *scanOp, NdbTransaction *trans);
int fetch_next(NdbScanOperation *op);
int fetch_next_pushed();
int set_auto_inc(THD *thd, Field *field);
int set_auto_inc_val(THD *thd, Uint64 value);
int next_result(uchar *buf);
int close_scan();
int unpack_record(uchar *dst_row, const uchar *src_row);
int unpack_record_and_set_generated_fields(uchar *dst_row,
const uchar *src_row);
void set_dbname(const char *pathname);
void set_tabname(const char *pathname);
const NdbDictionary::Column *get_hidden_key_column() {
return m_table->getColumn(m_table_map->get_hidden_key_column());
}
const NdbDictionary::Column *get_partition_id_column() {
return m_table->getColumn(m_table_map->get_partition_id_column());
}
uchar *get_buffer(Thd_ndb *thd_ndb, uint size);
uchar *copy_row_to_buffer(Thd_ndb *thd_ndb, const uchar *record);
int get_blob_values(const NdbOperation *ndb_op, uchar *dst_record,
const MY_BITMAP *bitmap);
int set_blob_values(const NdbOperation *ndb_op, ptrdiff_t row_offset,
const MY_BITMAP *bitmap, uint *set_count, bool batch);
friend int g_get_ndb_blobs_value(NdbBlob *ndb_blob, void *arg);
void release_blobs_buffer();
Uint32 setup_get_hidden_fields(NdbOperation::GetValueSpec gets[2]);
void get_hidden_fields_keyop(NdbOperation::OperationOptions *options,
NdbOperation::GetValueSpec gets[2]);
void get_hidden_fields_scan(NdbScanOperation::ScanOptions *options,
NdbOperation::GetValueSpec gets[2]);
void get_read_set(bool use_cursor, uint idx);
void eventSetAnyValue(THD *thd,
NdbOperation::OperationOptions *options) const;
bool check_index_fields_in_write_set(uint keyno);
int log_exclusive_read(const NdbRecord *key_rec, const uchar *key, uchar *buf,
Uint32 *ppartition_id);
int scan_log_exclusive_read(NdbScanOperation *, NdbTransaction *);
const NdbOperation *pk_unique_index_read_key(uint idx, const uchar *key,
uchar *buf,
NdbOperation::LockMode lm,
Uint32 *ppartition_id);
int pk_unique_index_read_key_pushed(uint idx, const uchar *key);
int read_multi_range_fetch_next();
int primary_key_cmp(const uchar *old_row, const uchar *new_row);
void get_auto_increment(ulonglong offset, ulonglong increment,
ulonglong number_of_desired_values,
ulonglong *first_value,
ulonglong *nb_reserved_values) override;
bool uses_blob_value(const MY_BITMAP *bitmap) const;
int check_ndb_connection(THD *thd) const;
void set_rec_per_key();
void no_uncommitted_rows_execute_failure();
void no_uncommitted_rows_update(int);
/* Ordered index statistics v4 */
int ndb_index_stat_query(uint inx, const key_range *min_key,
const key_range *max_key, NdbIndexStat::Stat &stat,
int from);
int ndb_index_stat_get_rir(uint inx, key_range *min_key, key_range *max_key,
ha_rows *rows_out);
int ndb_index_stat_set_rpk(uint inx);
int ndb_index_stat_analyze(uint *inx_list, uint inx_count);
NdbTransaction *start_transaction_part_id(uint32 part_id, int &error);
inline NdbTransaction *get_transaction_part_id(uint32 part_id, int &error) {
if (m_thd_ndb->trans) return m_thd_ndb->trans;
return start_transaction_part_id(part_id, error);
}
NdbTransaction *start_transaction(int &error);
inline NdbTransaction *get_transaction(int &error) {
if (m_thd_ndb->trans) return m_thd_ndb->trans;
return start_transaction(error);
}
NdbTransaction *start_transaction_row(const NdbRecord *ndb_record,
const uchar *record, int &error);
NdbTransaction *start_transaction_key(uint index, const uchar *key_data,
int &error);
friend int check_completed_operations_pre_commit(Thd_ndb *, NdbTransaction *,
const NdbOperation *,
uint *ignore_count);
friend int ndbcluster_commit(handlerton *, THD *thd, bool all);
int start_statement(THD *thd, Thd_ndb *thd_ndb, uint table_count);
int init_handler_for_statement(THD *thd);
/*
Implementing Partition_handler API.
*/
Partition_handler *get_partition_handler() override {
return static_cast<Partition_handler *>(this);
}
uint alter_flags(uint flags) const override;
void get_dynamic_partition_info(ha_statistics *stat_info,
ha_checksum *checksum, uint part_id) override;
int get_default_num_partitions(HA_CREATE_INFO *info) override;
bool get_num_parts(const char *name, uint *num_parts) override;
void set_auto_partitions(partition_info *part_info) override;
void set_part_info(partition_info *part_info, bool early) override;
/* End of Partition_handler API */
Ndb_table_map *m_table_map;
Thd_ndb *m_thd_ndb;
NdbScanOperation *m_active_cursor;
const NdbDictionary::Table *m_table;
/*
Normal NdbRecord for accessing rows, with all fields including hidden
fields (hidden primary key, user-defined partitioning function value).
*/
NdbRecord *m_ndb_record;
/* NdbRecord for accessing tuple by hidden Uint64 primary key. */
NdbRecord *m_ndb_hidden_key_record;
/* Bitmap used for NdbRecord operation column mask. */
MY_BITMAP m_bitmap;
my_bitmap_map
m_bitmap_buf[(NDB_MAX_ATTRIBUTES_IN_TABLE + 8 * sizeof(my_bitmap_map) -
1) /
(8 * sizeof(my_bitmap_map))]; // Buffer for m_bitmap
/* Bitmap with bit set for all primary key columns. */
MY_BITMAP *m_pk_bitmap_p;
my_bitmap_map
m_pk_bitmap_buf[(NDB_MAX_ATTRIBUTES_IN_TABLE + 8 * sizeof(my_bitmap_map) -
1) /
(8 * sizeof(my_bitmap_map))]; // Buffer for m_pk_bitmap
struct Ndb_local_table_statistics *m_table_info;
struct Ndb_local_table_statistics m_table_info_instance;
char m_dbname[FN_HEADLEN];
// char m_schemaname[FN_HEADLEN];
char m_tabname[FN_HEADLEN];
THR_LOCK_DATA m_lock;
bool m_lock_tuple;
NDB_SHARE *m_share;
NDB_INDEX_DATA m_index[MAX_KEY];
static const size_t fk_root_block_size = 1024;
MEM_ROOT m_fk_mem_root;
struct Ndb_fk_data *m_fk_data;
/*
Pointer to row returned from scan nextResult().
*/
union {
const char *_m_next_row;
const uchar *m_next_row;
};
/* For read_multi_range scans, the get_range_no() of current row. */
int m_current_range_no;
/* For multi range read, return from last mrr_funcs.next() call. */
int m_range_res;
MY_BITMAP **m_key_fields;
// NdbRecAttr has no reference to blob
NdbValue m_value[NDB_MAX_ATTRIBUTES_IN_TABLE];
Uint64 m_ref;
partition_info *m_part_info;
uint32 m_part_id;
bool m_user_defined_partitioning;
bool m_use_partition_pruning;
bool m_sorted;
bool m_use_write;
bool m_ignore_dup_key;
bool m_has_unique_index;
bool m_ignore_no_key;
bool m_read_before_write_removal_possible;
bool m_read_before_write_removal_used;
ha_rows m_rows_updated;
ha_rows m_rows_deleted;
ha_rows m_rows_to_insert; // TODO: merge it with
// handler::estimation_rows_to_insert?
ha_rows m_rows_inserted;
bool m_delete_cannot_batch;
bool m_update_cannot_batch;
uint m_bytes_per_write;
bool m_skip_auto_increment;
bool m_blobs_pending;
bool m_slow_path;
bool m_is_bulk_delete;
/* State for setActiveHook() callback for reading blob data. */
uint m_blob_counter;
uint m_blob_expected_count_per_row;
uchar *m_blob_destination_record;
Uint64 m_blobs_row_total_size; /* Bytes needed for all blobs in current row */
// memory for blobs in one tuple
uchar *m_blobs_buffer;
Uint64 m_blobs_buffer_size;
uint m_dupkey;
// set from thread variables at external lock
ha_rows m_autoincrement_prefetch;
// Joins pushed to NDB.
const ndb_pushed_join
*m_pushed_join_member; // Pushed join def. I am member of
int m_pushed_join_operation; // Op. id. in above pushed join
static const int PUSHED_ROOT = 0; // Op. id. if I'm root
bool m_disable_pushed_join; // Pushed execution allowed?
NdbQuery *m_active_query; // Pushed query instance executing
NdbQueryOperation *m_pushed_operation; // Pushed operation instance
/* In case we failed to push a 'pushed_cond', the handler will evaluate it */
ha_ndbcluster_cond m_cond;
bool m_disable_multi_read;
uchar *m_multi_range_result_ptr;
NdbIndexScanOperation *m_multi_cursor;
Ndb *get_ndb(THD *thd) const;
int update_stats(THD *thd, bool do_read_stat, uint part_id = ~(uint)0);
int add_handler_to_open_tables(THD *, Thd_ndb *, ha_ndbcluster *handler);
};
// Global handler synchronization
extern mysql_mutex_t ndbcluster_mutex;
extern mysql_cond_t ndbcluster_cond;
extern int ndb_setup_complete;
static const int NDB_INVALID_SCHEMA_OBJECT = 241;
int ndb_to_mysql_error(const NdbError *ndberr);
#endif