Merge mysql.com:/home/gluh/MySQL/Merge/5.1
into mysql.com:/home/gluh/MySQL/Merge/5.1-opt
This commit is contained in:
commit
53c5511a9b
@ -80,6 +80,7 @@ extern FILE *_db_fp_(void);
|
||||
#define DBUG_ASSERT(A) assert(A)
|
||||
#define DBUG_EXPLAIN(buf,len) _db_explain_(0, (buf),(len))
|
||||
#define DBUG_EXPLAIN_INITIAL(buf,len) _db_explain_init_((buf),(len))
|
||||
#define IF_DBUG(A) A
|
||||
#else /* No debugger */
|
||||
|
||||
#define DBUG_ENTER(a1)
|
||||
@ -106,6 +107,7 @@ extern FILE *_db_fp_(void);
|
||||
#define DBUG_UNLOCK_FILE
|
||||
#define DBUG_EXPLAIN(buf,len)
|
||||
#define DBUG_EXPLAIN_INITIAL(buf,len)
|
||||
#define IF_DBUG(A)
|
||||
#endif
|
||||
#ifdef __cplusplus
|
||||
}
|
||||
|
@ -409,24 +409,24 @@ sub main () {
|
||||
{
|
||||
next if $test->{skip};
|
||||
|
||||
$need_ndbcluster||= $test->{ndb_test};
|
||||
$need_im||= $test->{component_id} eq 'im';
|
||||
|
||||
# Count max number of slaves used by a test case
|
||||
if ( $test->{slave_num} > $max_slave_num)
|
||||
if (!$opt_extern)
|
||||
{
|
||||
$max_slave_num= $test->{slave_num};
|
||||
mtr_error("Too many slaves") if $max_slave_num > 3;
|
||||
}
|
||||
$need_ndbcluster||= $test->{ndb_test};
|
||||
$need_im||= $test->{component_id} eq 'im';
|
||||
|
||||
# Count max number of masters used by a test case
|
||||
if ( $test->{master_num} > $max_master_num)
|
||||
{
|
||||
$max_master_num= $test->{master_num};
|
||||
mtr_error("Too many masters") if $max_master_num > 2;
|
||||
mtr_error("Too few masters") if $max_master_num < 1;
|
||||
}
|
||||
# Count max number of slaves used by a test case
|
||||
if ( $test->{slave_num} > $max_slave_num) {
|
||||
$max_slave_num= $test->{slave_num};
|
||||
mtr_error("Too many slaves") if $max_slave_num > 3;
|
||||
}
|
||||
|
||||
# Count max number of masters used by a test case
|
||||
if ( $test->{master_num} > $max_master_num) {
|
||||
$max_master_num= $test->{master_num};
|
||||
mtr_error("Too many masters") if $max_master_num > 2;
|
||||
mtr_error("Too few masters") if $max_master_num < 1;
|
||||
}
|
||||
}
|
||||
$use_innodb||= $test->{'innodb_test'};
|
||||
}
|
||||
|
||||
|
@ -24,8 +24,8 @@
|
||||
|
||||
User::User(const LEX_STRING *user_name_arg, const char *password)
|
||||
{
|
||||
user_length= strmake(user, user_name_arg->str, USERNAME_LENGTH + 1) - user;
|
||||
|
||||
user_length= (uint8) (strmake(user, user_name_arg->str,
|
||||
USERNAME_LENGTH + 1) - user);
|
||||
set_password(password);
|
||||
}
|
||||
|
||||
@ -59,7 +59,7 @@ int User::init(const char *line)
|
||||
password= name_end + 1;
|
||||
}
|
||||
|
||||
user_length= name_end - name_begin;
|
||||
user_length= (uint8) (name_end - name_begin);
|
||||
if (user_length > USERNAME_LENGTH)
|
||||
{
|
||||
log_error("User name is too long (%d). Max length: %d. "
|
||||
@ -70,7 +70,7 @@ int User::init(const char *line)
|
||||
return 1;
|
||||
}
|
||||
|
||||
password_length= strlen(password);
|
||||
password_length= (int) strlen(password);
|
||||
if (password_length > SCRAMBLED_PASSWORD_CHAR_LENGTH)
|
||||
{
|
||||
log_error("Password is too long (%d). Max length: %d."
|
||||
|
@ -6297,9 +6297,9 @@ int Field_string::cmp(const char *a_ptr, const char *b_ptr)
|
||||
|
||||
void Field_string::sort_string(char *to,uint length)
|
||||
{
|
||||
uint tmp= my_strnxfrm(field_charset,
|
||||
(uchar*) to, length,
|
||||
(uchar*) ptr, field_length);
|
||||
IF_DBUG(uint tmp=) my_strnxfrm(field_charset,
|
||||
(uchar*) to, length,
|
||||
(uchar*) ptr, field_length);
|
||||
DBUG_ASSERT(tmp == length);
|
||||
}
|
||||
|
||||
|
@ -2764,10 +2764,12 @@ int ha_ndbcluster::write_row(byte *record)
|
||||
{
|
||||
Ndb *ndb= get_ndb();
|
||||
Uint64 next_val= (Uint64) table->next_number_field->val_int() + 1;
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22];
|
||||
DBUG_PRINT("info",
|
||||
("Trying to set next auto increment value to %s",
|
||||
llstr(next_val, buff)));
|
||||
#endif
|
||||
Ndb_tuple_id_range_guard g(m_share);
|
||||
if (ndb->setAutoIncrementValue(m_table, g.range, next_val, TRUE)
|
||||
== -1)
|
||||
@ -3999,7 +4001,7 @@ int ha_ndbcluster::end_bulk_insert()
|
||||
}
|
||||
else
|
||||
{
|
||||
int res= trans->restart();
|
||||
IF_DBUG(int res=) trans->restart();
|
||||
DBUG_ASSERT(res == 0);
|
||||
}
|
||||
}
|
||||
@ -4717,7 +4719,9 @@ static int create_ndb_column(NDBCOL &col,
|
||||
// Set autoincrement
|
||||
if (field->flags & AUTO_INCREMENT_FLAG)
|
||||
{
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22];
|
||||
#endif
|
||||
col.setAutoIncrement(TRUE);
|
||||
ulonglong value= info->auto_increment_value ?
|
||||
info->auto_increment_value : (ulonglong) 1;
|
||||
@ -5388,7 +5392,7 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
|
||||
{
|
||||
DBUG_PRINT("NDB_SHARE", ("%s temporary use_count: %u",
|
||||
share->key, share->use_count));
|
||||
int r= rename_share(share, to);
|
||||
IF_DBUG(int r=) rename_share(share, to);
|
||||
DBUG_ASSERT(r == 0);
|
||||
}
|
||||
#endif
|
||||
@ -5409,8 +5413,8 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
|
||||
#ifdef HAVE_NDB_BINLOG
|
||||
if (share)
|
||||
{
|
||||
int r= rename_share(share, from);
|
||||
DBUG_ASSERT(r == 0);
|
||||
IF_DBUG(int ret=) rename_share(share, from);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
/* ndb_share reference temporary free */
|
||||
DBUG_PRINT("NDB_SHARE", ("%s temporary free use_count: %u",
|
||||
share->key, share->use_count));
|
||||
@ -7268,7 +7272,9 @@ uint ndb_get_commitcount(THD *thd, char *dbname, char *tabname,
|
||||
if (share->commit_count != 0)
|
||||
{
|
||||
*commit_count= share->commit_count;
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22];
|
||||
#endif
|
||||
DBUG_PRINT("info", ("Getting commit_count: %s from share",
|
||||
llstr(share->commit_count, buff)));
|
||||
pthread_mutex_unlock(&share->mutex);
|
||||
@ -7304,7 +7310,9 @@ uint ndb_get_commitcount(THD *thd, char *dbname, char *tabname,
|
||||
pthread_mutex_lock(&share->mutex);
|
||||
if (share->commit_count_lock == lock)
|
||||
{
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22];
|
||||
#endif
|
||||
DBUG_PRINT("info", ("Setting commit_count to %s",
|
||||
llstr(stat.commit_count, buff)));
|
||||
share->commit_count= stat.commit_count;
|
||||
@ -7363,7 +7371,9 @@ ndbcluster_cache_retrieval_allowed(THD *thd,
|
||||
bool is_autocommit= !(thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN));
|
||||
char *dbname= full_name;
|
||||
char *tabname= dbname+strlen(dbname)+1;
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22], buff2[22];
|
||||
#endif
|
||||
DBUG_ENTER("ndbcluster_cache_retrieval_allowed");
|
||||
DBUG_PRINT("enter", ("dbname: %s, tabname: %s, is_autocommit: %d",
|
||||
dbname, tabname, is_autocommit));
|
||||
@ -7430,7 +7440,9 @@ ha_ndbcluster::register_query_cache_table(THD *thd,
|
||||
ulonglong *engine_data)
|
||||
{
|
||||
Uint64 commit_count;
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22];
|
||||
#endif
|
||||
bool is_autocommit= !(thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN));
|
||||
DBUG_ENTER("ha_ndbcluster::register_query_cache_table");
|
||||
DBUG_PRINT("enter",("dbname: %s, tabname: %s, is_autocommit: %d",
|
||||
@ -7875,7 +7887,9 @@ ndb_get_table_statistics(ha_ndbcluster* file, bool report_error, Ndb* ndb, const
|
||||
int retries= 10;
|
||||
int reterr= 0;
|
||||
int retry_sleep= 30 * 1000; /* 30 milliseconds */
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22], buff2[22], buff3[22], buff4[22];
|
||||
#endif
|
||||
DBUG_ENTER("ndb_get_table_statistics");
|
||||
DBUG_PRINT("enter", ("table: %s", ndbtab->getName()));
|
||||
|
||||
@ -8693,7 +8707,9 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused)))
|
||||
ndb_get_table_statistics(NULL, FALSE, ndb,
|
||||
ndbtab_g.get_table(), &stat) == 0)
|
||||
{
|
||||
#ifndef DBUG_OFF
|
||||
char buff[22], buff2[22];
|
||||
#endif
|
||||
DBUG_PRINT("info",
|
||||
("Table: %s commit_count: %s rows: %s",
|
||||
share->key,
|
||||
@ -9545,8 +9561,8 @@ void ndb_serialize_cond(const Item *item, void *arg)
|
||||
DBUG_PRINT("info", ("INT_ITEM"));
|
||||
if (context->expecting(Item::INT_ITEM))
|
||||
{
|
||||
Item_int *int_item= (Item_int *) item;
|
||||
DBUG_PRINT("info", ("value %ld", (long) int_item->value));
|
||||
DBUG_PRINT("info", ("value %ld",
|
||||
(long) ((Item_int*) item)->value));
|
||||
NDB_ITEM_QUALIFICATION q;
|
||||
q.value_type= Item::INT_ITEM;
|
||||
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
||||
@ -9572,8 +9588,7 @@ void ndb_serialize_cond(const Item *item, void *arg)
|
||||
DBUG_PRINT("info", ("REAL_ITEM"));
|
||||
if (context->expecting(Item::REAL_ITEM))
|
||||
{
|
||||
Item_float *float_item= (Item_float *) item;
|
||||
DBUG_PRINT("info", ("value %f", float_item->value));
|
||||
DBUG_PRINT("info", ("value %f", ((Item_float*) item)->value));
|
||||
NDB_ITEM_QUALIFICATION q;
|
||||
q.value_type= Item::REAL_ITEM;
|
||||
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
||||
@ -9620,8 +9635,8 @@ void ndb_serialize_cond(const Item *item, void *arg)
|
||||
DBUG_PRINT("info", ("DECIMAL_ITEM"));
|
||||
if (context->expecting(Item::DECIMAL_ITEM))
|
||||
{
|
||||
Item_decimal *decimal_item= (Item_decimal *) item;
|
||||
DBUG_PRINT("info", ("value %f", decimal_item->val_real()));
|
||||
DBUG_PRINT("info", ("value %f",
|
||||
((Item_decimal*) item)->val_real()));
|
||||
NDB_ITEM_QUALIFICATION q;
|
||||
q.value_type= Item::DECIMAL_ITEM;
|
||||
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
||||
|
@ -1829,15 +1829,15 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
|
||||
// fall through
|
||||
case SOT_CREATE_TABLE:
|
||||
pthread_mutex_lock(&LOCK_open);
|
||||
if (ndbcluster_check_if_local_table(schema->db, schema->name))
|
||||
{
|
||||
DBUG_PRINT("info", ("NDB binlog: Skipping locally defined table '%s.%s'",
|
||||
schema->db, schema->name));
|
||||
if (ndbcluster_check_if_local_table(schema->db, schema->name))
|
||||
{
|
||||
DBUG_PRINT("info", ("NDB binlog: Skipping locally defined table '%s.%s'",
|
||||
schema->db, schema->name));
|
||||
sql_print_error("NDB binlog: Skipping locally defined table '%s.%s' from "
|
||||
"binlog schema event '%s' from node %d. ",
|
||||
schema->db, schema->name, schema->query,
|
||||
schema->node_id);
|
||||
}
|
||||
}
|
||||
else if (ndb_create_table_from_engine(thd, schema->db, schema->name))
|
||||
{
|
||||
sql_print_error("NDB binlog: Could not discover table '%s.%s' from "
|
||||
@ -1854,27 +1854,27 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
|
||||
log_query= 1;
|
||||
break;
|
||||
case SOT_DROP_DB:
|
||||
/* Drop the database locally if it only contains ndb tables */
|
||||
if (! ndbcluster_check_if_local_tables_in_db(thd, schema->db))
|
||||
{
|
||||
run_query(thd, schema->query,
|
||||
schema->query + schema->query_length,
|
||||
TRUE, /* print error */
|
||||
TRUE); /* don't binlog the query */
|
||||
/* binlog dropping database after any table operations */
|
||||
post_epoch_log_list->push_back(schema, mem_root);
|
||||
/* acknowledge this query _after_ epoch completion */
|
||||
post_epoch_unlock= 1;
|
||||
}
|
||||
else
|
||||
{
|
||||
/* Database contained local tables, leave it */
|
||||
sql_print_error("NDB binlog: Skipping drop database '%s' since it contained local tables "
|
||||
/* Drop the database locally if it only contains ndb tables */
|
||||
if (! ndbcluster_check_if_local_tables_in_db(thd, schema->db))
|
||||
{
|
||||
run_query(thd, schema->query,
|
||||
schema->query + schema->query_length,
|
||||
TRUE, /* print error */
|
||||
TRUE); /* don't binlog the query */
|
||||
/* binlog dropping database after any table operations */
|
||||
post_epoch_log_list->push_back(schema, mem_root);
|
||||
/* acknowledge this query _after_ epoch completion */
|
||||
post_epoch_unlock= 1;
|
||||
}
|
||||
else
|
||||
{
|
||||
/* Database contained local tables, leave it */
|
||||
sql_print_error("NDB binlog: Skipping drop database '%s' since it contained local tables "
|
||||
"binlog schema event '%s' from node %d. ",
|
||||
schema->db, schema->query,
|
||||
schema->node_id);
|
||||
log_query= 1;
|
||||
}
|
||||
log_query= 1;
|
||||
}
|
||||
break;
|
||||
case SOT_CREATE_DB:
|
||||
/* fall through */
|
||||
@ -2121,18 +2121,18 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
|
||||
share= 0;
|
||||
}
|
||||
pthread_mutex_lock(&LOCK_open);
|
||||
if (ndbcluster_check_if_local_table(schema->db, schema->name))
|
||||
{
|
||||
DBUG_PRINT("info", ("NDB binlog: Skipping locally defined table '%s.%s'",
|
||||
schema->db, schema->name));
|
||||
if (ndbcluster_check_if_local_table(schema->db, schema->name))
|
||||
{
|
||||
DBUG_PRINT("info", ("NDB binlog: Skipping locally defined table '%s.%s'",
|
||||
schema->db, schema->name));
|
||||
sql_print_error("NDB binlog: Skipping locally defined table '%s.%s' from "
|
||||
"binlog schema event '%s' from node %d. ",
|
||||
schema->db, schema->name, schema->query,
|
||||
schema->node_id);
|
||||
}
|
||||
}
|
||||
else if (ndb_create_table_from_engine(thd, schema->db, schema->name))
|
||||
{
|
||||
sql_print_error("NDB binlog: Could not discover table '%s.%s' from "
|
||||
{
|
||||
sql_print_error("NDB binlog: Could not discover table '%s.%s' from "
|
||||
"binlog schema event '%s' from node %d. my_errno: %d",
|
||||
schema->db, schema->name, schema->query,
|
||||
schema->node_id, my_errno);
|
||||
@ -2260,7 +2260,7 @@ int ndb_add_ndb_binlog_index(THD *thd, void *_row)
|
||||
{
|
||||
TABLE_LIST *p_binlog_tables= &binlog_tables;
|
||||
close_tables_for_reopen(thd, &p_binlog_tables);
|
||||
ndb_binlog_index= 0;
|
||||
ndb_binlog_index= 0;
|
||||
continue;
|
||||
}
|
||||
sql_print_error("NDB Binlog: Unable to lock table ndb_binlog_index");
|
||||
@ -3225,15 +3225,17 @@ ndb_binlog_thread_handle_data_event(Ndb *ndb, NdbEventOperation *pOp,
|
||||
if (share->flags & NSF_BLOB_FLAG)
|
||||
{
|
||||
my_ptrdiff_t ptrdiff= 0;
|
||||
int ret= get_ndb_blobs_value(table, share->ndb_value[0],
|
||||
blobs_buffer[0], blobs_buffer_size[0],
|
||||
ptrdiff);
|
||||
IF_DBUG(int ret =) get_ndb_blobs_value(table, share->ndb_value[0],
|
||||
blobs_buffer[0],
|
||||
blobs_buffer_size[0],
|
||||
ptrdiff);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
ndb_unpack_record(table, share->ndb_value[0], &b, table->record[0]);
|
||||
int ret= trans.write_row(::server_id,
|
||||
injector::transaction::table(table, TRUE),
|
||||
&b, n_fields, table->record[0]);
|
||||
IF_DBUG(int ret=) trans.write_row(::server_id,
|
||||
injector::transaction::table(table,
|
||||
TRUE),
|
||||
&b, n_fields, table->record[0]);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
break;
|
||||
@ -3251,27 +3253,29 @@ ndb_binlog_thread_handle_data_event(Ndb *ndb, NdbEventOperation *pOp,
|
||||
n= 0; /*
|
||||
use the primary key only as it save time and space and
|
||||
it is the only thing needed to log the delete
|
||||
*/
|
||||
*/
|
||||
else
|
||||
n= 1; /*
|
||||
we use the before values since we don't have a primary key
|
||||
since the mysql server does not handle the hidden primary
|
||||
key
|
||||
*/
|
||||
*/
|
||||
|
||||
if (share->flags & NSF_BLOB_FLAG)
|
||||
{
|
||||
my_ptrdiff_t ptrdiff= table->record[n] - table->record[0];
|
||||
int ret= get_ndb_blobs_value(table, share->ndb_value[n],
|
||||
blobs_buffer[n], blobs_buffer_size[n],
|
||||
ptrdiff);
|
||||
IF_DBUG(int ret =) get_ndb_blobs_value(table, share->ndb_value[n],
|
||||
blobs_buffer[n],
|
||||
blobs_buffer_size[n],
|
||||
ptrdiff);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
ndb_unpack_record(table, share->ndb_value[n], &b, table->record[n]);
|
||||
DBUG_EXECUTE("info", print_records(table, table->record[n]););
|
||||
int ret= trans.delete_row(::server_id,
|
||||
injector::transaction::table(table, TRUE),
|
||||
&b, n_fields, table->record[n]);
|
||||
IF_DBUG(int ret =) trans.delete_row(::server_id,
|
||||
injector::transaction::table(table,
|
||||
TRUE),
|
||||
&b, n_fields, table->record[n]);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
break;
|
||||
@ -3283,9 +3287,10 @@ ndb_binlog_thread_handle_data_event(Ndb *ndb, NdbEventOperation *pOp,
|
||||
if (share->flags & NSF_BLOB_FLAG)
|
||||
{
|
||||
my_ptrdiff_t ptrdiff= 0;
|
||||
int ret= get_ndb_blobs_value(table, share->ndb_value[0],
|
||||
blobs_buffer[0], blobs_buffer_size[0],
|
||||
ptrdiff);
|
||||
IF_DBUG(int ret =) get_ndb_blobs_value(table, share->ndb_value[0],
|
||||
blobs_buffer[0],
|
||||
blobs_buffer_size[0],
|
||||
ptrdiff);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
ndb_unpack_record(table, share->ndb_value[0],
|
||||
@ -3296,7 +3301,7 @@ ndb_binlog_thread_handle_data_event(Ndb *ndb, NdbEventOperation *pOp,
|
||||
/*
|
||||
since table has a primary key, we can do a write
|
||||
using only after values
|
||||
*/
|
||||
*/
|
||||
trans.write_row(::server_id, injector::transaction::table(table, TRUE),
|
||||
&b, n_fields, table->record[0]);// after values
|
||||
}
|
||||
@ -3305,22 +3310,24 @@ ndb_binlog_thread_handle_data_event(Ndb *ndb, NdbEventOperation *pOp,
|
||||
/*
|
||||
mysql server cannot handle the ndb hidden key and
|
||||
therefore needs the before image as well
|
||||
*/
|
||||
*/
|
||||
if (share->flags & NSF_BLOB_FLAG)
|
||||
{
|
||||
my_ptrdiff_t ptrdiff= table->record[1] - table->record[0];
|
||||
int ret= get_ndb_blobs_value(table, share->ndb_value[1],
|
||||
blobs_buffer[1], blobs_buffer_size[1],
|
||||
ptrdiff);
|
||||
IF_DBUG(int ret =) get_ndb_blobs_value(table, share->ndb_value[1],
|
||||
blobs_buffer[1],
|
||||
blobs_buffer_size[1],
|
||||
ptrdiff);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
ndb_unpack_record(table, share->ndb_value[1], &b, table->record[1]);
|
||||
DBUG_EXECUTE("info", print_records(table, table->record[1]););
|
||||
int ret= trans.update_row(::server_id,
|
||||
injector::transaction::table(table, TRUE),
|
||||
&b, n_fields,
|
||||
table->record[1], // before values
|
||||
table->record[0]);// after values
|
||||
IF_DBUG(int ret =) trans.update_row(::server_id,
|
||||
injector::transaction::table(table,
|
||||
TRUE),
|
||||
&b, n_fields,
|
||||
table->record[1], // before values
|
||||
table->record[0]);// after values
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
}
|
||||
@ -3850,7 +3857,9 @@ restart:
|
||||
continue;
|
||||
}
|
||||
TABLE *table= share->table;
|
||||
#ifndef DBUG_OFF
|
||||
const LEX_STRING &name= table->s->table_name;
|
||||
#endif
|
||||
if ((event_types & (NdbDictionary::Event::TE_INSERT |
|
||||
NdbDictionary::Event::TE_UPDATE |
|
||||
NdbDictionary::Event::TE_DELETE)) == 0)
|
||||
@ -3867,7 +3876,7 @@ restart:
|
||||
}
|
||||
DBUG_PRINT("info", ("use_table: %.*s", name.length, name.str));
|
||||
injector::transaction::table tbl(table, TRUE);
|
||||
int ret= trans.use_table(::server_id, tbl);
|
||||
IF_DBUG(int ret=) trans.use_table(::server_id, tbl);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
}
|
||||
}
|
||||
@ -3877,10 +3886,12 @@ restart:
|
||||
{
|
||||
TABLE *table= ndb_apply_status_share->table;
|
||||
|
||||
const LEX_STRING& name=table->s->table_name;
|
||||
#ifndef DBUG_OFF
|
||||
const LEX_STRING& name= table->s->table_name;
|
||||
DBUG_PRINT("info", ("use_table: %.*s", name.length, name.str));
|
||||
#endif
|
||||
injector::transaction::table tbl(table, TRUE);
|
||||
int ret= trans.use_table(::server_id, tbl);
|
||||
IF_DBUG(int ret=) trans.use_table(::server_id, tbl);
|
||||
DBUG_ASSERT(ret == 0);
|
||||
|
||||
// Set all fields non-null.
|
||||
@ -3945,7 +3956,7 @@ restart:
|
||||
else
|
||||
{
|
||||
// set injector_ndb database/schema from table internal name
|
||||
int ret=
|
||||
IF_DBUG(int ret=)
|
||||
i_ndb->setDatabaseAndSchemaName(pOp->getEvent()->getTable());
|
||||
DBUG_ASSERT(ret == 0);
|
||||
ndb_binlog_thread_handle_non_data_event(thd, i_ndb, pOp, row);
|
||||
@ -3979,7 +3990,7 @@ restart:
|
||||
/*
|
||||
note! pOp is not referring to an event in the next epoch
|
||||
or is == 0
|
||||
*/
|
||||
*/
|
||||
#ifdef RUN_NDB_BINLOG_TIMER
|
||||
write_timer.stop();
|
||||
#endif
|
||||
|
@ -1742,7 +1742,7 @@ void setup_windows_event_source()
|
||||
|
||||
/* Register EventMessageFile */
|
||||
dwError = RegSetValueEx(hRegKey, "EventMessageFile", 0, REG_EXPAND_SZ,
|
||||
(PBYTE) szPath, strlen(szPath)+1);
|
||||
(PBYTE) szPath, (DWORD) (strlen(szPath) + 1));
|
||||
|
||||
/* Register supported event types */
|
||||
dwTypes= (EVENTLOG_ERROR_TYPE | EVENTLOG_WARNING_TYPE |
|
||||
|
@ -284,12 +284,14 @@ public:
|
||||
*/
|
||||
int check_state(enum_state const target_state)
|
||||
{
|
||||
#ifndef DBUG_OFF
|
||||
static char const *state_name[] = {
|
||||
"START_STATE", "TABLE_STATE", "ROW_STATE", "STATE_COUNT"
|
||||
};
|
||||
|
||||
DBUG_ASSERT(0 <= target_state && target_state <= STATE_COUNT);
|
||||
DBUG_PRINT("info", ("In state %s", state_name[m_state]));
|
||||
#endif
|
||||
|
||||
if (m_state <= target_state && target_state <= m_state + 1 &&
|
||||
m_state < STATE_COUNT)
|
||||
|
@ -163,7 +163,7 @@ void mysql_client_binlog_statement(THD* thd)
|
||||
(ulong) uint4korr(bufptr+EVENT_LEN_OFFSET)));
|
||||
#endif
|
||||
ev->thd= thd;
|
||||
if (int err= ev->exec_event(thd->rli_fake))
|
||||
if (IF_DBUG(int err= ) ev->exec_event(thd->rli_fake))
|
||||
{
|
||||
DBUG_PRINT("error", ("exec_event() returned: %d", err));
|
||||
/*
|
||||
|
@ -3268,7 +3268,9 @@ end_with_restore_list:
|
||||
We WANT to write and we CAN write.
|
||||
! we write after unlocking the table.
|
||||
*/
|
||||
/* Presumably, RESET and binlog writing doesn't require synchronization */
|
||||
/*
|
||||
Presumably, RESET and binlog writing doesn't require synchronization
|
||||
*/
|
||||
if (!lex->no_write_to_binlog && write_to_binlog)
|
||||
{
|
||||
if (mysql_bin_log.is_open())
|
||||
|
@ -23,7 +23,9 @@
|
||||
|
||||
int max_binlog_dump_events = 0; // unlimited
|
||||
my_bool opt_sporadic_binlog_dump_fail = 0;
|
||||
#ifndef DBUG_OFF
|
||||
static int binlog_dump_count = 0;
|
||||
#endif
|
||||
|
||||
/*
|
||||
fake_rotate_event() builds a fake (=which does not exist physically in any
|
||||
|
@ -41,7 +41,7 @@ static int copy_data_between_tables(TABLE *from,TABLE *to,
|
||||
|
||||
static bool prepare_blob_field(THD *thd, create_field *sql_field);
|
||||
static bool check_engine(THD *thd, const char *table_name,
|
||||
HA_CREATE_INFO *create_info);
|
||||
HA_CREATE_INFO *create_info);
|
||||
static int mysql_prepare_table(THD *thd, HA_CREATE_INFO *create_info,
|
||||
List<create_field> *fields,
|
||||
List<Key> *keys, bool tmp_table,
|
||||
@ -4818,7 +4818,8 @@ bool mysql_create_like_table(THD* thd, TABLE_LIST* table,
|
||||
else
|
||||
unlock_dst_table= TRUE;
|
||||
|
||||
int result= store_create_info(thd, table, &query, create_info);
|
||||
IF_DBUG(int result=) store_create_info(thd, table, &query,
|
||||
create_info);
|
||||
|
||||
DBUG_ASSERT(result == 0); // store_create_info() always return 0
|
||||
write_bin_log(thd, TRUE, query.ptr(), query.length());
|
||||
@ -6645,7 +6646,8 @@ view_err:
|
||||
thd->query, thd->query_length,
|
||||
db, table_name);
|
||||
|
||||
DBUG_ASSERT(!(mysql_bin_log.is_open() && thd->current_stmt_binlog_row_based &&
|
||||
DBUG_ASSERT(!(mysql_bin_log.is_open() &&
|
||||
thd->current_stmt_binlog_row_based &&
|
||||
(create_info->options & HA_LEX_CREATE_TMP_TABLE)));
|
||||
write_bin_log(thd, TRUE, thd->query, thd->query_length);
|
||||
|
||||
|
@ -1017,7 +1017,7 @@ YY_RULE_SETUP
|
||||
yylval = sym_tab_add_bound_lit(pars_sym_tab_global,
|
||||
yytext + 1, &type);
|
||||
|
||||
return(type);
|
||||
return((int) type);
|
||||
}
|
||||
YY_BREAK
|
||||
case 4:
|
||||
|
@ -460,6 +460,7 @@ void ha_myisammrg::update_create_info(HA_CREATE_INFO *create_info)
|
||||
{
|
||||
TABLE_LIST *ptr;
|
||||
LEX_STRING db, name;
|
||||
LINT_INIT(db.str);
|
||||
|
||||
if (!(ptr = (TABLE_LIST *) thd->calloc(sizeof(TABLE_LIST))))
|
||||
goto err;
|
||||
@ -570,6 +571,8 @@ void ha_myisammrg::append_create_info(String *packet)
|
||||
open_table++)
|
||||
{
|
||||
LEX_STRING db, name;
|
||||
LINT_INIT(db.str);
|
||||
|
||||
split_file_name(open_table->table->filename, &db, &name);
|
||||
if (open_table != first)
|
||||
packet->append(',');
|
||||
|
@ -1338,6 +1338,7 @@ operator<<(NdbOut& out, const Gci_container& gci)
|
||||
return out;
|
||||
}
|
||||
|
||||
#ifdef VM_TRACE
|
||||
static
|
||||
NdbOut&
|
||||
operator<<(NdbOut& out, const Gci_container_pod& gci)
|
||||
@ -1346,7 +1347,7 @@ operator<<(NdbOut& out, const Gci_container_pod& gci)
|
||||
out << *ptr;
|
||||
return out;
|
||||
}
|
||||
|
||||
#endif
|
||||
|
||||
static
|
||||
Gci_container*
|
||||
|
@ -322,6 +322,7 @@ static int my_strncasecmp_ucs2(CHARSET_INFO *cs,
|
||||
const char *te=t+len;
|
||||
MY_UNICASE_INFO **uni_plane= cs->caseinfo;
|
||||
LINT_INIT(s_wc);
|
||||
LINT_INIT(t_wc);
|
||||
|
||||
while ( s < se && t < te )
|
||||
{
|
||||
@ -1385,6 +1386,7 @@ int my_strnncoll_ucs2_bin(CHARSET_INFO *cs,
|
||||
const uchar *se=s+slen;
|
||||
const uchar *te=t+tlen;
|
||||
LINT_INIT(s_wc);
|
||||
LINT_INIT(t_wc);
|
||||
|
||||
while ( s < se && t < te )
|
||||
{
|
||||
|
@ -2313,6 +2313,7 @@ static int my_strnncoll_utf8(CHARSET_INFO *cs,
|
||||
const uchar *te=t+tlen;
|
||||
MY_UNICASE_INFO **uni_plane= cs->caseinfo;
|
||||
LINT_INIT(s_wc);
|
||||
LINT_INIT(t_wc);
|
||||
|
||||
while ( s < se && t < te )
|
||||
{
|
||||
@ -2383,6 +2384,7 @@ static int my_strnncollsp_utf8(CHARSET_INFO *cs,
|
||||
const uchar *se= s+slen, *te= t+tlen;
|
||||
MY_UNICASE_INFO **uni_plane= cs->caseinfo;
|
||||
LINT_INIT(s_wc);
|
||||
LINT_INIT(t_wc);
|
||||
|
||||
#ifndef VARCHAR_WITH_DIFF_ENDSPACE_ARE_DIFFERENT_FOR_UNIQUE
|
||||
diff_if_only_endspace_difference= 0;
|
||||
|
@ -19,6 +19,7 @@ sql_yacc.cc : .*switch statement contains 'default' but no 'case' labels.*
|
||||
#
|
||||
pars0grm.tab.c: .*'yyerrorlab' : unreferenced label.*
|
||||
_flex_tmp.c: .*not enough actual parameters for macro 'yywrap'.*
|
||||
pars0lex.l: .*conversion from 'ulint' to 'int', possible loss of data.*
|
||||
|
||||
#
|
||||
# bdb is not critical to keep up to date
|
||||
@ -33,17 +34,34 @@ db_vrfy.c : .*comparison is always false due to limited range of data type.*
|
||||
# Ignore all conversion warnings on windows 64
|
||||
# (Is safe as we are not yet supporting strings >= 2G)
|
||||
#
|
||||
.* : conversion from 'size_t' to .*int'.*
|
||||
.* : conversion from '__int64' to .*int'.*
|
||||
.* : conversion from '__int64' to 'uint8'.*
|
||||
.* : conversion from '__int64' to 'uint32'.*
|
||||
.* : conversion from '__int64' to 'u.*long'.*
|
||||
.* : conversion from '__int64' to 'long'.*
|
||||
.* : conversion from '__int64' to 'off_t'.*
|
||||
.* : conversion from '.*size_t' to .*int'.*
|
||||
.* : conversion from '.*size_t' to 'TaoCrypt::word32'.*
|
||||
.* : conversion from '.*size_t' to 'u.*long'.*
|
||||
.* : conversion from '.*size_t' to 'uint32'.*
|
||||
.* : conversion from '.*size_t' to 'off_t'.*
|
||||
.* : conversion from '.*size_t' to 'size_s'.*
|
||||
|
||||
#
|
||||
# The following should be fixed by the ndb team
|
||||
#
|
||||
.*/ndb/.* : .*used uninitialized in this function.*
|
||||
.*/ndb/.* : .*unused variable.*
|
||||
.*/ndb/.* : .*defined but not used.*
|
||||
|
||||
#
|
||||
# Unexplanable (?) stuff
|
||||
#
|
||||
listener.cc : .*conversion from 'SOCKET' to 'int'.*
|
||||
net_serv.c : .*conversion from 'SOCKET' to 'int'.*
|
||||
net_serv.cc : .*conversion from 'SOCKET' to 'int'.*
|
||||
mi_packrec.c : .*result of 32-bit shift implicitly converted to 64 bits.* : 567
|
||||
|
||||
#
|
||||
# Wrong compiler warnings
|
||||
#
|
||||
.* : .*no matching operator delete found; memory will not be freed if initialization throws an exception.*
|
||||
|
@ -88,5 +88,5 @@ may be necessary to clean the build tree to remove any stale objects.
|
||||
|
||||
2. To use Visual C++ Express Edition you also need to install the Platform SDK.
|
||||
Please see this link: http://msdn.microsoft.com/vstudio/express/visualc/usingpsdk/
|
||||
At step 4 you only need to add the libraries advapi32.lib and user32.lib to
|
||||
At step 5 you only need to add the libraries advapi32.lib and user32.lib to
|
||||
the file "corewin_express.vsprops" in order to avoid link errors.
|
||||
|
Loading…
x
Reference in New Issue
Block a user