Commit b96c4289 authored by unknown's avatar unknown

removing the usage if the ndb local dict cache to adress the 241 errors we get now and then

parent a8138ec7
...@@ -62,4 +62,6 @@ t4 ...@@ -62,4 +62,6 @@ t4
drop table t1, t2, t3, t4; drop table t1, t2, t3, t4;
drop table if exists t1, t3, t4; drop table if exists t1, t3, t4;
Warnings: Warnings:
Error 155 Table 'test.t1' doesn't exist
Error 155 Table 'test.t3' doesn't exist Error 155 Table 'test.t3' doesn't exist
Error 155 Table 'test.t4' doesn't exist
...@@ -141,7 +141,7 @@ static int rename_share(NDB_SHARE *share, const char *new_key); ...@@ -141,7 +141,7 @@ static int rename_share(NDB_SHARE *share, const char *new_key);
#endif #endif
static void ndb_set_fragmentation(NDBTAB &tab, TABLE *table, uint pk_len); static void ndb_set_fragmentation(NDBTAB &tab, TABLE *table, uint pk_len);
static int ndb_get_table_statistics(Ndb*, const char *, static int ndb_get_table_statistics(Ndb*, const NDBTAB *,
struct Ndb_statistics *); struct Ndb_statistics *);
...@@ -431,7 +431,7 @@ void ha_ndbcluster::records_update() ...@@ -431,7 +431,7 @@ void ha_ndbcluster::records_update()
Ndb *ndb= get_ndb(); Ndb *ndb= get_ndb();
ndb->setDatabaseName(m_dbname); ndb->setDatabaseName(m_dbname);
struct Ndb_statistics stat; struct Ndb_statistics stat;
if (ndb_get_table_statistics(ndb, m_tabname, &stat) == 0){ if (ndb_get_table_statistics(ndb, m_table, &stat) == 0){
mean_rec_length= stat.row_size; mean_rec_length= stat.row_size;
data_file_length= stat.fragment_memory; data_file_length= stat.fragment_memory;
info->records= stat.row_count; info->records= stat.row_count;
...@@ -479,92 +479,6 @@ void ha_ndbcluster::no_uncommitted_rows_reset(THD *thd) ...@@ -479,92 +479,6 @@ void ha_ndbcluster::no_uncommitted_rows_reset(THD *thd)
DBUG_VOID_RETURN; DBUG_VOID_RETURN;
} }
/*
Take care of the error that occured in NDB
RETURN
0 No error
# The mapped error code
*/
int ha_ndbcluster::invalidate_dictionary_cache(bool global, const NDBTAB *ndbtab)
{
NDBDICT *dict= get_ndb()->getDictionary();
DBUG_ENTER("invalidate_dictionary_cache");
DBUG_PRINT("info", ("m_tabname: %s global: %d", m_tabname, global));
if (global)
{
#ifdef HAVE_NDB_BINLOG
if (current_thd != injector_thd)
{
char key[FN_REFLEN];
build_table_filename(key, sizeof(key), m_dbname, m_tabname, "");
DBUG_PRINT("info", ("Getting ndbcluster mutex"));
pthread_mutex_lock(&ndbcluster_mutex);
NDB_SHARE *ndb_share= (NDB_SHARE*)hash_search(&ndbcluster_open_tables,
(byte*) key, strlen(key));
// Only binlog_thread is allowed to globally invalidate a table
if (ndb_share && ndb_share->op)
{
pthread_mutex_unlock(&ndbcluster_mutex);
DBUG_PRINT("info", ("Released ndbcluster mutex"));
DBUG_RETURN(1);
}
pthread_mutex_unlock(&ndbcluster_mutex);
DBUG_PRINT("info", ("Released ndbcluster mutex"));
}
#endif
if (!ndbtab)
{
ndbtab= dict->getTable(m_tabname);
if (!ndbtab)
DBUG_RETURN(1);
}
dict->invalidateTable(ndbtab);
table_share->version= 0L; /* Free when thread is ready */
}
else if (ndbtab)
dict->removeCachedTable(ndbtab);
else
dict->removeCachedTable(m_tabname);
/* Invalidate indexes */
for (uint i= 0; i < table_share->keys; i++)
{
NDBINDEX *index = (NDBINDEX *) m_index[i].index;
NDBINDEX *unique_index = (NDBINDEX *) m_index[i].unique_index;
if (!index && !unique_index)
continue;
NDB_INDEX_TYPE idx_type= m_index[i].type;
switch (idx_type) {
case PRIMARY_KEY_ORDERED_INDEX:
case ORDERED_INDEX:
if (global)
dict->invalidateIndex(index->getName(), m_tabname);
else
dict->removeCachedIndex(index->getName(), m_tabname);
break;
case UNIQUE_ORDERED_INDEX:
if (global)
dict->invalidateIndex(index->getName(), m_tabname);
else
dict->removeCachedIndex(index->getName(), m_tabname);
case UNIQUE_INDEX:
if (global)
dict->invalidateIndex(unique_index->getName(), m_tabname);
else
dict->removeCachedIndex(unique_index->getName(), m_tabname);
break;
case PRIMARY_KEY_INDEX:
case UNDEFINED_INDEX:
break;
}
}
DBUG_RETURN(0);
}
int ha_ndbcluster::ndb_err(NdbTransaction *trans) int ha_ndbcluster::ndb_err(NdbTransaction *trans)
{ {
int res; int res;
...@@ -575,32 +489,14 @@ int ha_ndbcluster::ndb_err(NdbTransaction *trans) ...@@ -575,32 +489,14 @@ int ha_ndbcluster::ndb_err(NdbTransaction *trans)
switch (err.classification) { switch (err.classification) {
case NdbError::SchemaError: case NdbError::SchemaError:
{ {
// TODO perhaps we need to do more here, invalidate also in the cache
m_table->setStatusInvalid();
/* Close other open handlers not used by any thread */ /* Close other open handlers not used by any thread */
TABLE_LIST table_list; TABLE_LIST table_list;
bzero((char*) &table_list,sizeof(table_list)); bzero((char*) &table_list,sizeof(table_list));
table_list.db= m_dbname; table_list.db= m_dbname;
table_list.alias= table_list.table_name= m_tabname; table_list.alias= table_list.table_name= m_tabname;
close_cached_tables(current_thd, 0, &table_list); close_cached_tables(current_thd, 0, &table_list);
invalidate_dictionary_cache(TRUE, m_table);
if (err.code==284)
{
/*
Check if the table is _really_ gone or if the table has
been alterend and thus changed table id
*/
NDBDICT *dict= get_ndb()->getDictionary();
DBUG_PRINT("info", ("Check if table %s is really gone", m_tabname));
if (!(dict->getTable(m_tabname)))
{
err= dict->getNdbError();
DBUG_PRINT("info", ("Table not found, error: %d", err.code));
if (err.code != 709 && err.code != 723)
DBUG_RETURN(1);
}
DBUG_PRINT("info", ("Table exists but must have changed"));
}
break; break;
} }
default: default:
...@@ -1046,24 +942,15 @@ int ha_ndbcluster::get_metadata(const char *path) ...@@ -1046,24 +942,15 @@ int ha_ndbcluster::get_metadata(const char *path)
NDBDICT *dict= ndb->getDictionary(); NDBDICT *dict= ndb->getDictionary();
const NDBTAB *tab; const NDBTAB *tab;
int error; int error;
bool invalidating_ndb_table= FALSE;
DBUG_ENTER("get_metadata"); DBUG_ENTER("get_metadata");
DBUG_PRINT("enter", ("m_tabname: %s, path: %s", m_tabname, path)); DBUG_PRINT("enter", ("m_tabname: %s, path: %s", m_tabname, path));
do { DBUG_ASSERT(m_table == NULL);
DBUG_ASSERT(m_table_info == NULL);
const void *data, *pack_data; const void *data, *pack_data;
uint length, pack_length; uint length, pack_length;
if (!(tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
// Check if thread has stale local cache
if (tab->getObjectStatus() == NdbDictionary::Object::Invalid)
{
invalidate_dictionary_cache(FALSE, tab);
if (!(tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
DBUG_PRINT("info", ("Table schema version: %d", tab->getObjectVersion()));
}
/* /*
Compare FrmData in NDB with frm file from disk. Compare FrmData in NDB with frm file from disk.
*/ */
...@@ -1076,16 +963,12 @@ int ha_ndbcluster::get_metadata(const char *path) ...@@ -1076,16 +963,12 @@ int ha_ndbcluster::get_metadata(const char *path)
DBUG_RETURN(1); DBUG_RETURN(1);
} }
Ndb_table_guard ndbtab_g(dict, m_tabname);
if (!(tab= ndbtab_g.get_table()))
ERR_RETURN(dict->getNdbError());
if (get_ndb_share_state(m_share) != NSS_ALTERED if (get_ndb_share_state(m_share) != NSS_ALTERED
&& cmp_frm(tab, pack_data, pack_length)) && cmp_frm(tab, pack_data, pack_length))
{
if (!invalidating_ndb_table)
{
DBUG_PRINT("info", ("Invalidating table"));
invalidate_dictionary_cache(TRUE, tab);
invalidating_ndb_table= TRUE;
}
else
{ {
DBUG_PRINT("error", DBUG_PRINT("error",
("metadata, pack_length: %d getFrmLength: %d memcmp: %d", ("metadata, pack_length: %d getFrmLength: %d memcmp: %d",
...@@ -1094,25 +977,24 @@ int ha_ndbcluster::get_metadata(const char *path) ...@@ -1094,25 +977,24 @@ int ha_ndbcluster::get_metadata(const char *path)
DBUG_DUMP("pack_data", (char*)pack_data, pack_length); DBUG_DUMP("pack_data", (char*)pack_data, pack_length);
DBUG_DUMP("frm", (char*)tab->getFrmData(), tab->getFrmLength()); DBUG_DUMP("frm", (char*)tab->getFrmData(), tab->getFrmLength());
error= HA_ERR_TABLE_DEF_CHANGED; error= HA_ERR_TABLE_DEF_CHANGED;
invalidating_ndb_table= FALSE;
}
}
else
{
invalidating_ndb_table= FALSE;
} }
my_free((char*)data, MYF(0)); my_free((char*)data, MYF(0));
my_free((char*)pack_data, MYF(0)); my_free((char*)pack_data, MYF(0));
} while (invalidating_ndb_table);
if (error) if (error)
DBUG_RETURN(error); goto err;
m_table_version= tab->getObjectVersion(); DBUG_PRINT("info", ("fetched table %s", tab->getName()));
m_table= tab; m_table= tab;
m_table_info= NULL; // Set in external lock if ((error= open_indexes(ndb, table, FALSE)) == 0)
{
DBUG_RETURN(open_indexes(ndb, table, FALSE)); ndbtab_g.release();
DBUG_RETURN(0);
}
err:
ndbtab_g.invalidate();
m_table= NULL;
DBUG_RETURN(error);
} }
static int fix_unique_index_attr_order(NDB_INDEX_DATA &data, static int fix_unique_index_attr_order(NDB_INDEX_DATA &data,
...@@ -1149,36 +1031,6 @@ static int fix_unique_index_attr_order(NDB_INDEX_DATA &data, ...@@ -1149,36 +1031,6 @@ static int fix_unique_index_attr_order(NDB_INDEX_DATA &data,
DBUG_RETURN(0); DBUG_RETURN(0);
} }
int ha_ndbcluster::table_changed(const void *pack_frm_data, uint pack_frm_len)
{
Ndb *ndb;
NDBDICT *dict;
const NDBTAB *orig_tab;
NdbDictionary::Table new_tab;
int result;
DBUG_ENTER("ha_ndbcluster::table_changed");
DBUG_PRINT("info", ("Modifying frm for table %s", m_tabname));
if (check_ndb_connection())
DBUG_RETURN(my_errno= HA_ERR_NO_CONNECTION);
ndb= get_ndb();
dict= ndb->getDictionary();
if (!(orig_tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
// Check if thread has stale local cache
if (orig_tab->getObjectStatus() == NdbDictionary::Object::Invalid)
{
dict->removeCachedTable(orig_tab);
if (!(orig_tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
}
new_tab= *orig_tab;
new_tab.setFrm(pack_frm_data, pack_frm_len);
if (dict->alterTable(new_tab) != 0)
ERR_RETURN(dict->getNdbError());
DBUG_RETURN(0);
}
/* /*
Create all the indexes for a table. Create all the indexes for a table.
If any index should fail to be created, If any index should fail to be created,
...@@ -1246,7 +1098,7 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info, ...@@ -1246,7 +1098,7 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
const NDBINDEX *index; const NDBINDEX *index;
do do
{ {
index= dict->getIndex(index_name, m_tabname); index= dict->getIndexGlobal(index_name, *m_table);
if (!index) if (!index)
ERR_RETURN(dict->getNdbError()); ERR_RETURN(dict->getNdbError());
DBUG_PRINT("info", ("index: 0x%x id: %d version: %d.%d status: %d", DBUG_PRINT("info", ("index: 0x%x id: %d version: %d.%d status: %d",
...@@ -1255,14 +1107,11 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info, ...@@ -1255,14 +1107,11 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
index->getObjectVersion() & 0xFFFFFF, index->getObjectVersion() & 0xFFFFFF,
index->getObjectVersion() >> 24, index->getObjectVersion() >> 24,
index->getObjectStatus())); index->getObjectStatus()));
if (index->getObjectStatus() != NdbDictionary::Object::Retrieved) DBUG_ASSERT(index->getObjectStatus() ==
{ NdbDictionary::Object::Retrieved);
dict->removeCachedIndex(index);
continue;
}
break; break;
} while (1); } while (1);
m_index[index_no].index= (void *) index; m_index[index_no].index= index;
// ordered index - add stats // ordered index - add stats
NDB_INDEX_DATA& d=m_index[index_no]; NDB_INDEX_DATA& d=m_index[index_no];
delete d.index_stat; delete d.index_stat;
...@@ -1293,7 +1142,7 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info, ...@@ -1293,7 +1142,7 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
const NDBINDEX *index; const NDBINDEX *index;
do do
{ {
index= dict->getIndex(unique_index_name, m_tabname); index= dict->getIndexGlobal(unique_index_name, *m_table);
if (!index) if (!index)
ERR_RETURN(dict->getNdbError()); ERR_RETURN(dict->getNdbError());
DBUG_PRINT("info", ("index: 0x%x id: %d version: %d.%d status: %d", DBUG_PRINT("info", ("index: 0x%x id: %d version: %d.%d status: %d",
...@@ -1302,14 +1151,11 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info, ...@@ -1302,14 +1151,11 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
index->getObjectVersion() & 0xFFFFFF, index->getObjectVersion() & 0xFFFFFF,
index->getObjectVersion() >> 24, index->getObjectVersion() >> 24,
index->getObjectStatus())); index->getObjectStatus()));
if (index->getObjectStatus() != NdbDictionary::Object::Retrieved) DBUG_ASSERT(index->getObjectStatus() ==
{ NdbDictionary::Object::Retrieved);
dict->removeCachedIndex(index);
continue;
}
break; break;
} while (1); } while (1);
m_index[index_no].unique_index= (void *) index; m_index[index_no].unique_index= index;
error= fix_unique_index_attr_order(m_index[index_no], index, key_info); error= fix_unique_index_attr_order(m_index[index_no], index, key_info);
} }
if (!error) if (!error)
...@@ -1341,6 +1187,26 @@ int ha_ndbcluster::open_indexes(Ndb *ndb, TABLE *tab, bool ignore_error) ...@@ -1341,6 +1187,26 @@ int ha_ndbcluster::open_indexes(Ndb *ndb, TABLE *tab, bool ignore_error)
break; break;
} }
if (error && !ignore_error)
{
while (i > 0)
{
i--;
if (m_index[i].index)
{
dict->removeIndexGlobal(*m_index[i].index, 1);
m_index[i].index= NULL;
}
if (m_index[i].unique_index)
{
dict->removeIndexGlobal(*m_index[i].unique_index, 1);
m_index[i].unique_index= NULL;
}
}
}
DBUG_ASSERT(error == 0 || error == 4243);
DBUG_RETURN(error); DBUG_RETURN(error);
} }
...@@ -1400,26 +1266,32 @@ int ha_ndbcluster::drop_indexes(Ndb *ndb, TABLE *tab) ...@@ -1400,26 +1266,32 @@ int ha_ndbcluster::drop_indexes(Ndb *ndb, TABLE *tab)
m_index[i].type= idx_type; m_index[i].type= idx_type;
if (m_index[i].status == TO_BE_DROPPED) if (m_index[i].status == TO_BE_DROPPED)
{ {
NdbDictionary::Index *index= const NdbDictionary::Index *index= m_index[i].index;
(NdbDictionary::Index *) m_index[i].index; const NdbDictionary::Index *unique_index= m_index[i].unique_index;
NdbDictionary::Index *unique_index=
(NdbDictionary::Index *) m_index[i].unique_index;
if (index) if (index)
{ {
index_name= index->getName(); index_name= index->getName();
DBUG_PRINT("info", ("Dropping index %u: %s", i, index_name)); DBUG_PRINT("info", ("Dropping index %u: %s", i, index_name));
// Drop ordered index from ndb // Drop ordered index from ndb
error= drop_ndb_index(index_name); error= dict->dropIndexGlobal(*index);
}
if (!error) if (!error)
{
dict->removeIndexGlobal(*index, 1);
m_index[i].index= NULL; m_index[i].index= NULL;
}
}
if (!error && unique_index) if (!error && unique_index)
{ {
index_name= unique_index->getName(); index_name= unique_index->getName();
DBUG_PRINT("info", ("Dropping unique index %u: %s", i, index_name)); DBUG_PRINT("info", ("Dropping unique index %u: %s", i, index_name));
// Drop unique index from ndb // Drop unique index from ndb
error= drop_ndb_index(index_name); error= dict->dropIndexGlobal(*unique_index);
if (!error)
{
dict->removeIndexGlobal(*unique_index, 1);
m_index[i].unique_index= NULL;
}
} }
if (error) if (error)
DBUG_RETURN(error); DBUG_RETURN(error);
...@@ -1476,30 +1348,57 @@ int ha_ndbcluster::check_index_fields_not_null(uint inx) ...@@ -1476,30 +1348,57 @@ int ha_ndbcluster::check_index_fields_not_null(uint inx)
DBUG_RETURN(0); DBUG_RETURN(0);
} }
void ha_ndbcluster::release_metadata() void ha_ndbcluster::release_metadata(THD *thd, Ndb *ndb)
{ {
uint i; uint i;
DBUG_ENTER("release_metadata"); DBUG_ENTER("release_metadata");
DBUG_PRINT("enter", ("m_tabname: %s", m_tabname)); DBUG_PRINT("enter", ("m_tabname: %s", m_tabname));
m_table= NULL; NDBDICT *dict= ndb->getDictionary();
int invalidate_indexes= 0;
if (thd && thd->lex && thd->lex->sql_command == SQLCOM_FLUSH)
{
invalidate_indexes = 1;
}
if (m_table != NULL)
{
if (m_table->getObjectStatus() == NdbDictionary::Object::Invalid)
invalidate_indexes= 1;
dict->removeTableGlobal(*m_table, invalidate_indexes);
}
// TODO investigate
DBUG_ASSERT(m_table_info == NULL);
m_table_info= NULL; m_table_info= NULL;
// Release index list // Release index list
for (i= 0; i < MAX_KEY; i++) for (i= 0; i < MAX_KEY; i++)
{ {
if (m_index[i].unique_index)
{
DBUG_ASSERT(m_table != NULL);
dict->removeIndexGlobal(*m_index[i].unique_index, invalidate_indexes);
m_index[i].unique_index= NULL; m_index[i].unique_index= NULL;
}
if (m_index[i].index)
{
DBUG_ASSERT(m_table != NULL);
dict->removeIndexGlobal(*m_index[i].index, invalidate_indexes);
m_index[i].index= NULL; m_index[i].index= NULL;
}
if (m_index[i].unique_index_attrid_map) if (m_index[i].unique_index_attrid_map)
{ {
my_free((char *)m_index[i].unique_index_attrid_map, MYF(0)); my_free((char *)m_index[i].unique_index_attrid_map, MYF(0));
m_index[i].unique_index_attrid_map= NULL; m_index[i].unique_index_attrid_map= NULL;
} }
if (m_index[i].index_stat)
{
delete m_index[i].index_stat; delete m_index[i].index_stat;
m_index[i].index_stat=NULL; m_index[i].index_stat= NULL;
}
} }
m_table= NULL;
DBUG_VOID_RETURN; DBUG_VOID_RETURN;
} }
...@@ -1971,11 +1870,10 @@ int ha_ndbcluster::peek_indexed_rows(const byte *record) ...@@ -1971,11 +1870,10 @@ int ha_ndbcluster::peek_indexed_rows(const byte *record)
{ {
// A unique index is defined on table // A unique index is defined on table
NdbIndexOperation *iop; NdbIndexOperation *iop;
NDBINDEX *unique_index = (NDBINDEX *) m_index[i].unique_index; const NDBINDEX *unique_index = m_index[i].unique_index;
key_part= key_info->key_part; key_part= key_info->key_part;
end= key_part + key_info->key_parts; end= key_part + key_info->key_parts;
if (!(iop= trans->getNdbIndexOperation(unique_index, if (!(iop= trans->getNdbIndexOperation(unique_index, m_table)) ||
(const NDBTAB *) m_table)) ||
iop->readTuple(lm) != 0) iop->readTuple(lm) != 0)
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
...@@ -2024,9 +1922,8 @@ int ha_ndbcluster::unique_index_read(const byte *key, ...@@ -2024,9 +1922,8 @@ int ha_ndbcluster::unique_index_read(const byte *key,
NdbOperation::LockMode lm= NdbOperation::LockMode lm=
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type); (NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
if (!(op= trans->getNdbIndexOperation((NDBINDEX *) if (!(op= trans->getNdbIndexOperation(m_index[active_index].unique_index,
m_index[active_index].unique_index, m_table)) ||
(const NDBTAB *) m_table)) ||
op->readTuple(lm) != 0) op->readTuple(lm) != 0)
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
...@@ -2365,9 +2262,8 @@ int ha_ndbcluster::ordered_index_scan(const key_range *start_key, ...@@ -2365,9 +2262,8 @@ int ha_ndbcluster::ordered_index_scan(const key_range *start_key,
restart= FALSE; restart= FALSE;
NdbOperation::LockMode lm= NdbOperation::LockMode lm=
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type); (NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
if (!(op= trans->getNdbIndexScanOperation((NDBINDEX *) if (!(op= trans->getNdbIndexScanOperation(m_index[active_index].index,
m_index[active_index].index, m_table)) ||
(const NDBTAB *) m_table)) ||
op->readTuples(lm, 0, parallelism, sorted, descending)) op->readTuples(lm, 0, parallelism, sorted, descending))
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
if (m_use_partition_function && part_spec != NULL && if (m_use_partition_function && part_spec != NULL &&
...@@ -2437,7 +2333,7 @@ int ha_ndbcluster::full_table_scan(byte *buf) ...@@ -2437,7 +2333,7 @@ int ha_ndbcluster::full_table_scan(byte *buf)
NdbOperation::LockMode lm= NdbOperation::LockMode lm=
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type); (NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
if (!(op=trans->getNdbScanOperation((const NDBTAB *) m_table)) || if (!(op=trans->getNdbScanOperation(m_table)) ||
op->readTuples(lm, 0, parallelism)) op->readTuples(lm, 0, parallelism))
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
m_active_cursor= op; m_active_cursor= op;
...@@ -2539,7 +2435,7 @@ int ha_ndbcluster::write_row(byte *record) ...@@ -2539,7 +2435,7 @@ int ha_ndbcluster::write_row(byte *record)
if (table->timestamp_field_type & TIMESTAMP_AUTO_SET_ON_INSERT) if (table->timestamp_field_type & TIMESTAMP_AUTO_SET_ON_INSERT)
table->timestamp_field->set_time(); table->timestamp_field->set_time();
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table))) if (!(op= trans->getNdbOperation(m_table)))
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
res= (m_use_write) ? op->writeTuple() :op->insertTuple(); res= (m_use_write) ? op->writeTuple() :op->insertTuple();
...@@ -2565,7 +2461,7 @@ int ha_ndbcluster::write_row(byte *record) ...@@ -2565,7 +2461,7 @@ int ha_ndbcluster::write_row(byte *record)
Uint64 auto_value= NDB_FAILED_AUTO_INCREMENT; Uint64 auto_value= NDB_FAILED_AUTO_INCREMENT;
uint retries= NDB_AUTO_INCREMENT_RETRIES; uint retries= NDB_AUTO_INCREMENT_RETRIES;
do { do {
auto_value= ndb->getAutoIncrementValue((const NDBTAB *) m_table); auto_value= ndb->getAutoIncrementValue(m_table);
} while (auto_value == NDB_FAILED_AUTO_INCREMENT && } while (auto_value == NDB_FAILED_AUTO_INCREMENT &&
--retries && --retries &&
ndb->getNdbError().status == NdbError::TemporaryError); ndb->getNdbError().status == NdbError::TemporaryError);
...@@ -2666,7 +2562,7 @@ int ha_ndbcluster::write_row(byte *record) ...@@ -2666,7 +2562,7 @@ int ha_ndbcluster::write_row(byte *record)
DBUG_PRINT("info", DBUG_PRINT("info",
("Trying to set next auto increment value to %lu", ("Trying to set next auto increment value to %lu",
(ulong) next_val)); (ulong) next_val));
if (ndb->setAutoIncrementValue((const NDBTAB *) m_table, next_val, TRUE)) if (ndb->setAutoIncrementValue(m_table, next_val, TRUE))
DBUG_PRINT("info", DBUG_PRINT("info",
("Setting next auto increment value to %u", next_val)); ("Setting next auto increment value to %u", next_val));
} }
...@@ -2816,7 +2712,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data) ...@@ -2816,7 +2712,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
} }
else else
{ {
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) || if (!(op= trans->getNdbOperation(m_table)) ||
op->updateTuple() != 0) op->updateTuple() != 0)
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
...@@ -2925,7 +2821,7 @@ int ha_ndbcluster::delete_row(const byte *record) ...@@ -2925,7 +2821,7 @@ int ha_ndbcluster::delete_row(const byte *record)
else else
{ {
if (!(op=trans->getNdbOperation((const NDBTAB *) m_table)) || if (!(op=trans->getNdbOperation(m_table)) ||
op->deleteTuple() != 0) op->deleteTuple() != 0)
ERR_RETURN(trans->getNdbError()); ERR_RETURN(trans->getNdbError());
...@@ -3078,7 +2974,7 @@ void ha_ndbcluster::unpack_record(byte *buf) ...@@ -3078,7 +2974,7 @@ void ha_ndbcluster::unpack_record(byte *buf)
{ {
// Table with hidden primary key // Table with hidden primary key
int hidden_no= table_share->fields; int hidden_no= table_share->fields;
const NDBTAB *tab= (const NDBTAB *) m_table; const NDBTAB *tab= m_table;
const NDBCOL *hidden_col= tab->getColumn(hidden_no); const NDBCOL *hidden_col= tab->getColumn(hidden_no);
const NdbRecAttr* rec= m_value[hidden_no].rec; const NdbRecAttr* rec= m_value[hidden_no].rec;
DBUG_ASSERT(rec); DBUG_ASSERT(rec);
...@@ -3554,7 +3450,7 @@ void ha_ndbcluster::position(const byte *record) ...@@ -3554,7 +3450,7 @@ void ha_ndbcluster::position(const byte *record)
key_length= ref_length; key_length= ref_length;
#ifndef DBUG_OFF #ifndef DBUG_OFF
int hidden_no= table->s->fields; int hidden_no= table->s->fields;
const NDBTAB *tab= (const NDBTAB *) m_table; const NDBTAB *tab= m_table;
const NDBCOL *hidden_col= tab->getColumn(hidden_no); const NDBCOL *hidden_col= tab->getColumn(hidden_no);
DBUG_ASSERT(hidden_col->getPrimaryKey() && DBUG_ASSERT(hidden_col->getPrimaryKey() &&
hidden_col->getAutoIncrement() && hidden_col->getAutoIncrement() &&
...@@ -3600,7 +3496,7 @@ void ha_ndbcluster::info(uint flag) ...@@ -3600,7 +3496,7 @@ void ha_ndbcluster::info(uint flag)
ndb->setDatabaseName(m_dbname); ndb->setDatabaseName(m_dbname);
struct Ndb_statistics stat; struct Ndb_statistics stat;
if (current_thd->variables.ndb_use_exact_count && if (current_thd->variables.ndb_use_exact_count &&
ndb_get_table_statistics(ndb, m_tabname, &stat) == 0) ndb_get_table_statistics(ndb, m_table, &stat) == 0)
{ {
mean_rec_length= stat.row_size; mean_rec_length= stat.row_size;
data_file_length= stat.fragment_memory; data_file_length= stat.fragment_memory;
...@@ -3631,7 +3527,7 @@ void ha_ndbcluster::info(uint flag) ...@@ -3631,7 +3527,7 @@ void ha_ndbcluster::info(uint flag)
Ndb *ndb= get_ndb(); Ndb *ndb= get_ndb();
auto_increment_value= auto_increment_value=
ndb->readAutoIncrementValue((const NDBTAB *) m_table); ndb->readAutoIncrementValue(m_table);
} }
} }
DBUG_VOID_RETURN; DBUG_VOID_RETURN;
...@@ -3715,7 +3611,7 @@ int ha_ndbcluster::extra(enum ha_extra_function operation) ...@@ -3715,7 +3611,7 @@ int ha_ndbcluster::extra(enum ha_extra_function operation)
void ha_ndbcluster::start_bulk_insert(ha_rows rows) void ha_ndbcluster::start_bulk_insert(ha_rows rows)
{ {
int bytes, batch; int bytes, batch;
const NDBTAB *tab= (const NDBTAB *) m_table; const NDBTAB *tab= m_table;
DBUG_ENTER("start_bulk_insert"); DBUG_ENTER("start_bulk_insert");
DBUG_PRINT("enter", ("rows: %d", (int)rows)); DBUG_PRINT("enter", ("rows: %d", (int)rows));
...@@ -3994,41 +3890,8 @@ int ha_ndbcluster::external_lock(THD *thd, int lock_type) ...@@ -3994,41 +3890,8 @@ int ha_ndbcluster::external_lock(THD *thd, int lock_type)
// Start of transaction // Start of transaction
m_rows_changed= 0; m_rows_changed= 0;
m_ops_pending= 0; m_ops_pending= 0;
{
NDBDICT *dict= ndb->getDictionary(); // TODO remove double pointers...
const NDBTAB *tab;
if (!(tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
DBUG_PRINT("info", ("Table schema version: %d",
tab->getObjectVersion()));
// Check if thread has stale local cache
// New transaction must not use old tables... (trans != 0)
// Running might...
if ((trans && tab->getObjectStatus() != NdbDictionary::Object::Retrieved)
|| tab->getObjectStatus() == NdbDictionary::Object::Invalid)
{
invalidate_dictionary_cache(FALSE, tab);
if (!(tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
DBUG_PRINT("info", ("Table schema version: %d",
tab->getObjectVersion()));
}
if (m_table_version < tab->getObjectVersion())
{
/*
The table has been altered, caller has to retry
*/
NdbError err= ndb->getNdbError(NDB_INVALID_SCHEMA_OBJECT);
DBUG_RETURN(ndb_to_mysql_error(&err));
}
if (m_table != (void *)tab)
{
m_table= tab;
m_table_version = tab->getObjectVersion();
if (!(my_errno= open_indexes(ndb, table, FALSE)))
DBUG_RETURN(my_errno);
}
}
m_thd_ndb_share= thd_ndb->get_open_table(thd, m_table); m_thd_ndb_share= thd_ndb->get_open_table(thd, m_table);
m_table_info= &m_thd_ndb_share->stat; m_table_info= &m_thd_ndb_share->stat;
} }
...@@ -4673,6 +4536,14 @@ int ha_ndbcluster::create(const char *name, ...@@ -4673,6 +4536,14 @@ int ha_ndbcluster::create(const char *name,
my_errno= ndb_to_mysql_error(&err); my_errno= ndb_to_mysql_error(&err);
DBUG_RETURN(my_errno); DBUG_RETURN(my_errno);
} }
Ndb_table_guard ndbtab_g(dict, m_tabname);
// temporary set m_table during create
// reset at return
m_table= ndbtab_g.get_table();
// TODO check also that we have the same frm...
DBUG_ASSERT(m_table != 0);
DBUG_PRINT("info", ("Table %s/%s created successfully", DBUG_PRINT("info", ("Table %s/%s created successfully",
m_dbname, m_tabname)); m_dbname, m_tabname));
...@@ -4687,7 +4558,10 @@ int ha_ndbcluster::create(const char *name, ...@@ -4687,7 +4558,10 @@ int ha_ndbcluster::create(const char *name,
Failed to create an index, Failed to create an index,
drop the table (and all it's indexes) drop the table (and all it's indexes)
*/ */
drop_ndb_table(); if (dict->dropTableGlobal(*m_table) == 0)
{
m_table = 0;
}
} }
#ifdef HAVE_NDB_BINLOG #ifdef HAVE_NDB_BINLOG
...@@ -4717,7 +4591,6 @@ int ha_ndbcluster::create(const char *name, ...@@ -4717,7 +4591,6 @@ int ha_ndbcluster::create(const char *name,
while (!IS_TMP_PREFIX(m_tabname)) while (!IS_TMP_PREFIX(m_tabname))
{ {
const NDBTAB *t= dict->getTable(m_tabname);
String event_name(INJECTOR_EVENT_LEN); String event_name(INJECTOR_EVENT_LEN);
ndb_rep_event_name(&event_name,m_dbname,m_tabname); ndb_rep_event_name(&event_name,m_dbname,m_tabname);
int do_event_op= ndb_binlog_running; int do_event_op= ndb_binlog_running;
...@@ -4731,14 +4604,14 @@ int ha_ndbcluster::create(const char *name, ...@@ -4731,14 +4604,14 @@ int ha_ndbcluster::create(const char *name,
Always create an event for the table, as other mysql servers Always create an event for the table, as other mysql servers
expect it to be there. expect it to be there.
*/ */
if (!ndbcluster_create_event(ndb, t, event_name.c_ptr(), share, if (!ndbcluster_create_event(ndb, m_table, event_name.c_ptr(), share,
share && do_event_op /* push warning */)) share && do_event_op /* push warning */))
{ {
if (ndb_extra_logging) if (ndb_extra_logging)
sql_print_information("NDB Binlog: CREATE TABLE Event: %s", sql_print_information("NDB Binlog: CREATE TABLE Event: %s",
event_name.c_ptr()); event_name.c_ptr());
if (share && do_event_op && if (share && do_event_op &&
ndbcluster_create_event_ops(share, t, event_name.c_ptr())) ndbcluster_create_event_ops(share, m_table, event_name.c_ptr()))
{ {
sql_print_error("NDB Binlog: FAILED CREATE TABLE event operations." sql_print_error("NDB Binlog: FAILED CREATE TABLE event operations."
" Event: %s", name); " Event: %s", name);
...@@ -4754,13 +4627,15 @@ int ha_ndbcluster::create(const char *name, ...@@ -4754,13 +4627,15 @@ int ha_ndbcluster::create(const char *name,
ndbcluster_log_schema_op(current_thd, share, ndbcluster_log_schema_op(current_thd, share,
current_thd->query, current_thd->query_length, current_thd->query, current_thd->query_length,
share->db, share->table_name, share->db, share->table_name,
0, 0, m_table->getObjectId(),
m_table->getObjectVersion(),
SOT_CREATE_TABLE); SOT_CREATE_TABLE);
break; break;
} }
} }
#endif /* HAVE_NDB_BINLOG */ #endif /* HAVE_NDB_BINLOG */
m_table= 0;
DBUG_RETURN(my_errno); DBUG_RETURN(my_errno);
} }
...@@ -4790,10 +4665,15 @@ int ha_ndbcluster::create_handler_files(const char *file, ...@@ -4790,10 +4665,15 @@ int ha_ndbcluster::create_handler_files(const char *file,
NDBDICT *dict= ndb->getDictionary(); NDBDICT *dict= ndb->getDictionary();
if (!info->frm_only) if (!info->frm_only)
DBUG_RETURN(0); // Must be a create, ignore since frm is saved in create DBUG_RETURN(0); // Must be a create, ignore since frm is saved in create
// TODO handle this
DBUG_ASSERT(m_table != 0);
set_dbname(file); set_dbname(file);
set_tabname(file); set_tabname(file);
Ndb_table_guard ndbtab_g(dict, m_tabname);
DBUG_PRINT("info", ("m_dbname: %s, m_tabname: %s", m_dbname, m_tabname)); DBUG_PRINT("info", ("m_dbname: %s, m_tabname: %s", m_dbname, m_tabname));
if (!(tab= dict->getTable(m_tabname))) if (!(tab= ndbtab_g.get_table()))
DBUG_RETURN(0); // Unkown table, must be temporary table DBUG_RETURN(0); // Unkown table, must be temporary table
DBUG_ASSERT(get_ndb_share_state(m_share) == NSS_ALTERED); DBUG_ASSERT(get_ndb_share_state(m_share) == NSS_ALTERED);
...@@ -4809,7 +4689,12 @@ int ha_ndbcluster::create_handler_files(const char *file, ...@@ -4809,7 +4689,12 @@ int ha_ndbcluster::create_handler_files(const char *file,
{ {
DBUG_PRINT("info", ("Table %s has changed, altering frm in ndb", DBUG_PRINT("info", ("Table %s has changed, altering frm in ndb",
m_tabname)); m_tabname));
error= table_changed(pack_data, pack_length); NdbDictionary::Table new_tab= *tab;
new_tab.setFrm(pack_data, pack_length);
if (dict->alterTableGlobal(*tab, new_tab))
{
error= ndb_to_mysql_error(&dict->getNdbError());
}
my_free((char*)data, MYF(MY_ALLOW_ZERO_PTR)); my_free((char*)data, MYF(MY_ALLOW_ZERO_PTR));
my_free((char*)pack_data, MYF(MY_ALLOW_ZERO_PTR)); my_free((char*)pack_data, MYF(MY_ALLOW_ZERO_PTR));
} }
...@@ -4912,7 +4797,7 @@ int ha_ndbcluster::create_ndb_index(const char *name, ...@@ -4912,7 +4797,7 @@ int ha_ndbcluster::create_ndb_index(const char *name,
ndb_index.addColumnName(field->field_name); ndb_index.addColumnName(field->field_name);
} }
if (dict->createIndex(ndb_index)) if (dict->createIndex(ndb_index, *m_table))
ERR_RETURN(dict->getNdbError()); ERR_RETURN(dict->getNdbError());
// Success // Success
...@@ -4965,18 +4850,6 @@ int ha_ndbcluster::add_index(TABLE *table_arg, ...@@ -4965,18 +4850,6 @@ int ha_ndbcluster::add_index(TABLE *table_arg,
DBUG_RETURN(error); DBUG_RETURN(error);
} }
/*
Drop an index in ndb
*/
int ha_ndbcluster::drop_ndb_index(const char *name)
{
DBUG_ENTER("ha_ndbcluster::drop_index");
DBUG_PRINT("enter", ("name: %s ", name));
Ndb *ndb= get_ndb();
NdbDictionary::Dictionary *dict= ndb->getDictionary();
DBUG_RETURN(dict->dropIndex(name, m_tabname));
}
/* /*
Mark one or several indexes for deletion. and Mark one or several indexes for deletion. and
renumber the remaining indexes renumber the remaining indexes
...@@ -5045,16 +4918,14 @@ int ha_ndbcluster::rename_table(const char *from, const char *to) ...@@ -5045,16 +4918,14 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
Ndb *ndb= get_ndb(); Ndb *ndb= get_ndb();
ndb->setDatabaseName(old_dbname); ndb->setDatabaseName(old_dbname);
dict= ndb->getDictionary(); dict= ndb->getDictionary();
if (!(orig_tab= dict->getTable(m_tabname))) Ndb_table_guard ndbtab_g(dict, m_tabname);
if (!(orig_tab= ndbtab_g.get_table()))
ERR_RETURN(dict->getNdbError()); ERR_RETURN(dict->getNdbError());
// Check if thread has stale local cache
if (orig_tab->getObjectStatus() == NdbDictionary::Object::Invalid)
{
dict->removeCachedTable(orig_tab);
if (!(orig_tab= dict->getTable(m_tabname)))
ERR_RETURN(dict->getNdbError());
}
#ifdef HAVE_NDB_BINLOG #ifdef HAVE_NDB_BINLOG
int ndb_table_id= orig_tab->getObjectId();
int ndb_table_version= orig_tab->getObjectVersion();
NDB_SHARE *share= get_share(from, 0, false); NDB_SHARE *share= get_share(from, 0, false);
if (share) if (share)
{ {
...@@ -5062,13 +4933,15 @@ int ha_ndbcluster::rename_table(const char *from, const char *to) ...@@ -5062,13 +4933,15 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
} }
#endif #endif
m_table= orig_tab;
// Change current database to that of target table // Change current database to that of target table
set_dbname(to); set_dbname(to);
ndb->setDatabaseName(m_dbname); ndb->setDatabaseName(m_dbname);
if ((result= alter_table_name(new_tabname))) NdbDictionary::Table new_tab= *orig_tab;
new_tab.setName(new_tabname);
if (dict->alterTableGlobal(*orig_tab, new_tab) != 0)
{ {
NdbError ndb_error= dict->getNdbError();
#ifdef HAVE_NDB_BINLOG #ifdef HAVE_NDB_BINLOG
if (share) if (share)
{ {
...@@ -5077,7 +4950,7 @@ int ha_ndbcluster::rename_table(const char *from, const char *to) ...@@ -5077,7 +4950,7 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
free_share(&share); free_share(&share);
} }
#endif #endif
DBUG_RETURN(result); ERR_RETURN(ndb_error);
} }
// Rename .ndb file // Rename .ndb file
...@@ -5111,7 +4984,8 @@ int ha_ndbcluster::rename_table(const char *from, const char *to) ...@@ -5111,7 +4984,8 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
/* always create an event for the table */ /* always create an event for the table */
String event_name(INJECTOR_EVENT_LEN); String event_name(INJECTOR_EVENT_LEN);
ndb_rep_event_name(&event_name, to + sizeof(share_prefix) - 1, 0); ndb_rep_event_name(&event_name, to + sizeof(share_prefix) - 1, 0);
const NDBTAB *ndbtab= dict->getTable(new_tabname); Ndb_table_guard ndbtab_g2(dict, new_tabname);
const NDBTAB *ndbtab= ndbtab_g2.get_table();
if (!ndbcluster_create_event(ndb, ndbtab, event_name.c_ptr(), share, if (!ndbcluster_create_event(ndb, ndbtab, event_name.c_ptr(), share,
share && ndb_binlog_running /* push warning */)) share && ndb_binlog_running /* push warning */))
...@@ -5134,10 +5008,10 @@ int ha_ndbcluster::rename_table(const char *from, const char *to) ...@@ -5134,10 +5008,10 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
if (!is_old_table_tmpfile) if (!is_old_table_tmpfile)
ndbcluster_log_schema_op(current_thd, share, ndbcluster_log_schema_op(current_thd, share,
current_thd->query, current_thd->query_length, current_thd->query, current_thd->query_length,
m_dbname, new_tabname, old_dbname, m_tabname,
0, 0, ndb_table_id, ndb_table_version,
SOT_RENAME_TABLE, SOT_RENAME_TABLE,
old_dbname, m_tabname); m_dbname, new_tabname);
} }
if (share) if (share)
free_share(&share); free_share(&share);
...@@ -5147,30 +5021,6 @@ int ha_ndbcluster::rename_table(const char *from, const char *to) ...@@ -5147,30 +5021,6 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
} }
/*
Rename a table in NDB Cluster using alter table
*/
int ha_ndbcluster::alter_table_name(const char *to)
{
Ndb *ndb= get_ndb();
NDBDICT *dict= ndb->getDictionary();
const NDBTAB *orig_tab= (const NDBTAB *) m_table;
DBUG_ENTER("alter_table_name");
DBUG_PRINT("info", ("from: %s to: %s", orig_tab->getName(), to));
NdbDictionary::Table new_tab= *orig_tab;
new_tab.setName(to);
if (dict->alterTable(new_tab) != 0)
ERR_RETURN(dict->getNdbError());
m_table= NULL;
m_table_info= NULL;
DBUG_RETURN(0);
}
/* /*
Delete table from NDB Cluster Delete table from NDB Cluster
...@@ -5187,6 +5037,8 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb, ...@@ -5187,6 +5037,8 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
DBUG_ENTER("ha_ndbcluster::ndbcluster_delete_table"); DBUG_ENTER("ha_ndbcluster::ndbcluster_delete_table");
NDBDICT *dict= ndb->getDictionary(); NDBDICT *dict= ndb->getDictionary();
#ifdef HAVE_NDB_BINLOG #ifdef HAVE_NDB_BINLOG
int ndb_table_id= 0;
int ndb_table_version= 0;
/* /*
Don't allow drop table unless Don't allow drop table unless
schema distribution table is setup schema distribution table is setup
...@@ -5202,14 +5054,45 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb, ...@@ -5202,14 +5054,45 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
/* Drop the table from NDB */ /* Drop the table from NDB */
int res; int res;
if (h) if (h && h->m_table)
{ {
res= h->drop_ndb_table(); if (dict->dropTableGlobal(*h->m_table))
res= ndb_to_mysql_error(&dict->getNdbError());
#ifdef HAVE_NDB_BINLOG
if (res == 0)
{
ndb_table_id= h->m_table->getObjectId();
ndb_table_version= h->m_table->getObjectVersion();
}
#endif
h->release_metadata(current_thd, ndb);
} }
else else
{ {
ndb->setDatabaseName(db); ndb->setDatabaseName(db);
res= dict->dropTable(table_name); while (1)
{
Ndb_table_guard ndbtab_g(dict, table_name);
if (ndbtab_g.get_table())
{
if (dict->dropTableGlobal(*ndbtab_g.get_table()) == 0)
{
#ifdef HAVE_NDB_BINLOG
ndb_table_id= ndbtab_g.get_table()->getObjectId();
ndb_table_version= ndbtab_g.get_table()->getObjectVersion();
#endif
res= 0;
}
else if (dict->getNdbError().code == NDB_INVALID_SCHEMA_OBJECT)
{
ndbtab_g.invalidate();
continue;
}
}
else
res= ndb_to_mysql_error(&dict->getNdbError());
break;
}
} }
if (res) if (res)
...@@ -5251,7 +5134,7 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb, ...@@ -5251,7 +5134,7 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
ndbcluster_log_schema_op(current_thd, share, ndbcluster_log_schema_op(current_thd, share,
current_thd->query, current_thd->query_length, current_thd->query, current_thd->query_length,
share->db, share->table_name, share->db, share->table_name,
0, 0, ndb_table_id, ndb_table_version,
SOT_DROP_TABLE); SOT_DROP_TABLE);
} }
else if (table_dropped && share && share->op) /* ndbcluster_log_schema_op else if (table_dropped && share && share->op) /* ndbcluster_log_schema_op
...@@ -5315,24 +5198,6 @@ int ha_ndbcluster::delete_table(const char *name) ...@@ -5315,24 +5198,6 @@ int ha_ndbcluster::delete_table(const char *name)
} }
/*
Drop table in NDB Cluster
*/
int ha_ndbcluster::drop_ndb_table()
{
Ndb *ndb= get_ndb();
NdbDictionary::Dictionary *dict= ndb->getDictionary();
DBUG_ENTER("intern_drop_table");
DBUG_PRINT("enter", ("Deleting %s", m_tabname));
release_metadata();
if (dict->dropTable(m_tabname))
ERR_RETURN(dict->getNdbError());
DBUG_RETURN(0);
}
ulonglong ha_ndbcluster::get_auto_increment() ulonglong ha_ndbcluster::get_auto_increment()
{ {
int cache_size; int cache_size;
...@@ -5356,8 +5221,8 @@ ulonglong ha_ndbcluster::get_auto_increment() ...@@ -5356,8 +5221,8 @@ ulonglong ha_ndbcluster::get_auto_increment()
do { do {
auto_value= auto_value=
(m_skip_auto_increment) ? (m_skip_auto_increment) ?
ndb->readAutoIncrementValue((const NDBTAB *) m_table) ndb->readAutoIncrementValue(m_table)
: ndb->getAutoIncrementValue((const NDBTAB *) m_table, cache_size); : ndb->getAutoIncrementValue(m_table, cache_size);
} while (auto_value == NDB_FAILED_AUTO_INCREMENT && } while (auto_value == NDB_FAILED_AUTO_INCREMENT &&
--retries && --retries &&
ndb->getNdbError().status == NdbError::TemporaryError); ndb->getNdbError().status == NdbError::TemporaryError);
...@@ -5391,7 +5256,6 @@ ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg): ...@@ -5391,7 +5256,6 @@ ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg):
m_active_trans(NULL), m_active_trans(NULL),
m_active_cursor(NULL), m_active_cursor(NULL),
m_table(NULL), m_table(NULL),
m_table_version(-1),
m_table_info(NULL), m_table_info(NULL),
m_table_flags(HA_NDBCLUSTER_TABLE_FLAGS), m_table_flags(HA_NDBCLUSTER_TABLE_FLAGS),
m_share(0), m_share(0),
...@@ -5437,19 +5301,31 @@ ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg): ...@@ -5437,19 +5301,31 @@ ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg):
} }
int ha_ndbcluster::ha_initialise()
{
DBUG_ENTER("ha_ndbcluster::ha_initialise");
if (check_ndb_in_thd(current_thd))
{
DBUG_RETURN(FALSE);
}
DBUG_RETURN(TRUE);
}
/* /*
Destructor for NDB Cluster table handler Destructor for NDB Cluster table handler
*/ */
ha_ndbcluster::~ha_ndbcluster() ha_ndbcluster::~ha_ndbcluster()
{ {
THD *thd= current_thd;
Ndb *ndb= thd ? check_ndb_in_thd(thd) : g_ndb;
DBUG_ENTER("~ha_ndbcluster"); DBUG_ENTER("~ha_ndbcluster");
if (m_share) if (m_share)
{ {
free_share(&m_share); free_share(&m_share);
} }
release_metadata(); release_metadata(thd, ndb);
my_free(m_blobs_buffer, MYF(MY_ALLOW_ZERO_PTR)); my_free(m_blobs_buffer, MYF(MY_ALLOW_ZERO_PTR));
m_blobs_buffer= 0; m_blobs_buffer= 0;
...@@ -5565,9 +5441,11 @@ void ha_ndbcluster::set_part_info(partition_info *part_info) ...@@ -5565,9 +5441,11 @@ void ha_ndbcluster::set_part_info(partition_info *part_info)
int ha_ndbcluster::close(void) int ha_ndbcluster::close(void)
{ {
DBUG_ENTER("close"); DBUG_ENTER("close");
THD *thd= current_thd;
Ndb *ndb= thd ? check_ndb_in_thd(thd) : g_ndb;
free_share(&m_share); free_share(&m_share);
m_share= 0; m_share= 0;
release_metadata(); release_metadata(thd, ndb);
DBUG_RETURN(0); DBUG_RETURN(0);
} }
...@@ -5655,9 +5533,10 @@ static int ndbcluster_close_connection(THD *thd) ...@@ -5655,9 +5533,10 @@ static int ndbcluster_close_connection(THD *thd)
int ndbcluster_discover(THD* thd, const char *db, const char *name, int ndbcluster_discover(THD* thd, const char *db, const char *name,
const void** frmblob, uint* frmlen) const void** frmblob, uint* frmlen)
{ {
int error= 0;
NdbError ndb_error;
uint len; uint len;
const void* data; const void* data;
const NDBTAB* tab;
Ndb* ndb; Ndb* ndb;
char key[FN_REFLEN]; char key[FN_REFLEN];
DBUG_ENTER("ndbcluster_discover"); DBUG_ENTER("ndbcluster_discover");
...@@ -5667,7 +5546,6 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name, ...@@ -5667,7 +5546,6 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
DBUG_RETURN(HA_ERR_NO_CONNECTION); DBUG_RETURN(HA_ERR_NO_CONNECTION);
ndb->setDatabaseName(db); ndb->setDatabaseName(db);
NDBDICT* dict= ndb->getDictionary(); NDBDICT* dict= ndb->getDictionary();
dict->invalidateTable(name);
build_table_filename(key, sizeof(key), db, name, ""); build_table_filename(key, sizeof(key), db, name, "");
NDB_SHARE *share= get_share(key, 0, false); NDB_SHARE *share= get_share(key, 0, false);
if (share && get_ndb_share_state(share) == NSS_ALTERED) if (share && get_ndb_share_state(share) == NSS_ALTERED)
...@@ -5676,21 +5554,22 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name, ...@@ -5676,21 +5554,22 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
if (readfrm(key, &data, &len)) if (readfrm(key, &data, &len))
{ {
DBUG_PRINT("error", ("Could not read frm")); DBUG_PRINT("error", ("Could not read frm"));
if (share) error= 1;
free_share(&share); goto err;
DBUG_RETURN(1);
} }
} }
else else
{ {
if (!(tab= dict->getTable(name))) Ndb_table_guard ndbtab_g(dict, name);
const NDBTAB *tab= ndbtab_g.get_table();
if (!tab)
{ {
const NdbError err= dict->getNdbError(); const NdbError err= dict->getNdbError();
if (share)
free_share(&share);
if (err.code == 709 || err.code == 723) if (err.code == 709 || err.code == 723)
DBUG_RETURN(-1); error= -1;
ERR_RETURN(err); else
ndb_error= err;
goto err;
} }
DBUG_PRINT("info", ("Found table %s", tab->getName())); DBUG_PRINT("info", ("Found table %s", tab->getName()));
...@@ -5698,17 +5577,15 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name, ...@@ -5698,17 +5577,15 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
if (len == 0 || tab->getFrmData() == NULL) if (len == 0 || tab->getFrmData() == NULL)
{ {
DBUG_PRINT("error", ("No frm data found.")); DBUG_PRINT("error", ("No frm data found."));
if (share) error= 1;
free_share(&share); goto err;
DBUG_RETURN(1);
} }
if (unpackfrm(&data, &len, tab->getFrmData())) if (unpackfrm(&data, &len, tab->getFrmData()))
{ {
DBUG_PRINT("error", ("Could not unpack table")); DBUG_PRINT("error", ("Could not unpack table"));
if (share) error= 1;
free_share(&share); goto err;
DBUG_RETURN(1);
} }
} }
...@@ -5719,6 +5596,14 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name, ...@@ -5719,6 +5596,14 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
free_share(&share); free_share(&share);
DBUG_RETURN(0); DBUG_RETURN(0);
err:
if (share)
free_share(&share);
if (ndb_error.code)
{
ERR_RETURN(ndb_error);
}
DBUG_RETURN(error);
} }
/* /*
...@@ -5726,29 +5611,32 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name, ...@@ -5726,29 +5611,32 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
*/ */
int ndbcluster_table_exists_in_engine(THD* thd, const char *db, const char *name) int ndbcluster_table_exists_in_engine(THD* thd, const char *db,
const char *name)
{ {
const NDBTAB* tab;
Ndb* ndb; Ndb* ndb;
DBUG_ENTER("ndbcluster_table_exists_in_engine"); DBUG_ENTER("ndbcluster_table_exists_in_engine");
DBUG_PRINT("enter", ("db: %s, name: %s", db, name)); DBUG_PRINT("enter", ("db: %s, name: %s", db, name));
if (!(ndb= check_ndb_in_thd(thd))) if (!(ndb= check_ndb_in_thd(thd)))
DBUG_RETURN(HA_ERR_NO_CONNECTION); DBUG_RETURN(HA_ERR_NO_CONNECTION);
ndb->setDatabaseName(db);
NDBDICT* dict= ndb->getDictionary(); NDBDICT* dict= ndb->getDictionary();
dict->invalidateTable(name); NdbDictionary::Dictionary::List list;
if (!(tab= dict->getTable(name))) if (dict->listObjects(list, NdbDictionary::Object::UserTable) != 0)
ERR_RETURN(dict->getNdbError());
for (int i= 0 ; i < list.count ; i++)
{ {
const NdbError err= dict->getNdbError(); NdbDictionary::Dictionary::List::Element& elmt= list.elements[i];
if (err.code == 709 || err.code == 723) if (my_strcasecmp(system_charset_info, elmt.database, db))
DBUG_RETURN(0); continue;
ERR_RETURN(err); if (my_strcasecmp(system_charset_info, elmt.name, name))
} continue;
// table found
DBUG_PRINT("info", ("Found table %s", tab->getName())); DBUG_PRINT("info", ("Found table"));
DBUG_RETURN(1); DBUG_RETURN(1);
}
DBUG_RETURN(0);
} }
...@@ -5902,9 +5790,9 @@ int ndbcluster_find_all_files(THD *thd) ...@@ -5902,9 +5790,9 @@ int ndbcluster_find_all_files(THD *thd)
} }
ndb->setDatabaseName(elmt.database); ndb->setDatabaseName(elmt.database);
const NDBTAB *ndbtab; Ndb_table_guard ndbtab_g(dict, elmt.name);
const NDBTAB *ndbtab= ndbtab_g.get_table();
if (!(ndbtab= dict->getTable(elmt.name))) if (!ndbtab)
{ {
if (retries == 0) if (retries == 0)
sql_print_error("NDB: failed to setup table %s.%s, error: %d, %s", sql_print_error("NDB: failed to setup table %s.%s, error: %d, %s",
...@@ -6184,11 +6072,13 @@ static int connect_callback() ...@@ -6184,11 +6072,13 @@ static int connect_callback()
return 0; return 0;
} }
extern int ndb_dictionary_is_mysqld;
static bool ndbcluster_init() static bool ndbcluster_init()
{ {
int res; int res;
DBUG_ENTER("ndbcluster_init"); DBUG_ENTER("ndbcluster_init");
ndb_dictionary_is_mysqld= 1;
if (have_ndbcluster != SHOW_OPTION_YES) if (have_ndbcluster != SHOW_OPTION_YES)
goto ndbcluster_init_error; goto ndbcluster_init_error;
...@@ -6320,6 +6210,24 @@ static int ndbcluster_end(ha_panic_function type) ...@@ -6320,6 +6210,24 @@ static int ndbcluster_end(ha_panic_function type)
if (!ndbcluster_inited) if (!ndbcluster_inited)
DBUG_RETURN(0); DBUG_RETURN(0);
#ifdef HAVE_NDB_BINLOG
{
pthread_mutex_lock(&ndbcluster_mutex);
for (uint i= 0; i < ndbcluster_open_tables.records; i++)
{
NDB_SHARE *share=
(NDB_SHARE*) hash_element(&ndbcluster_open_tables, i);
#ifndef DBUG_OFF
fprintf(stderr, "NDB: table share %s with use_count %d not freed\n",
share->key, share->use_count);
#endif
real_free_share(&share);
}
pthread_mutex_unlock(&ndbcluster_mutex);
}
#endif
hash_free(&ndbcluster_open_tables);
if (g_ndb) if (g_ndb)
{ {
#ifndef DBUG_OFF #ifndef DBUG_OFF
...@@ -6341,23 +6249,6 @@ static int ndbcluster_end(ha_panic_function type) ...@@ -6341,23 +6249,6 @@ static int ndbcluster_end(ha_panic_function type)
delete g_ndb_cluster_connection; delete g_ndb_cluster_connection;
g_ndb_cluster_connection= NULL; g_ndb_cluster_connection= NULL;
#ifdef HAVE_NDB_BINLOG
{
pthread_mutex_lock(&ndbcluster_mutex);
for (uint i= 0; i < ndbcluster_open_tables.records; i++)
{
NDB_SHARE *share=
(NDB_SHARE*) hash_element(&ndbcluster_open_tables, i);
#ifndef DBUG_OFF
fprintf(stderr, "NDB: table share %s with use_count %d not freed\n",
share->key, share->use_count);
#endif
real_free_share(&share);
}
pthread_mutex_unlock(&ndbcluster_mutex);
}
#endif
hash_free(&ndbcluster_open_tables);
pthread_mutex_destroy(&ndbcluster_mutex); pthread_mutex_destroy(&ndbcluster_mutex);
pthread_mutex_destroy(&LOCK_ndb_util_thread); pthread_mutex_destroy(&LOCK_ndb_util_thread);
pthread_cond_destroy(&COND_ndb_util_thread); pthread_cond_destroy(&COND_ndb_util_thread);
...@@ -6518,7 +6409,7 @@ ha_ndbcluster::records_in_range(uint inx, key_range *min_key, ...@@ -6518,7 +6409,7 @@ ha_ndbcluster::records_in_range(uint inx, key_range *min_key,
m_index[inx].index_stat != NULL) m_index[inx].index_stat != NULL)
{ {
NDB_INDEX_DATA& d=m_index[inx]; NDB_INDEX_DATA& d=m_index[inx];
NDBINDEX* index=(NDBINDEX*)d.index; const NDBINDEX* index= d.index;
Ndb* ndb=get_ndb(); Ndb* ndb=get_ndb();
NdbTransaction* trans=NULL; NdbTransaction* trans=NULL;
NdbIndexScanOperation* op=NULL; NdbIndexScanOperation* op=NULL;
...@@ -6538,7 +6429,7 @@ ha_ndbcluster::records_in_range(uint inx, key_range *min_key, ...@@ -6538,7 +6429,7 @@ ha_ndbcluster::records_in_range(uint inx, key_range *min_key,
else else
{ {
Ndb_statistics stat; Ndb_statistics stat;
if ((res=ndb_get_table_statistics(ndb, m_tabname, &stat)) != 0) if ((res=ndb_get_table_statistics(ndb, m_table, &stat)) != 0)
break; break;
table_rows=stat.row_count; table_rows=stat.row_count;
DBUG_PRINT("info", ("use db row_count: %llu", table_rows)); DBUG_PRINT("info", ("use db row_count: %llu", table_rows));
...@@ -6699,11 +6590,15 @@ uint ndb_get_commitcount(THD *thd, char *dbname, char *tabname, ...@@ -6699,11 +6590,15 @@ uint ndb_get_commitcount(THD *thd, char *dbname, char *tabname,
pthread_mutex_unlock(&share->mutex); pthread_mutex_unlock(&share->mutex);
struct Ndb_statistics stat; struct Ndb_statistics stat;
if (ndb_get_table_statistics(ndb, tabname, &stat)) {
Ndb_table_guard ndbtab_g(ndb->getDictionary(), tabname);
if (ndbtab_g.get_table() == 0
|| ndb_get_table_statistics(ndb, ndbtab_g.get_table(), &stat))
{ {
free_share(&share); free_share(&share);
DBUG_RETURN(1); DBUG_RETURN(1);
} }
}
pthread_mutex_lock(&share->mutex); pthread_mutex_lock(&share->mutex);
if (share->commit_count_lock == lock) if (share->commit_count_lock == lock)
...@@ -6916,7 +6811,11 @@ int handle_trailing_share(NDB_SHARE *share) ...@@ -6916,7 +6811,11 @@ int handle_trailing_share(NDB_SHARE *share)
++share->use_count; ++share->use_count;
pthread_mutex_unlock(&ndbcluster_mutex); pthread_mutex_unlock(&ndbcluster_mutex);
close_cached_tables((THD*) 0, 0, (TABLE_LIST*) 0, TRUE); TABLE_LIST table_list;
bzero((char*) &table_list,sizeof(table_list));
table_list.db= share->db;
table_list.alias= table_list.table_name= share->table_name;
close_cached_tables(current_thd, 0, &table_list, TRUE);
pthread_mutex_lock(&ndbcluster_mutex); pthread_mutex_lock(&ndbcluster_mutex);
if (!--share->use_count) if (!--share->use_count)
...@@ -7262,15 +7161,17 @@ void ndbcluster_free_share(NDB_SHARE **share, bool have_lock) ...@@ -7262,15 +7161,17 @@ void ndbcluster_free_share(NDB_SHARE **share, bool have_lock)
static static
int int
ndb_get_table_statistics(Ndb* ndb, const char * table, ndb_get_table_statistics(Ndb* ndb, const NDBTAB *ndbtab,
struct Ndb_statistics * ndbstat) struct Ndb_statistics * ndbstat)
{ {
DBUG_ENTER("ndb_get_table_statistics"); DBUG_ENTER("ndb_get_table_statistics");
DBUG_PRINT("enter", ("table: %s", table)); DBUG_PRINT("enter", ("table: %s", ndbtab->getName()));
NdbTransaction* pTrans; NdbTransaction* pTrans;
int retries= 10; int retries= 10;
int retry_sleep= 30 * 1000; /* 30 milliseconds */ int retry_sleep= 30 * 1000; /* 30 milliseconds */
DBUG_ASSERT(ndbtab != 0);
do do
{ {
pTrans= ndb->startTransaction(); pTrans= ndb->startTransaction();
...@@ -7285,7 +7186,7 @@ ndb_get_table_statistics(Ndb* ndb, const char * table, ...@@ -7285,7 +7186,7 @@ ndb_get_table_statistics(Ndb* ndb, const char * table,
ERR_RETURN(ndb->getNdbError()); ERR_RETURN(ndb->getNdbError());
} }
NdbScanOperation* pOp= pTrans->getNdbScanOperation(table); NdbScanOperation* pOp= pTrans->getNdbScanOperation(ndbtab);
if (pOp == NULL) if (pOp == NULL)
break; break;
...@@ -7446,9 +7347,9 @@ ha_ndbcluster::read_multi_range_first(KEY_MULTI_RANGE **found_range_p, ...@@ -7446,9 +7347,9 @@ ha_ndbcluster::read_multi_range_first(KEY_MULTI_RANGE **found_range_p,
byte *end_of_buffer= (byte*)buffer->buffer_end; byte *end_of_buffer= (byte*)buffer->buffer_end;
NdbOperation::LockMode lm= NdbOperation::LockMode lm=
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type); (NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
const NDBTAB *tab= (const NDBTAB *) m_table; const NDBTAB *tab= m_table;
const NDBINDEX *unique_idx= (NDBINDEX *) m_index[active_index].unique_index; const NDBINDEX *unique_idx= m_index[active_index].unique_index;
const NDBINDEX *idx= (NDBINDEX *) m_index[active_index].index; const NDBINDEX *idx= m_index[active_index].index;
const NdbOperation* lastOp= m_active_trans->getLastDefinedOperation(); const NdbOperation* lastOp= m_active_trans->getLastDefinedOperation();
NdbIndexScanOperation* scanOp= 0; NdbIndexScanOperation* scanOp= 0;
for (; multi_range_curr<multi_range_end && curr+reclength <= end_of_buffer; for (; multi_range_curr<multi_range_end && curr+reclength <= end_of_buffer;
...@@ -7797,11 +7698,8 @@ ha_ndbcluster::update_table_comment( ...@@ -7797,11 +7698,8 @@ ha_ndbcluster::update_table_comment(
ndb->setDatabaseName(m_dbname); ndb->setDatabaseName(m_dbname);
NDBDICT* dict= ndb->getDictionary(); NDBDICT* dict= ndb->getDictionary();
const NDBTAB* tab; const NDBTAB* tab= m_table;
if (!(tab= dict->getTable(m_tabname))) DBUG_ASSERT(tab != NULL);
{
return((char*)comment);
}
char *str; char *str;
const char *fmt="%s%snumber_of_replicas: %d"; const char *fmt="%s%snumber_of_replicas: %d";
...@@ -7987,7 +7885,10 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused))) ...@@ -7987,7 +7885,10 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused)))
lock= share->commit_count_lock; lock= share->commit_count_lock;
pthread_mutex_unlock(&share->mutex); pthread_mutex_unlock(&share->mutex);
if (ndb_get_table_statistics(ndb, share->table_name, &stat) == 0) {
Ndb_table_guard ndbtab_g(ndb->getDictionary(), share->table_name);
if (ndbtab_g.get_table() &&
ndb_get_table_statistics(ndb, ndbtab_g.get_table(), &stat) == 0)
{ {
DBUG_PRINT("ndb_util_thread", DBUG_PRINT("ndb_util_thread",
("Table: %s, commit_count: %llu, rows: %llu", ("Table: %s, commit_count: %llu, rows: %llu",
...@@ -8000,6 +7901,7 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused))) ...@@ -8000,6 +7901,7 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused)))
share->key)); share->key));
stat.commit_count= 0; stat.commit_count= 0;
} }
}
pthread_mutex_lock(&share->mutex); pthread_mutex_lock(&share->mutex);
if (share->commit_count_lock == lock) if (share->commit_count_lock == lock)
...@@ -9390,12 +9292,8 @@ char* ha_ndbcluster::get_tablespace_name(THD *thd) ...@@ -9390,12 +9292,8 @@ char* ha_ndbcluster::get_tablespace_name(THD *thd)
NdbError ndberr; NdbError ndberr;
Uint32 id; Uint32 id;
ndb->setDatabaseName(m_dbname); ndb->setDatabaseName(m_dbname);
const NDBTAB *ndbtab= ndbdict->getTable(m_tabname); const NDBTAB *ndbtab= m_table;
if (ndbtab == 0) DBUG_ASSERT(ndbtab != NULL);
{
ndberr= ndbdict->getNdbError();
goto err;
}
if (!ndbtab->getTablespace(&id)) if (!ndbtab->getTablespace(&id))
{ {
return 0; return 0;
...@@ -10060,17 +9958,10 @@ bool ha_ndbcluster::get_no_parts(const char *name, uint *no_parts) ...@@ -10060,17 +9958,10 @@ bool ha_ndbcluster::get_no_parts(const char *name, uint *no_parts)
} }
ndb= get_ndb(); ndb= get_ndb();
ndb->setDatabaseName(m_dbname); ndb->setDatabaseName(m_dbname);
dict= ndb->getDictionary(); Ndb_table_guard ndbtab_g(dict= ndb->getDictionary(), m_tabname);
if (!(tab= dict->getTable(m_tabname))) if (!ndbtab_g.get_table())
ERR_BREAK(dict->getNdbError(), err); ERR_BREAK(dict->getNdbError(), err);
// Check if thread has stale local cache *no_parts= ndbtab_g.get_table()->getFragmentCount();
if (tab->getObjectStatus() == NdbDictionary::Object::Invalid)
{
invalidate_dictionary_cache(FALSE, tab);
if (!(tab= dict->getTable(m_tabname)))
ERR_BREAK(dict->getNdbError(), err);
}
*no_parts= tab->getFragmentCount();
DBUG_RETURN(FALSE); DBUG_RETURN(FALSE);
} while (1); } while (1);
......
...@@ -70,8 +70,8 @@ typedef enum ndb_index_status { ...@@ -70,8 +70,8 @@ typedef enum ndb_index_status {
typedef struct ndb_index_data { typedef struct ndb_index_data {
NDB_INDEX_TYPE type; NDB_INDEX_TYPE type;
NDB_INDEX_STATUS status; NDB_INDEX_STATUS status;
void *index; const NdbDictionary::Index *index;
void *unique_index; const NdbDictionary::Index *unique_index;
unsigned char *unique_index_attrid_map; unsigned char *unique_index_attrid_map;
// In this version stats are not shared between threads // In this version stats are not shared between threads
NdbIndexStat* index_stat; NdbIndexStat* index_stat;
...@@ -560,6 +560,7 @@ class ha_ndbcluster: public handler ...@@ -560,6 +560,7 @@ class ha_ndbcluster: public handler
ha_ndbcluster(TABLE_SHARE *table); ha_ndbcluster(TABLE_SHARE *table);
~ha_ndbcluster(); ~ha_ndbcluster();
int ha_initialise();
int open(const char *name, int mode, uint test_if_locked); int open(const char *name, int mode, uint test_if_locked);
int close(void); int close(void);
...@@ -708,19 +709,15 @@ static void set_tabname(const char *pathname, char *tabname); ...@@ -708,19 +709,15 @@ static void set_tabname(const char *pathname, char *tabname);
Ndb *ndb, NdbEventOperation *pOp, Ndb *ndb, NdbEventOperation *pOp,
NDB_SHARE *share); NDB_SHARE *share);
int alter_table_name(const char *to);
static int delete_table(ha_ndbcluster *h, Ndb *ndb, static int delete_table(ha_ndbcluster *h, Ndb *ndb,
const char *path, const char *path,
const char *db, const char *db,
const char *table_name); const char *table_name);
int drop_ndb_table();
int create_ndb_index(const char *name, KEY *key_info, bool unique); int create_ndb_index(const char *name, KEY *key_info, bool unique);
int create_ordered_index(const char *name, KEY *key_info); int create_ordered_index(const char *name, KEY *key_info);
int create_unique_index(const char *name, KEY *key_info); int create_unique_index(const char *name, KEY *key_info);
int create_index(const char *name, KEY *key_info, int create_index(const char *name, KEY *key_info,
NDB_INDEX_TYPE idx_type, uint idx_no); NDB_INDEX_TYPE idx_type, uint idx_no);
int drop_ndb_index(const char *name);
int table_changed(const void *pack_frm_data, uint pack_frm_len);
// Index list management // Index list management
int create_indexes(Ndb *ndb, TABLE *tab); int create_indexes(Ndb *ndb, TABLE *tab);
void clear_index(int i); void clear_index(int i);
...@@ -732,7 +729,7 @@ static void set_tabname(const char *pathname, char *tabname); ...@@ -732,7 +729,7 @@ static void set_tabname(const char *pathname, char *tabname);
KEY *key_info, const char *index_name, uint index_no); KEY *key_info, const char *index_name, uint index_no);
int initialize_autoincrement(const void *table); int initialize_autoincrement(const void *table);
int get_metadata(const char* path); int get_metadata(const char* path);
void release_metadata(); void release_metadata(THD *thd, Ndb *ndb);
NDB_INDEX_TYPE get_index_type(uint idx_no) const; NDB_INDEX_TYPE get_index_type(uint idx_no) const;
NDB_INDEX_TYPE get_index_type_from_table(uint index_no) const; NDB_INDEX_TYPE get_index_type_from_table(uint index_no) const;
NDB_INDEX_TYPE get_index_type_from_key(uint index_no, KEY *key_info, NDB_INDEX_TYPE get_index_type_from_key(uint index_no, KEY *key_info,
...@@ -795,8 +792,6 @@ static void set_tabname(const char *pathname, char *tabname); ...@@ -795,8 +792,6 @@ static void set_tabname(const char *pathname, char *tabname);
void print_results(); void print_results();
ulonglong get_auto_increment(); ulonglong get_auto_increment();
int invalidate_dictionary_cache(bool global,
const NdbDictionary::Table *ndbtab);
int ndb_err(NdbTransaction*); int ndb_err(NdbTransaction*);
bool uses_blob_value(); bool uses_blob_value();
...@@ -834,7 +829,6 @@ static void set_tabname(const char *pathname, char *tabname); ...@@ -834,7 +829,6 @@ static void set_tabname(const char *pathname, char *tabname);
NdbTransaction *m_active_trans; NdbTransaction *m_active_trans;
NdbScanOperation *m_active_cursor; NdbScanOperation *m_active_cursor;
const NdbDictionary::Table *m_table; const NdbDictionary::Table *m_table;
int m_table_version;
struct Ndb_local_table_statistics *m_table_info; struct Ndb_local_table_statistics *m_table_info;
char m_dbname[FN_HEADLEN]; char m_dbname[FN_HEADLEN];
//char m_schemaname[FN_HEADLEN]; //char m_schemaname[FN_HEADLEN];
......
...@@ -986,7 +986,7 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -986,7 +986,7 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
uint32 ndb_table_id, uint32 ndb_table_id,
uint32 ndb_table_version, uint32 ndb_table_version,
enum SCHEMA_OP_TYPE type, enum SCHEMA_OP_TYPE type,
const char *old_db, const char *old_table_name) const char *new_db, const char *new_table_name)
{ {
DBUG_ENTER("ndbcluster_log_schema_op"); DBUG_ENTER("ndbcluster_log_schema_op");
Thd_ndb *thd_ndb= get_thd_ndb(thd); Thd_ndb *thd_ndb= get_thd_ndb(thd);
...@@ -1026,8 +1026,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -1026,8 +1026,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
/* redo the rename table query as is may contain several tables */ /* redo the rename table query as is may contain several tables */
query= tmp_buf2; query= tmp_buf2;
query_length= (uint) (strxmov(tmp_buf2, "rename table `", query_length= (uint) (strxmov(tmp_buf2, "rename table `",
old_db, ".", old_table_name, "` to `", db, ".", table_name, "` to `",
db, ".", table_name, "`", NullS) - tmp_buf2); new_db, ".", new_table_name, "`", NullS) - tmp_buf2);
type_str= "rename table"; type_str= "rename table";
break; break;
case SOT_CREATE_TABLE: case SOT_CREATE_TABLE:
...@@ -1067,6 +1067,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -1067,6 +1067,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
Uint64 epoch= 0; Uint64 epoch= 0;
MY_BITMAP schema_subscribers; MY_BITMAP schema_subscribers;
uint32 bitbuf[sizeof(ndb_schema_object->slock)/4]; uint32 bitbuf[sizeof(ndb_schema_object->slock)/4];
uint32 bitbuf_e[sizeof(bitbuf)];
bzero((char *)bitbuf_e, sizeof(bitbuf_e));
{ {
int i, updated= 0; int i, updated= 0;
int no_storage_nodes= g_ndb_cluster_connection->no_db_nodes(); int no_storage_nodes= g_ndb_cluster_connection->no_db_nodes();
...@@ -1110,7 +1112,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -1110,7 +1112,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
char tmp_buf[FN_REFLEN]; char tmp_buf[FN_REFLEN];
NDBDICT *dict= ndb->getDictionary(); NDBDICT *dict= ndb->getDictionary();
ndb->setDatabaseName(NDB_REP_DB); ndb->setDatabaseName(NDB_REP_DB);
const NDBTAB *ndbtab= dict->getTable(NDB_SCHEMA_TABLE); Ndb_table_guard ndbtab_g(dict, NDB_SCHEMA_TABLE);
const NDBTAB *ndbtab= ndbtab_g.get_table();
NdbTransaction *trans= 0; NdbTransaction *trans= 0;
int retries= 100; int retries= 100;
const NDBCOL *col[SCHEMA_SIZE]; const NDBCOL *col[SCHEMA_SIZE];
...@@ -1141,8 +1144,13 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -1141,8 +1144,13 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
while (1) while (1)
{ {
const char *log_db= db;
const char *log_tab= table_name;
const char *log_subscribers= (char*)schema_subscribers.bitmap;
uint32 log_type= (uint32)type;
if ((trans= ndb->startTransaction()) == 0) if ((trans= ndb->startTransaction()) == 0)
goto err; goto err;
while (1)
{ {
NdbOperation *op= 0; NdbOperation *op= 0;
int r= 0; int r= 0;
...@@ -1152,17 +1160,17 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -1152,17 +1160,17 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
/* db */ /* db */
ndb_pack_varchar(col[SCHEMA_DB_I], tmp_buf, db, strlen(db)); ndb_pack_varchar(col[SCHEMA_DB_I], tmp_buf, log_db, strlen(log_db));
r|= op->equal(SCHEMA_DB_I, tmp_buf); r|= op->equal(SCHEMA_DB_I, tmp_buf);
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
/* name */ /* name */
ndb_pack_varchar(col[SCHEMA_NAME_I], tmp_buf, table_name, ndb_pack_varchar(col[SCHEMA_NAME_I], tmp_buf, log_tab,
strlen(table_name)); strlen(log_tab));
r|= op->equal(SCHEMA_NAME_I, tmp_buf); r|= op->equal(SCHEMA_NAME_I, tmp_buf);
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
/* slock */ /* slock */
DBUG_ASSERT(sz[SCHEMA_SLOCK_I] == sizeof(bitbuf)); DBUG_ASSERT(sz[SCHEMA_SLOCK_I] == sizeof(bitbuf));
r|= op->setValue(SCHEMA_SLOCK_I, (char*)schema_subscribers.bitmap); r|= op->setValue(SCHEMA_SLOCK_I, log_subscribers);
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
/* query */ /* query */
{ {
...@@ -1186,8 +1194,17 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -1186,8 +1194,17 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
r|= op->setValue(SCHEMA_VERSION_I, ndb_table_version); r|= op->setValue(SCHEMA_VERSION_I, ndb_table_version);
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
/* type */ /* type */
r|= op->setValue(SCHEMA_TYPE_I, (uint32)type); r|= op->setValue(SCHEMA_TYPE_I, log_type);
DBUG_ASSERT(r == 0); DBUG_ASSERT(r == 0);
if (log_db != new_db && new_db && new_table_name)
{
log_db= new_db;
log_tab= new_table_name;
log_subscribers= (const char *)bitbuf_e; // no ack expected on this
log_type= (uint32)SOT_RENAME_TABLE_NEW;
continue;
}
break;
} }
if (trans->execute(NdbTransaction::Commit) == 0) if (trans->execute(NdbTransaction::Commit) == 0)
{ {
...@@ -1306,7 +1323,8 @@ ndbcluster_update_slock(THD *thd, ...@@ -1306,7 +1323,8 @@ ndbcluster_update_slock(THD *thd,
char tmp_buf[FN_REFLEN]; char tmp_buf[FN_REFLEN];
NDBDICT *dict= ndb->getDictionary(); NDBDICT *dict= ndb->getDictionary();
ndb->setDatabaseName(NDB_REP_DB); ndb->setDatabaseName(NDB_REP_DB);
const NDBTAB *ndbtab= dict->getTable(NDB_SCHEMA_TABLE); Ndb_table_guard ndbtab_g(dict, NDB_SCHEMA_TABLE);
const NDBTAB *ndbtab= ndbtab_g.get_table();
NdbTransaction *trans= 0; NdbTransaction *trans= 0;
int retries= 100; int retries= 100;
const NDBCOL *col[SCHEMA_SIZE]; const NDBCOL *col[SCHEMA_SIZE];
...@@ -1452,30 +1470,27 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp, ...@@ -1452,30 +1470,27 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
{ {
if (pOp->tableFrmChanged()) if (pOp->tableFrmChanged())
{ {
DBUG_PRINT("info", ("NDBEVENT::TE_ALTER: table frm changed"));
is_online_alter_table= TRUE; is_online_alter_table= TRUE;
} }
else else
{ {
DBUG_PRINT("info", ("NDBEVENT::TE_ALTER: name changed"));
DBUG_ASSERT(pOp->tableNameChanged()); DBUG_ASSERT(pOp->tableNameChanged());
is_rename_table= TRUE; is_rename_table= TRUE;
} }
} }
/* {
Refresh local dictionary cache by
invalidating table and all it's indexes
*/
ndb->setDatabaseName(dbname); ndb->setDatabaseName(dbname);
Thd_ndb *thd_ndb= get_thd_ndb(thd); Ndb_table_guard ndbtab_g(ndb->getDictionary(), tabname);
DBUG_ASSERT(thd_ndb != NULL); const NDBTAB *ev_tab= pOp->getTable();
Ndb* old_ndb= thd_ndb->ndb; const NDBTAB *cache_tab= ndbtab_g.get_table();
thd_ndb->ndb= ndb; if (cache_tab &&
ha_ndbcluster table_handler(table_share); cache_tab->getObjectId() == ev_tab->getObjectId() &&
(void)strxmov(table_handler.m_dbname, dbname, NullS); cache_tab->getObjectVersion() <= ev_tab->getObjectVersion())
(void)strxmov(table_handler.m_tabname, tabname, NullS); ndbtab_g.invalidate();
table_handler.open_indexes(ndb, table, TRUE); }
table_handler.invalidate_dictionary_cache(TRUE, 0);
thd_ndb->ndb= old_ndb;
/* /*
Refresh local frm file and dictionary cache if Refresh local frm file and dictionary cache if
...@@ -1505,7 +1520,8 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp, ...@@ -1505,7 +1520,8 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
DBUG_DUMP("frm", (char*)altered_table->getFrmData(), DBUG_DUMP("frm", (char*)altered_table->getFrmData(),
altered_table->getFrmLength()); altered_table->getFrmLength());
pthread_mutex_lock(&LOCK_open); pthread_mutex_lock(&LOCK_open);
const NDBTAB *old= dict->getTable(tabname); Ndb_table_guard ndbtab_g(dict, tabname);
const NDBTAB *old= ndbtab_g.get_table();
if (!old && if (!old &&
old->getObjectVersion() != altered_table->getObjectVersion()) old->getObjectVersion() != altered_table->getObjectVersion())
dict->putTable(altered_table); dict->putTable(altered_table);
...@@ -1517,7 +1533,13 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp, ...@@ -1517,7 +1533,13 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
dbname, tabname, error); dbname, tabname, error);
} }
ndbcluster_binlog_close_table(thd, share); ndbcluster_binlog_close_table(thd, share);
close_cached_tables((THD*) 0, 0, (TABLE_LIST*) 0, TRUE);
TABLE_LIST table_list;
bzero((char*) &table_list,sizeof(table_list));
table_list.db= (char *)dbname;
table_list.alias= table_list.table_name= (char *)tabname;
close_cached_tables(thd, 0, &table_list, TRUE);
if ((error= ndbcluster_binlog_open_table(thd, share, if ((error= ndbcluster_binlog_open_table(thd, share,
table_share, table))) table_share, table)))
sql_print_information("NDB: Failed to re-open table %s.%s", sql_print_information("NDB: Failed to re-open table %s.%s",
...@@ -1545,26 +1567,22 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp, ...@@ -1545,26 +1567,22 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
share_prefix, share->table->s->db.str, share_prefix, share->table->s->db.str,
share->table->s->table_name.str, share->table->s->table_name.str,
share->key); share->key);
{
ndb->setDatabaseName(share->table->s->db.str);
Ndb_table_guard ndbtab_g(ndb->getDictionary(),
share->table->s->table_name.str);
const NDBTAB *ev_tab= pOp->getTable();
const NDBTAB *cache_tab= ndbtab_g.get_table();
if (cache_tab &&
cache_tab->getObjectId() == ev_tab->getObjectId() &&
cache_tab->getObjectVersion() <= ev_tab->getObjectVersion())
ndbtab_g.invalidate();
}
/* do the rename of the table in the share */ /* do the rename of the table in the share */
share->table->s->db.str= share->db; share->table->s->db.str= share->db;
share->table->s->db.length= strlen(share->db); share->table->s->db.length= strlen(share->db);
share->table->s->table_name.str= share->table_name; share->table->s->table_name.str= share->table_name;
share->table->s->table_name.length= strlen(share->table_name); share->table->s->table_name.length= strlen(share->table_name);
/*
Refresh local dictionary cache by invalidating any
old table with same name and all it's indexes
*/
ndb->setDatabaseName(dbname);
Thd_ndb *thd_ndb= get_thd_ndb(thd);
DBUG_ASSERT(thd_ndb != NULL);
Ndb* old_ndb= thd_ndb->ndb;
thd_ndb->ndb= ndb;
ha_ndbcluster table_handler(table_share);
table_handler.set_dbname(share->key);
table_handler.set_tabname(share->key);
table_handler.open_indexes(ndb, table, TRUE);
table_handler.invalidate_dictionary_cache(TRUE, 0);
thd_ndb->ndb= old_ndb;
} }
DBUG_ASSERT(share->op == pOp || share->op_old == pOp); DBUG_ASSERT(share->op == pOp || share->op_old == pOp);
if (share->op_old == pOp) if (share->op_old == pOp)
...@@ -1582,14 +1600,19 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp, ...@@ -1582,14 +1600,19 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
if (is_remote_change && share && share->state != NSS_DROPPED) if (is_remote_change && share && share->state != NSS_DROPPED)
{ {
DBUG_PRINT("info", ("remote change")); DBUG_PRINT("info", ("remote change"));
share->state= NSS_DROPPED;
if (share->use_count != 1) if (share->use_count != 1)
do_close_cached_tables= TRUE; do_close_cached_tables= TRUE;
share->state= NSS_DROPPED; else
{
free_share(&share, TRUE); free_share(&share, TRUE);
share= 0;
} }
}
else
share= 0;
pthread_mutex_unlock(&ndbcluster_mutex); pthread_mutex_unlock(&ndbcluster_mutex);
share= 0;
pOp->setCustomData(0); pOp->setCustomData(0);
pthread_mutex_lock(&injector_mutex); pthread_mutex_lock(&injector_mutex);
...@@ -1598,7 +1621,14 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp, ...@@ -1598,7 +1621,14 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
pthread_mutex_unlock(&injector_mutex); pthread_mutex_unlock(&injector_mutex);
if (do_close_cached_tables) if (do_close_cached_tables)
close_cached_tables((THD*) 0, 0, (TABLE_LIST*) 0); {
TABLE_LIST table_list;
bzero((char*) &table_list,sizeof(table_list));
table_list.db= (char *)dbname;
table_list.alias= table_list.table_name= (char *)tabname;
close_cached_tables(thd, 0, &table_list);
free_share(&share);
}
DBUG_RETURN(0); DBUG_RETURN(0);
} }
...@@ -1630,53 +1660,27 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb, ...@@ -1630,53 +1660,27 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
if (schema->node_id != node_id) if (schema->node_id != node_id)
{ {
int log_query= 0, post_epoch_unlock= 0; int log_query= 0, post_epoch_unlock= 0;
DBUG_PRINT("info", ("log query_length: %d query: '%s'", DBUG_PRINT("info",
schema->query_length, schema->query)); ("%s.%s: log query_length: %d query: '%s' type: %d",
schema->db, schema->name,
schema->query_length, schema->query,
schema->type));
char key[FN_REFLEN]; char key[FN_REFLEN];
build_table_filename(key, sizeof(key), schema->db, schema->name, ""); build_table_filename(key, sizeof(key), schema->db, schema->name, "");
NDB_SHARE *share= get_share(key, 0, false, false);
switch ((enum SCHEMA_OP_TYPE)schema->type) switch ((enum SCHEMA_OP_TYPE)schema->type)
{ {
case SOT_DROP_TABLE: case SOT_DROP_TABLE:
/* binlog dropping table after any table operations */ // fall through
if (share && share->op)
{
post_epoch_log_list->push_back(schema, mem_root);
/* acknowledge this query _after_ epoch completion */
post_epoch_unlock= 1;
}
/* table is either ignored or logging is postponed to later */
log_query= 0;
break;
case SOT_RENAME_TABLE: case SOT_RENAME_TABLE:
if (share && share->op) // fall through
{ case SOT_RENAME_TABLE_NEW:
post_epoch_log_list->push_back(schema, mem_root); // fall through
/* acknowledge this query _after_ epoch completion */
post_epoch_unlock= 1;
break; /* discovery will be handled by binlog */
}
goto sot_create_table;
case SOT_ALTER_TABLE: case SOT_ALTER_TABLE:
if (share && share->op)
{
post_epoch_log_list->push_back(schema, mem_root); post_epoch_log_list->push_back(schema, mem_root);
/* acknowledge this query _after_ epoch completion */ /* acknowledge this query _after_ epoch completion */
post_epoch_unlock= 1; post_epoch_unlock= 1;
break; /* discovery will be handled by binlog */ break;
}
goto sot_create_table;
case SOT_CREATE_TABLE: case SOT_CREATE_TABLE:
sot_create_table:
/*
we need to free any share here as command below
may need to call handle_trailing_share
*/
if (share)
{
free_share(&share);
share= 0;
}
pthread_mutex_lock(&LOCK_open); pthread_mutex_lock(&LOCK_open);
if (ndb_create_table_from_engine(thd, schema->db, schema->name)) if (ndb_create_table_from_engine(thd, schema->db, schema->name))
{ {
...@@ -1694,12 +1698,9 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb, ...@@ -1694,12 +1698,9 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
TRUE, /* print error */ TRUE, /* print error */
TRUE); /* don't binlog the query */ TRUE); /* don't binlog the query */
/* binlog dropping database after any table operations */ /* binlog dropping database after any table operations */
if (ndb_binlog_running)
{
post_epoch_log_list->push_back(schema, mem_root); post_epoch_log_list->push_back(schema, mem_root);
/* acknowledge this query _after_ epoch completion */ /* acknowledge this query _after_ epoch completion */
post_epoch_unlock= 1; post_epoch_unlock= 1;
}
break; break;
case SOT_CREATE_DB: case SOT_CREATE_DB:
/* fall through */ /* fall through */
...@@ -1726,8 +1727,6 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb, ...@@ -1726,8 +1727,6 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
pthread_mutex_unlock(&ndb_schema_object->mutex); pthread_mutex_unlock(&ndb_schema_object->mutex);
pthread_cond_signal(&injector_cond); pthread_cond_signal(&injector_cond);
} }
if (share)
free_share(&share, TRUE);
pthread_mutex_unlock(&ndbcluster_mutex); pthread_mutex_unlock(&ndbcluster_mutex);
DBUG_RETURN(0); DBUG_RETURN(0);
} }
...@@ -1736,11 +1735,6 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb, ...@@ -1736,11 +1735,6 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
log_query= 1; log_query= 1;
break; break;
} }
if (share)
{
free_share(&share);
share= 0;
}
if (log_query && ndb_binlog_running) if (log_query && ndb_binlog_running)
{ {
char *thd_db_save= thd->db; char *thd_db_save= thd->db;
...@@ -1864,26 +1858,69 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd, ...@@ -1864,26 +1858,69 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
List<Cluster_schema> List<Cluster_schema>
*post_epoch_unlock_list) *post_epoch_unlock_list)
{ {
if (post_epoch_log_list->elements == 0)
return;
DBUG_ENTER("ndb_binlog_thread_handle_schema_event_post_epoch"); DBUG_ENTER("ndb_binlog_thread_handle_schema_event_post_epoch");
Cluster_schema *schema; Cluster_schema *schema;
while ((schema= post_epoch_log_list->pop())) while ((schema= post_epoch_log_list->pop()))
{ {
DBUG_PRINT("info", ("log query_length: %d query: '%s'", DBUG_PRINT("info",
schema->query_length, schema->query)); ("%s.%s: log query_length: %d query: '%s' type: %d",
schema->db, schema->name,
schema->query_length, schema->query,
schema->type));
int log_query= 0;
{ {
char key[FN_REFLEN]; char key[FN_REFLEN];
build_table_filename(key, sizeof(key), schema->db, schema->name, ""); build_table_filename(key, sizeof(key), schema->db, schema->name, "");
NDB_SHARE *share= get_share(key, 0, false, false); NDB_SHARE *share= get_share(key, 0, false, false);
switch ((enum SCHEMA_OP_TYPE)schema->type) enum SCHEMA_OP_TYPE schema_type= (enum SCHEMA_OP_TYPE)schema->type;
switch (schema_type)
{ {
case SOT_DROP_DB: case SOT_DROP_DB:
log_query= 1;
break;
case SOT_DROP_TABLE: case SOT_DROP_TABLE:
// invalidation already handled by binlog thread
if (share && share->op)
{
log_query= 1;
break; break;
}
// fall through
case SOT_RENAME_TABLE: case SOT_RENAME_TABLE:
// fall through
case SOT_ALTER_TABLE: case SOT_ALTER_TABLE:
if (share && share->op) // invalidation already handled by binlog thread
if (!share || !share->op)
{
{
injector_ndb->setDatabaseName(schema->db);
Ndb_table_guard ndbtab_g(injector_ndb->getDictionary(),
schema->name);
ndbtab_g.invalidate();
}
TABLE_LIST table_list;
bzero((char*) &table_list,sizeof(table_list));
table_list.db= schema->db;
table_list.alias= table_list.table_name= schema->name;
close_cached_tables(thd, 0, &table_list, FALSE);
}
if (schema_type != SOT_ALTER_TABLE)
break;
// fall through
case SOT_RENAME_TABLE_NEW:
log_query= 1;
if (ndb_binlog_running)
{
/*
we need to free any share here as command below
may need to call handle_trailing_share
*/
if (share)
{ {
break; /* discovery handled by binlog */ free_share(&share);
share= 0;
} }
pthread_mutex_lock(&LOCK_open); pthread_mutex_lock(&LOCK_open);
if (ndb_create_table_from_engine(thd, schema->db, schema->name)) if (ndb_create_table_from_engine(thd, schema->db, schema->name))
...@@ -1894,6 +1931,8 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd, ...@@ -1894,6 +1931,8 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
schema->node_id); schema->node_id);
} }
pthread_mutex_unlock(&LOCK_open); pthread_mutex_unlock(&LOCK_open);
}
break;
default: default:
DBUG_ASSERT(false); DBUG_ASSERT(false);
} }
...@@ -1903,6 +1942,7 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd, ...@@ -1903,6 +1942,7 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
share= 0; share= 0;
} }
} }
if (ndb_binlog_running && log_query)
{ {
char *thd_db_save= thd->db; char *thd_db_save= thd->db;
thd->db= schema->db; thd->db= schema->db;
...@@ -2186,7 +2226,8 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key, ...@@ -2186,7 +2226,8 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key,
ndb->setDatabaseName(db); ndb->setDatabaseName(db);
NDBDICT *dict= ndb->getDictionary(); NDBDICT *dict= ndb->getDictionary();
const NDBTAB *ndbtab= dict->getTable(table_name); Ndb_table_guard ndbtab_g(dict, table_name);
const NDBTAB *ndbtab= ndbtab_g.get_table();
if (ndbtab == 0) if (ndbtab == 0)
{ {
if (ndb_extra_logging) if (ndb_extra_logging)
...@@ -2201,7 +2242,8 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key, ...@@ -2201,7 +2242,8 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key,
event should have been created by someone else, event should have been created by someone else,
but let's make sure, and create if it doesn't exist but let's make sure, and create if it doesn't exist
*/ */
if (!dict->getEvent(event_name.c_ptr())) const NDBEVENT *ev= dict->getEvent(event_name.c_ptr());
if (!ev)
{ {
if (ndbcluster_create_event(ndb, ndbtab, event_name.c_ptr(), share)) if (ndbcluster_create_event(ndb, ndbtab, event_name.c_ptr(), share))
{ {
...@@ -2216,9 +2258,12 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key, ...@@ -2216,9 +2258,12 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key,
event_name.c_ptr()); event_name.c_ptr());
} }
else else
{
delete ev;
if (ndb_extra_logging) if (ndb_extra_logging)
sql_print_information("NDB Binlog: DISCOVER TABLE Event: %s", sql_print_information("NDB Binlog: DISCOVER TABLE Event: %s",
event_name.c_ptr()); event_name.c_ptr());
}
/* /*
create the event operations for receiving logging events create the event operations for receiving logging events
...@@ -2328,8 +2373,10 @@ ndbcluster_create_event(Ndb *ndb, const NDBTAB *ndbtab, ...@@ -2328,8 +2373,10 @@ ndbcluster_create_event(Ndb *ndb, const NDBTAB *ndbtab,
try retrieving the event, if table version/id matches, we will get try retrieving the event, if table version/id matches, we will get
a valid event. Otherwise we have a trailing event from before a valid event. Otherwise we have a trailing event from before
*/ */
if (dict->getEvent(event_name)) const NDBEVENT *ev;
if ((ev= dict->getEvent(event_name)))
{ {
delete ev;
DBUG_RETURN(0); DBUG_RETURN(0);
} }
......
...@@ -41,14 +41,15 @@ enum SCHEMA_OP_TYPE ...@@ -41,14 +41,15 @@ enum SCHEMA_OP_TYPE
{ {
SOT_DROP_TABLE= 0, SOT_DROP_TABLE= 0,
SOT_CREATE_TABLE= 1, SOT_CREATE_TABLE= 1,
SOT_RENAME_TABLE= 2, SOT_RENAME_TABLE_NEW= 2,
SOT_ALTER_TABLE= 3, SOT_ALTER_TABLE= 3,
SOT_DROP_DB= 4, SOT_DROP_DB= 4,
SOT_CREATE_DB= 5, SOT_CREATE_DB= 5,
SOT_ALTER_DB= 6, SOT_ALTER_DB= 6,
SOT_CLEAR_SLOCK= 7, SOT_CLEAR_SLOCK= 7,
SOT_TABLESPACE= 8, SOT_TABLESPACE= 8,
SOT_LOGFILE_GROUP= 9 SOT_LOGFILE_GROUP= 9,
SOT_RENAME_TABLE= 10
}; };
const uint max_ndb_nodes= 64; /* multiple of 32 */ const uint max_ndb_nodes= 64; /* multiple of 32 */
...@@ -56,6 +57,45 @@ const uint max_ndb_nodes= 64; /* multiple of 32 */ ...@@ -56,6 +57,45 @@ const uint max_ndb_nodes= 64; /* multiple of 32 */
static const char *ha_ndb_ext=".ndb"; static const char *ha_ndb_ext=".ndb";
static const char share_prefix[]= "./"; static const char share_prefix[]= "./";
class Ndb_table_guard
{
public:
Ndb_table_guard(NDBDICT *dict, const char *tabname)
: m_dict(dict)
{
DBUG_ENTER("Ndb_table_guard");
m_ndbtab= m_dict->getTableGlobal(tabname);
m_invalidate= 0;
DBUG_PRINT("info", ("m_ndbtab: %p", m_ndbtab));
DBUG_VOID_RETURN;
}
~Ndb_table_guard()
{
DBUG_ENTER("~Ndb_table_guard");
if (m_ndbtab)
{
DBUG_PRINT("info", ("m_ndbtab: %p m_invalidate: %d",
m_ndbtab, m_invalidate));
m_dict->removeTableGlobal(*m_ndbtab, m_invalidate);
}
DBUG_VOID_RETURN;
}
const NDBTAB *get_table() { return m_ndbtab; }
void invalidate() { m_invalidate= 1; }
const NDBTAB *release()
{
DBUG_ENTER("Ndb_table_guard::release");
const NDBTAB *tmp= m_ndbtab;
DBUG_PRINT("info", ("m_ndbtab: %p", m_ndbtab));
m_ndbtab = 0;
DBUG_RETURN(tmp);
}
private:
const NDBTAB *m_ndbtab;
NDBDICT *m_dict;
int m_invalidate;
};
#ifdef HAVE_NDB_BINLOG #ifdef HAVE_NDB_BINLOG
extern pthread_t ndb_binlog_thread; extern pthread_t ndb_binlog_thread;
extern pthread_mutex_t injector_mutex; extern pthread_mutex_t injector_mutex;
...@@ -98,8 +138,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share, ...@@ -98,8 +138,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
uint32 ndb_table_id, uint32 ndb_table_id,
uint32 ndb_table_version, uint32 ndb_table_version,
enum SCHEMA_OP_TYPE type, enum SCHEMA_OP_TYPE type,
const char *old_db= 0, const char *new_db= 0,
const char *old_table_name= 0); const char *new_table_name= 0);
int ndbcluster_handle_drop_table(Ndb *ndb, const char *event_name, int ndbcluster_handle_drop_table(Ndb *ndb, const char *event_name,
NDB_SHARE *share, NDB_SHARE *share,
const char *type_str); const char *type_str);
......
...@@ -1470,6 +1470,8 @@ public: ...@@ -1470,6 +1470,8 @@ public:
* *
* @return tuple id or 0 on error * @return tuple id or 0 on error
*/ */
int initAutoIncrement();
Uint64 getAutoIncrementValue(const char* aTableName, Uint64 getAutoIncrementValue(const char* aTableName,
Uint32 cacheSize = 1); Uint32 cacheSize = 1);
Uint64 getAutoIncrementValue(const NdbDictionary::Table * aTable, Uint64 getAutoIncrementValue(const NdbDictionary::Table * aTable,
...@@ -1694,6 +1696,7 @@ private: ...@@ -1694,6 +1696,7 @@ private:
// The tupleId is retreived from DB the // The tupleId is retreived from DB the
// tupleId is unique for each tableid. // tupleId is unique for each tableid.
const NdbDictionary::Table *m_sys_tab_0;
Uint64 theFirstTupleId[2048]; Uint64 theFirstTupleId[2048];
Uint64 theLastTupleId[2048]; Uint64 theLastTupleId[2048];
......
...@@ -798,6 +798,7 @@ public: ...@@ -798,6 +798,7 @@ public:
* Get object status * Get object status
*/ */
virtual Object::Status getObjectStatus() const; virtual Object::Status getObjectStatus() const;
void setStatusInvalid() const;
/** /**
* Get object version * Get object version
...@@ -1734,6 +1735,7 @@ public: ...@@ -1734,6 +1735,7 @@ public:
* @return 0 if successful otherwise -1. * @return 0 if successful otherwise -1.
*/ */
int createIndex(const Index &index); int createIndex(const Index &index);
int createIndex(const Index &index, const Table &table);
/** /**
* Drop index with given name * Drop index with given name
...@@ -1805,6 +1807,15 @@ public: ...@@ -1805,6 +1807,15 @@ public:
#ifndef DOXYGEN_SHOULD_SKIP_INTERNAL #ifndef DOXYGEN_SHOULD_SKIP_INTERNAL
const Table * getTable(const char * name, void **data) const; const Table * getTable(const char * name, void **data) const;
void set_local_table_data_size(unsigned sz); void set_local_table_data_size(unsigned sz);
const Index * getIndexGlobal(const char * indexName,
const Table &ndbtab) const;
const Table * getTableGlobal(const char * tableName) const;
int alterTableGlobal(const Table &f, const Table &t);
int dropTableGlobal(const Table &ndbtab);
int dropIndexGlobal(const Index &index);
int removeIndexGlobal(const Index &ndbidx, int invalidate) const;
int removeTableGlobal(const Table &ndbtab, int invalidate) const;
#endif #endif
}; };
}; };
......
...@@ -56,7 +56,7 @@ public: ...@@ -56,7 +56,7 @@ public:
* multiplied by a percentage obtained from the cache (result zero is * multiplied by a percentage obtained from the cache (result zero is
* returned as 1). * returned as 1).
*/ */
int records_in_range(NdbDictionary::Index* index, int records_in_range(const NdbDictionary::Index* index,
NdbIndexScanOperation* op, NdbIndexScanOperation* op,
Uint64 table_rows, Uint64 table_rows,
Uint64* count, Uint64* count,
......
...@@ -63,6 +63,7 @@ LocalDictCache::~LocalDictCache(){ ...@@ -63,6 +63,7 @@ LocalDictCache::~LocalDictCache(){
Ndb_local_table_info * Ndb_local_table_info *
LocalDictCache::get(const char * name){ LocalDictCache::get(const char * name){
ASSERT_NOT_MYSQLD;
assert(! is_ndb_blob_table(name)); assert(! is_ndb_blob_table(name));
const Uint32 len = strlen(name); const Uint32 len = strlen(name);
return m_tableHash.getData(name, len); return m_tableHash.getData(name, len);
...@@ -70,6 +71,7 @@ LocalDictCache::get(const char * name){ ...@@ -70,6 +71,7 @@ LocalDictCache::get(const char * name){
void void
LocalDictCache::put(const char * name, Ndb_local_table_info * tab_info){ LocalDictCache::put(const char * name, Ndb_local_table_info * tab_info){
ASSERT_NOT_MYSQLD;
assert(! is_ndb_blob_table(name)); assert(! is_ndb_blob_table(name));
const Uint32 id = tab_info->m_table_impl->m_id; const Uint32 id = tab_info->m_table_impl->m_id;
m_tableHash.insertKey(name, strlen(name), id, tab_info); m_tableHash.insertKey(name, strlen(name), id, tab_info);
...@@ -77,6 +79,7 @@ LocalDictCache::put(const char * name, Ndb_local_table_info * tab_info){ ...@@ -77,6 +79,7 @@ LocalDictCache::put(const char * name, Ndb_local_table_info * tab_info){
void void
LocalDictCache::drop(const char * name){ LocalDictCache::drop(const char * name){
ASSERT_NOT_MYSQLD;
assert(! is_ndb_blob_table(name)); assert(! is_ndb_blob_table(name));
Ndb_local_table_info *info= m_tableHash.deleteKey(name, strlen(name)); Ndb_local_table_info *info= m_tableHash.deleteKey(name, strlen(name));
DBUG_ASSERT(info != 0); DBUG_ASSERT(info != 0);
...@@ -100,9 +103,16 @@ GlobalDictCache::~GlobalDictCache(){ ...@@ -100,9 +103,16 @@ GlobalDictCache::~GlobalDictCache(){
Vector<TableVersion> * vers = curr->theData; Vector<TableVersion> * vers = curr->theData;
const unsigned sz = vers->size(); const unsigned sz = vers->size();
for(unsigned i = 0; i<sz ; i++){ for(unsigned i = 0; i<sz ; i++){
if((* vers)[i].m_impl != 0) TableVersion tv= (*vers)[i];
DBUG_PRINT(" ", ("vers[%d]: ver: %d, refCount: %d, status: %d",
i, tv.m_version, tv.m_refCount, tv.m_status));
if(tv.m_impl != 0)
{
DBUG_PRINT(" ", ("m_impl: internalname: %s",
tv.m_impl->m_internalName.c_str()));
delete (* vers)[i].m_impl; delete (* vers)[i].m_impl;
} }
}
delete curr->theData; delete curr->theData;
curr->theData= NULL; curr->theData= NULL;
curr = m_tableHash.getNext(curr); curr = m_tableHash.getNext(curr);
...@@ -164,8 +174,15 @@ GlobalDictCache::get(const char * name) ...@@ -164,8 +174,15 @@ GlobalDictCache::get(const char * name)
TableVersion * ver = & versions->back(); TableVersion * ver = & versions->back();
switch(ver->m_status){ switch(ver->m_status){
case OK: case OK:
if (ver->m_impl->m_status == NdbDictionary::Object::Invalid)
{
ver->m_status = DROPPED;
retreive = true; // Break loop
break;
}
ver->m_refCount++; ver->m_refCount++;
DBUG_PRINT("info", ("Table OK version=%x.%x refCount=%u", DBUG_PRINT("info", ("Table OK tab: %p version=%x.%x refCount=%u",
ver->m_impl,
ver->m_impl->m_version & 0xFFFFFF, ver->m_impl->m_version & 0xFFFFFF,
ver->m_impl->m_version >> 24, ver->m_impl->m_version >> 24,
ver->m_refCount)); ver->m_refCount));
...@@ -197,8 +214,8 @@ NdbTableImpl * ...@@ -197,8 +214,8 @@ NdbTableImpl *
GlobalDictCache::put(const char * name, NdbTableImpl * tab) GlobalDictCache::put(const char * name, NdbTableImpl * tab)
{ {
DBUG_ENTER("GlobalDictCache::put"); DBUG_ENTER("GlobalDictCache::put");
DBUG_PRINT("enter", ("name: %s, internal_name: %s version: %x.%x", DBUG_PRINT("enter", ("tab: %p name: %s, internal_name: %s version: %x.%x",
name, tab, name,
tab ? tab->m_internalName.c_str() : "tab NULL", tab ? tab->m_internalName.c_str() : "tab NULL",
tab ? tab->m_version & 0xFFFFFF : 0, tab ? tab->m_version & 0xFFFFFF : 0,
tab ? tab->m_version >> 24 : 0)); tab ? tab->m_version >> 24 : 0));
...@@ -264,10 +281,11 @@ GlobalDictCache::put(const char * name, NdbTableImpl * tab) ...@@ -264,10 +281,11 @@ GlobalDictCache::put(const char * name, NdbTableImpl * tab)
} }
void void
GlobalDictCache::drop(NdbTableImpl * tab) GlobalDictCache::release(NdbTableImpl * tab, int invalidate)
{ {
DBUG_ENTER("GlobalDictCache::drop"); DBUG_ENTER("GlobalDictCache::release");
DBUG_PRINT("enter", ("internal_name: %s", tab->m_internalName.c_str())); DBUG_PRINT("enter", ("tab: %p internal_name: %s",
tab, tab->m_internalName.c_str()));
assert(! is_ndb_blob_table(tab)); assert(! is_ndb_blob_table(tab));
unsigned i; unsigned i;
...@@ -292,15 +310,19 @@ GlobalDictCache::drop(NdbTableImpl * tab) ...@@ -292,15 +310,19 @@ GlobalDictCache::drop(NdbTableImpl * tab)
if(ver.m_impl == tab){ if(ver.m_impl == tab){
if(ver.m_refCount == 0 || ver.m_status == RETREIVING || if(ver.m_refCount == 0 || ver.m_status == RETREIVING ||
ver.m_version != tab->m_version){ ver.m_version != tab->m_version){
DBUG_PRINT("info", ("Dropping with refCount=%d status=%d impl=%p", DBUG_PRINT("info", ("Releasing with refCount=%d status=%d impl=%p",
ver.m_refCount, ver.m_status, ver.m_impl)); ver.m_refCount, ver.m_status, ver.m_impl));
break; break;
} }
DBUG_PRINT("info", ("Found table to drop, i: %d, name: %s",
i, ver.m_impl->m_internalName.c_str()));
ver.m_refCount--; ver.m_refCount--;
if (ver.m_impl->m_status == NdbDictionary::Object::Invalid || invalidate)
{
ver.m_impl->m_status = NdbDictionary::Object::Invalid;
ver.m_status = DROPPED; ver.m_status = DROPPED;
if(ver.m_refCount == 0){ }
if (ver.m_refCount == 0 && ver.m_status == DROPPED)
{
DBUG_PRINT("info", ("refCount is zero, deleting m_impl")); DBUG_PRINT("info", ("refCount is zero, deleting m_impl"));
delete ver.m_impl; delete ver.m_impl;
vers->erase(i); vers->erase(i);
...@@ -319,61 +341,13 @@ GlobalDictCache::drop(NdbTableImpl * tab) ...@@ -319,61 +341,13 @@ GlobalDictCache::drop(NdbTableImpl * tab)
abort(); abort();
} }
void
GlobalDictCache::release(NdbTableImpl * tab)
{
DBUG_ENTER("GlobalDictCache::release");
DBUG_PRINT("enter", ("internal_name: %s", tab->m_internalName.c_str()));
assert(! is_ndb_blob_table(tab));
unsigned i;
const Uint32 len = strlen(tab->m_internalName.c_str());
Vector<TableVersion> * vers =
m_tableHash.getData(tab->m_internalName.c_str(), len);
if(vers == 0){
// Should always tried to retreive it first
// and thus there should be a record
abort();
}
const Uint32 sz = vers->size();
if(sz == 0){
// Should always tried to retreive it first
// and thus there should be a record
abort();
}
for(i = 0; i < sz; i++){
TableVersion & ver = (* vers)[i];
if(ver.m_impl == tab){
if(ver.m_refCount == 0 || ver.m_status == RETREIVING ||
ver.m_version != tab->m_version){
DBUG_PRINT("info", ("Releasing with refCount=%d status=%d impl=%p",
ver.m_refCount, ver.m_status, ver.m_impl));
break;
}
ver.m_refCount--;
DBUG_VOID_RETURN;
}
}
for(i = 0; i<sz; i++){
TableVersion & ver = (* vers)[i];
ndbout_c("%d: version: %d refCount: %d status: %d impl: %p",
i, ver.m_version, ver.m_refCount,
ver.m_status, ver.m_impl);
}
abort();
}
void void
GlobalDictCache::alter_table_rep(const char * name, GlobalDictCache::alter_table_rep(const char * name,
Uint32 tableId, Uint32 tableId,
Uint32 tableVersion, Uint32 tableVersion,
bool altered) bool altered)
{ {
DBUG_ENTER("GlobalDictCache::alter_table_rep");
assert(! is_ndb_blob_table(name)); assert(! is_ndb_blob_table(name));
const Uint32 len = strlen(name); const Uint32 len = strlen(name);
Vector<TableVersion> * vers = Vector<TableVersion> * vers =
...@@ -381,13 +355,13 @@ GlobalDictCache::alter_table_rep(const char * name, ...@@ -381,13 +355,13 @@ GlobalDictCache::alter_table_rep(const char * name,
if(vers == 0) if(vers == 0)
{ {
return; DBUG_VOID_RETURN;
} }
const Uint32 sz = vers->size(); const Uint32 sz = vers->size();
if(sz == 0) if(sz == 0)
{ {
return; DBUG_VOID_RETURN;
} }
for(Uint32 i = 0; i < sz; i++) for(Uint32 i = 0; i < sz; i++)
...@@ -399,15 +373,16 @@ GlobalDictCache::alter_table_rep(const char * name, ...@@ -399,15 +373,16 @@ GlobalDictCache::alter_table_rep(const char * name,
ver.m_status = DROPPED; ver.m_status = DROPPED;
ver.m_impl->m_status = altered ? ver.m_impl->m_status = altered ?
NdbDictionary::Object::Altered : NdbDictionary::Object::Invalid; NdbDictionary::Object::Altered : NdbDictionary::Object::Invalid;
return; DBUG_VOID_RETURN;
} }
if(i == sz - 1 && ver.m_status == RETREIVING) if(i == sz - 1 && ver.m_status == RETREIVING)
{ {
ver.m_impl = altered ? &f_altered_table : &f_invalid_table; ver.m_impl = altered ? &f_altered_table : &f_invalid_table;
return; DBUG_VOID_RETURN;
} }
} }
DBUG_VOID_RETURN;
} }
template class Vector<GlobalDictCache::TableVersion>; template class Vector<GlobalDictCache::TableVersion>;
...@@ -63,11 +63,11 @@ public: ...@@ -63,11 +63,11 @@ public:
GlobalDictCache(); GlobalDictCache();
~GlobalDictCache(); ~GlobalDictCache();
NdbTableImpl * get(NdbTableImpl *tab);
NdbTableImpl * get(const char * name); NdbTableImpl * get(const char * name);
NdbTableImpl* put(const char * name, NdbTableImpl *); NdbTableImpl* put(const char * name, NdbTableImpl *);
void drop(NdbTableImpl *); void release(NdbTableImpl *, int invalidate = 0);
void release(NdbTableImpl *);
void alter_table_rep(const char * name, void alter_table_rep(const char * name,
Uint32 tableId, Uint32 tableVersion, bool altered); Uint32 tableId, Uint32 tableVersion, bool altered);
......
...@@ -901,6 +901,27 @@ Ndb::setTupleIdInNdb(Uint32 aTableId, Uint64 val, bool increase ) ...@@ -901,6 +901,27 @@ Ndb::setTupleIdInNdb(Uint32 aTableId, Uint64 val, bool increase )
DBUG_RETURN((opTupleIdOnNdb(aTableId, val, 1) == val)); DBUG_RETURN((opTupleIdOnNdb(aTableId, val, 1) == val));
} }
int Ndb::initAutoIncrement()
{
if (m_sys_tab_0)
return 0;
BaseString currentDb(getDatabaseName());
BaseString currentSchema(getDatabaseSchemaName());
setDatabaseName("sys");
setDatabaseSchemaName("def");
m_sys_tab_0 = getDictionary()->getTableGlobal("SYSTAB_0");
// Restore current name space
setDatabaseName(currentDb.c_str());
setDatabaseSchemaName(currentSchema.c_str());
return (m_sys_tab_0 == NULL);
}
Uint64 Uint64
Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op) Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op)
{ {
...@@ -916,19 +937,14 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op) ...@@ -916,19 +937,14 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op)
CHECK_STATUS_MACRO_ZERO; CHECK_STATUS_MACRO_ZERO;
BaseString currentDb(getDatabaseName()); if (initAutoIncrement())
BaseString currentSchema(getDatabaseSchemaName()); goto error_return;
setDatabaseName("sys");
setDatabaseSchemaName("def");
tConnection = this->startTransaction(); tConnection = this->startTransaction();
if (tConnection == NULL) if (tConnection == NULL)
goto error_return; goto error_return;
if (usingFullyQualifiedNames()) tOperation = tConnection->getNdbOperation(m_sys_tab_0);
tOperation = tConnection->getNdbOperation("SYSTAB_0");
else
tOperation = tConnection->getNdbOperation("sys/def/SYSTAB_0");
if (tOperation == NULL) if (tOperation == NULL)
goto error_handler; goto error_handler;
...@@ -997,20 +1013,12 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op) ...@@ -997,20 +1013,12 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op)
this->closeTransaction(tConnection); this->closeTransaction(tConnection);
// Restore current name space
setDatabaseName(currentDb.c_str());
setDatabaseSchemaName(currentSchema.c_str());
DBUG_RETURN(ret); DBUG_RETURN(ret);
error_handler: error_handler:
theError.code = tConnection->theError.code; theError.code = tConnection->theError.code;
this->closeTransaction(tConnection); this->closeTransaction(tConnection);
error_return: error_return:
// Restore current name space
setDatabaseName(currentDb.c_str());
setDatabaseSchemaName(currentSchema.c_str());
DBUG_PRINT("error", ("ndb=%d con=%d op=%d", DBUG_PRINT("error", ("ndb=%d con=%d op=%d",
theError.code, theError.code,
tConnection ? tConnection->theError.code : -1, tConnection ? tConnection->theError.code : -1,
......
...@@ -559,6 +559,11 @@ NdbDictionary::Table::getObjectStatus() const { ...@@ -559,6 +559,11 @@ NdbDictionary::Table::getObjectStatus() const {
return m_impl.m_status; return m_impl.m_status;
} }
void
NdbDictionary::Table::setStatusInvalid() const {
m_impl.m_status = NdbDictionary::Object::Invalid;
}
int int
NdbDictionary::Table::getObjectVersion() const { NdbDictionary::Table::getObjectVersion() const {
return m_impl.m_version; return m_impl.m_version;
...@@ -1330,6 +1335,11 @@ NdbDictionary::Dictionary::dropTable(Table & t){ ...@@ -1330,6 +1335,11 @@ NdbDictionary::Dictionary::dropTable(Table & t){
return m_impl.dropTable(NdbTableImpl::getImpl(t)); return m_impl.dropTable(NdbTableImpl::getImpl(t));
} }
int
NdbDictionary::Dictionary::dropTableGlobal(const Table & t){
return m_impl.dropTableGlobal(NdbTableImpl::getImpl(t));
}
int int
NdbDictionary::Dictionary::dropTable(const char * name){ NdbDictionary::Dictionary::dropTable(const char * name){
return m_impl.dropTable(name); return m_impl.dropTable(name);
...@@ -1340,6 +1350,14 @@ NdbDictionary::Dictionary::alterTable(const Table & t){ ...@@ -1340,6 +1350,14 @@ NdbDictionary::Dictionary::alterTable(const Table & t){
return m_impl.alterTable(NdbTableImpl::getImpl(t)); return m_impl.alterTable(NdbTableImpl::getImpl(t));
} }
int
NdbDictionary::Dictionary::alterTableGlobal(const Table & f,
const Table & t)
{
return m_impl.alterTableGlobal(NdbTableImpl::getImpl(f),
NdbTableImpl::getImpl(t));
}
const NdbDictionary::Table * const NdbDictionary::Table *
NdbDictionary::Dictionary::getTable(const char * name, void **data) const NdbDictionary::Dictionary::getTable(const char * name, void **data) const
{ {
...@@ -1349,6 +1367,40 @@ NdbDictionary::Dictionary::getTable(const char * name, void **data) const ...@@ -1349,6 +1367,40 @@ NdbDictionary::Dictionary::getTable(const char * name, void **data) const
return 0; return 0;
} }
const NdbDictionary::Index *
NdbDictionary::Dictionary::getIndexGlobal(const char * indexName,
const Table &ndbtab) const
{
NdbIndexImpl * i = m_impl.getIndexGlobal(indexName,
NdbTableImpl::getImpl(ndbtab));
if(i)
return i->m_facade;
return 0;
}
const NdbDictionary::Table *
NdbDictionary::Dictionary::getTableGlobal(const char * name) const
{
NdbTableImpl * t = m_impl.getTableGlobal(name);
if(t)
return t->m_facade;
return 0;
}
int
NdbDictionary::Dictionary::removeIndexGlobal(const Index &ndbidx,
int invalidate) const
{
return m_impl.releaseIndexGlobal(NdbIndexImpl::getImpl(ndbidx), invalidate);
}
int
NdbDictionary::Dictionary::removeTableGlobal(const Table &ndbtab,
int invalidate) const
{
return m_impl.releaseTableGlobal(NdbTableImpl::getImpl(ndbtab), invalidate);
}
void NdbDictionary::Dictionary::putTable(const NdbDictionary::Table * table) void NdbDictionary::Dictionary::putTable(const NdbDictionary::Table * table)
{ {
NdbDictionary::Table *copy_table = new NdbDictionary::Table; NdbDictionary::Table *copy_table = new NdbDictionary::Table;
...@@ -1420,6 +1472,13 @@ NdbDictionary::Dictionary::createIndex(const Index & ind) ...@@ -1420,6 +1472,13 @@ NdbDictionary::Dictionary::createIndex(const Index & ind)
return m_impl.createIndex(NdbIndexImpl::getImpl(ind)); return m_impl.createIndex(NdbIndexImpl::getImpl(ind));
} }
int
NdbDictionary::Dictionary::createIndex(const Index & ind, const Table & tab)
{
return m_impl.createIndex(NdbIndexImpl::getImpl(ind),
NdbTableImpl::getImpl(tab));
}
int int
NdbDictionary::Dictionary::dropIndex(const char * indexName, NdbDictionary::Dictionary::dropIndex(const char * indexName,
const char * tableName) const char * tableName)
...@@ -1427,6 +1486,12 @@ NdbDictionary::Dictionary::dropIndex(const char * indexName, ...@@ -1427,6 +1486,12 @@ NdbDictionary::Dictionary::dropIndex(const char * indexName,
return m_impl.dropIndex(indexName, tableName); return m_impl.dropIndex(indexName, tableName);
} }
int
NdbDictionary::Dictionary::dropIndexGlobal(const Index &ind)
{
return m_impl.dropIndexGlobal(NdbIndexImpl::getImpl(ind));
}
const NdbDictionary::Index * const NdbDictionary::Index *
NdbDictionary::Dictionary::getIndex(const char * indexName, NdbDictionary::Dictionary::getIndex(const char * indexName,
const char * tableName) const const char * tableName) const
......
...@@ -50,7 +50,14 @@ ...@@ -50,7 +50,14 @@
#define DICT_WAITFOR_TIMEOUT (7*24*60*60*1000) #define DICT_WAITFOR_TIMEOUT (7*24*60*60*1000)
#define ERR_RETURN(a,b) \
{\
DBUG_PRINT("exit", ("error %d", (a).code));\
DBUG_RETURN(b);\
}
extern Uint64 g_latest_trans_gci; extern Uint64 g_latest_trans_gci;
int ndb_dictionary_is_mysqld = 0;
bool bool
is_ndb_blob_table(const char* name, Uint32* ptab_id, Uint32* pcol_no) is_ndb_blob_table(const char* name, Uint32* ptab_id, Uint32* pcol_no)
...@@ -1288,44 +1295,30 @@ NdbDictionaryImpl::~NdbDictionaryImpl() ...@@ -1288,44 +1295,30 @@ NdbDictionaryImpl::~NdbDictionaryImpl()
} }
} }
Ndb_local_table_info * NdbTableImpl *
NdbDictionaryImpl::fetchGlobalTableImpl(const BaseString& internalTableName) NdbDictionaryImpl::fetchGlobalTableImplRef(const GlobalCacheInitObject &obj)
{ {
DBUG_ENTER("fetchGlobalTableImplRef");
NdbTableImpl *impl; NdbTableImpl *impl;
m_globalHash->lock(); m_globalHash->lock();
impl = m_globalHash->get(internalTableName.c_str()); impl = m_globalHash->get(obj.m_name.c_str());
m_globalHash->unlock(); m_globalHash->unlock();
if (impl == 0){ if (impl == 0){
impl = m_receiver.getTable(internalTableName, impl = m_receiver.getTable(obj.m_name.c_str(),
m_ndb.usingFullyQualifiedNames()); m_ndb.usingFullyQualifiedNames());
if (impl != 0) { if (impl != 0 && obj.init(*impl))
int ret = getBlobTables(*impl); {
if (ret != 0) {
delete impl; delete impl;
impl = 0; impl = 0;
} }
}
m_globalHash->lock(); m_globalHash->lock();
m_globalHash->put(internalTableName.c_str(), impl); m_globalHash->put(obj.m_name.c_str(), impl);
m_globalHash->unlock(); m_globalHash->unlock();
if(impl == 0){
return 0;
} }
}
Ndb_local_table_info *info=
Ndb_local_table_info::create(impl, m_local_table_data_size);
m_localHash.put(internalTableName.c_str(), info);
m_ndb.theFirstTupleId[impl->getTableId()] = ~0; DBUG_RETURN(impl);
m_ndb.theLastTupleId[impl->getTableId()] = ~0;
return info;
} }
void void
...@@ -2276,18 +2269,30 @@ int NdbDictionaryImpl::alterTable(NdbTableImpl &impl) ...@@ -2276,18 +2269,30 @@ int NdbDictionaryImpl::alterTable(NdbTableImpl &impl)
} }
// Alter the table // Alter the table
int ret = m_receiver.alterTable(m_ndb, impl); int ret = alterTableGlobal(*local->m_table_impl, impl);
if(ret == 0){ if(ret == 0)
// Remove cached information and let it be refreshed at next access {
m_globalHash->lock(); m_globalHash->lock();
local->m_table_impl->m_status = NdbDictionary::Object::Invalid; m_globalHash->release(local->m_table_impl, 1);
m_globalHash->drop(local->m_table_impl);
m_globalHash->unlock(); m_globalHash->unlock();
m_localHash.drop(originalInternalName); m_localHash.drop(originalInternalName);
} }
DBUG_RETURN(ret); DBUG_RETURN(ret);
} }
int NdbDictionaryImpl::alterTableGlobal(NdbTableImpl &old_impl,
NdbTableImpl &impl)
{
DBUG_ENTER("NdbDictionaryImpl::alterTableGlobal");
// Alter the table
int ret = m_receiver.alterTable(m_ndb, impl);
old_impl.m_status = NdbDictionary::Object::Invalid;
if(ret == 0){
DBUG_RETURN(ret);
}
ERR_RETURN(getNdbError(), ret);
}
int int
NdbDictInterface::alterTable(Ndb & ndb, NdbDictInterface::alterTable(Ndb & ndb,
NdbTableImpl & impl) NdbTableImpl & impl)
...@@ -2731,6 +2736,7 @@ NdbDictionaryImpl::dropTable(const char * name) ...@@ -2731,6 +2736,7 @@ NdbDictionaryImpl::dropTable(const char * name)
{ {
DBUG_ENTER("NdbDictionaryImpl::dropTable"); DBUG_ENTER("NdbDictionaryImpl::dropTable");
DBUG_PRINT("enter",("name: %s", name)); DBUG_PRINT("enter",("name: %s", name));
ASSERT_NOT_MYSQLD;
NdbTableImpl * tab = getTable(name); NdbTableImpl * tab = getTable(name);
if(tab == 0){ if(tab == 0){
DBUG_RETURN(-1); DBUG_RETURN(-1);
...@@ -2743,8 +2749,7 @@ NdbDictionaryImpl::dropTable(const char * name) ...@@ -2743,8 +2749,7 @@ NdbDictionaryImpl::dropTable(const char * name)
DBUG_PRINT("info",("INCOMPATIBLE_VERSION internal_name: %s", internalTableName.c_str())); DBUG_PRINT("info",("INCOMPATIBLE_VERSION internal_name: %s", internalTableName.c_str()));
m_localHash.drop(internalTableName.c_str()); m_localHash.drop(internalTableName.c_str());
m_globalHash->lock(); m_globalHash->lock();
tab->m_status = NdbDictionary::Object::Invalid; m_globalHash->release(tab, 1);
m_globalHash->drop(tab);
m_globalHash->unlock(); m_globalHash->unlock();
DBUG_RETURN(dropTable(name)); DBUG_RETURN(dropTable(name));
} }
...@@ -2792,8 +2797,7 @@ NdbDictionaryImpl::dropTable(NdbTableImpl & impl) ...@@ -2792,8 +2797,7 @@ NdbDictionaryImpl::dropTable(NdbTableImpl & impl)
m_localHash.drop(internalTableName); m_localHash.drop(internalTableName);
m_globalHash->lock(); m_globalHash->lock();
impl.m_status = NdbDictionary::Object::Invalid; m_globalHash->release(&impl, 1);
m_globalHash->drop(&impl);
m_globalHash->unlock(); m_globalHash->unlock();
return 0; return 0;
...@@ -2802,6 +2806,50 @@ NdbDictionaryImpl::dropTable(NdbTableImpl & impl) ...@@ -2802,6 +2806,50 @@ NdbDictionaryImpl::dropTable(NdbTableImpl & impl)
return ret; return ret;
} }
int
NdbDictionaryImpl::dropTableGlobal(NdbTableImpl & impl)
{
int res;
const char * name = impl.getName();
DBUG_ENTER("NdbDictionaryImpl::dropTableGlobal");
DBUG_ASSERT(impl.m_status != NdbDictionary::Object::New);
DBUG_ASSERT(impl.m_indexType == NdbDictionary::Object::TypeUndefined);
List list;
if ((res = listIndexes(list, impl.m_id)) == -1){
ERR_RETURN(getNdbError(), -1);
}
for (unsigned i = 0; i < list.count; i++) {
const List::Element& element = list.elements[i];
NdbIndexImpl *idx= getIndexGlobal(element.name, impl);
if (idx == NULL)
{
ERR_RETURN(getNdbError(), -1);
}
if ((res = dropIndexGlobal(*idx)) == -1)
{
releaseIndexGlobal(*idx, 1);
ERR_RETURN(getNdbError(), -1);
}
releaseIndexGlobal(*idx, 1);
}
if (impl.m_noOfBlobs != 0) {
if (dropBlobTables(impl) != 0){
ERR_RETURN(getNdbError(), -1);
}
}
int ret = m_receiver.dropTable(impl);
impl.m_status = NdbDictionary::Object::Invalid;
if(ret == 0 || m_error.code == 709 || m_error.code == 723)
{
DBUG_RETURN(0);
}
ERR_RETURN(getNdbError(), ret);
}
int int
NdbDictionaryImpl::dropBlobTables(NdbTableImpl & t) NdbDictionaryImpl::dropBlobTables(NdbTableImpl & t)
{ {
...@@ -2822,7 +2870,7 @@ NdbDictionaryImpl::dropBlobTables(NdbTableImpl & t) ...@@ -2822,7 +2870,7 @@ NdbDictionaryImpl::dropBlobTables(NdbTableImpl & t)
DBUG_PRINT("info", ("col %s: blob table %s: error %d", DBUG_PRINT("info", ("col %s: blob table %s: error %d",
c.m_name.c_str(), bt->m_internalName.c_str(), m_error.code)); c.m_name.c_str(), bt->m_internalName.c_str(), m_error.code));
if (! (ret == 709 || ret == 723)) // "force" mode on if (! (ret == 709 || ret == 723)) // "force" mode on
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
// leave c.m_blobTable defined // leave c.m_blobTable defined
} }
...@@ -2891,8 +2939,7 @@ NdbDictionaryImpl::invalidateObject(NdbTableImpl & impl) ...@@ -2891,8 +2939,7 @@ NdbDictionaryImpl::invalidateObject(NdbTableImpl & impl)
m_localHash.drop(internalTableName); m_localHash.drop(internalTableName);
m_globalHash->lock(); m_globalHash->lock();
impl.m_status = NdbDictionary::Object::Invalid; m_globalHash->release(&impl, 1);
m_globalHash->drop(&impl);
m_globalHash->unlock(); m_globalHash->unlock();
DBUG_RETURN(0); DBUG_RETURN(0);
} }
...@@ -2918,6 +2965,7 @@ NdbIndexImpl* ...@@ -2918,6 +2965,7 @@ NdbIndexImpl*
NdbDictionaryImpl::getIndexImpl(const char * externalName, NdbDictionaryImpl::getIndexImpl(const char * externalName,
const BaseString& internalName) const BaseString& internalName)
{ {
ASSERT_NOT_MYSQLD;
Ndb_local_table_info * info = get_local_table_info(internalName); Ndb_local_table_info * info = get_local_table_info(internalName);
if(info == 0){ if(info == 0){
m_error.code = 4243; m_error.code = 4243;
...@@ -2938,26 +2986,41 @@ NdbDictionaryImpl::getIndexImpl(const char * externalName, ...@@ -2938,26 +2986,41 @@ NdbDictionaryImpl::getIndexImpl(const char * externalName,
return 0; return 0;
} }
return getIndexImpl(externalName, internalName, *tab, *prim);
}
NdbIndexImpl*
NdbDictionaryImpl::getIndexImpl(const char * externalName,
const BaseString& internalName,
NdbTableImpl &tab,
NdbTableImpl &prim)
{
DBUG_ENTER("NdbDictionaryImpl::getIndexImpl");
DBUG_ASSERT(tab.m_indexType != NdbDictionary::Object::TypeUndefined);
/** /**
* Create index impl * Create index impl
*/ */
NdbIndexImpl* idx; NdbIndexImpl* idx;
if(NdbDictInterface::create_index_obj_from_table(&idx, tab, prim) == 0){ if(NdbDictInterface::create_index_obj_from_table(&idx, &tab, &prim) == 0){
idx->m_table = tab; idx->m_table = &tab;
idx->m_externalName.assign(externalName); idx->m_externalName.assign(externalName);
idx->m_internalName.assign(internalName); idx->m_internalName.assign(internalName);
idx->m_table_id = prim.getObjectId();
idx->m_table_version = prim.getObjectVersion();
// TODO Assign idx to tab->m_index // TODO Assign idx to tab->m_index
// Don't do it right now since assign can't asign a table with index // Don't do it right now since assign can't asign a table with index
// tab->m_index = idx; // tab->m_index = idx;
return idx; DBUG_RETURN(idx);
} }
return 0; DBUG_RETURN(0);
} }
int int
NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst, NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst,
NdbTableImpl* tab, NdbTableImpl* tab,
const NdbTableImpl* prim){ const NdbTableImpl* prim)
{
DBUG_ENTER("NdbDictInterface::create_index_obj_from_table");
NdbIndexImpl *idx = new NdbIndexImpl(); NdbIndexImpl *idx = new NdbIndexImpl();
idx->m_version = tab->m_version; idx->m_version = tab->m_version;
idx->m_status = tab->m_status; idx->m_status = tab->m_status;
...@@ -3010,8 +3073,8 @@ NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst, ...@@ -3010,8 +3073,8 @@ NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst,
} }
* dst = idx; * dst = idx;
DBUG_PRINT("exit", ("m_id: %d m_version: %d", idx->m_id, idx->m_version));
return 0; DBUG_RETURN(0);
} }
/***************************************************************** /*****************************************************************
...@@ -3020,6 +3083,7 @@ NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst, ...@@ -3020,6 +3083,7 @@ NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst,
int int
NdbDictionaryImpl::createIndex(NdbIndexImpl &ix) NdbDictionaryImpl::createIndex(NdbIndexImpl &ix)
{ {
ASSERT_NOT_MYSQLD;
NdbTableImpl* tab = getTable(ix.getTable()); NdbTableImpl* tab = getTable(ix.getTable());
if(tab == 0){ if(tab == 0){
m_error.code = 4249; m_error.code = 4249;
...@@ -3029,6 +3093,12 @@ NdbDictionaryImpl::createIndex(NdbIndexImpl &ix) ...@@ -3029,6 +3093,12 @@ NdbDictionaryImpl::createIndex(NdbIndexImpl &ix)
return m_receiver.createIndex(m_ndb, ix, * tab); return m_receiver.createIndex(m_ndb, ix, * tab);
} }
int
NdbDictionaryImpl::createIndex(NdbIndexImpl &ix, NdbTableImpl &tab)
{
return m_receiver.createIndex(m_ndb, ix, tab);
}
int int
NdbDictInterface::createIndex(Ndb & ndb, NdbDictInterface::createIndex(Ndb & ndb,
const NdbIndexImpl & impl, const NdbIndexImpl & impl,
...@@ -3135,6 +3205,7 @@ int ...@@ -3135,6 +3205,7 @@ int
NdbDictionaryImpl::dropIndex(const char * indexName, NdbDictionaryImpl::dropIndex(const char * indexName,
const char * tableName) const char * tableName)
{ {
ASSERT_NOT_MYSQLD;
NdbIndexImpl * idx = getIndex(indexName, tableName); NdbIndexImpl * idx = getIndex(indexName, tableName);
if (idx == 0) { if (idx == 0) {
m_error.code = 4243; m_error.code = 4243;
...@@ -3152,8 +3223,7 @@ NdbDictionaryImpl::dropIndex(const char * indexName, ...@@ -3152,8 +3223,7 @@ NdbDictionaryImpl::dropIndex(const char * indexName,
m_localHash.drop(internalIndexName.c_str()); m_localHash.drop(internalIndexName.c_str());
m_globalHash->lock(); m_globalHash->lock();
idx->m_table->m_status = NdbDictionary::Object::Invalid; m_globalHash->release(idx->m_table, 1);
m_globalHash->drop(idx->m_table);
m_globalHash->unlock(); m_globalHash->unlock();
return dropIndex(indexName, tableName); return dropIndex(indexName, tableName);
} }
...@@ -3183,13 +3253,13 @@ NdbDictionaryImpl::dropIndex(NdbIndexImpl & impl, const char * tableName) ...@@ -3183,13 +3253,13 @@ NdbDictionaryImpl::dropIndex(NdbIndexImpl & impl, const char * tableName)
return dropIndex(indexName, tableName); return dropIndex(indexName, tableName);
} }
int ret = m_receiver.dropIndex(impl, *timpl); int ret= dropIndexGlobal(impl);
if(ret == 0){ if (ret == 0)
m_localHash.drop(internalIndexName.c_str()); {
m_globalHash->lock(); m_globalHash->lock();
impl.m_table->m_status = NdbDictionary::Object::Invalid; m_globalHash->release(impl.m_table, 1);
m_globalHash->drop(impl.m_table);
m_globalHash->unlock(); m_globalHash->unlock();
m_localHash.drop(internalIndexName.c_str());
} }
return ret; return ret;
} }
...@@ -3198,10 +3268,26 @@ NdbDictionaryImpl::dropIndex(NdbIndexImpl & impl, const char * tableName) ...@@ -3198,10 +3268,26 @@ NdbDictionaryImpl::dropIndex(NdbIndexImpl & impl, const char * tableName)
return -1; return -1;
} }
int
NdbDictionaryImpl::dropIndexGlobal(NdbIndexImpl & impl)
{
DBUG_ENTER("NdbDictionaryImpl::dropIndexGlobal");
int ret = m_receiver.dropIndex(impl, *impl.m_table);
impl.m_status = NdbDictionary::Object::Invalid;
if(ret == 0)
{
DBUG_RETURN(0);
}
ERR_RETURN(getNdbError(), ret);
}
int int
NdbDictInterface::dropIndex(const NdbIndexImpl & impl, NdbDictInterface::dropIndex(const NdbIndexImpl & impl,
const NdbTableImpl & timpl) const NdbTableImpl & timpl)
{ {
DBUG_ENTER("NdbDictInterface::dropIndex");
DBUG_PRINT("enter", ("indexId: %d indexVersion: %d",
timpl.m_id, timpl.m_version));
NdbApiSignal tSignal(m_reference); NdbApiSignal tSignal(m_reference);
tSignal.theReceiversBlockNumber = DBDICT; tSignal.theReceiversBlockNumber = DBDICT;
tSignal.theVerId_signalNumber = GSN_DROP_INDX_REQ; tSignal.theVerId_signalNumber = GSN_DROP_INDX_REQ;
...@@ -3223,9 +3309,9 @@ NdbDictInterface::dropIndex(const NdbIndexImpl & impl, ...@@ -3223,9 +3309,9 @@ NdbDictInterface::dropIndex(const NdbIndexImpl & impl,
errCodes); errCodes);
if(m_error.code == DropIndxRef::InvalidIndexVersion) { if(m_error.code == DropIndxRef::InvalidIndexVersion) {
// Clear caches and try again // Clear caches and try again
return INCOMPATIBLE_VERSION; ERR_RETURN(m_error, INCOMPATIBLE_VERSION);
} }
return r; ERR_RETURN(m_error, r);
} }
void void
...@@ -3262,7 +3348,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt) ...@@ -3262,7 +3348,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
if(tab == 0){ if(tab == 0){
DBUG_PRINT("info",("NdbDictionaryImpl::createEvent: table not found: %s", DBUG_PRINT("info",("NdbDictionaryImpl::createEvent: table not found: %s",
evnt.getTableName())); evnt.getTableName()));
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
evnt.setTable(tab); evnt.setTable(tab);
} }
...@@ -3281,7 +3367,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt) ...@@ -3281,7 +3367,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
ndbout_c("Attr id %u in table %s not found", evnt.m_attrIds[i], ndbout_c("Attr id %u in table %s not found", evnt.m_attrIds[i],
evnt.getTableName()); evnt.getTableName());
m_error.code= 4713; m_error.code= 4713;
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
} }
...@@ -3302,7 +3388,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt) ...@@ -3302,7 +3388,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
table.getColumn(evnt.m_columns[i]->m_name.c_str()); table.getColumn(evnt.m_columns[i]->m_name.c_str());
if(col == 0){ if(col == 0){
m_error.code= 4247; m_error.code= 4247;
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
// Copy column definition // Copy column definition
*evnt.m_columns[i] = *col; *evnt.m_columns[i] = *col;
...@@ -3328,7 +3414,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt) ...@@ -3328,7 +3414,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
for(i = 1; i<attributeList_sz; i++) { for(i = 1; i<attributeList_sz; i++) {
if (evnt.m_columns[i-1]->m_attrId == evnt.m_columns[i]->m_attrId) { if (evnt.m_columns[i-1]->m_attrId == evnt.m_columns[i]->m_attrId) {
m_error.code= 4258; m_error.code= 4258;
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
} }
...@@ -3340,14 +3426,14 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt) ...@@ -3340,14 +3426,14 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
// NdbDictInterface m_receiver; // NdbDictInterface m_receiver;
if (m_receiver.createEvent(m_ndb, evnt, 0 /* getFlag unset */) != 0) if (m_receiver.createEvent(m_ndb, evnt, 0 /* getFlag unset */) != 0)
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
// Create blob events // Create blob events
if (evnt.m_mergeEvents && createBlobEvents(evnt) != 0) { if (evnt.m_mergeEvents && createBlobEvents(evnt) != 0) {
int save_code = m_error.code; int save_code = m_error.code;
(void)dropEvent(evnt.m_name.c_str()); (void)dropEvent(evnt.m_name.c_str());
m_error.code = save_code; m_error.code = save_code;
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
DBUG_RETURN(0); DBUG_RETURN(0);
} }
...@@ -3367,7 +3453,7 @@ NdbDictionaryImpl::createBlobEvents(NdbEventImpl& evnt) ...@@ -3367,7 +3453,7 @@ NdbDictionaryImpl::createBlobEvents(NdbEventImpl& evnt)
NdbEventImpl blob_evnt; NdbEventImpl blob_evnt;
NdbBlob::getBlobEvent(blob_evnt, &evnt, &c); NdbBlob::getBlobEvent(blob_evnt, &evnt, &c);
if (createEvent(blob_evnt) != 0) if (createEvent(blob_evnt) != 0)
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
DBUG_RETURN(0); DBUG_RETURN(0);
} }
...@@ -3418,7 +3504,7 @@ NdbDictInterface::createEvent(class Ndb & ndb, ...@@ -3418,7 +3504,7 @@ NdbDictInterface::createEvent(class Ndb & ndb,
const size_t len = strlen(evnt.m_name.c_str()) + 1; const size_t len = strlen(evnt.m_name.c_str()) + 1;
if(len > MAX_TAB_NAME_SIZE) { if(len > MAX_TAB_NAME_SIZE) {
m_error.code= 4241; m_error.code= 4241;
DBUG_RETURN(-1); ERR_RETURN(getNdbError(), -1);
} }
w.add(SimpleProperties::StringValue, evnt.m_name.c_str()); w.add(SimpleProperties::StringValue, evnt.m_name.c_str());
...@@ -3442,7 +3528,7 @@ NdbDictInterface::createEvent(class Ndb & ndb, ...@@ -3442,7 +3528,7 @@ NdbDictInterface::createEvent(class Ndb & ndb,
0, -1); 0, -1);
if (ret) { if (ret) {
DBUG_RETURN(ret); ERR_RETURN(getNdbError(), ret);
} }
char *dataPtr = (char *)m_buffer.get_data(); char *dataPtr = (char *)m_buffer.get_data();
...@@ -3468,7 +3554,7 @@ NdbDictInterface::createEvent(class Ndb & ndb, ...@@ -3468,7 +3554,7 @@ NdbDictInterface::createEvent(class Ndb & ndb,
//evnt.m_attrListBitmask != evntConf->getAttrListBitmask() || //evnt.m_attrListBitmask != evntConf->getAttrListBitmask() ||
evnt.mi_type != evntConf->getEventType()) { evnt.mi_type != evntConf->getEventType()) {
ndbout_c("ERROR*************"); ndbout_c("ERROR*************");
DBUG_RETURN(1); ERR_RETURN(getNdbError(), 1);
} }
} }
...@@ -3569,35 +3655,36 @@ NdbDictionaryImpl::getEvent(const char * eventName, NdbTableImpl* tab) ...@@ -3569,35 +3655,36 @@ NdbDictionaryImpl::getEvent(const char * eventName, NdbTableImpl* tab)
DBUG_RETURN(NULL); DBUG_RETURN(NULL);
} }
if (tab == NULL) {
// We only have the table name with internal name // We only have the table name with internal name
DBUG_PRINT("info",("table %s", ev->getTableName())); DBUG_PRINT("info",("table %s", ev->getTableName()));
Ndb_local_table_info *info; if (tab == NULL)
info= get_local_table_info(ev->getTableName()); {
if (info == 0) tab= fetchGlobalTableImplRef(InitTable(this, ev->getTableName()));
if (tab == 0)
{ {
DBUG_PRINT("error",("unable to find table %s", ev->getTableName())); DBUG_PRINT("error",("unable to find table %s", ev->getTableName()));
delete ev; delete ev;
DBUG_RETURN(NULL); DBUG_RETURN(NULL);
} }
if ((info->m_table_impl->m_status != NdbDictionary::Object::Retrieved) || if ((tab->m_status != NdbDictionary::Object::Retrieved) ||
(info->m_table_impl->m_id != ev->m_table_id) || (tab->m_id != ev->m_table_id) ||
(table_version_major(info->m_table_impl->m_version) != (table_version_major(tab->m_version) !=
table_version_major(ev->m_table_version))) table_version_major(ev->m_table_version)))
{ {
removeCachedObject(*info->m_table_impl); DBUG_PRINT("info", ("mismatch on verison in cache"));
info= get_local_table_info(ev->getTableName()); releaseTableGlobal(*tab, 1);
if (info == 0) tab= fetchGlobalTableImplRef(InitTable(this, ev->getTableName()));
if (tab == 0)
{ {
DBUG_PRINT("error",("unable to find table %s", ev->getTableName())); DBUG_PRINT("error",("unable to find table %s", ev->getTableName()));
delete ev; delete ev;
DBUG_RETURN(NULL); DBUG_RETURN(NULL);
} }
} }
tab = info->m_table_impl;
} }
ev->setTable(tab); ev->setTable(tab);
ev->setTable(m_ndb.externalizeTableName(ev->getTableName())); ev->setTable(m_ndb.externalizeTableName(ev->getTableName()));
// get the columns from the attrListBitmask // get the columns from the attrListBitmask
NdbTableImpl &table = *ev->m_tableImpl; NdbTableImpl &table = *ev->m_tableImpl;
......
...@@ -35,6 +35,9 @@ is_ndb_blob_table(const char* name, Uint32* ptab_id = 0, Uint32* pcol_no = 0); ...@@ -35,6 +35,9 @@ is_ndb_blob_table(const char* name, Uint32* ptab_id = 0, Uint32* pcol_no = 0);
bool bool
is_ndb_blob_table(const class NdbTableImpl* t); is_ndb_blob_table(const class NdbTableImpl* t);
extern int ndb_dictionary_is_mysqld;
#define ASSERT_NOT_MYSQLD assert(ndb_dictionary_is_mysqld == 0)
class NdbDictObjectImpl { class NdbDictObjectImpl {
public: public:
int m_id; int m_id;
...@@ -253,6 +256,8 @@ public: ...@@ -253,6 +256,8 @@ public:
BaseString m_internalName; BaseString m_internalName;
BaseString m_externalName; BaseString m_externalName;
BaseString m_tableName; BaseString m_tableName;
Uint32 m_table_id;
Uint32 m_table_version;
Vector<NdbColumnImpl *> m_columns; Vector<NdbColumnImpl *> m_columns;
Vector<int> m_key_ids; Vector<int> m_key_ids;
...@@ -539,6 +544,21 @@ private: ...@@ -539,6 +544,21 @@ private:
UtilBuffer m_buffer; UtilBuffer m_buffer;
}; };
class NdbDictionaryImpl;
class GlobalCacheInitObject
{
public:
NdbDictionaryImpl *m_dict;
const BaseString &m_name;
GlobalCacheInitObject(NdbDictionaryImpl *dict,
const BaseString &name) :
m_dict(dict),
m_name(name)
{}
virtual ~GlobalCacheInitObject() {}
virtual int init(NdbTableImpl &tab) const = 0;
};
class NdbDictionaryImpl : public NdbDictionary::Dictionary { class NdbDictionaryImpl : public NdbDictionary::Dictionary {
public: public:
NdbDictionaryImpl(Ndb &ndb); NdbDictionaryImpl(Ndb &ndb);
...@@ -558,6 +578,7 @@ public: ...@@ -558,6 +578,7 @@ public:
int removeCachedObject(NdbTableImpl &); int removeCachedObject(NdbTableImpl &);
int createIndex(NdbIndexImpl &ix); int createIndex(NdbIndexImpl &ix);
int createIndex(NdbIndexImpl &ix, NdbTableImpl & tab);
int dropIndex(const char * indexName, int dropIndex(const char * indexName,
const char * tableName); const char * tableName);
int dropIndex(NdbIndexImpl &, const char * tableName); int dropIndex(NdbIndexImpl &, const char * tableName);
...@@ -578,6 +599,15 @@ public: ...@@ -578,6 +599,15 @@ public:
int listObjects(List& list, NdbDictionary::Object::Type type); int listObjects(List& list, NdbDictionary::Object::Type type);
int listIndexes(List& list, Uint32 indexId); int listIndexes(List& list, Uint32 indexId);
NdbTableImpl * getTableGlobal(const char * tableName);
NdbIndexImpl * getIndexGlobal(const char * indexName,
NdbTableImpl &ndbtab);
int alterTableGlobal(NdbTableImpl &orig_impl, NdbTableImpl &impl);
int dropTableGlobal(NdbTableImpl &);
int dropIndexGlobal(NdbIndexImpl & impl);
int releaseTableGlobal(NdbTableImpl & impl, int invalidate);
int releaseIndexGlobal(NdbIndexImpl & impl, int invalidate);
NdbTableImpl * getTable(const char * tableName, void **data= 0); NdbTableImpl * getTable(const char * tableName, void **data= 0);
NdbTableImpl * getBlobTable(const NdbTableImpl&, uint col_no); NdbTableImpl * getBlobTable(const NdbTableImpl&, uint col_no);
NdbTableImpl * getBlobTable(uint tab_id, uint col_no); NdbTableImpl * getBlobTable(uint tab_id, uint col_no);
...@@ -616,10 +646,14 @@ public: ...@@ -616,10 +646,14 @@ public:
NdbDictInterface m_receiver; NdbDictInterface m_receiver;
Ndb & m_ndb; Ndb & m_ndb;
private: NdbIndexImpl* getIndexImpl(const char * externalName,
const BaseString& internalName,
NdbTableImpl &tab,
NdbTableImpl &prim);
NdbIndexImpl * getIndexImpl(const char * name, NdbIndexImpl * getIndexImpl(const char * name,
const BaseString& internalName); const BaseString& internalName);
Ndb_local_table_info * fetchGlobalTableImpl(const BaseString& internalName); private:
NdbTableImpl * fetchGlobalTableImplRef(const GlobalCacheInitObject &obj);
}; };
inline inline
...@@ -852,6 +886,27 @@ NdbDictionaryImpl::getImpl(const NdbDictionary::Dictionary & t){ ...@@ -852,6 +886,27 @@ NdbDictionaryImpl::getImpl(const NdbDictionary::Dictionary & t){
* Inline:d getters * Inline:d getters
*/ */
class InitTable : public GlobalCacheInitObject
{
public:
InitTable(NdbDictionaryImpl *dict,
const BaseString &name) :
GlobalCacheInitObject(dict, name)
{}
int init(NdbTableImpl &tab) const
{
return m_dict->getBlobTables(tab);
}
};
inline
NdbTableImpl *
NdbDictionaryImpl::getTableGlobal(const char * table_name)
{
const BaseString internal_tabname(m_ndb.internalize_table_name(table_name));
return fetchGlobalTableImplRef(InitTable(this, internal_tabname));
}
inline inline
NdbTableImpl * NdbTableImpl *
NdbDictionaryImpl::getTable(const char * table_name, void **data) NdbDictionaryImpl::getTable(const char * table_name, void **data)
...@@ -885,21 +940,134 @@ NdbDictionaryImpl::get_local_table_info(const BaseString& internalTableName) ...@@ -885,21 +940,134 @@ NdbDictionaryImpl::get_local_table_info(const BaseString& internalTableName)
DBUG_PRINT("enter", ("table: %s", internalTableName.c_str())); DBUG_PRINT("enter", ("table: %s", internalTableName.c_str()));
Ndb_local_table_info *info= m_localHash.get(internalTableName.c_str()); Ndb_local_table_info *info= m_localHash.get(internalTableName.c_str());
if (info == 0) { if (info == 0)
info= fetchGlobalTableImpl(internalTableName); {
if (info == 0) { NdbTableImpl *tab=
DBUG_RETURN(0); fetchGlobalTableImplRef(InitTable(this, internalTableName));
if (tab)
{
info= Ndb_local_table_info::create(tab, m_local_table_data_size);
if (info)
{
m_localHash.put(internalTableName.c_str(), info);
m_ndb.theFirstTupleId[tab->getTableId()] = ~0;
m_ndb.theLastTupleId[tab->getTableId()] = ~0;
}
} }
} }
DBUG_RETURN(info); // autoincrement already initialized DBUG_RETURN(info); // autoincrement already initialized
} }
class InitIndexGlobal : public GlobalCacheInitObject
{
public:
const char *m_index_name;
NdbTableImpl &m_prim;
InitIndexGlobal(NdbDictionaryImpl *dict,
const BaseString &internal_indexname,
const char *index_name,
NdbTableImpl &prim) :
GlobalCacheInitObject(dict, internal_indexname),
m_index_name(index_name),
m_prim(prim)
{}
int init(NdbTableImpl &tab) const
{
tab.m_index= m_dict->getIndexImpl(m_index_name, m_name, tab, m_prim);
if (tab.m_index == 0)
return 1;
tab.m_index->m_table= &tab;
return 0;
}
};
class InitIndex : public GlobalCacheInitObject
{
public:
const char *m_index_name;
InitIndex(NdbDictionaryImpl *dict,
const BaseString &internal_indexname,
const char *index_name) :
GlobalCacheInitObject(dict, internal_indexname),
m_index_name(index_name)
{}
int init(NdbTableImpl &tab) const
{
DBUG_ASSERT(tab.m_index == 0);
tab.m_index= m_dict->getIndexImpl(m_index_name, m_name);
if (tab.m_index)
{
tab.m_index->m_table= &tab;
return 0;
}
return 1;
}
};
inline
NdbIndexImpl *
NdbDictionaryImpl::getIndexGlobal(const char * index_name,
NdbTableImpl &ndbtab)
{
DBUG_ENTER("NdbDictionaryImpl::getIndexGlobal");
const BaseString
internal_indexname(m_ndb.internalize_index_name(&ndbtab, index_name));
int retry= 2;
while (retry)
{
NdbTableImpl *tab=
fetchGlobalTableImplRef(InitIndexGlobal(this, internal_indexname,
index_name, ndbtab));
if (tab)
{
// tab->m_index sould be set. otherwise tab == 0
NdbIndexImpl *idx= tab->m_index;
if (idx->m_table_id != ndbtab.getObjectId() ||
idx->m_table_version != ndbtab.getObjectVersion())
{
releaseIndexGlobal(*idx, 1);
retry--;
continue;
}
DBUG_RETURN(idx);
}
break;
}
m_error.code= 4243;
DBUG_RETURN(0);
}
inline int
NdbDictionaryImpl::releaseTableGlobal(NdbTableImpl & impl, int invalidate)
{
DBUG_ENTER("NdbDictionaryImpl::releaseTableGlobal");
DBUG_PRINT("enter", ("internal_name: %s", impl.m_internalName.c_str()));
m_globalHash->lock();
m_globalHash->release(&impl, invalidate);
m_globalHash->unlock();
DBUG_RETURN(0);
}
inline int
NdbDictionaryImpl::releaseIndexGlobal(NdbIndexImpl & impl, int invalidate)
{
DBUG_ENTER("NdbDictionaryImpl::releaseIndexGlobal");
DBUG_PRINT("enter", ("internal_name: %s", impl.m_internalName.c_str()));
m_globalHash->lock();
m_globalHash->release(impl.m_table, invalidate);
m_globalHash->unlock();
DBUG_RETURN(0);
}
inline inline
NdbIndexImpl * NdbIndexImpl *
NdbDictionaryImpl::getIndex(const char * index_name, NdbDictionaryImpl::getIndex(const char * index_name,
const char * table_name) const char * table_name)
{ {
if (table_name || m_ndb.usingFullyQualifiedNames()) while (table_name || m_ndb.usingFullyQualifiedNames())
{ {
const BaseString internal_indexname( const BaseString internal_indexname(
(table_name) (table_name)
...@@ -910,18 +1078,28 @@ NdbDictionaryImpl::getIndex(const char * index_name, ...@@ -910,18 +1078,28 @@ NdbDictionaryImpl::getIndex(const char * index_name,
if (internal_indexname.length()) if (internal_indexname.length())
{ {
Ndb_local_table_info * info= Ndb_local_table_info *info= m_localHash.get(internal_indexname.c_str());
get_local_table_info(internal_indexname); NdbTableImpl *tab;
if (info) if (info == 0)
{ {
NdbTableImpl * tab= info->m_table_impl; tab= fetchGlobalTableImplRef(InitIndex(this, internal_indexname,
if (tab->m_index == 0) index_name));
tab->m_index= getIndexImpl(index_name, internal_indexname); if (tab)
if (tab->m_index != 0) {
tab->m_index->m_table= tab; info= Ndb_local_table_info::create(tab, 0);
return tab->m_index; if (info)
m_localHash.put(internal_indexname.c_str(), info);
else
break;
} }
else
break;
}
else
tab= info->m_table_impl;
return tab->m_index;
} }
break;
} }
m_error.code= 4243; m_error.code= 4243;
......
...@@ -377,7 +377,7 @@ NdbIndexStat::stat_select(const Uint32* key1, Uint32 keylen1, const Uint32* key2 ...@@ -377,7 +377,7 @@ NdbIndexStat::stat_select(const Uint32* key1, Uint32 keylen1, const Uint32* key2
} }
int int
NdbIndexStat::records_in_range(NdbDictionary::Index* index, NdbIndexScanOperation* op, Uint64 table_rows, Uint64* count, int flags) NdbIndexStat::records_in_range(const NdbDictionary::Index* index, NdbIndexScanOperation* op, Uint64 table_rows, Uint64* count, int flags)
{ {
DBUG_ENTER("NdbIndexStat::records_in_range"); DBUG_ENTER("NdbIndexStat::records_in_range");
Uint64 rows; Uint64 rows;
......
...@@ -99,6 +99,7 @@ void Ndb::setup(Ndb_cluster_connection *ndb_cluster_connection, ...@@ -99,6 +99,7 @@ void Ndb::setup(Ndb_cluster_connection *ndb_cluster_connection,
for (i = 0; i < MAX_NDB_NODES ; i++) { for (i = 0; i < MAX_NDB_NODES ; i++) {
theConnectionArray[i] = NULL; theConnectionArray[i] = NULL;
}//forg }//forg
m_sys_tab_0 = NULL;
for (i = 0; i < 2048 ; i++) { for (i = 0; i < 2048 ; i++) {
theFirstTupleId[i] = 0; theFirstTupleId[i] = 0;
theLastTupleId[i] = 0; theLastTupleId[i] = 0;
...@@ -137,6 +138,9 @@ Ndb::~Ndb() ...@@ -137,6 +138,9 @@ Ndb::~Ndb()
DBUG_ENTER("Ndb::~Ndb()"); DBUG_ENTER("Ndb::~Ndb()");
DBUG_PRINT("enter",("this=0x%x",this)); DBUG_PRINT("enter",("this=0x%x",this));
if (m_sys_tab_0)
getDictionary()->removeTableGlobal(*m_sys_tab_0, 0);
assert(theImpl->m_ev_op == 0); // user should return NdbEventOperation's assert(theImpl->m_ev_op == 0); // user should return NdbEventOperation's
for (NdbEventOperationImpl *op= theImpl->m_ev_op; op; op=op->m_next) for (NdbEventOperationImpl *op= theImpl->m_ev_op; op; op=op->m_next)
{ {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment