Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
M
MariaDB
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
nexedi
MariaDB
Commits
b96c4289
Commit
b96c4289
authored
May 04, 2006
by
unknown
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
removing the usage if the ndb local dict cache to adress the 241 errors we get now and then
parent
a8138ec7
Changes
16
Show whitespace changes
Inline
Side-by-side
Showing
16 changed files
with
1079 additions
and
774 deletions
+1079
-774
mysql-test/r/ndb_multi_row.result
mysql-test/r/ndb_multi_row.result
+2
-0
sql/ha_ndbcluster.cc
sql/ha_ndbcluster.cc
+353
-462
sql/ha_ndbcluster.h
sql/ha_ndbcluster.h
+4
-10
sql/ha_ndbcluster_binlog.cc
sql/ha_ndbcluster_binlog.cc
+162
-115
sql/ha_ndbcluster_binlog.h
sql/ha_ndbcluster_binlog.h
+44
-4
storage/ndb/include/ndbapi/Ndb.hpp
storage/ndb/include/ndbapi/Ndb.hpp
+3
-0
storage/ndb/include/ndbapi/NdbDictionary.hpp
storage/ndb/include/ndbapi/NdbDictionary.hpp
+11
-0
storage/ndb/include/ndbapi/NdbIndexStat.hpp
storage/ndb/include/ndbapi/NdbIndexStat.hpp
+1
-1
storage/ndb/src/ndbapi/DictCache.cpp
storage/ndb/src/ndbapi/DictCache.cpp
+44
-69
storage/ndb/src/ndbapi/DictCache.hpp
storage/ndb/src/ndbapi/DictCache.hpp
+2
-2
storage/ndb/src/ndbapi/Ndb.cpp
storage/ndb/src/ndbapi/Ndb.cpp
+24
-16
storage/ndb/src/ndbapi/NdbDictionary.cpp
storage/ndb/src/ndbapi/NdbDictionary.cpp
+65
-0
storage/ndb/src/ndbapi/NdbDictionaryImpl.cpp
storage/ndb/src/ndbapi/NdbDictionaryImpl.cpp
+165
-78
storage/ndb/src/ndbapi/NdbDictionaryImpl.hpp
storage/ndb/src/ndbapi/NdbDictionaryImpl.hpp
+194
-16
storage/ndb/src/ndbapi/NdbIndexStat.cpp
storage/ndb/src/ndbapi/NdbIndexStat.cpp
+1
-1
storage/ndb/src/ndbapi/Ndbinit.cpp
storage/ndb/src/ndbapi/Ndbinit.cpp
+4
-0
No files found.
mysql-test/r/ndb_multi_row.result
View file @
b96c4289
...
...
@@ -62,4 +62,6 @@ t4
drop table t1, t2, t3, t4;
drop table if exists t1, t3, t4;
Warnings:
Error 155 Table 'test.t1' doesn't exist
Error 155 Table 'test.t3' doesn't exist
Error 155 Table 'test.t4' doesn't exist
sql/ha_ndbcluster.cc
View file @
b96c4289
...
...
@@ -141,7 +141,7 @@ static int rename_share(NDB_SHARE *share, const char *new_key);
#endif
static
void
ndb_set_fragmentation
(
NDBTAB
&
tab
,
TABLE
*
table
,
uint
pk_len
);
static
int
ndb_get_table_statistics
(
Ndb
*
,
const
char
*
,
static
int
ndb_get_table_statistics
(
Ndb
*
,
const
NDBTAB
*
,
struct
Ndb_statistics
*
);
...
...
@@ -431,7 +431,7 @@ void ha_ndbcluster::records_update()
Ndb
*
ndb
=
get_ndb
();
ndb
->
setDatabaseName
(
m_dbname
);
struct
Ndb_statistics
stat
;
if
(
ndb_get_table_statistics
(
ndb
,
m_tab
nam
e
,
&
stat
)
==
0
){
if
(
ndb_get_table_statistics
(
ndb
,
m_tab
l
e
,
&
stat
)
==
0
){
mean_rec_length
=
stat
.
row_size
;
data_file_length
=
stat
.
fragment_memory
;
info
->
records
=
stat
.
row_count
;
...
...
@@ -479,92 +479,6 @@ void ha_ndbcluster::no_uncommitted_rows_reset(THD *thd)
DBUG_VOID_RETURN
;
}
/*
Take care of the error that occured in NDB
RETURN
0 No error
# The mapped error code
*/
int
ha_ndbcluster
::
invalidate_dictionary_cache
(
bool
global
,
const
NDBTAB
*
ndbtab
)
{
NDBDICT
*
dict
=
get_ndb
()
->
getDictionary
();
DBUG_ENTER
(
"invalidate_dictionary_cache"
);
DBUG_PRINT
(
"info"
,
(
"m_tabname: %s global: %d"
,
m_tabname
,
global
));
if
(
global
)
{
#ifdef HAVE_NDB_BINLOG
if
(
current_thd
!=
injector_thd
)
{
char
key
[
FN_REFLEN
];
build_table_filename
(
key
,
sizeof
(
key
),
m_dbname
,
m_tabname
,
""
);
DBUG_PRINT
(
"info"
,
(
"Getting ndbcluster mutex"
));
pthread_mutex_lock
(
&
ndbcluster_mutex
);
NDB_SHARE
*
ndb_share
=
(
NDB_SHARE
*
)
hash_search
(
&
ndbcluster_open_tables
,
(
byte
*
)
key
,
strlen
(
key
));
// Only binlog_thread is allowed to globally invalidate a table
if
(
ndb_share
&&
ndb_share
->
op
)
{
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
DBUG_PRINT
(
"info"
,
(
"Released ndbcluster mutex"
));
DBUG_RETURN
(
1
);
}
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
DBUG_PRINT
(
"info"
,
(
"Released ndbcluster mutex"
));
}
#endif
if
(
!
ndbtab
)
{
ndbtab
=
dict
->
getTable
(
m_tabname
);
if
(
!
ndbtab
)
DBUG_RETURN
(
1
);
}
dict
->
invalidateTable
(
ndbtab
);
table_share
->
version
=
0L
;
/* Free when thread is ready */
}
else
if
(
ndbtab
)
dict
->
removeCachedTable
(
ndbtab
);
else
dict
->
removeCachedTable
(
m_tabname
);
/* Invalidate indexes */
for
(
uint
i
=
0
;
i
<
table_share
->
keys
;
i
++
)
{
NDBINDEX
*
index
=
(
NDBINDEX
*
)
m_index
[
i
].
index
;
NDBINDEX
*
unique_index
=
(
NDBINDEX
*
)
m_index
[
i
].
unique_index
;
if
(
!
index
&&
!
unique_index
)
continue
;
NDB_INDEX_TYPE
idx_type
=
m_index
[
i
].
type
;
switch
(
idx_type
)
{
case
PRIMARY_KEY_ORDERED_INDEX
:
case
ORDERED_INDEX
:
if
(
global
)
dict
->
invalidateIndex
(
index
->
getName
(),
m_tabname
);
else
dict
->
removeCachedIndex
(
index
->
getName
(),
m_tabname
);
break
;
case
UNIQUE_ORDERED_INDEX
:
if
(
global
)
dict
->
invalidateIndex
(
index
->
getName
(),
m_tabname
);
else
dict
->
removeCachedIndex
(
index
->
getName
(),
m_tabname
);
case
UNIQUE_INDEX
:
if
(
global
)
dict
->
invalidateIndex
(
unique_index
->
getName
(),
m_tabname
);
else
dict
->
removeCachedIndex
(
unique_index
->
getName
(),
m_tabname
);
break
;
case
PRIMARY_KEY_INDEX
:
case
UNDEFINED_INDEX
:
break
;
}
}
DBUG_RETURN
(
0
);
}
int
ha_ndbcluster
::
ndb_err
(
NdbTransaction
*
trans
)
{
int
res
;
...
...
@@ -575,32 +489,14 @@ int ha_ndbcluster::ndb_err(NdbTransaction *trans)
switch
(
err
.
classification
)
{
case
NdbError
:
:
SchemaError
:
{
// TODO perhaps we need to do more here, invalidate also in the cache
m_table
->
setStatusInvalid
();
/* Close other open handlers not used by any thread */
TABLE_LIST
table_list
;
bzero
((
char
*
)
&
table_list
,
sizeof
(
table_list
));
table_list
.
db
=
m_dbname
;
table_list
.
alias
=
table_list
.
table_name
=
m_tabname
;
close_cached_tables
(
current_thd
,
0
,
&
table_list
);
invalidate_dictionary_cache
(
TRUE
,
m_table
);
if
(
err
.
code
==
284
)
{
/*
Check if the table is _really_ gone or if the table has
been alterend and thus changed table id
*/
NDBDICT
*
dict
=
get_ndb
()
->
getDictionary
();
DBUG_PRINT
(
"info"
,
(
"Check if table %s is really gone"
,
m_tabname
));
if
(
!
(
dict
->
getTable
(
m_tabname
)))
{
err
=
dict
->
getNdbError
();
DBUG_PRINT
(
"info"
,
(
"Table not found, error: %d"
,
err
.
code
));
if
(
err
.
code
!=
709
&&
err
.
code
!=
723
)
DBUG_RETURN
(
1
);
}
DBUG_PRINT
(
"info"
,
(
"Table exists but must have changed"
));
}
break
;
}
default:
...
...
@@ -1046,24 +942,15 @@ int ha_ndbcluster::get_metadata(const char *path)
NDBDICT
*
dict
=
ndb
->
getDictionary
();
const
NDBTAB
*
tab
;
int
error
;
bool
invalidating_ndb_table
=
FALSE
;
DBUG_ENTER
(
"get_metadata"
);
DBUG_PRINT
(
"enter"
,
(
"m_tabname: %s, path: %s"
,
m_tabname
,
path
));
do
{
DBUG_ASSERT
(
m_table
==
NULL
);
DBUG_ASSERT
(
m_table_info
==
NULL
);
const
void
*
data
,
*
pack_data
;
uint
length
,
pack_length
;
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
// Check if thread has stale local cache
if
(
tab
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Invalid
)
{
invalidate_dictionary_cache
(
FALSE
,
tab
);
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_PRINT
(
"info"
,
(
"Table schema version: %d"
,
tab
->
getObjectVersion
()));
}
/*
Compare FrmData in NDB with frm file from disk.
*/
...
...
@@ -1076,16 +963,12 @@ int ha_ndbcluster::get_metadata(const char *path)
DBUG_RETURN
(
1
);
}
Ndb_table_guard
ndbtab_g
(
dict
,
m_tabname
);
if
(
!
(
tab
=
ndbtab_g
.
get_table
()))
ERR_RETURN
(
dict
->
getNdbError
());
if
(
get_ndb_share_state
(
m_share
)
!=
NSS_ALTERED
&&
cmp_frm
(
tab
,
pack_data
,
pack_length
))
{
if
(
!
invalidating_ndb_table
)
{
DBUG_PRINT
(
"info"
,
(
"Invalidating table"
));
invalidate_dictionary_cache
(
TRUE
,
tab
);
invalidating_ndb_table
=
TRUE
;
}
else
{
DBUG_PRINT
(
"error"
,
(
"metadata, pack_length: %d getFrmLength: %d memcmp: %d"
,
...
...
@@ -1094,25 +977,24 @@ int ha_ndbcluster::get_metadata(const char *path)
DBUG_DUMP
(
"pack_data"
,
(
char
*
)
pack_data
,
pack_length
);
DBUG_DUMP
(
"frm"
,
(
char
*
)
tab
->
getFrmData
(),
tab
->
getFrmLength
());
error
=
HA_ERR_TABLE_DEF_CHANGED
;
invalidating_ndb_table
=
FALSE
;
}
}
else
{
invalidating_ndb_table
=
FALSE
;
}
my_free
((
char
*
)
data
,
MYF
(
0
));
my_free
((
char
*
)
pack_data
,
MYF
(
0
));
}
while
(
invalidating_ndb_table
);
if
(
error
)
DBUG_RETURN
(
error
)
;
goto
err
;
m_table_version
=
tab
->
getObjectVersion
(
);
DBUG_PRINT
(
"info"
,
(
"fetched table %s"
,
tab
->
getName
())
);
m_table
=
tab
;
m_table_info
=
NULL
;
// Set in external lock
DBUG_RETURN
(
open_indexes
(
ndb
,
table
,
FALSE
));
if
((
error
=
open_indexes
(
ndb
,
table
,
FALSE
))
==
0
)
{
ndbtab_g
.
release
();
DBUG_RETURN
(
0
);
}
err:
ndbtab_g
.
invalidate
();
m_table
=
NULL
;
DBUG_RETURN
(
error
);
}
static
int
fix_unique_index_attr_order
(
NDB_INDEX_DATA
&
data
,
...
...
@@ -1149,36 +1031,6 @@ static int fix_unique_index_attr_order(NDB_INDEX_DATA &data,
DBUG_RETURN
(
0
);
}
int
ha_ndbcluster
::
table_changed
(
const
void
*
pack_frm_data
,
uint
pack_frm_len
)
{
Ndb
*
ndb
;
NDBDICT
*
dict
;
const
NDBTAB
*
orig_tab
;
NdbDictionary
::
Table
new_tab
;
int
result
;
DBUG_ENTER
(
"ha_ndbcluster::table_changed"
);
DBUG_PRINT
(
"info"
,
(
"Modifying frm for table %s"
,
m_tabname
));
if
(
check_ndb_connection
())
DBUG_RETURN
(
my_errno
=
HA_ERR_NO_CONNECTION
);
ndb
=
get_ndb
();
dict
=
ndb
->
getDictionary
();
if
(
!
(
orig_tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
// Check if thread has stale local cache
if
(
orig_tab
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Invalid
)
{
dict
->
removeCachedTable
(
orig_tab
);
if
(
!
(
orig_tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
}
new_tab
=
*
orig_tab
;
new_tab
.
setFrm
(
pack_frm_data
,
pack_frm_len
);
if
(
dict
->
alterTable
(
new_tab
)
!=
0
)
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_RETURN
(
0
);
}
/*
Create all the indexes for a table.
If any index should fail to be created,
...
...
@@ -1246,7 +1098,7 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
const
NDBINDEX
*
index
;
do
{
index
=
dict
->
getIndex
(
index_name
,
m_tabnam
e
);
index
=
dict
->
getIndex
Global
(
index_name
,
*
m_tabl
e
);
if
(
!
index
)
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_PRINT
(
"info"
,
(
"index: 0x%x id: %d version: %d.%d status: %d"
,
...
...
@@ -1255,14 +1107,11 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
index
->
getObjectVersion
()
&
0xFFFFFF
,
index
->
getObjectVersion
()
>>
24
,
index
->
getObjectStatus
()));
if
(
index
->
getObjectStatus
()
!=
NdbDictionary
::
Object
::
Retrieved
)
{
dict
->
removeCachedIndex
(
index
);
continue
;
}
DBUG_ASSERT
(
index
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Retrieved
);
break
;
}
while
(
1
);
m_index
[
index_no
].
index
=
(
void
*
)
index
;
m_index
[
index_no
].
index
=
index
;
// ordered index - add stats
NDB_INDEX_DATA
&
d
=
m_index
[
index_no
];
delete
d
.
index_stat
;
...
...
@@ -1293,7 +1142,7 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
const
NDBINDEX
*
index
;
do
{
index
=
dict
->
getIndex
(
unique_index_name
,
m_tabnam
e
);
index
=
dict
->
getIndex
Global
(
unique_index_name
,
*
m_tabl
e
);
if
(
!
index
)
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_PRINT
(
"info"
,
(
"index: 0x%x id: %d version: %d.%d status: %d"
,
...
...
@@ -1302,14 +1151,11 @@ int ha_ndbcluster::add_index_handle(THD *thd, NDBDICT *dict, KEY *key_info,
index
->
getObjectVersion
()
&
0xFFFFFF
,
index
->
getObjectVersion
()
>>
24
,
index
->
getObjectStatus
()));
if
(
index
->
getObjectStatus
()
!=
NdbDictionary
::
Object
::
Retrieved
)
{
dict
->
removeCachedIndex
(
index
);
continue
;
}
DBUG_ASSERT
(
index
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Retrieved
);
break
;
}
while
(
1
);
m_index
[
index_no
].
unique_index
=
(
void
*
)
index
;
m_index
[
index_no
].
unique_index
=
index
;
error
=
fix_unique_index_attr_order
(
m_index
[
index_no
],
index
,
key_info
);
}
if
(
!
error
)
...
...
@@ -1341,6 +1187,26 @@ int ha_ndbcluster::open_indexes(Ndb *ndb, TABLE *tab, bool ignore_error)
break
;
}
if
(
error
&&
!
ignore_error
)
{
while
(
i
>
0
)
{
i
--
;
if
(
m_index
[
i
].
index
)
{
dict
->
removeIndexGlobal
(
*
m_index
[
i
].
index
,
1
);
m_index
[
i
].
index
=
NULL
;
}
if
(
m_index
[
i
].
unique_index
)
{
dict
->
removeIndexGlobal
(
*
m_index
[
i
].
unique_index
,
1
);
m_index
[
i
].
unique_index
=
NULL
;
}
}
}
DBUG_ASSERT
(
error
==
0
||
error
==
4243
);
DBUG_RETURN
(
error
);
}
...
...
@@ -1400,26 +1266,32 @@ int ha_ndbcluster::drop_indexes(Ndb *ndb, TABLE *tab)
m_index
[
i
].
type
=
idx_type
;
if
(
m_index
[
i
].
status
==
TO_BE_DROPPED
)
{
NdbDictionary
::
Index
*
index
=
(
NdbDictionary
::
Index
*
)
m_index
[
i
].
index
;
NdbDictionary
::
Index
*
unique_index
=
(
NdbDictionary
::
Index
*
)
m_index
[
i
].
unique_index
;
const
NdbDictionary
::
Index
*
index
=
m_index
[
i
].
index
;
const
NdbDictionary
::
Index
*
unique_index
=
m_index
[
i
].
unique_index
;
if
(
index
)
{
index_name
=
index
->
getName
();
DBUG_PRINT
(
"info"
,
(
"Dropping index %u: %s"
,
i
,
index_name
));
// Drop ordered index from ndb
error
=
drop_ndb_index
(
index_name
);
}
error
=
dict
->
dropIndexGlobal
(
*
index
);
if
(
!
error
)
{
dict
->
removeIndexGlobal
(
*
index
,
1
);
m_index
[
i
].
index
=
NULL
;
}
}
if
(
!
error
&&
unique_index
)
{
index_name
=
unique_index
->
getName
();
DBUG_PRINT
(
"info"
,
(
"Dropping unique index %u: %s"
,
i
,
index_name
));
// Drop unique index from ndb
error
=
drop_ndb_index
(
index_name
);
error
=
dict
->
dropIndexGlobal
(
*
unique_index
);
if
(
!
error
)
{
dict
->
removeIndexGlobal
(
*
unique_index
,
1
);
m_index
[
i
].
unique_index
=
NULL
;
}
}
if
(
error
)
DBUG_RETURN
(
error
);
...
...
@@ -1476,30 +1348,57 @@ int ha_ndbcluster::check_index_fields_not_null(uint inx)
DBUG_RETURN
(
0
);
}
void
ha_ndbcluster
::
release_metadata
()
void
ha_ndbcluster
::
release_metadata
(
THD
*
thd
,
Ndb
*
ndb
)
{
uint
i
;
DBUG_ENTER
(
"release_metadata"
);
DBUG_PRINT
(
"enter"
,
(
"m_tabname: %s"
,
m_tabname
));
m_table
=
NULL
;
NDBDICT
*
dict
=
ndb
->
getDictionary
();
int
invalidate_indexes
=
0
;
if
(
thd
&&
thd
->
lex
&&
thd
->
lex
->
sql_command
==
SQLCOM_FLUSH
)
{
invalidate_indexes
=
1
;
}
if
(
m_table
!=
NULL
)
{
if
(
m_table
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Invalid
)
invalidate_indexes
=
1
;
dict
->
removeTableGlobal
(
*
m_table
,
invalidate_indexes
);
}
// TODO investigate
DBUG_ASSERT
(
m_table_info
==
NULL
);
m_table_info
=
NULL
;
// Release index list
for
(
i
=
0
;
i
<
MAX_KEY
;
i
++
)
{
if
(
m_index
[
i
].
unique_index
)
{
DBUG_ASSERT
(
m_table
!=
NULL
);
dict
->
removeIndexGlobal
(
*
m_index
[
i
].
unique_index
,
invalidate_indexes
);
m_index
[
i
].
unique_index
=
NULL
;
}
if
(
m_index
[
i
].
index
)
{
DBUG_ASSERT
(
m_table
!=
NULL
);
dict
->
removeIndexGlobal
(
*
m_index
[
i
].
index
,
invalidate_indexes
);
m_index
[
i
].
index
=
NULL
;
}
if
(
m_index
[
i
].
unique_index_attrid_map
)
{
my_free
((
char
*
)
m_index
[
i
].
unique_index_attrid_map
,
MYF
(
0
));
m_index
[
i
].
unique_index_attrid_map
=
NULL
;
}
if
(
m_index
[
i
].
index_stat
)
{
delete
m_index
[
i
].
index_stat
;
m_index
[
i
].
index_stat
=
NULL
;
m_index
[
i
].
index_stat
=
NULL
;
}
}
m_table
=
NULL
;
DBUG_VOID_RETURN
;
}
...
...
@@ -1971,11 +1870,10 @@ int ha_ndbcluster::peek_indexed_rows(const byte *record)
{
// A unique index is defined on table
NdbIndexOperation
*
iop
;
NDBINDEX
*
unique_index
=
(
NDBINDEX
*
)
m_index
[
i
].
unique_index
;
const
NDBINDEX
*
unique_index
=
m_index
[
i
].
unique_index
;
key_part
=
key_info
->
key_part
;
end
=
key_part
+
key_info
->
key_parts
;
if
(
!
(
iop
=
trans
->
getNdbIndexOperation
(
unique_index
,
(
const
NDBTAB
*
)
m_table
))
||
if
(
!
(
iop
=
trans
->
getNdbIndexOperation
(
unique_index
,
m_table
))
||
iop
->
readTuple
(
lm
)
!=
0
)
ERR_RETURN
(
trans
->
getNdbError
());
...
...
@@ -2024,9 +1922,8 @@ int ha_ndbcluster::unique_index_read(const byte *key,
NdbOperation
::
LockMode
lm
=
(
NdbOperation
::
LockMode
)
get_ndb_lock_type
(
m_lock
.
type
);
if
(
!
(
op
=
trans
->
getNdbIndexOperation
((
NDBINDEX
*
)
m_index
[
active_index
].
unique_index
,
(
const
NDBTAB
*
)
m_table
))
||
if
(
!
(
op
=
trans
->
getNdbIndexOperation
(
m_index
[
active_index
].
unique_index
,
m_table
))
||
op
->
readTuple
(
lm
)
!=
0
)
ERR_RETURN
(
trans
->
getNdbError
());
...
...
@@ -2365,9 +2262,8 @@ int ha_ndbcluster::ordered_index_scan(const key_range *start_key,
restart
=
FALSE
;
NdbOperation
::
LockMode
lm
=
(
NdbOperation
::
LockMode
)
get_ndb_lock_type
(
m_lock
.
type
);
if
(
!
(
op
=
trans
->
getNdbIndexScanOperation
((
NDBINDEX
*
)
m_index
[
active_index
].
index
,
(
const
NDBTAB
*
)
m_table
))
||
if
(
!
(
op
=
trans
->
getNdbIndexScanOperation
(
m_index
[
active_index
].
index
,
m_table
))
||
op
->
readTuples
(
lm
,
0
,
parallelism
,
sorted
,
descending
))
ERR_RETURN
(
trans
->
getNdbError
());
if
(
m_use_partition_function
&&
part_spec
!=
NULL
&&
...
...
@@ -2437,7 +2333,7 @@ int ha_ndbcluster::full_table_scan(byte *buf)
NdbOperation
::
LockMode
lm
=
(
NdbOperation
::
LockMode
)
get_ndb_lock_type
(
m_lock
.
type
);
if
(
!
(
op
=
trans
->
getNdbScanOperation
(
(
const
NDBTAB
*
)
m_table
))
||
if
(
!
(
op
=
trans
->
getNdbScanOperation
(
m_table
))
||
op
->
readTuples
(
lm
,
0
,
parallelism
))
ERR_RETURN
(
trans
->
getNdbError
());
m_active_cursor
=
op
;
...
...
@@ -2539,7 +2435,7 @@ int ha_ndbcluster::write_row(byte *record)
if
(
table
->
timestamp_field_type
&
TIMESTAMP_AUTO_SET_ON_INSERT
)
table
->
timestamp_field
->
set_time
();
if
(
!
(
op
=
trans
->
getNdbOperation
(
(
const
NDBTAB
*
)
m_table
)))
if
(
!
(
op
=
trans
->
getNdbOperation
(
m_table
)))
ERR_RETURN
(
trans
->
getNdbError
());
res
=
(
m_use_write
)
?
op
->
writeTuple
()
:
op
->
insertTuple
();
...
...
@@ -2565,7 +2461,7 @@ int ha_ndbcluster::write_row(byte *record)
Uint64
auto_value
=
NDB_FAILED_AUTO_INCREMENT
;
uint
retries
=
NDB_AUTO_INCREMENT_RETRIES
;
do
{
auto_value
=
ndb
->
getAutoIncrementValue
(
(
const
NDBTAB
*
)
m_table
);
auto_value
=
ndb
->
getAutoIncrementValue
(
m_table
);
}
while
(
auto_value
==
NDB_FAILED_AUTO_INCREMENT
&&
--
retries
&&
ndb
->
getNdbError
().
status
==
NdbError
::
TemporaryError
);
...
...
@@ -2666,7 +2562,7 @@ int ha_ndbcluster::write_row(byte *record)
DBUG_PRINT
(
"info"
,
(
"Trying to set next auto increment value to %lu"
,
(
ulong
)
next_val
));
if
(
ndb
->
setAutoIncrementValue
(
(
const
NDBTAB
*
)
m_table
,
next_val
,
TRUE
))
if
(
ndb
->
setAutoIncrementValue
(
m_table
,
next_val
,
TRUE
))
DBUG_PRINT
(
"info"
,
(
"Setting next auto increment value to %u"
,
next_val
));
}
...
...
@@ -2816,7 +2712,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
}
else
{
if
(
!
(
op
=
trans
->
getNdbOperation
(
(
const
NDBTAB
*
)
m_table
))
||
if
(
!
(
op
=
trans
->
getNdbOperation
(
m_table
))
||
op
->
updateTuple
()
!=
0
)
ERR_RETURN
(
trans
->
getNdbError
());
...
...
@@ -2925,7 +2821,7 @@ int ha_ndbcluster::delete_row(const byte *record)
else
{
if
(
!
(
op
=
trans
->
getNdbOperation
(
(
const
NDBTAB
*
)
m_table
))
||
if
(
!
(
op
=
trans
->
getNdbOperation
(
m_table
))
||
op
->
deleteTuple
()
!=
0
)
ERR_RETURN
(
trans
->
getNdbError
());
...
...
@@ -3078,7 +2974,7 @@ void ha_ndbcluster::unpack_record(byte *buf)
{
// Table with hidden primary key
int
hidden_no
=
table_share
->
fields
;
const
NDBTAB
*
tab
=
(
const
NDBTAB
*
)
m_table
;
const
NDBTAB
*
tab
=
m_table
;
const
NDBCOL
*
hidden_col
=
tab
->
getColumn
(
hidden_no
);
const
NdbRecAttr
*
rec
=
m_value
[
hidden_no
].
rec
;
DBUG_ASSERT
(
rec
);
...
...
@@ -3554,7 +3450,7 @@ void ha_ndbcluster::position(const byte *record)
key_length
=
ref_length
;
#ifndef DBUG_OFF
int
hidden_no
=
table
->
s
->
fields
;
const
NDBTAB
*
tab
=
(
const
NDBTAB
*
)
m_table
;
const
NDBTAB
*
tab
=
m_table
;
const
NDBCOL
*
hidden_col
=
tab
->
getColumn
(
hidden_no
);
DBUG_ASSERT
(
hidden_col
->
getPrimaryKey
()
&&
hidden_col
->
getAutoIncrement
()
&&
...
...
@@ -3600,7 +3496,7 @@ void ha_ndbcluster::info(uint flag)
ndb
->
setDatabaseName
(
m_dbname
);
struct
Ndb_statistics
stat
;
if
(
current_thd
->
variables
.
ndb_use_exact_count
&&
ndb_get_table_statistics
(
ndb
,
m_tab
nam
e
,
&
stat
)
==
0
)
ndb_get_table_statistics
(
ndb
,
m_tab
l
e
,
&
stat
)
==
0
)
{
mean_rec_length
=
stat
.
row_size
;
data_file_length
=
stat
.
fragment_memory
;
...
...
@@ -3631,7 +3527,7 @@ void ha_ndbcluster::info(uint flag)
Ndb
*
ndb
=
get_ndb
();
auto_increment_value
=
ndb
->
readAutoIncrementValue
(
(
const
NDBTAB
*
)
m_table
);
ndb
->
readAutoIncrementValue
(
m_table
);
}
}
DBUG_VOID_RETURN
;
...
...
@@ -3715,7 +3611,7 @@ int ha_ndbcluster::extra(enum ha_extra_function operation)
void
ha_ndbcluster
::
start_bulk_insert
(
ha_rows
rows
)
{
int
bytes
,
batch
;
const
NDBTAB
*
tab
=
(
const
NDBTAB
*
)
m_table
;
const
NDBTAB
*
tab
=
m_table
;
DBUG_ENTER
(
"start_bulk_insert"
);
DBUG_PRINT
(
"enter"
,
(
"rows: %d"
,
(
int
)
rows
));
...
...
@@ -3994,41 +3890,8 @@ int ha_ndbcluster::external_lock(THD *thd, int lock_type)
// Start of transaction
m_rows_changed
=
0
;
m_ops_pending
=
0
;
{
NDBDICT
*
dict
=
ndb
->
getDictionary
();
const
NDBTAB
*
tab
;
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_PRINT
(
"info"
,
(
"Table schema version: %d"
,
tab
->
getObjectVersion
()));
// Check if thread has stale local cache
// New transaction must not use old tables... (trans != 0)
// Running might...
if
((
trans
&&
tab
->
getObjectStatus
()
!=
NdbDictionary
::
Object
::
Retrieved
)
||
tab
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Invalid
)
{
invalidate_dictionary_cache
(
FALSE
,
tab
);
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_PRINT
(
"info"
,
(
"Table schema version: %d"
,
tab
->
getObjectVersion
()));
}
if
(
m_table_version
<
tab
->
getObjectVersion
())
{
/*
The table has been altered, caller has to retry
*/
NdbError
err
=
ndb
->
getNdbError
(
NDB_INVALID_SCHEMA_OBJECT
);
DBUG_RETURN
(
ndb_to_mysql_error
(
&
err
));
}
if
(
m_table
!=
(
void
*
)
tab
)
{
m_table
=
tab
;
m_table_version
=
tab
->
getObjectVersion
();
if
(
!
(
my_errno
=
open_indexes
(
ndb
,
table
,
FALSE
)))
DBUG_RETURN
(
my_errno
);
}
}
// TODO remove double pointers...
m_thd_ndb_share
=
thd_ndb
->
get_open_table
(
thd
,
m_table
);
m_table_info
=
&
m_thd_ndb_share
->
stat
;
}
...
...
@@ -4673,6 +4536,14 @@ int ha_ndbcluster::create(const char *name,
my_errno
=
ndb_to_mysql_error
(
&
err
);
DBUG_RETURN
(
my_errno
);
}
Ndb_table_guard
ndbtab_g
(
dict
,
m_tabname
);
// temporary set m_table during create
// reset at return
m_table
=
ndbtab_g
.
get_table
();
// TODO check also that we have the same frm...
DBUG_ASSERT
(
m_table
!=
0
);
DBUG_PRINT
(
"info"
,
(
"Table %s/%s created successfully"
,
m_dbname
,
m_tabname
));
...
...
@@ -4687,7 +4558,10 @@ int ha_ndbcluster::create(const char *name,
Failed to create an index,
drop the table (and all it's indexes)
*/
drop_ndb_table
();
if
(
dict
->
dropTableGlobal
(
*
m_table
)
==
0
)
{
m_table
=
0
;
}
}
#ifdef HAVE_NDB_BINLOG
...
...
@@ -4717,7 +4591,6 @@ int ha_ndbcluster::create(const char *name,
while
(
!
IS_TMP_PREFIX
(
m_tabname
))
{
const
NDBTAB
*
t
=
dict
->
getTable
(
m_tabname
);
String
event_name
(
INJECTOR_EVENT_LEN
);
ndb_rep_event_name
(
&
event_name
,
m_dbname
,
m_tabname
);
int
do_event_op
=
ndb_binlog_running
;
...
...
@@ -4731,14 +4604,14 @@ int ha_ndbcluster::create(const char *name,
Always create an event for the table, as other mysql servers
expect it to be there.
*/
if
(
!
ndbcluster_create_event
(
ndb
,
t
,
event_name
.
c_ptr
(),
share
,
if
(
!
ndbcluster_create_event
(
ndb
,
m_table
,
event_name
.
c_ptr
(),
share
,
share
&&
do_event_op
/* push warning */
))
{
if
(
ndb_extra_logging
)
sql_print_information
(
"NDB Binlog: CREATE TABLE Event: %s"
,
event_name
.
c_ptr
());
if
(
share
&&
do_event_op
&&
ndbcluster_create_event_ops
(
share
,
t
,
event_name
.
c_ptr
()))
ndbcluster_create_event_ops
(
share
,
m_table
,
event_name
.
c_ptr
()))
{
sql_print_error
(
"NDB Binlog: FAILED CREATE TABLE event operations."
" Event: %s"
,
name
);
...
...
@@ -4754,13 +4627,15 @@ int ha_ndbcluster::create(const char *name,
ndbcluster_log_schema_op
(
current_thd
,
share
,
current_thd
->
query
,
current_thd
->
query_length
,
share
->
db
,
share
->
table_name
,
0
,
0
,
m_table
->
getObjectId
(),
m_table
->
getObjectVersion
(),
SOT_CREATE_TABLE
);
break
;
}
}
#endif
/* HAVE_NDB_BINLOG */
m_table
=
0
;
DBUG_RETURN
(
my_errno
);
}
...
...
@@ -4790,10 +4665,15 @@ int ha_ndbcluster::create_handler_files(const char *file,
NDBDICT
*
dict
=
ndb
->
getDictionary
();
if
(
!
info
->
frm_only
)
DBUG_RETURN
(
0
);
// Must be a create, ignore since frm is saved in create
// TODO handle this
DBUG_ASSERT
(
m_table
!=
0
);
set_dbname
(
file
);
set_tabname
(
file
);
Ndb_table_guard
ndbtab_g
(
dict
,
m_tabname
);
DBUG_PRINT
(
"info"
,
(
"m_dbname: %s, m_tabname: %s"
,
m_dbname
,
m_tabname
));
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
if
(
!
(
tab
=
ndbtab_g
.
get_table
(
)))
DBUG_RETURN
(
0
);
// Unkown table, must be temporary table
DBUG_ASSERT
(
get_ndb_share_state
(
m_share
)
==
NSS_ALTERED
);
...
...
@@ -4809,7 +4689,12 @@ int ha_ndbcluster::create_handler_files(const char *file,
{
DBUG_PRINT
(
"info"
,
(
"Table %s has changed, altering frm in ndb"
,
m_tabname
));
error
=
table_changed
(
pack_data
,
pack_length
);
NdbDictionary
::
Table
new_tab
=
*
tab
;
new_tab
.
setFrm
(
pack_data
,
pack_length
);
if
(
dict
->
alterTableGlobal
(
*
tab
,
new_tab
))
{
error
=
ndb_to_mysql_error
(
&
dict
->
getNdbError
());
}
my_free
((
char
*
)
data
,
MYF
(
MY_ALLOW_ZERO_PTR
));
my_free
((
char
*
)
pack_data
,
MYF
(
MY_ALLOW_ZERO_PTR
));
}
...
...
@@ -4912,7 +4797,7 @@ int ha_ndbcluster::create_ndb_index(const char *name,
ndb_index
.
addColumnName
(
field
->
field_name
);
}
if
(
dict
->
createIndex
(
ndb_index
))
if
(
dict
->
createIndex
(
ndb_index
,
*
m_table
))
ERR_RETURN
(
dict
->
getNdbError
());
// Success
...
...
@@ -4965,18 +4850,6 @@ int ha_ndbcluster::add_index(TABLE *table_arg,
DBUG_RETURN
(
error
);
}
/*
Drop an index in ndb
*/
int
ha_ndbcluster
::
drop_ndb_index
(
const
char
*
name
)
{
DBUG_ENTER
(
"ha_ndbcluster::drop_index"
);
DBUG_PRINT
(
"enter"
,
(
"name: %s "
,
name
));
Ndb
*
ndb
=
get_ndb
();
NdbDictionary
::
Dictionary
*
dict
=
ndb
->
getDictionary
();
DBUG_RETURN
(
dict
->
dropIndex
(
name
,
m_tabname
));
}
/*
Mark one or several indexes for deletion. and
renumber the remaining indexes
...
...
@@ -5045,16 +4918,14 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
Ndb
*
ndb
=
get_ndb
();
ndb
->
setDatabaseName
(
old_dbname
);
dict
=
ndb
->
getDictionary
();
if
(
!
(
orig_tab
=
dict
->
getTable
(
m_tabname
)))
Ndb_table_guard
ndbtab_g
(
dict
,
m_tabname
);
if
(
!
(
orig_tab
=
ndbtab_g
.
get_table
()))
ERR_RETURN
(
dict
->
getNdbError
());
// Check if thread has stale local cache
if
(
orig_tab
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Invalid
)
{
dict
->
removeCachedTable
(
orig_tab
);
if
(
!
(
orig_tab
=
dict
->
getTable
(
m_tabname
)))
ERR_RETURN
(
dict
->
getNdbError
());
}
#ifdef HAVE_NDB_BINLOG
int
ndb_table_id
=
orig_tab
->
getObjectId
();
int
ndb_table_version
=
orig_tab
->
getObjectVersion
();
NDB_SHARE
*
share
=
get_share
(
from
,
0
,
false
);
if
(
share
)
{
...
...
@@ -5062,13 +4933,15 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
DBUG_ASSERT
(
r
==
0
);
}
#endif
m_table
=
orig_tab
;
// Change current database to that of target table
set_dbname
(
to
);
ndb
->
setDatabaseName
(
m_dbname
);
if
((
result
=
alter_table_name
(
new_tabname
)))
NdbDictionary
::
Table
new_tab
=
*
orig_tab
;
new_tab
.
setName
(
new_tabname
);
if
(
dict
->
alterTableGlobal
(
*
orig_tab
,
new_tab
)
!=
0
)
{
NdbError
ndb_error
=
dict
->
getNdbError
();
#ifdef HAVE_NDB_BINLOG
if
(
share
)
{
...
...
@@ -5077,7 +4950,7 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
free_share
(
&
share
);
}
#endif
DBUG_RETURN
(
result
);
ERR_RETURN
(
ndb_error
);
}
// Rename .ndb file
...
...
@@ -5111,7 +4984,8 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
/* always create an event for the table */
String
event_name
(
INJECTOR_EVENT_LEN
);
ndb_rep_event_name
(
&
event_name
,
to
+
sizeof
(
share_prefix
)
-
1
,
0
);
const
NDBTAB
*
ndbtab
=
dict
->
getTable
(
new_tabname
);
Ndb_table_guard
ndbtab_g2
(
dict
,
new_tabname
);
const
NDBTAB
*
ndbtab
=
ndbtab_g2
.
get_table
();
if
(
!
ndbcluster_create_event
(
ndb
,
ndbtab
,
event_name
.
c_ptr
(),
share
,
share
&&
ndb_binlog_running
/* push warning */
))
...
...
@@ -5134,10 +5008,10 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
if
(
!
is_old_table_tmpfile
)
ndbcluster_log_schema_op
(
current_thd
,
share
,
current_thd
->
query
,
current_thd
->
query_length
,
m_dbname
,
new
_tabname
,
0
,
0
,
old_dbname
,
m
_tabname
,
ndb_table_id
,
ndb_table_version
,
SOT_RENAME_TABLE
,
old_dbname
,
m
_tabname
);
m_dbname
,
new
_tabname
);
}
if
(
share
)
free_share
(
&
share
);
...
...
@@ -5147,30 +5021,6 @@ int ha_ndbcluster::rename_table(const char *from, const char *to)
}
/*
Rename a table in NDB Cluster using alter table
*/
int
ha_ndbcluster
::
alter_table_name
(
const
char
*
to
)
{
Ndb
*
ndb
=
get_ndb
();
NDBDICT
*
dict
=
ndb
->
getDictionary
();
const
NDBTAB
*
orig_tab
=
(
const
NDBTAB
*
)
m_table
;
DBUG_ENTER
(
"alter_table_name"
);
DBUG_PRINT
(
"info"
,
(
"from: %s to: %s"
,
orig_tab
->
getName
(),
to
));
NdbDictionary
::
Table
new_tab
=
*
orig_tab
;
new_tab
.
setName
(
to
);
if
(
dict
->
alterTable
(
new_tab
)
!=
0
)
ERR_RETURN
(
dict
->
getNdbError
());
m_table
=
NULL
;
m_table_info
=
NULL
;
DBUG_RETURN
(
0
);
}
/*
Delete table from NDB Cluster
...
...
@@ -5187,6 +5037,8 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
DBUG_ENTER
(
"ha_ndbcluster::ndbcluster_delete_table"
);
NDBDICT
*
dict
=
ndb
->
getDictionary
();
#ifdef HAVE_NDB_BINLOG
int
ndb_table_id
=
0
;
int
ndb_table_version
=
0
;
/*
Don't allow drop table unless
schema distribution table is setup
...
...
@@ -5202,14 +5054,45 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
/* Drop the table from NDB */
int
res
;
if
(
h
)
if
(
h
&&
h
->
m_table
)
{
res
=
h
->
drop_ndb_table
();
if
(
dict
->
dropTableGlobal
(
*
h
->
m_table
))
res
=
ndb_to_mysql_error
(
&
dict
->
getNdbError
());
#ifdef HAVE_NDB_BINLOG
if
(
res
==
0
)
{
ndb_table_id
=
h
->
m_table
->
getObjectId
();
ndb_table_version
=
h
->
m_table
->
getObjectVersion
();
}
#endif
h
->
release_metadata
(
current_thd
,
ndb
);
}
else
{
ndb
->
setDatabaseName
(
db
);
res
=
dict
->
dropTable
(
table_name
);
while
(
1
)
{
Ndb_table_guard
ndbtab_g
(
dict
,
table_name
);
if
(
ndbtab_g
.
get_table
())
{
if
(
dict
->
dropTableGlobal
(
*
ndbtab_g
.
get_table
())
==
0
)
{
#ifdef HAVE_NDB_BINLOG
ndb_table_id
=
ndbtab_g
.
get_table
()
->
getObjectId
();
ndb_table_version
=
ndbtab_g
.
get_table
()
->
getObjectVersion
();
#endif
res
=
0
;
}
else
if
(
dict
->
getNdbError
().
code
==
NDB_INVALID_SCHEMA_OBJECT
)
{
ndbtab_g
.
invalidate
();
continue
;
}
}
else
res
=
ndb_to_mysql_error
(
&
dict
->
getNdbError
());
break
;
}
}
if
(
res
)
...
...
@@ -5251,7 +5134,7 @@ ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
ndbcluster_log_schema_op
(
current_thd
,
share
,
current_thd
->
query
,
current_thd
->
query_length
,
share
->
db
,
share
->
table_name
,
0
,
0
,
ndb_table_id
,
ndb_table_version
,
SOT_DROP_TABLE
);
}
else
if
(
table_dropped
&&
share
&&
share
->
op
)
/* ndbcluster_log_schema_op
...
...
@@ -5315,24 +5198,6 @@ int ha_ndbcluster::delete_table(const char *name)
}
/*
Drop table in NDB Cluster
*/
int
ha_ndbcluster
::
drop_ndb_table
()
{
Ndb
*
ndb
=
get_ndb
();
NdbDictionary
::
Dictionary
*
dict
=
ndb
->
getDictionary
();
DBUG_ENTER
(
"intern_drop_table"
);
DBUG_PRINT
(
"enter"
,
(
"Deleting %s"
,
m_tabname
));
release_metadata
();
if
(
dict
->
dropTable
(
m_tabname
))
ERR_RETURN
(
dict
->
getNdbError
());
DBUG_RETURN
(
0
);
}
ulonglong
ha_ndbcluster
::
get_auto_increment
()
{
int
cache_size
;
...
...
@@ -5356,8 +5221,8 @@ ulonglong ha_ndbcluster::get_auto_increment()
do
{
auto_value
=
(
m_skip_auto_increment
)
?
ndb
->
readAutoIncrementValue
(
(
const
NDBTAB
*
)
m_table
)
:
ndb
->
getAutoIncrementValue
(
(
const
NDBTAB
*
)
m_table
,
cache_size
);
ndb
->
readAutoIncrementValue
(
m_table
)
:
ndb
->
getAutoIncrementValue
(
m_table
,
cache_size
);
}
while
(
auto_value
==
NDB_FAILED_AUTO_INCREMENT
&&
--
retries
&&
ndb
->
getNdbError
().
status
==
NdbError
::
TemporaryError
);
...
...
@@ -5391,7 +5256,6 @@ ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg):
m_active_trans
(
NULL
),
m_active_cursor
(
NULL
),
m_table
(
NULL
),
m_table_version
(
-
1
),
m_table_info
(
NULL
),
m_table_flags
(
HA_NDBCLUSTER_TABLE_FLAGS
),
m_share
(
0
),
...
...
@@ -5437,19 +5301,31 @@ ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg):
}
int
ha_ndbcluster
::
ha_initialise
()
{
DBUG_ENTER
(
"ha_ndbcluster::ha_initialise"
);
if
(
check_ndb_in_thd
(
current_thd
))
{
DBUG_RETURN
(
FALSE
);
}
DBUG_RETURN
(
TRUE
);
}
/*
Destructor for NDB Cluster table handler
*/
ha_ndbcluster
::~
ha_ndbcluster
()
{
THD
*
thd
=
current_thd
;
Ndb
*
ndb
=
thd
?
check_ndb_in_thd
(
thd
)
:
g_ndb
;
DBUG_ENTER
(
"~ha_ndbcluster"
);
if
(
m_share
)
{
free_share
(
&
m_share
);
}
release_metadata
();
release_metadata
(
thd
,
ndb
);
my_free
(
m_blobs_buffer
,
MYF
(
MY_ALLOW_ZERO_PTR
));
m_blobs_buffer
=
0
;
...
...
@@ -5565,9 +5441,11 @@ void ha_ndbcluster::set_part_info(partition_info *part_info)
int
ha_ndbcluster
::
close
(
void
)
{
DBUG_ENTER
(
"close"
);
THD
*
thd
=
current_thd
;
Ndb
*
ndb
=
thd
?
check_ndb_in_thd
(
thd
)
:
g_ndb
;
free_share
(
&
m_share
);
m_share
=
0
;
release_metadata
();
release_metadata
(
thd
,
ndb
);
DBUG_RETURN
(
0
);
}
...
...
@@ -5655,9 +5533,10 @@ static int ndbcluster_close_connection(THD *thd)
int
ndbcluster_discover
(
THD
*
thd
,
const
char
*
db
,
const
char
*
name
,
const
void
**
frmblob
,
uint
*
frmlen
)
{
int
error
=
0
;
NdbError
ndb_error
;
uint
len
;
const
void
*
data
;
const
NDBTAB
*
tab
;
Ndb
*
ndb
;
char
key
[
FN_REFLEN
];
DBUG_ENTER
(
"ndbcluster_discover"
);
...
...
@@ -5667,7 +5546,6 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
DBUG_RETURN
(
HA_ERR_NO_CONNECTION
);
ndb
->
setDatabaseName
(
db
);
NDBDICT
*
dict
=
ndb
->
getDictionary
();
dict
->
invalidateTable
(
name
);
build_table_filename
(
key
,
sizeof
(
key
),
db
,
name
,
""
);
NDB_SHARE
*
share
=
get_share
(
key
,
0
,
false
);
if
(
share
&&
get_ndb_share_state
(
share
)
==
NSS_ALTERED
)
...
...
@@ -5676,21 +5554,22 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
if
(
readfrm
(
key
,
&
data
,
&
len
))
{
DBUG_PRINT
(
"error"
,
(
"Could not read frm"
));
if
(
share
)
free_share
(
&
share
);
DBUG_RETURN
(
1
);
error
=
1
;
goto
err
;
}
}
else
{
if
(
!
(
tab
=
dict
->
getTable
(
name
)))
Ndb_table_guard
ndbtab_g
(
dict
,
name
);
const
NDBTAB
*
tab
=
ndbtab_g
.
get_table
();
if
(
!
tab
)
{
const
NdbError
err
=
dict
->
getNdbError
();
if
(
share
)
free_share
(
&
share
);
if
(
err
.
code
==
709
||
err
.
code
==
723
)
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
err
);
error
=
-
1
;
else
ndb_error
=
err
;
goto
err
;
}
DBUG_PRINT
(
"info"
,
(
"Found table %s"
,
tab
->
getName
()));
...
...
@@ -5698,17 +5577,15 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
if
(
len
==
0
||
tab
->
getFrmData
()
==
NULL
)
{
DBUG_PRINT
(
"error"
,
(
"No frm data found."
));
if
(
share
)
free_share
(
&
share
);
DBUG_RETURN
(
1
);
error
=
1
;
goto
err
;
}
if
(
unpackfrm
(
&
data
,
&
len
,
tab
->
getFrmData
()))
{
DBUG_PRINT
(
"error"
,
(
"Could not unpack table"
));
if
(
share
)
free_share
(
&
share
);
DBUG_RETURN
(
1
);
error
=
1
;
goto
err
;
}
}
...
...
@@ -5719,6 +5596,14 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
free_share
(
&
share
);
DBUG_RETURN
(
0
);
err:
if
(
share
)
free_share
(
&
share
);
if
(
ndb_error
.
code
)
{
ERR_RETURN
(
ndb_error
);
}
DBUG_RETURN
(
error
);
}
/*
...
...
@@ -5726,29 +5611,32 @@ int ndbcluster_discover(THD* thd, const char *db, const char *name,
*/
int
ndbcluster_table_exists_in_engine
(
THD
*
thd
,
const
char
*
db
,
const
char
*
name
)
int
ndbcluster_table_exists_in_engine
(
THD
*
thd
,
const
char
*
db
,
const
char
*
name
)
{
const
NDBTAB
*
tab
;
Ndb
*
ndb
;
DBUG_ENTER
(
"ndbcluster_table_exists_in_engine"
);
DBUG_PRINT
(
"enter"
,
(
"db: %s, name: %s"
,
db
,
name
));
if
(
!
(
ndb
=
check_ndb_in_thd
(
thd
)))
DBUG_RETURN
(
HA_ERR_NO_CONNECTION
);
ndb
->
setDatabaseName
(
db
);
NDBDICT
*
dict
=
ndb
->
getDictionary
();
dict
->
invalidateTable
(
name
);
if
(
!
(
tab
=
dict
->
getTable
(
name
)))
NdbDictionary
::
Dictionary
::
List
list
;
if
(
dict
->
listObjects
(
list
,
NdbDictionary
::
Object
::
UserTable
)
!=
0
)
ERR_RETURN
(
dict
->
getNdbError
());
for
(
int
i
=
0
;
i
<
list
.
count
;
i
++
)
{
const
NdbError
err
=
dict
->
getNdbError
()
;
if
(
err
.
code
==
709
||
err
.
code
==
723
)
DBUG_RETURN
(
0
)
;
ERR_RETURN
(
err
);
}
DBUG_PRINT
(
"info"
,
(
"Found table %s"
,
tab
->
getName
()
));
NdbDictionary
::
Dictionary
::
List
::
Element
&
elmt
=
list
.
elements
[
i
]
;
if
(
my_strcasecmp
(
system_charset_info
,
elmt
.
database
,
db
)
)
continue
;
if
(
my_strcasecmp
(
system_charset_info
,
elmt
.
name
,
name
))
continue
;
// table found
DBUG_PRINT
(
"info"
,
(
"Found table"
));
DBUG_RETURN
(
1
);
}
DBUG_RETURN
(
0
);
}
...
...
@@ -5902,9 +5790,9 @@ int ndbcluster_find_all_files(THD *thd)
}
ndb
->
setDatabaseName
(
elmt
.
database
);
const
NDBTAB
*
ndbtab
;
if
(
!
(
ndbtab
=
dict
->
getTable
(
elmt
.
name
))
)
Ndb_table_guard
ndbtab_g
(
dict
,
elmt
.
name
)
;
const
NDBTAB
*
ndbtab
=
ndbtab_g
.
get_table
();
if
(
!
ndbtab
)
{
if
(
retries
==
0
)
sql_print_error
(
"NDB: failed to setup table %s.%s, error: %d, %s"
,
...
...
@@ -6184,11 +6072,13 @@ static int connect_callback()
return
0
;
}
extern
int
ndb_dictionary_is_mysqld
;
static
bool
ndbcluster_init
()
{
int
res
;
DBUG_ENTER
(
"ndbcluster_init"
);
ndb_dictionary_is_mysqld
=
1
;
if
(
have_ndbcluster
!=
SHOW_OPTION_YES
)
goto
ndbcluster_init_error
;
...
...
@@ -6320,6 +6210,24 @@ static int ndbcluster_end(ha_panic_function type)
if
(
!
ndbcluster_inited
)
DBUG_RETURN
(
0
);
#ifdef HAVE_NDB_BINLOG
{
pthread_mutex_lock
(
&
ndbcluster_mutex
);
for
(
uint
i
=
0
;
i
<
ndbcluster_open_tables
.
records
;
i
++
)
{
NDB_SHARE
*
share
=
(
NDB_SHARE
*
)
hash_element
(
&
ndbcluster_open_tables
,
i
);
#ifndef DBUG_OFF
fprintf
(
stderr
,
"NDB: table share %s with use_count %d not freed
\n
"
,
share
->
key
,
share
->
use_count
);
#endif
real_free_share
(
&
share
);
}
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
}
#endif
hash_free
(
&
ndbcluster_open_tables
);
if
(
g_ndb
)
{
#ifndef DBUG_OFF
...
...
@@ -6341,23 +6249,6 @@ static int ndbcluster_end(ha_panic_function type)
delete
g_ndb_cluster_connection
;
g_ndb_cluster_connection
=
NULL
;
#ifdef HAVE_NDB_BINLOG
{
pthread_mutex_lock
(
&
ndbcluster_mutex
);
for
(
uint
i
=
0
;
i
<
ndbcluster_open_tables
.
records
;
i
++
)
{
NDB_SHARE
*
share
=
(
NDB_SHARE
*
)
hash_element
(
&
ndbcluster_open_tables
,
i
);
#ifndef DBUG_OFF
fprintf
(
stderr
,
"NDB: table share %s with use_count %d not freed
\n
"
,
share
->
key
,
share
->
use_count
);
#endif
real_free_share
(
&
share
);
}
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
}
#endif
hash_free
(
&
ndbcluster_open_tables
);
pthread_mutex_destroy
(
&
ndbcluster_mutex
);
pthread_mutex_destroy
(
&
LOCK_ndb_util_thread
);
pthread_cond_destroy
(
&
COND_ndb_util_thread
);
...
...
@@ -6518,7 +6409,7 @@ ha_ndbcluster::records_in_range(uint inx, key_range *min_key,
m_index
[
inx
].
index_stat
!=
NULL
)
{
NDB_INDEX_DATA
&
d
=
m_index
[
inx
];
NDBINDEX
*
index
=
(
NDBINDEX
*
)
d
.
index
;
const
NDBINDEX
*
index
=
d
.
index
;
Ndb
*
ndb
=
get_ndb
();
NdbTransaction
*
trans
=
NULL
;
NdbIndexScanOperation
*
op
=
NULL
;
...
...
@@ -6538,7 +6429,7 @@ ha_ndbcluster::records_in_range(uint inx, key_range *min_key,
else
{
Ndb_statistics
stat
;
if
((
res
=
ndb_get_table_statistics
(
ndb
,
m_tab
nam
e
,
&
stat
))
!=
0
)
if
((
res
=
ndb_get_table_statistics
(
ndb
,
m_tab
l
e
,
&
stat
))
!=
0
)
break
;
table_rows
=
stat
.
row_count
;
DBUG_PRINT
(
"info"
,
(
"use db row_count: %llu"
,
table_rows
));
...
...
@@ -6699,11 +6590,15 @@ uint ndb_get_commitcount(THD *thd, char *dbname, char *tabname,
pthread_mutex_unlock
(
&
share
->
mutex
);
struct
Ndb_statistics
stat
;
if
(
ndb_get_table_statistics
(
ndb
,
tabname
,
&
stat
))
{
Ndb_table_guard
ndbtab_g
(
ndb
->
getDictionary
(),
tabname
);
if
(
ndbtab_g
.
get_table
()
==
0
||
ndb_get_table_statistics
(
ndb
,
ndbtab_g
.
get_table
(),
&
stat
))
{
free_share
(
&
share
);
DBUG_RETURN
(
1
);
}
}
pthread_mutex_lock
(
&
share
->
mutex
);
if
(
share
->
commit_count_lock
==
lock
)
...
...
@@ -6916,7 +6811,11 @@ int handle_trailing_share(NDB_SHARE *share)
++
share
->
use_count
;
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
close_cached_tables
((
THD
*
)
0
,
0
,
(
TABLE_LIST
*
)
0
,
TRUE
);
TABLE_LIST
table_list
;
bzero
((
char
*
)
&
table_list
,
sizeof
(
table_list
));
table_list
.
db
=
share
->
db
;
table_list
.
alias
=
table_list
.
table_name
=
share
->
table_name
;
close_cached_tables
(
current_thd
,
0
,
&
table_list
,
TRUE
);
pthread_mutex_lock
(
&
ndbcluster_mutex
);
if
(
!--
share
->
use_count
)
...
...
@@ -7262,15 +7161,17 @@ void ndbcluster_free_share(NDB_SHARE **share, bool have_lock)
static
int
ndb_get_table_statistics
(
Ndb
*
ndb
,
const
char
*
table
,
ndb_get_table_statistics
(
Ndb
*
ndb
,
const
NDBTAB
*
ndbtab
,
struct
Ndb_statistics
*
ndbstat
)
{
DBUG_ENTER
(
"ndb_get_table_statistics"
);
DBUG_PRINT
(
"enter"
,
(
"table: %s"
,
table
));
DBUG_PRINT
(
"enter"
,
(
"table: %s"
,
ndbtab
->
getName
()
));
NdbTransaction
*
pTrans
;
int
retries
=
10
;
int
retry_sleep
=
30
*
1000
;
/* 30 milliseconds */
DBUG_ASSERT
(
ndbtab
!=
0
);
do
{
pTrans
=
ndb
->
startTransaction
();
...
...
@@ -7285,7 +7186,7 @@ ndb_get_table_statistics(Ndb* ndb, const char * table,
ERR_RETURN
(
ndb
->
getNdbError
());
}
NdbScanOperation
*
pOp
=
pTrans
->
getNdbScanOperation
(
table
);
NdbScanOperation
*
pOp
=
pTrans
->
getNdbScanOperation
(
ndbtab
);
if
(
pOp
==
NULL
)
break
;
...
...
@@ -7446,9 +7347,9 @@ ha_ndbcluster::read_multi_range_first(KEY_MULTI_RANGE **found_range_p,
byte
*
end_of_buffer
=
(
byte
*
)
buffer
->
buffer_end
;
NdbOperation
::
LockMode
lm
=
(
NdbOperation
::
LockMode
)
get_ndb_lock_type
(
m_lock
.
type
);
const
NDBTAB
*
tab
=
(
const
NDBTAB
*
)
m_table
;
const
NDBINDEX
*
unique_idx
=
(
NDBINDEX
*
)
m_index
[
active_index
].
unique_index
;
const
NDBINDEX
*
idx
=
(
NDBINDEX
*
)
m_index
[
active_index
].
index
;
const
NDBTAB
*
tab
=
m_table
;
const
NDBINDEX
*
unique_idx
=
m_index
[
active_index
].
unique_index
;
const
NDBINDEX
*
idx
=
m_index
[
active_index
].
index
;
const
NdbOperation
*
lastOp
=
m_active_trans
->
getLastDefinedOperation
();
NdbIndexScanOperation
*
scanOp
=
0
;
for
(;
multi_range_curr
<
multi_range_end
&&
curr
+
reclength
<=
end_of_buffer
;
...
...
@@ -7797,11 +7698,8 @@ ha_ndbcluster::update_table_comment(
ndb
->
setDatabaseName
(
m_dbname
);
NDBDICT
*
dict
=
ndb
->
getDictionary
();
const
NDBTAB
*
tab
;
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
{
return
((
char
*
)
comment
);
}
const
NDBTAB
*
tab
=
m_table
;
DBUG_ASSERT
(
tab
!=
NULL
);
char
*
str
;
const
char
*
fmt
=
"%s%snumber_of_replicas: %d"
;
...
...
@@ -7987,7 +7885,10 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused)))
lock
=
share
->
commit_count_lock
;
pthread_mutex_unlock
(
&
share
->
mutex
);
if
(
ndb_get_table_statistics
(
ndb
,
share
->
table_name
,
&
stat
)
==
0
)
{
Ndb_table_guard
ndbtab_g
(
ndb
->
getDictionary
(),
share
->
table_name
);
if
(
ndbtab_g
.
get_table
()
&&
ndb_get_table_statistics
(
ndb
,
ndbtab_g
.
get_table
(),
&
stat
)
==
0
)
{
DBUG_PRINT
(
"ndb_util_thread"
,
(
"Table: %s, commit_count: %llu, rows: %llu"
,
...
...
@@ -8000,6 +7901,7 @@ pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused)))
share
->
key
));
stat
.
commit_count
=
0
;
}
}
pthread_mutex_lock
(
&
share
->
mutex
);
if
(
share
->
commit_count_lock
==
lock
)
...
...
@@ -9390,12 +9292,8 @@ char* ha_ndbcluster::get_tablespace_name(THD *thd)
NdbError
ndberr
;
Uint32
id
;
ndb
->
setDatabaseName
(
m_dbname
);
const
NDBTAB
*
ndbtab
=
ndbdict
->
getTable
(
m_tabname
);
if
(
ndbtab
==
0
)
{
ndberr
=
ndbdict
->
getNdbError
();
goto
err
;
}
const
NDBTAB
*
ndbtab
=
m_table
;
DBUG_ASSERT
(
ndbtab
!=
NULL
);
if
(
!
ndbtab
->
getTablespace
(
&
id
))
{
return
0
;
...
...
@@ -10060,17 +9958,10 @@ bool ha_ndbcluster::get_no_parts(const char *name, uint *no_parts)
}
ndb
=
get_ndb
();
ndb
->
setDatabaseName
(
m_dbname
);
dict
=
ndb
->
getDictionary
(
);
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)
))
Ndb_table_guard
ndbtab_g
(
dict
=
ndb
->
getDictionary
(),
m_tabname
);
if
(
!
ndbtab_g
.
get_table
(
))
ERR_BREAK
(
dict
->
getNdbError
(),
err
);
// Check if thread has stale local cache
if
(
tab
->
getObjectStatus
()
==
NdbDictionary
::
Object
::
Invalid
)
{
invalidate_dictionary_cache
(
FALSE
,
tab
);
if
(
!
(
tab
=
dict
->
getTable
(
m_tabname
)))
ERR_BREAK
(
dict
->
getNdbError
(),
err
);
}
*
no_parts
=
tab
->
getFragmentCount
();
*
no_parts
=
ndbtab_g
.
get_table
()
->
getFragmentCount
();
DBUG_RETURN
(
FALSE
);
}
while
(
1
);
...
...
sql/ha_ndbcluster.h
View file @
b96c4289
...
...
@@ -70,8 +70,8 @@ typedef enum ndb_index_status {
typedef
struct
ndb_index_data
{
NDB_INDEX_TYPE
type
;
NDB_INDEX_STATUS
status
;
void
*
index
;
void
*
unique_index
;
const
NdbDictionary
::
Index
*
index
;
const
NdbDictionary
::
Index
*
unique_index
;
unsigned
char
*
unique_index_attrid_map
;
// In this version stats are not shared between threads
NdbIndexStat
*
index_stat
;
...
...
@@ -560,6 +560,7 @@ class ha_ndbcluster: public handler
ha_ndbcluster
(
TABLE_SHARE
*
table
);
~
ha_ndbcluster
();
int
ha_initialise
();
int
open
(
const
char
*
name
,
int
mode
,
uint
test_if_locked
);
int
close
(
void
);
...
...
@@ -708,19 +709,15 @@ static void set_tabname(const char *pathname, char *tabname);
Ndb
*
ndb
,
NdbEventOperation
*
pOp
,
NDB_SHARE
*
share
);
int
alter_table_name
(
const
char
*
to
);
static
int
delete_table
(
ha_ndbcluster
*
h
,
Ndb
*
ndb
,
const
char
*
path
,
const
char
*
db
,
const
char
*
table_name
);
int
drop_ndb_table
();
int
create_ndb_index
(
const
char
*
name
,
KEY
*
key_info
,
bool
unique
);
int
create_ordered_index
(
const
char
*
name
,
KEY
*
key_info
);
int
create_unique_index
(
const
char
*
name
,
KEY
*
key_info
);
int
create_index
(
const
char
*
name
,
KEY
*
key_info
,
NDB_INDEX_TYPE
idx_type
,
uint
idx_no
);
int
drop_ndb_index
(
const
char
*
name
);
int
table_changed
(
const
void
*
pack_frm_data
,
uint
pack_frm_len
);
// Index list management
int
create_indexes
(
Ndb
*
ndb
,
TABLE
*
tab
);
void
clear_index
(
int
i
);
...
...
@@ -732,7 +729,7 @@ static void set_tabname(const char *pathname, char *tabname);
KEY
*
key_info
,
const
char
*
index_name
,
uint
index_no
);
int
initialize_autoincrement
(
const
void
*
table
);
int
get_metadata
(
const
char
*
path
);
void
release_metadata
();
void
release_metadata
(
THD
*
thd
,
Ndb
*
ndb
);
NDB_INDEX_TYPE
get_index_type
(
uint
idx_no
)
const
;
NDB_INDEX_TYPE
get_index_type_from_table
(
uint
index_no
)
const
;
NDB_INDEX_TYPE
get_index_type_from_key
(
uint
index_no
,
KEY
*
key_info
,
...
...
@@ -795,8 +792,6 @@ static void set_tabname(const char *pathname, char *tabname);
void
print_results
();
ulonglong
get_auto_increment
();
int
invalidate_dictionary_cache
(
bool
global
,
const
NdbDictionary
::
Table
*
ndbtab
);
int
ndb_err
(
NdbTransaction
*
);
bool
uses_blob_value
();
...
...
@@ -834,7 +829,6 @@ static void set_tabname(const char *pathname, char *tabname);
NdbTransaction
*
m_active_trans
;
NdbScanOperation
*
m_active_cursor
;
const
NdbDictionary
::
Table
*
m_table
;
int
m_table_version
;
struct
Ndb_local_table_statistics
*
m_table_info
;
char
m_dbname
[
FN_HEADLEN
];
//char m_schemaname[FN_HEADLEN];
...
...
sql/ha_ndbcluster_binlog.cc
View file @
b96c4289
...
...
@@ -986,7 +986,7 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
uint32
ndb_table_id
,
uint32
ndb_table_version
,
enum
SCHEMA_OP_TYPE
type
,
const
char
*
old_db
,
const
char
*
old
_table_name
)
const
char
*
new_db
,
const
char
*
new
_table_name
)
{
DBUG_ENTER
(
"ndbcluster_log_schema_op"
);
Thd_ndb
*
thd_ndb
=
get_thd_ndb
(
thd
);
...
...
@@ -1026,8 +1026,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
/* redo the rename table query as is may contain several tables */
query
=
tmp_buf2
;
query_length
=
(
uint
)
(
strxmov
(
tmp_buf2
,
"rename table `"
,
old_db
,
"."
,
old_
table_name
,
"` to `"
,
db
,
"."
,
table_name
,
"`"
,
NullS
)
-
tmp_buf2
);
db
,
"."
,
table_name
,
"` to `"
,
new_db
,
"."
,
new_
table_name
,
"`"
,
NullS
)
-
tmp_buf2
);
type_str
=
"rename table"
;
break
;
case
SOT_CREATE_TABLE
:
...
...
@@ -1067,6 +1067,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
Uint64
epoch
=
0
;
MY_BITMAP
schema_subscribers
;
uint32
bitbuf
[
sizeof
(
ndb_schema_object
->
slock
)
/
4
];
uint32
bitbuf_e
[
sizeof
(
bitbuf
)];
bzero
((
char
*
)
bitbuf_e
,
sizeof
(
bitbuf_e
));
{
int
i
,
updated
=
0
;
int
no_storage_nodes
=
g_ndb_cluster_connection
->
no_db_nodes
();
...
...
@@ -1110,7 +1112,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
char
tmp_buf
[
FN_REFLEN
];
NDBDICT
*
dict
=
ndb
->
getDictionary
();
ndb
->
setDatabaseName
(
NDB_REP_DB
);
const
NDBTAB
*
ndbtab
=
dict
->
getTable
(
NDB_SCHEMA_TABLE
);
Ndb_table_guard
ndbtab_g
(
dict
,
NDB_SCHEMA_TABLE
);
const
NDBTAB
*
ndbtab
=
ndbtab_g
.
get_table
();
NdbTransaction
*
trans
=
0
;
int
retries
=
100
;
const
NDBCOL
*
col
[
SCHEMA_SIZE
];
...
...
@@ -1141,8 +1144,13 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
while
(
1
)
{
const
char
*
log_db
=
db
;
const
char
*
log_tab
=
table_name
;
const
char
*
log_subscribers
=
(
char
*
)
schema_subscribers
.
bitmap
;
uint32
log_type
=
(
uint32
)
type
;
if
((
trans
=
ndb
->
startTransaction
())
==
0
)
goto
err
;
while
(
1
)
{
NdbOperation
*
op
=
0
;
int
r
=
0
;
...
...
@@ -1152,17 +1160,17 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
DBUG_ASSERT
(
r
==
0
);
/* db */
ndb_pack_varchar
(
col
[
SCHEMA_DB_I
],
tmp_buf
,
db
,
strlen
(
db
));
ndb_pack_varchar
(
col
[
SCHEMA_DB_I
],
tmp_buf
,
log_db
,
strlen
(
log_
db
));
r
|=
op
->
equal
(
SCHEMA_DB_I
,
tmp_buf
);
DBUG_ASSERT
(
r
==
0
);
/* name */
ndb_pack_varchar
(
col
[
SCHEMA_NAME_I
],
tmp_buf
,
table_name
,
strlen
(
table_name
));
ndb_pack_varchar
(
col
[
SCHEMA_NAME_I
],
tmp_buf
,
log_tab
,
strlen
(
log_tab
));
r
|=
op
->
equal
(
SCHEMA_NAME_I
,
tmp_buf
);
DBUG_ASSERT
(
r
==
0
);
/* slock */
DBUG_ASSERT
(
sz
[
SCHEMA_SLOCK_I
]
==
sizeof
(
bitbuf
));
r
|=
op
->
setValue
(
SCHEMA_SLOCK_I
,
(
char
*
)
schema_subscribers
.
bitmap
);
r
|=
op
->
setValue
(
SCHEMA_SLOCK_I
,
log_subscribers
);
DBUG_ASSERT
(
r
==
0
);
/* query */
{
...
...
@@ -1186,8 +1194,17 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
r
|=
op
->
setValue
(
SCHEMA_VERSION_I
,
ndb_table_version
);
DBUG_ASSERT
(
r
==
0
);
/* type */
r
|=
op
->
setValue
(
SCHEMA_TYPE_I
,
(
uint32
)
type
);
r
|=
op
->
setValue
(
SCHEMA_TYPE_I
,
log_
type
);
DBUG_ASSERT
(
r
==
0
);
if
(
log_db
!=
new_db
&&
new_db
&&
new_table_name
)
{
log_db
=
new_db
;
log_tab
=
new_table_name
;
log_subscribers
=
(
const
char
*
)
bitbuf_e
;
// no ack expected on this
log_type
=
(
uint32
)
SOT_RENAME_TABLE_NEW
;
continue
;
}
break
;
}
if
(
trans
->
execute
(
NdbTransaction
::
Commit
)
==
0
)
{
...
...
@@ -1306,7 +1323,8 @@ ndbcluster_update_slock(THD *thd,
char
tmp_buf
[
FN_REFLEN
];
NDBDICT
*
dict
=
ndb
->
getDictionary
();
ndb
->
setDatabaseName
(
NDB_REP_DB
);
const
NDBTAB
*
ndbtab
=
dict
->
getTable
(
NDB_SCHEMA_TABLE
);
Ndb_table_guard
ndbtab_g
(
dict
,
NDB_SCHEMA_TABLE
);
const
NDBTAB
*
ndbtab
=
ndbtab_g
.
get_table
();
NdbTransaction
*
trans
=
0
;
int
retries
=
100
;
const
NDBCOL
*
col
[
SCHEMA_SIZE
];
...
...
@@ -1452,30 +1470,27 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
{
if
(
pOp
->
tableFrmChanged
())
{
DBUG_PRINT
(
"info"
,
(
"NDBEVENT::TE_ALTER: table frm changed"
));
is_online_alter_table
=
TRUE
;
}
else
{
DBUG_PRINT
(
"info"
,
(
"NDBEVENT::TE_ALTER: name changed"
));
DBUG_ASSERT
(
pOp
->
tableNameChanged
());
is_rename_table
=
TRUE
;
}
}
/*
Refresh local dictionary cache by
invalidating table and all it's indexes
*/
{
ndb
->
setDatabaseName
(
dbname
);
Thd_ndb
*
thd_ndb
=
get_thd_ndb
(
thd
);
DBUG_ASSERT
(
thd_ndb
!=
NULL
);
Ndb
*
old_ndb
=
thd_ndb
->
ndb
;
thd_ndb
->
ndb
=
ndb
;
ha_ndbcluster
table_handler
(
table_share
);
(
void
)
strxmov
(
table_handler
.
m_dbname
,
dbname
,
NullS
);
(
void
)
strxmov
(
table_handler
.
m_tabname
,
tabname
,
NullS
);
table_handler
.
open_indexes
(
ndb
,
table
,
TRUE
);
table_handler
.
invalidate_dictionary_cache
(
TRUE
,
0
);
thd_ndb
->
ndb
=
old_ndb
;
Ndb_table_guard
ndbtab_g
(
ndb
->
getDictionary
(),
tabname
);
const
NDBTAB
*
ev_tab
=
pOp
->
getTable
();
const
NDBTAB
*
cache_tab
=
ndbtab_g
.
get_table
();
if
(
cache_tab
&&
cache_tab
->
getObjectId
()
==
ev_tab
->
getObjectId
()
&&
cache_tab
->
getObjectVersion
()
<=
ev_tab
->
getObjectVersion
())
ndbtab_g
.
invalidate
();
}
/*
Refresh local frm file and dictionary cache if
...
...
@@ -1505,7 +1520,8 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
DBUG_DUMP
(
"frm"
,
(
char
*
)
altered_table
->
getFrmData
(),
altered_table
->
getFrmLength
());
pthread_mutex_lock
(
&
LOCK_open
);
const
NDBTAB
*
old
=
dict
->
getTable
(
tabname
);
Ndb_table_guard
ndbtab_g
(
dict
,
tabname
);
const
NDBTAB
*
old
=
ndbtab_g
.
get_table
();
if
(
!
old
&&
old
->
getObjectVersion
()
!=
altered_table
->
getObjectVersion
())
dict
->
putTable
(
altered_table
);
...
...
@@ -1517,7 +1533,13 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
dbname
,
tabname
,
error
);
}
ndbcluster_binlog_close_table
(
thd
,
share
);
close_cached_tables
((
THD
*
)
0
,
0
,
(
TABLE_LIST
*
)
0
,
TRUE
);
TABLE_LIST
table_list
;
bzero
((
char
*
)
&
table_list
,
sizeof
(
table_list
));
table_list
.
db
=
(
char
*
)
dbname
;
table_list
.
alias
=
table_list
.
table_name
=
(
char
*
)
tabname
;
close_cached_tables
(
thd
,
0
,
&
table_list
,
TRUE
);
if
((
error
=
ndbcluster_binlog_open_table
(
thd
,
share
,
table_share
,
table
)))
sql_print_information
(
"NDB: Failed to re-open table %s.%s"
,
...
...
@@ -1545,26 +1567,22 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
share_prefix
,
share
->
table
->
s
->
db
.
str
,
share
->
table
->
s
->
table_name
.
str
,
share
->
key
);
{
ndb
->
setDatabaseName
(
share
->
table
->
s
->
db
.
str
);
Ndb_table_guard
ndbtab_g
(
ndb
->
getDictionary
(),
share
->
table
->
s
->
table_name
.
str
);
const
NDBTAB
*
ev_tab
=
pOp
->
getTable
();
const
NDBTAB
*
cache_tab
=
ndbtab_g
.
get_table
();
if
(
cache_tab
&&
cache_tab
->
getObjectId
()
==
ev_tab
->
getObjectId
()
&&
cache_tab
->
getObjectVersion
()
<=
ev_tab
->
getObjectVersion
())
ndbtab_g
.
invalidate
();
}
/* do the rename of the table in the share */
share
->
table
->
s
->
db
.
str
=
share
->
db
;
share
->
table
->
s
->
db
.
length
=
strlen
(
share
->
db
);
share
->
table
->
s
->
table_name
.
str
=
share
->
table_name
;
share
->
table
->
s
->
table_name
.
length
=
strlen
(
share
->
table_name
);
/*
Refresh local dictionary cache by invalidating any
old table with same name and all it's indexes
*/
ndb
->
setDatabaseName
(
dbname
);
Thd_ndb
*
thd_ndb
=
get_thd_ndb
(
thd
);
DBUG_ASSERT
(
thd_ndb
!=
NULL
);
Ndb
*
old_ndb
=
thd_ndb
->
ndb
;
thd_ndb
->
ndb
=
ndb
;
ha_ndbcluster
table_handler
(
table_share
);
table_handler
.
set_dbname
(
share
->
key
);
table_handler
.
set_tabname
(
share
->
key
);
table_handler
.
open_indexes
(
ndb
,
table
,
TRUE
);
table_handler
.
invalidate_dictionary_cache
(
TRUE
,
0
);
thd_ndb
->
ndb
=
old_ndb
;
}
DBUG_ASSERT
(
share
->
op
==
pOp
||
share
->
op_old
==
pOp
);
if
(
share
->
op_old
==
pOp
)
...
...
@@ -1582,14 +1600,19 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
if
(
is_remote_change
&&
share
&&
share
->
state
!=
NSS_DROPPED
)
{
DBUG_PRINT
(
"info"
,
(
"remote change"
));
share
->
state
=
NSS_DROPPED
;
if
(
share
->
use_count
!=
1
)
do_close_cached_tables
=
TRUE
;
share
->
state
=
NSS_DROPPED
;
else
{
free_share
(
&
share
,
TRUE
);
share
=
0
;
}
}
else
share
=
0
;
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
share
=
0
;
pOp
->
setCustomData
(
0
);
pthread_mutex_lock
(
&
injector_mutex
);
...
...
@@ -1598,7 +1621,14 @@ ndb_handle_schema_change(THD *thd, Ndb *ndb, NdbEventOperation *pOp,
pthread_mutex_unlock
(
&
injector_mutex
);
if
(
do_close_cached_tables
)
close_cached_tables
((
THD
*
)
0
,
0
,
(
TABLE_LIST
*
)
0
);
{
TABLE_LIST
table_list
;
bzero
((
char
*
)
&
table_list
,
sizeof
(
table_list
));
table_list
.
db
=
(
char
*
)
dbname
;
table_list
.
alias
=
table_list
.
table_name
=
(
char
*
)
tabname
;
close_cached_tables
(
thd
,
0
,
&
table_list
);
free_share
(
&
share
);
}
DBUG_RETURN
(
0
);
}
...
...
@@ -1630,53 +1660,27 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
if
(
schema
->
node_id
!=
node_id
)
{
int
log_query
=
0
,
post_epoch_unlock
=
0
;
DBUG_PRINT
(
"info"
,
(
"log query_length: %d query: '%s'"
,
schema
->
query_length
,
schema
->
query
));
DBUG_PRINT
(
"info"
,
(
"%s.%s: log query_length: %d query: '%s' type: %d"
,
schema
->
db
,
schema
->
name
,
schema
->
query_length
,
schema
->
query
,
schema
->
type
));
char
key
[
FN_REFLEN
];
build_table_filename
(
key
,
sizeof
(
key
),
schema
->
db
,
schema
->
name
,
""
);
NDB_SHARE
*
share
=
get_share
(
key
,
0
,
false
,
false
);
switch
((
enum
SCHEMA_OP_TYPE
)
schema
->
type
)
{
case
SOT_DROP_TABLE
:
/* binlog dropping table after any table operations */
if
(
share
&&
share
->
op
)
{
post_epoch_log_list
->
push_back
(
schema
,
mem_root
);
/* acknowledge this query _after_ epoch completion */
post_epoch_unlock
=
1
;
}
/* table is either ignored or logging is postponed to later */
log_query
=
0
;
break
;
// fall through
case
SOT_RENAME_TABLE
:
if
(
share
&&
share
->
op
)
{
post_epoch_log_list
->
push_back
(
schema
,
mem_root
);
/* acknowledge this query _after_ epoch completion */
post_epoch_unlock
=
1
;
break
;
/* discovery will be handled by binlog */
}
goto
sot_create_table
;
// fall through
case
SOT_RENAME_TABLE_NEW
:
// fall through
case
SOT_ALTER_TABLE
:
if
(
share
&&
share
->
op
)
{
post_epoch_log_list
->
push_back
(
schema
,
mem_root
);
/* acknowledge this query _after_ epoch completion */
post_epoch_unlock
=
1
;
break
;
/* discovery will be handled by binlog */
}
goto
sot_create_table
;
break
;
case
SOT_CREATE_TABLE
:
sot_create_table:
/*
we need to free any share here as command below
may need to call handle_trailing_share
*/
if
(
share
)
{
free_share
(
&
share
);
share
=
0
;
}
pthread_mutex_lock
(
&
LOCK_open
);
if
(
ndb_create_table_from_engine
(
thd
,
schema
->
db
,
schema
->
name
))
{
...
...
@@ -1694,12 +1698,9 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
TRUE
,
/* print error */
TRUE
);
/* don't binlog the query */
/* binlog dropping database after any table operations */
if
(
ndb_binlog_running
)
{
post_epoch_log_list
->
push_back
(
schema
,
mem_root
);
/* acknowledge this query _after_ epoch completion */
post_epoch_unlock
=
1
;
}
break
;
case
SOT_CREATE_DB
:
/* fall through */
...
...
@@ -1726,8 +1727,6 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
pthread_mutex_unlock
(
&
ndb_schema_object
->
mutex
);
pthread_cond_signal
(
&
injector_cond
);
}
if
(
share
)
free_share
(
&
share
,
TRUE
);
pthread_mutex_unlock
(
&
ndbcluster_mutex
);
DBUG_RETURN
(
0
);
}
...
...
@@ -1736,11 +1735,6 @@ ndb_binlog_thread_handle_schema_event(THD *thd, Ndb *ndb,
log_query
=
1
;
break
;
}
if
(
share
)
{
free_share
(
&
share
);
share
=
0
;
}
if
(
log_query
&&
ndb_binlog_running
)
{
char
*
thd_db_save
=
thd
->
db
;
...
...
@@ -1864,26 +1858,69 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
List
<
Cluster_schema
>
*
post_epoch_unlock_list
)
{
if
(
post_epoch_log_list
->
elements
==
0
)
return
;
DBUG_ENTER
(
"ndb_binlog_thread_handle_schema_event_post_epoch"
);
Cluster_schema
*
schema
;
while
((
schema
=
post_epoch_log_list
->
pop
()))
{
DBUG_PRINT
(
"info"
,
(
"log query_length: %d query: '%s'"
,
schema
->
query_length
,
schema
->
query
));
DBUG_PRINT
(
"info"
,
(
"%s.%s: log query_length: %d query: '%s' type: %d"
,
schema
->
db
,
schema
->
name
,
schema
->
query_length
,
schema
->
query
,
schema
->
type
));
int
log_query
=
0
;
{
char
key
[
FN_REFLEN
];
build_table_filename
(
key
,
sizeof
(
key
),
schema
->
db
,
schema
->
name
,
""
);
NDB_SHARE
*
share
=
get_share
(
key
,
0
,
false
,
false
);
switch
((
enum
SCHEMA_OP_TYPE
)
schema
->
type
)
enum
SCHEMA_OP_TYPE
schema_type
=
(
enum
SCHEMA_OP_TYPE
)
schema
->
type
;
switch
(
schema_type
)
{
case
SOT_DROP_DB
:
log_query
=
1
;
break
;
case
SOT_DROP_TABLE
:
// invalidation already handled by binlog thread
if
(
share
&&
share
->
op
)
{
log_query
=
1
;
break
;
}
// fall through
case
SOT_RENAME_TABLE
:
// fall through
case
SOT_ALTER_TABLE
:
if
(
share
&&
share
->
op
)
// invalidation already handled by binlog thread
if
(
!
share
||
!
share
->
op
)
{
{
injector_ndb
->
setDatabaseName
(
schema
->
db
);
Ndb_table_guard
ndbtab_g
(
injector_ndb
->
getDictionary
(),
schema
->
name
);
ndbtab_g
.
invalidate
();
}
TABLE_LIST
table_list
;
bzero
((
char
*
)
&
table_list
,
sizeof
(
table_list
));
table_list
.
db
=
schema
->
db
;
table_list
.
alias
=
table_list
.
table_name
=
schema
->
name
;
close_cached_tables
(
thd
,
0
,
&
table_list
,
FALSE
);
}
if
(
schema_type
!=
SOT_ALTER_TABLE
)
break
;
// fall through
case
SOT_RENAME_TABLE_NEW
:
log_query
=
1
;
if
(
ndb_binlog_running
)
{
/*
we need to free any share here as command below
may need to call handle_trailing_share
*/
if
(
share
)
{
break
;
/* discovery handled by binlog */
free_share
(
&
share
);
share
=
0
;
}
pthread_mutex_lock
(
&
LOCK_open
);
if
(
ndb_create_table_from_engine
(
thd
,
schema
->
db
,
schema
->
name
))
...
...
@@ -1894,6 +1931,8 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
schema
->
node_id
);
}
pthread_mutex_unlock
(
&
LOCK_open
);
}
break
;
default:
DBUG_ASSERT
(
false
);
}
...
...
@@ -1903,6 +1942,7 @@ ndb_binlog_thread_handle_schema_event_post_epoch(THD *thd,
share
=
0
;
}
}
if
(
ndb_binlog_running
&&
log_query
)
{
char
*
thd_db_save
=
thd
->
db
;
thd
->
db
=
schema
->
db
;
...
...
@@ -2186,7 +2226,8 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key,
ndb
->
setDatabaseName
(
db
);
NDBDICT
*
dict
=
ndb
->
getDictionary
();
const
NDBTAB
*
ndbtab
=
dict
->
getTable
(
table_name
);
Ndb_table_guard
ndbtab_g
(
dict
,
table_name
);
const
NDBTAB
*
ndbtab
=
ndbtab_g
.
get_table
();
if
(
ndbtab
==
0
)
{
if
(
ndb_extra_logging
)
...
...
@@ -2201,7 +2242,8 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key,
event should have been created by someone else,
but let's make sure, and create if it doesn't exist
*/
if
(
!
dict
->
getEvent
(
event_name
.
c_ptr
()))
const
NDBEVENT
*
ev
=
dict
->
getEvent
(
event_name
.
c_ptr
());
if
(
!
ev
)
{
if
(
ndbcluster_create_event
(
ndb
,
ndbtab
,
event_name
.
c_ptr
(),
share
))
{
...
...
@@ -2216,9 +2258,12 @@ int ndbcluster_create_binlog_setup(Ndb *ndb, const char *key,
event_name
.
c_ptr
());
}
else
{
delete
ev
;
if
(
ndb_extra_logging
)
sql_print_information
(
"NDB Binlog: DISCOVER TABLE Event: %s"
,
event_name
.
c_ptr
());
}
/*
create the event operations for receiving logging events
...
...
@@ -2328,8 +2373,10 @@ ndbcluster_create_event(Ndb *ndb, const NDBTAB *ndbtab,
try retrieving the event, if table version/id matches, we will get
a valid event. Otherwise we have a trailing event from before
*/
if
(
dict
->
getEvent
(
event_name
))
const
NDBEVENT
*
ev
;
if
((
ev
=
dict
->
getEvent
(
event_name
)))
{
delete
ev
;
DBUG_RETURN
(
0
);
}
...
...
sql/ha_ndbcluster_binlog.h
View file @
b96c4289
...
...
@@ -41,14 +41,15 @@ enum SCHEMA_OP_TYPE
{
SOT_DROP_TABLE
=
0
,
SOT_CREATE_TABLE
=
1
,
SOT_RENAME_TABLE
=
2
,
SOT_RENAME_TABLE
_NEW
=
2
,
SOT_ALTER_TABLE
=
3
,
SOT_DROP_DB
=
4
,
SOT_CREATE_DB
=
5
,
SOT_ALTER_DB
=
6
,
SOT_CLEAR_SLOCK
=
7
,
SOT_TABLESPACE
=
8
,
SOT_LOGFILE_GROUP
=
9
SOT_LOGFILE_GROUP
=
9
,
SOT_RENAME_TABLE
=
10
};
const
uint
max_ndb_nodes
=
64
;
/* multiple of 32 */
...
...
@@ -56,6 +57,45 @@ const uint max_ndb_nodes= 64; /* multiple of 32 */
static
const
char
*
ha_ndb_ext
=
".ndb"
;
static
const
char
share_prefix
[]
=
"./"
;
class
Ndb_table_guard
{
public:
Ndb_table_guard
(
NDBDICT
*
dict
,
const
char
*
tabname
)
:
m_dict
(
dict
)
{
DBUG_ENTER
(
"Ndb_table_guard"
);
m_ndbtab
=
m_dict
->
getTableGlobal
(
tabname
);
m_invalidate
=
0
;
DBUG_PRINT
(
"info"
,
(
"m_ndbtab: %p"
,
m_ndbtab
));
DBUG_VOID_RETURN
;
}
~
Ndb_table_guard
()
{
DBUG_ENTER
(
"~Ndb_table_guard"
);
if
(
m_ndbtab
)
{
DBUG_PRINT
(
"info"
,
(
"m_ndbtab: %p m_invalidate: %d"
,
m_ndbtab
,
m_invalidate
));
m_dict
->
removeTableGlobal
(
*
m_ndbtab
,
m_invalidate
);
}
DBUG_VOID_RETURN
;
}
const
NDBTAB
*
get_table
()
{
return
m_ndbtab
;
}
void
invalidate
()
{
m_invalidate
=
1
;
}
const
NDBTAB
*
release
()
{
DBUG_ENTER
(
"Ndb_table_guard::release"
);
const
NDBTAB
*
tmp
=
m_ndbtab
;
DBUG_PRINT
(
"info"
,
(
"m_ndbtab: %p"
,
m_ndbtab
));
m_ndbtab
=
0
;
DBUG_RETURN
(
tmp
);
}
private:
const
NDBTAB
*
m_ndbtab
;
NDBDICT
*
m_dict
;
int
m_invalidate
;
};
#ifdef HAVE_NDB_BINLOG
extern
pthread_t
ndb_binlog_thread
;
extern
pthread_mutex_t
injector_mutex
;
...
...
@@ -98,8 +138,8 @@ int ndbcluster_log_schema_op(THD *thd, NDB_SHARE *share,
uint32
ndb_table_id
,
uint32
ndb_table_version
,
enum
SCHEMA_OP_TYPE
type
,
const
char
*
old
_db
=
0
,
const
char
*
old
_table_name
=
0
);
const
char
*
new
_db
=
0
,
const
char
*
new
_table_name
=
0
);
int
ndbcluster_handle_drop_table
(
Ndb
*
ndb
,
const
char
*
event_name
,
NDB_SHARE
*
share
,
const
char
*
type_str
);
...
...
storage/ndb/include/ndbapi/Ndb.hpp
View file @
b96c4289
...
...
@@ -1470,6 +1470,8 @@ public:
*
* @return tuple id or 0 on error
*/
int
initAutoIncrement
();
Uint64
getAutoIncrementValue
(
const
char
*
aTableName
,
Uint32
cacheSize
=
1
);
Uint64
getAutoIncrementValue
(
const
NdbDictionary
::
Table
*
aTable
,
...
...
@@ -1694,6 +1696,7 @@ private:
// The tupleId is retreived from DB the
// tupleId is unique for each tableid.
const
NdbDictionary
::
Table
*
m_sys_tab_0
;
Uint64
theFirstTupleId
[
2048
];
Uint64
theLastTupleId
[
2048
];
...
...
storage/ndb/include/ndbapi/NdbDictionary.hpp
View file @
b96c4289
...
...
@@ -798,6 +798,7 @@ public:
* Get object status
*/
virtual
Object
::
Status
getObjectStatus
()
const
;
void
setStatusInvalid
()
const
;
/**
* Get object version
...
...
@@ -1734,6 +1735,7 @@ public:
* @return 0 if successful otherwise -1.
*/
int
createIndex
(
const
Index
&
index
);
int
createIndex
(
const
Index
&
index
,
const
Table
&
table
);
/**
* Drop index with given name
...
...
@@ -1805,6 +1807,15 @@ public:
#ifndef DOXYGEN_SHOULD_SKIP_INTERNAL
const
Table
*
getTable
(
const
char
*
name
,
void
**
data
)
const
;
void
set_local_table_data_size
(
unsigned
sz
);
const
Index
*
getIndexGlobal
(
const
char
*
indexName
,
const
Table
&
ndbtab
)
const
;
const
Table
*
getTableGlobal
(
const
char
*
tableName
)
const
;
int
alterTableGlobal
(
const
Table
&
f
,
const
Table
&
t
);
int
dropTableGlobal
(
const
Table
&
ndbtab
);
int
dropIndexGlobal
(
const
Index
&
index
);
int
removeIndexGlobal
(
const
Index
&
ndbidx
,
int
invalidate
)
const
;
int
removeTableGlobal
(
const
Table
&
ndbtab
,
int
invalidate
)
const
;
#endif
};
};
...
...
storage/ndb/include/ndbapi/NdbIndexStat.hpp
View file @
b96c4289
...
...
@@ -56,7 +56,7 @@ public:
* multiplied by a percentage obtained from the cache (result zero is
* returned as 1).
*/
int
records_in_range
(
NdbDictionary
::
Index
*
index
,
int
records_in_range
(
const
NdbDictionary
::
Index
*
index
,
NdbIndexScanOperation
*
op
,
Uint64
table_rows
,
Uint64
*
count
,
...
...
storage/ndb/src/ndbapi/DictCache.cpp
View file @
b96c4289
...
...
@@ -63,6 +63,7 @@ LocalDictCache::~LocalDictCache(){
Ndb_local_table_info
*
LocalDictCache
::
get
(
const
char
*
name
){
ASSERT_NOT_MYSQLD
;
assert
(
!
is_ndb_blob_table
(
name
));
const
Uint32
len
=
strlen
(
name
);
return
m_tableHash
.
getData
(
name
,
len
);
...
...
@@ -70,6 +71,7 @@ LocalDictCache::get(const char * name){
void
LocalDictCache
::
put
(
const
char
*
name
,
Ndb_local_table_info
*
tab_info
){
ASSERT_NOT_MYSQLD
;
assert
(
!
is_ndb_blob_table
(
name
));
const
Uint32
id
=
tab_info
->
m_table_impl
->
m_id
;
m_tableHash
.
insertKey
(
name
,
strlen
(
name
),
id
,
tab_info
);
...
...
@@ -77,6 +79,7 @@ LocalDictCache::put(const char * name, Ndb_local_table_info * tab_info){
void
LocalDictCache
::
drop
(
const
char
*
name
){
ASSERT_NOT_MYSQLD
;
assert
(
!
is_ndb_blob_table
(
name
));
Ndb_local_table_info
*
info
=
m_tableHash
.
deleteKey
(
name
,
strlen
(
name
));
DBUG_ASSERT
(
info
!=
0
);
...
...
@@ -100,9 +103,16 @@ GlobalDictCache::~GlobalDictCache(){
Vector
<
TableVersion
>
*
vers
=
curr
->
theData
;
const
unsigned
sz
=
vers
->
size
();
for
(
unsigned
i
=
0
;
i
<
sz
;
i
++
){
if
((
*
vers
)[
i
].
m_impl
!=
0
)
TableVersion
tv
=
(
*
vers
)[
i
];
DBUG_PRINT
(
" "
,
(
"vers[%d]: ver: %d, refCount: %d, status: %d"
,
i
,
tv
.
m_version
,
tv
.
m_refCount
,
tv
.
m_status
));
if
(
tv
.
m_impl
!=
0
)
{
DBUG_PRINT
(
" "
,
(
"m_impl: internalname: %s"
,
tv
.
m_impl
->
m_internalName
.
c_str
()));
delete
(
*
vers
)[
i
].
m_impl
;
}
}
delete
curr
->
theData
;
curr
->
theData
=
NULL
;
curr
=
m_tableHash
.
getNext
(
curr
);
...
...
@@ -164,8 +174,15 @@ GlobalDictCache::get(const char * name)
TableVersion
*
ver
=
&
versions
->
back
();
switch
(
ver
->
m_status
){
case
OK
:
if
(
ver
->
m_impl
->
m_status
==
NdbDictionary
::
Object
::
Invalid
)
{
ver
->
m_status
=
DROPPED
;
retreive
=
true
;
// Break loop
break
;
}
ver
->
m_refCount
++
;
DBUG_PRINT
(
"info"
,
(
"Table OK version=%x.%x refCount=%u"
,
DBUG_PRINT
(
"info"
,
(
"Table OK tab: %p version=%x.%x refCount=%u"
,
ver
->
m_impl
,
ver
->
m_impl
->
m_version
&
0xFFFFFF
,
ver
->
m_impl
->
m_version
>>
24
,
ver
->
m_refCount
));
...
...
@@ -197,8 +214,8 @@ NdbTableImpl *
GlobalDictCache
::
put
(
const
char
*
name
,
NdbTableImpl
*
tab
)
{
DBUG_ENTER
(
"GlobalDictCache::put"
);
DBUG_PRINT
(
"enter"
,
(
"name: %s, internal_name: %s version: %x.%x"
,
name
,
DBUG_PRINT
(
"enter"
,
(
"
tab: %p
name: %s, internal_name: %s version: %x.%x"
,
tab
,
name
,
tab
?
tab
->
m_internalName
.
c_str
()
:
"tab NULL"
,
tab
?
tab
->
m_version
&
0xFFFFFF
:
0
,
tab
?
tab
->
m_version
>>
24
:
0
));
...
...
@@ -264,10 +281,11 @@ GlobalDictCache::put(const char * name, NdbTableImpl * tab)
}
void
GlobalDictCache
::
drop
(
NdbTableImpl
*
tab
)
GlobalDictCache
::
release
(
NdbTableImpl
*
tab
,
int
invalidate
)
{
DBUG_ENTER
(
"GlobalDictCache::drop"
);
DBUG_PRINT
(
"enter"
,
(
"internal_name: %s"
,
tab
->
m_internalName
.
c_str
()));
DBUG_ENTER
(
"GlobalDictCache::release"
);
DBUG_PRINT
(
"enter"
,
(
"tab: %p internal_name: %s"
,
tab
,
tab
->
m_internalName
.
c_str
()));
assert
(
!
is_ndb_blob_table
(
tab
));
unsigned
i
;
...
...
@@ -292,15 +310,19 @@ GlobalDictCache::drop(NdbTableImpl * tab)
if
(
ver
.
m_impl
==
tab
){
if
(
ver
.
m_refCount
==
0
||
ver
.
m_status
==
RETREIVING
||
ver
.
m_version
!=
tab
->
m_version
){
DBUG_PRINT
(
"info"
,
(
"
Dropp
ing with refCount=%d status=%d impl=%p"
,
DBUG_PRINT
(
"info"
,
(
"
Releas
ing with refCount=%d status=%d impl=%p"
,
ver
.
m_refCount
,
ver
.
m_status
,
ver
.
m_impl
));
break
;
}
DBUG_PRINT
(
"info"
,
(
"Found table to drop, i: %d, name: %s"
,
i
,
ver
.
m_impl
->
m_internalName
.
c_str
()));
ver
.
m_refCount
--
;
if
(
ver
.
m_impl
->
m_status
==
NdbDictionary
::
Object
::
Invalid
||
invalidate
)
{
ver
.
m_impl
->
m_status
=
NdbDictionary
::
Object
::
Invalid
;
ver
.
m_status
=
DROPPED
;
if
(
ver
.
m_refCount
==
0
){
}
if
(
ver
.
m_refCount
==
0
&&
ver
.
m_status
==
DROPPED
)
{
DBUG_PRINT
(
"info"
,
(
"refCount is zero, deleting m_impl"
));
delete
ver
.
m_impl
;
vers
->
erase
(
i
);
...
...
@@ -319,61 +341,13 @@ GlobalDictCache::drop(NdbTableImpl * tab)
abort
();
}
void
GlobalDictCache
::
release
(
NdbTableImpl
*
tab
)
{
DBUG_ENTER
(
"GlobalDictCache::release"
);
DBUG_PRINT
(
"enter"
,
(
"internal_name: %s"
,
tab
->
m_internalName
.
c_str
()));
assert
(
!
is_ndb_blob_table
(
tab
));
unsigned
i
;
const
Uint32
len
=
strlen
(
tab
->
m_internalName
.
c_str
());
Vector
<
TableVersion
>
*
vers
=
m_tableHash
.
getData
(
tab
->
m_internalName
.
c_str
(),
len
);
if
(
vers
==
0
){
// Should always tried to retreive it first
// and thus there should be a record
abort
();
}
const
Uint32
sz
=
vers
->
size
();
if
(
sz
==
0
){
// Should always tried to retreive it first
// and thus there should be a record
abort
();
}
for
(
i
=
0
;
i
<
sz
;
i
++
){
TableVersion
&
ver
=
(
*
vers
)[
i
];
if
(
ver
.
m_impl
==
tab
){
if
(
ver
.
m_refCount
==
0
||
ver
.
m_status
==
RETREIVING
||
ver
.
m_version
!=
tab
->
m_version
){
DBUG_PRINT
(
"info"
,
(
"Releasing with refCount=%d status=%d impl=%p"
,
ver
.
m_refCount
,
ver
.
m_status
,
ver
.
m_impl
));
break
;
}
ver
.
m_refCount
--
;
DBUG_VOID_RETURN
;
}
}
for
(
i
=
0
;
i
<
sz
;
i
++
){
TableVersion
&
ver
=
(
*
vers
)[
i
];
ndbout_c
(
"%d: version: %d refCount: %d status: %d impl: %p"
,
i
,
ver
.
m_version
,
ver
.
m_refCount
,
ver
.
m_status
,
ver
.
m_impl
);
}
abort
();
}
void
GlobalDictCache
::
alter_table_rep
(
const
char
*
name
,
Uint32
tableId
,
Uint32
tableVersion
,
bool
altered
)
{
DBUG_ENTER
(
"GlobalDictCache::alter_table_rep"
);
assert
(
!
is_ndb_blob_table
(
name
));
const
Uint32
len
=
strlen
(
name
);
Vector
<
TableVersion
>
*
vers
=
...
...
@@ -381,13 +355,13 @@ GlobalDictCache::alter_table_rep(const char * name,
if
(
vers
==
0
)
{
return
;
DBUG_VOID_RETURN
;
}
const
Uint32
sz
=
vers
->
size
();
if
(
sz
==
0
)
{
return
;
DBUG_VOID_RETURN
;
}
for
(
Uint32
i
=
0
;
i
<
sz
;
i
++
)
...
...
@@ -399,15 +373,16 @@ GlobalDictCache::alter_table_rep(const char * name,
ver
.
m_status
=
DROPPED
;
ver
.
m_impl
->
m_status
=
altered
?
NdbDictionary
::
Object
::
Altered
:
NdbDictionary
::
Object
::
Invalid
;
return
;
DBUG_VOID_RETURN
;
}
if
(
i
==
sz
-
1
&&
ver
.
m_status
==
RETREIVING
)
{
ver
.
m_impl
=
altered
?
&
f_altered_table
:
&
f_invalid_table
;
return
;
DBUG_VOID_RETURN
;
}
}
DBUG_VOID_RETURN
;
}
template
class
Vector
<
GlobalDictCache
::
TableVersion
>;
storage/ndb/src/ndbapi/DictCache.hpp
View file @
b96c4289
...
...
@@ -63,11 +63,11 @@ public:
GlobalDictCache
();
~
GlobalDictCache
();
NdbTableImpl
*
get
(
NdbTableImpl
*
tab
);
NdbTableImpl
*
get
(
const
char
*
name
);
NdbTableImpl
*
put
(
const
char
*
name
,
NdbTableImpl
*
);
void
drop
(
NdbTableImpl
*
);
void
release
(
NdbTableImpl
*
);
void
release
(
NdbTableImpl
*
,
int
invalidate
=
0
);
void
alter_table_rep
(
const
char
*
name
,
Uint32
tableId
,
Uint32
tableVersion
,
bool
altered
);
...
...
storage/ndb/src/ndbapi/Ndb.cpp
View file @
b96c4289
...
...
@@ -901,6 +901,27 @@ Ndb::setTupleIdInNdb(Uint32 aTableId, Uint64 val, bool increase )
DBUG_RETURN
((
opTupleIdOnNdb
(
aTableId
,
val
,
1
)
==
val
));
}
int
Ndb
::
initAutoIncrement
()
{
if
(
m_sys_tab_0
)
return
0
;
BaseString
currentDb
(
getDatabaseName
());
BaseString
currentSchema
(
getDatabaseSchemaName
());
setDatabaseName
(
"sys"
);
setDatabaseSchemaName
(
"def"
);
m_sys_tab_0
=
getDictionary
()
->
getTableGlobal
(
"SYSTAB_0"
);
// Restore current name space
setDatabaseName
(
currentDb
.
c_str
());
setDatabaseSchemaName
(
currentSchema
.
c_str
());
return
(
m_sys_tab_0
==
NULL
);
}
Uint64
Ndb
::
opTupleIdOnNdb
(
Uint32
aTableId
,
Uint64
opValue
,
Uint32
op
)
{
...
...
@@ -916,19 +937,14 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op)
CHECK_STATUS_MACRO_ZERO
;
BaseString
currentDb
(
getDatabaseName
());
BaseString
currentSchema
(
getDatabaseSchemaName
())
;
if
(
initAutoIncrement
())
goto
error_return
;
setDatabaseName
(
"sys"
);
setDatabaseSchemaName
(
"def"
);
tConnection
=
this
->
startTransaction
();
if
(
tConnection
==
NULL
)
goto
error_return
;
if
(
usingFullyQualifiedNames
())
tOperation
=
tConnection
->
getNdbOperation
(
"SYSTAB_0"
);
else
tOperation
=
tConnection
->
getNdbOperation
(
"sys/def/SYSTAB_0"
);
tOperation
=
tConnection
->
getNdbOperation
(
m_sys_tab_0
);
if
(
tOperation
==
NULL
)
goto
error_handler
;
...
...
@@ -997,20 +1013,12 @@ Ndb::opTupleIdOnNdb(Uint32 aTableId, Uint64 opValue, Uint32 op)
this
->
closeTransaction
(
tConnection
);
// Restore current name space
setDatabaseName
(
currentDb
.
c_str
());
setDatabaseSchemaName
(
currentSchema
.
c_str
());
DBUG_RETURN
(
ret
);
error_handler:
theError
.
code
=
tConnection
->
theError
.
code
;
this
->
closeTransaction
(
tConnection
);
error_return:
// Restore current name space
setDatabaseName
(
currentDb
.
c_str
());
setDatabaseSchemaName
(
currentSchema
.
c_str
());
DBUG_PRINT
(
"error"
,
(
"ndb=%d con=%d op=%d"
,
theError
.
code
,
tConnection
?
tConnection
->
theError
.
code
:
-
1
,
...
...
storage/ndb/src/ndbapi/NdbDictionary.cpp
View file @
b96c4289
...
...
@@ -559,6 +559,11 @@ NdbDictionary::Table::getObjectStatus() const {
return
m_impl
.
m_status
;
}
void
NdbDictionary
::
Table
::
setStatusInvalid
()
const
{
m_impl
.
m_status
=
NdbDictionary
::
Object
::
Invalid
;
}
int
NdbDictionary
::
Table
::
getObjectVersion
()
const
{
return
m_impl
.
m_version
;
...
...
@@ -1330,6 +1335,11 @@ NdbDictionary::Dictionary::dropTable(Table & t){
return
m_impl
.
dropTable
(
NdbTableImpl
::
getImpl
(
t
));
}
int
NdbDictionary
::
Dictionary
::
dropTableGlobal
(
const
Table
&
t
){
return
m_impl
.
dropTableGlobal
(
NdbTableImpl
::
getImpl
(
t
));
}
int
NdbDictionary
::
Dictionary
::
dropTable
(
const
char
*
name
){
return
m_impl
.
dropTable
(
name
);
...
...
@@ -1340,6 +1350,14 @@ NdbDictionary::Dictionary::alterTable(const Table & t){
return
m_impl
.
alterTable
(
NdbTableImpl
::
getImpl
(
t
));
}
int
NdbDictionary
::
Dictionary
::
alterTableGlobal
(
const
Table
&
f
,
const
Table
&
t
)
{
return
m_impl
.
alterTableGlobal
(
NdbTableImpl
::
getImpl
(
f
),
NdbTableImpl
::
getImpl
(
t
));
}
const
NdbDictionary
::
Table
*
NdbDictionary
::
Dictionary
::
getTable
(
const
char
*
name
,
void
**
data
)
const
{
...
...
@@ -1349,6 +1367,40 @@ NdbDictionary::Dictionary::getTable(const char * name, void **data) const
return
0
;
}
const
NdbDictionary
::
Index
*
NdbDictionary
::
Dictionary
::
getIndexGlobal
(
const
char
*
indexName
,
const
Table
&
ndbtab
)
const
{
NdbIndexImpl
*
i
=
m_impl
.
getIndexGlobal
(
indexName
,
NdbTableImpl
::
getImpl
(
ndbtab
));
if
(
i
)
return
i
->
m_facade
;
return
0
;
}
const
NdbDictionary
::
Table
*
NdbDictionary
::
Dictionary
::
getTableGlobal
(
const
char
*
name
)
const
{
NdbTableImpl
*
t
=
m_impl
.
getTableGlobal
(
name
);
if
(
t
)
return
t
->
m_facade
;
return
0
;
}
int
NdbDictionary
::
Dictionary
::
removeIndexGlobal
(
const
Index
&
ndbidx
,
int
invalidate
)
const
{
return
m_impl
.
releaseIndexGlobal
(
NdbIndexImpl
::
getImpl
(
ndbidx
),
invalidate
);
}
int
NdbDictionary
::
Dictionary
::
removeTableGlobal
(
const
Table
&
ndbtab
,
int
invalidate
)
const
{
return
m_impl
.
releaseTableGlobal
(
NdbTableImpl
::
getImpl
(
ndbtab
),
invalidate
);
}
void
NdbDictionary
::
Dictionary
::
putTable
(
const
NdbDictionary
::
Table
*
table
)
{
NdbDictionary
::
Table
*
copy_table
=
new
NdbDictionary
::
Table
;
...
...
@@ -1420,6 +1472,13 @@ NdbDictionary::Dictionary::createIndex(const Index & ind)
return
m_impl
.
createIndex
(
NdbIndexImpl
::
getImpl
(
ind
));
}
int
NdbDictionary
::
Dictionary
::
createIndex
(
const
Index
&
ind
,
const
Table
&
tab
)
{
return
m_impl
.
createIndex
(
NdbIndexImpl
::
getImpl
(
ind
),
NdbTableImpl
::
getImpl
(
tab
));
}
int
NdbDictionary
::
Dictionary
::
dropIndex
(
const
char
*
indexName
,
const
char
*
tableName
)
...
...
@@ -1427,6 +1486,12 @@ NdbDictionary::Dictionary::dropIndex(const char * indexName,
return
m_impl
.
dropIndex
(
indexName
,
tableName
);
}
int
NdbDictionary
::
Dictionary
::
dropIndexGlobal
(
const
Index
&
ind
)
{
return
m_impl
.
dropIndexGlobal
(
NdbIndexImpl
::
getImpl
(
ind
));
}
const
NdbDictionary
::
Index
*
NdbDictionary
::
Dictionary
::
getIndex
(
const
char
*
indexName
,
const
char
*
tableName
)
const
...
...
storage/ndb/src/ndbapi/NdbDictionaryImpl.cpp
View file @
b96c4289
...
...
@@ -50,7 +50,14 @@
#define DICT_WAITFOR_TIMEOUT (7*24*60*60*1000)
#define ERR_RETURN(a,b) \
{\
DBUG_PRINT("exit", ("error %d", (a).code));\
DBUG_RETURN(b);\
}
extern
Uint64
g_latest_trans_gci
;
int
ndb_dictionary_is_mysqld
=
0
;
bool
is_ndb_blob_table
(
const
char
*
name
,
Uint32
*
ptab_id
,
Uint32
*
pcol_no
)
...
...
@@ -1288,44 +1295,30 @@ NdbDictionaryImpl::~NdbDictionaryImpl()
}
}
Ndb
_local_table_info
*
NdbDictionaryImpl
::
fetchGlobalTableImpl
(
const
BaseString
&
internalTableName
)
Ndb
TableImpl
*
NdbDictionaryImpl
::
fetchGlobalTableImpl
Ref
(
const
GlobalCacheInitObject
&
obj
)
{
DBUG_ENTER
(
"fetchGlobalTableImplRef"
);
NdbTableImpl
*
impl
;
m_globalHash
->
lock
();
impl
=
m_globalHash
->
get
(
internalTableN
ame
.
c_str
());
impl
=
m_globalHash
->
get
(
obj
.
m_n
ame
.
c_str
());
m_globalHash
->
unlock
();
if
(
impl
==
0
){
impl
=
m_receiver
.
getTable
(
internalTableName
,
impl
=
m_receiver
.
getTable
(
obj
.
m_name
.
c_str
()
,
m_ndb
.
usingFullyQualifiedNames
());
if
(
impl
!=
0
)
{
int
ret
=
getBlobTables
(
*
impl
);
if
(
ret
!=
0
)
{
if
(
impl
!=
0
&&
obj
.
init
(
*
impl
))
{
delete
impl
;
impl
=
0
;
}
}
m_globalHash
->
lock
();
m_globalHash
->
put
(
internalTableN
ame
.
c_str
(),
impl
);
m_globalHash
->
put
(
obj
.
m_n
ame
.
c_str
(),
impl
);
m_globalHash
->
unlock
();
if
(
impl
==
0
){
return
0
;
}
}
Ndb_local_table_info
*
info
=
Ndb_local_table_info
::
create
(
impl
,
m_local_table_data_size
);
m_localHash
.
put
(
internalTableName
.
c_str
(),
info
);
m_ndb
.
theFirstTupleId
[
impl
->
getTableId
()]
=
~
0
;
m_ndb
.
theLastTupleId
[
impl
->
getTableId
()]
=
~
0
;
return
info
;
DBUG_RETURN
(
impl
);
}
void
...
...
@@ -2276,18 +2269,30 @@ int NdbDictionaryImpl::alterTable(NdbTableImpl &impl)
}
// Alter the table
int
ret
=
m_receiver
.
alterTable
(
m_ndb
,
impl
);
if
(
ret
==
0
)
{
// Remove cached information and let it be refreshed at next access
int
ret
=
alterTableGlobal
(
*
local
->
m_table_impl
,
impl
);
if
(
ret
==
0
)
{
m_globalHash
->
lock
();
local
->
m_table_impl
->
m_status
=
NdbDictionary
::
Object
::
Invalid
;
m_globalHash
->
drop
(
local
->
m_table_impl
);
m_globalHash
->
release
(
local
->
m_table_impl
,
1
);
m_globalHash
->
unlock
();
m_localHash
.
drop
(
originalInternalName
);
}
DBUG_RETURN
(
ret
);
}
int
NdbDictionaryImpl
::
alterTableGlobal
(
NdbTableImpl
&
old_impl
,
NdbTableImpl
&
impl
)
{
DBUG_ENTER
(
"NdbDictionaryImpl::alterTableGlobal"
);
// Alter the table
int
ret
=
m_receiver
.
alterTable
(
m_ndb
,
impl
);
old_impl
.
m_status
=
NdbDictionary
::
Object
::
Invalid
;
if
(
ret
==
0
){
DBUG_RETURN
(
ret
);
}
ERR_RETURN
(
getNdbError
(),
ret
);
}
int
NdbDictInterface
::
alterTable
(
Ndb
&
ndb
,
NdbTableImpl
&
impl
)
...
...
@@ -2731,6 +2736,7 @@ NdbDictionaryImpl::dropTable(const char * name)
{
DBUG_ENTER
(
"NdbDictionaryImpl::dropTable"
);
DBUG_PRINT
(
"enter"
,(
"name: %s"
,
name
));
ASSERT_NOT_MYSQLD
;
NdbTableImpl
*
tab
=
getTable
(
name
);
if
(
tab
==
0
){
DBUG_RETURN
(
-
1
);
...
...
@@ -2743,8 +2749,7 @@ NdbDictionaryImpl::dropTable(const char * name)
DBUG_PRINT
(
"info"
,(
"INCOMPATIBLE_VERSION internal_name: %s"
,
internalTableName
.
c_str
()));
m_localHash
.
drop
(
internalTableName
.
c_str
());
m_globalHash
->
lock
();
tab
->
m_status
=
NdbDictionary
::
Object
::
Invalid
;
m_globalHash
->
drop
(
tab
);
m_globalHash
->
release
(
tab
,
1
);
m_globalHash
->
unlock
();
DBUG_RETURN
(
dropTable
(
name
));
}
...
...
@@ -2792,8 +2797,7 @@ NdbDictionaryImpl::dropTable(NdbTableImpl & impl)
m_localHash
.
drop
(
internalTableName
);
m_globalHash
->
lock
();
impl
.
m_status
=
NdbDictionary
::
Object
::
Invalid
;
m_globalHash
->
drop
(
&
impl
);
m_globalHash
->
release
(
&
impl
,
1
);
m_globalHash
->
unlock
();
return
0
;
...
...
@@ -2802,6 +2806,50 @@ NdbDictionaryImpl::dropTable(NdbTableImpl & impl)
return
ret
;
}
int
NdbDictionaryImpl
::
dropTableGlobal
(
NdbTableImpl
&
impl
)
{
int
res
;
const
char
*
name
=
impl
.
getName
();
DBUG_ENTER
(
"NdbDictionaryImpl::dropTableGlobal"
);
DBUG_ASSERT
(
impl
.
m_status
!=
NdbDictionary
::
Object
::
New
);
DBUG_ASSERT
(
impl
.
m_indexType
==
NdbDictionary
::
Object
::
TypeUndefined
);
List
list
;
if
((
res
=
listIndexes
(
list
,
impl
.
m_id
))
==
-
1
){
ERR_RETURN
(
getNdbError
(),
-
1
);
}
for
(
unsigned
i
=
0
;
i
<
list
.
count
;
i
++
)
{
const
List
::
Element
&
element
=
list
.
elements
[
i
];
NdbIndexImpl
*
idx
=
getIndexGlobal
(
element
.
name
,
impl
);
if
(
idx
==
NULL
)
{
ERR_RETURN
(
getNdbError
(),
-
1
);
}
if
((
res
=
dropIndexGlobal
(
*
idx
))
==
-
1
)
{
releaseIndexGlobal
(
*
idx
,
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
releaseIndexGlobal
(
*
idx
,
1
);
}
if
(
impl
.
m_noOfBlobs
!=
0
)
{
if
(
dropBlobTables
(
impl
)
!=
0
){
ERR_RETURN
(
getNdbError
(),
-
1
);
}
}
int
ret
=
m_receiver
.
dropTable
(
impl
);
impl
.
m_status
=
NdbDictionary
::
Object
::
Invalid
;
if
(
ret
==
0
||
m_error
.
code
==
709
||
m_error
.
code
==
723
)
{
DBUG_RETURN
(
0
);
}
ERR_RETURN
(
getNdbError
(),
ret
);
}
int
NdbDictionaryImpl
::
dropBlobTables
(
NdbTableImpl
&
t
)
{
...
...
@@ -2822,7 +2870,7 @@ NdbDictionaryImpl::dropBlobTables(NdbTableImpl & t)
DBUG_PRINT
(
"info"
,
(
"col %s: blob table %s: error %d"
,
c
.
m_name
.
c_str
(),
bt
->
m_internalName
.
c_str
(),
m_error
.
code
));
if
(
!
(
ret
==
709
||
ret
==
723
))
// "force" mode on
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
// leave c.m_blobTable defined
}
...
...
@@ -2891,8 +2939,7 @@ NdbDictionaryImpl::invalidateObject(NdbTableImpl & impl)
m_localHash
.
drop
(
internalTableName
);
m_globalHash
->
lock
();
impl
.
m_status
=
NdbDictionary
::
Object
::
Invalid
;
m_globalHash
->
drop
(
&
impl
);
m_globalHash
->
release
(
&
impl
,
1
);
m_globalHash
->
unlock
();
DBUG_RETURN
(
0
);
}
...
...
@@ -2918,6 +2965,7 @@ NdbIndexImpl*
NdbDictionaryImpl
::
getIndexImpl
(
const
char
*
externalName
,
const
BaseString
&
internalName
)
{
ASSERT_NOT_MYSQLD
;
Ndb_local_table_info
*
info
=
get_local_table_info
(
internalName
);
if
(
info
==
0
){
m_error
.
code
=
4243
;
...
...
@@ -2938,26 +2986,41 @@ NdbDictionaryImpl::getIndexImpl(const char * externalName,
return
0
;
}
return
getIndexImpl
(
externalName
,
internalName
,
*
tab
,
*
prim
);
}
NdbIndexImpl
*
NdbDictionaryImpl
::
getIndexImpl
(
const
char
*
externalName
,
const
BaseString
&
internalName
,
NdbTableImpl
&
tab
,
NdbTableImpl
&
prim
)
{
DBUG_ENTER
(
"NdbDictionaryImpl::getIndexImpl"
);
DBUG_ASSERT
(
tab
.
m_indexType
!=
NdbDictionary
::
Object
::
TypeUndefined
);
/**
* Create index impl
*/
NdbIndexImpl
*
idx
;
if
(
NdbDictInterface
::
create_index_obj_from_table
(
&
idx
,
tab
,
prim
)
==
0
){
idx
->
m_table
=
tab
;
if
(
NdbDictInterface
::
create_index_obj_from_table
(
&
idx
,
&
tab
,
&
prim
)
==
0
){
idx
->
m_table
=
&
tab
;
idx
->
m_externalName
.
assign
(
externalName
);
idx
->
m_internalName
.
assign
(
internalName
);
idx
->
m_table_id
=
prim
.
getObjectId
();
idx
->
m_table_version
=
prim
.
getObjectVersion
();
// TODO Assign idx to tab->m_index
// Don't do it right now since assign can't asign a table with index
// tab->m_index = idx;
return
idx
;
DBUG_RETURN
(
idx
)
;
}
return
0
;
DBUG_RETURN
(
0
)
;
}
int
NdbDictInterface
::
create_index_obj_from_table
(
NdbIndexImpl
**
dst
,
NdbTableImpl
*
tab
,
const
NdbTableImpl
*
prim
){
const
NdbTableImpl
*
prim
)
{
DBUG_ENTER
(
"NdbDictInterface::create_index_obj_from_table"
);
NdbIndexImpl
*
idx
=
new
NdbIndexImpl
();
idx
->
m_version
=
tab
->
m_version
;
idx
->
m_status
=
tab
->
m_status
;
...
...
@@ -3010,8 +3073,8 @@ NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst,
}
*
dst
=
idx
;
return
0
;
DBUG_PRINT
(
"exit"
,
(
"m_id: %d m_version: %d"
,
idx
->
m_id
,
idx
->
m_version
));
DBUG_RETURN
(
0
)
;
}
/*****************************************************************
...
...
@@ -3020,6 +3083,7 @@ NdbDictInterface::create_index_obj_from_table(NdbIndexImpl** dst,
int
NdbDictionaryImpl
::
createIndex
(
NdbIndexImpl
&
ix
)
{
ASSERT_NOT_MYSQLD
;
NdbTableImpl
*
tab
=
getTable
(
ix
.
getTable
());
if
(
tab
==
0
){
m_error
.
code
=
4249
;
...
...
@@ -3029,6 +3093,12 @@ NdbDictionaryImpl::createIndex(NdbIndexImpl &ix)
return
m_receiver
.
createIndex
(
m_ndb
,
ix
,
*
tab
);
}
int
NdbDictionaryImpl
::
createIndex
(
NdbIndexImpl
&
ix
,
NdbTableImpl
&
tab
)
{
return
m_receiver
.
createIndex
(
m_ndb
,
ix
,
tab
);
}
int
NdbDictInterface
::
createIndex
(
Ndb
&
ndb
,
const
NdbIndexImpl
&
impl
,
...
...
@@ -3135,6 +3205,7 @@ int
NdbDictionaryImpl
::
dropIndex
(
const
char
*
indexName
,
const
char
*
tableName
)
{
ASSERT_NOT_MYSQLD
;
NdbIndexImpl
*
idx
=
getIndex
(
indexName
,
tableName
);
if
(
idx
==
0
)
{
m_error
.
code
=
4243
;
...
...
@@ -3152,8 +3223,7 @@ NdbDictionaryImpl::dropIndex(const char * indexName,
m_localHash
.
drop
(
internalIndexName
.
c_str
());
m_globalHash
->
lock
();
idx
->
m_table
->
m_status
=
NdbDictionary
::
Object
::
Invalid
;
m_globalHash
->
drop
(
idx
->
m_table
);
m_globalHash
->
release
(
idx
->
m_table
,
1
);
m_globalHash
->
unlock
();
return
dropIndex
(
indexName
,
tableName
);
}
...
...
@@ -3183,13 +3253,13 @@ NdbDictionaryImpl::dropIndex(NdbIndexImpl & impl, const char * tableName)
return
dropIndex
(
indexName
,
tableName
);
}
int
ret
=
m_receiver
.
dropIndex
(
impl
,
*
t
impl
);
if
(
ret
==
0
){
m_localHash
.
drop
(
internalIndexName
.
c_str
());
int
ret
=
dropIndexGlobal
(
impl
);
if
(
ret
==
0
)
{
m_globalHash
->
lock
();
impl
.
m_table
->
m_status
=
NdbDictionary
::
Object
::
Invalid
;
m_globalHash
->
drop
(
impl
.
m_table
);
m_globalHash
->
release
(
impl
.
m_table
,
1
);
m_globalHash
->
unlock
();
m_localHash
.
drop
(
internalIndexName
.
c_str
());
}
return
ret
;
}
...
...
@@ -3198,10 +3268,26 @@ NdbDictionaryImpl::dropIndex(NdbIndexImpl & impl, const char * tableName)
return
-
1
;
}
int
NdbDictionaryImpl
::
dropIndexGlobal
(
NdbIndexImpl
&
impl
)
{
DBUG_ENTER
(
"NdbDictionaryImpl::dropIndexGlobal"
);
int
ret
=
m_receiver
.
dropIndex
(
impl
,
*
impl
.
m_table
);
impl
.
m_status
=
NdbDictionary
::
Object
::
Invalid
;
if
(
ret
==
0
)
{
DBUG_RETURN
(
0
);
}
ERR_RETURN
(
getNdbError
(),
ret
);
}
int
NdbDictInterface
::
dropIndex
(
const
NdbIndexImpl
&
impl
,
const
NdbTableImpl
&
timpl
)
{
DBUG_ENTER
(
"NdbDictInterface::dropIndex"
);
DBUG_PRINT
(
"enter"
,
(
"indexId: %d indexVersion: %d"
,
timpl
.
m_id
,
timpl
.
m_version
));
NdbApiSignal
tSignal
(
m_reference
);
tSignal
.
theReceiversBlockNumber
=
DBDICT
;
tSignal
.
theVerId_signalNumber
=
GSN_DROP_INDX_REQ
;
...
...
@@ -3223,9 +3309,9 @@ NdbDictInterface::dropIndex(const NdbIndexImpl & impl,
errCodes
);
if
(
m_error
.
code
==
DropIndxRef
::
InvalidIndexVersion
)
{
// Clear caches and try again
return
INCOMPATIBLE_VERSION
;
ERR_RETURN
(
m_error
,
INCOMPATIBLE_VERSION
)
;
}
return
r
;
ERR_RETURN
(
m_error
,
r
)
;
}
void
...
...
@@ -3262,7 +3348,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
if
(
tab
==
0
){
DBUG_PRINT
(
"info"
,(
"NdbDictionaryImpl::createEvent: table not found: %s"
,
evnt
.
getTableName
()));
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
evnt
.
setTable
(
tab
);
}
...
...
@@ -3281,7 +3367,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
ndbout_c
(
"Attr id %u in table %s not found"
,
evnt
.
m_attrIds
[
i
],
evnt
.
getTableName
());
m_error
.
code
=
4713
;
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
}
...
...
@@ -3302,7 +3388,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
table
.
getColumn
(
evnt
.
m_columns
[
i
]
->
m_name
.
c_str
());
if
(
col
==
0
){
m_error
.
code
=
4247
;
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
// Copy column definition
*
evnt
.
m_columns
[
i
]
=
*
col
;
...
...
@@ -3328,7 +3414,7 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
for
(
i
=
1
;
i
<
attributeList_sz
;
i
++
)
{
if
(
evnt
.
m_columns
[
i
-
1
]
->
m_attrId
==
evnt
.
m_columns
[
i
]
->
m_attrId
)
{
m_error
.
code
=
4258
;
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
}
...
...
@@ -3340,14 +3426,14 @@ NdbDictionaryImpl::createEvent(NdbEventImpl & evnt)
// NdbDictInterface m_receiver;
if
(
m_receiver
.
createEvent
(
m_ndb
,
evnt
,
0
/* getFlag unset */
)
!=
0
)
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
// Create blob events
if
(
evnt
.
m_mergeEvents
&&
createBlobEvents
(
evnt
)
!=
0
)
{
int
save_code
=
m_error
.
code
;
(
void
)
dropEvent
(
evnt
.
m_name
.
c_str
());
m_error
.
code
=
save_code
;
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
DBUG_RETURN
(
0
);
}
...
...
@@ -3367,7 +3453,7 @@ NdbDictionaryImpl::createBlobEvents(NdbEventImpl& evnt)
NdbEventImpl
blob_evnt
;
NdbBlob
::
getBlobEvent
(
blob_evnt
,
&
evnt
,
&
c
);
if
(
createEvent
(
blob_evnt
)
!=
0
)
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
DBUG_RETURN
(
0
);
}
...
...
@@ -3418,7 +3504,7 @@ NdbDictInterface::createEvent(class Ndb & ndb,
const
size_t
len
=
strlen
(
evnt
.
m_name
.
c_str
())
+
1
;
if
(
len
>
MAX_TAB_NAME_SIZE
)
{
m_error
.
code
=
4241
;
DBUG_RETURN
(
-
1
);
ERR_RETURN
(
getNdbError
(),
-
1
);
}
w
.
add
(
SimpleProperties
::
StringValue
,
evnt
.
m_name
.
c_str
());
...
...
@@ -3442,7 +3528,7 @@ NdbDictInterface::createEvent(class Ndb & ndb,
0
,
-
1
);
if
(
ret
)
{
DBUG_RETURN
(
ret
);
ERR_RETURN
(
getNdbError
(),
ret
);
}
char
*
dataPtr
=
(
char
*
)
m_buffer
.
get_data
();
...
...
@@ -3468,7 +3554,7 @@ NdbDictInterface::createEvent(class Ndb & ndb,
//evnt.m_attrListBitmask != evntConf->getAttrListBitmask() ||
evnt
.
mi_type
!=
evntConf
->
getEventType
())
{
ndbout_c
(
"ERROR*************"
);
DBUG_RETURN
(
1
);
ERR_RETURN
(
getNdbError
(),
1
);
}
}
...
...
@@ -3569,35 +3655,36 @@ NdbDictionaryImpl::getEvent(const char * eventName, NdbTableImpl* tab)
DBUG_RETURN
(
NULL
);
}
if
(
tab
==
NULL
)
{
// We only have the table name with internal name
DBUG_PRINT
(
"info"
,(
"table %s"
,
ev
->
getTableName
()));
Ndb_local_table_info
*
info
;
info
=
get_local_table_info
(
ev
->
getTableName
());
if
(
info
==
0
)
if
(
tab
==
NULL
)
{
tab
=
fetchGlobalTableImplRef
(
InitTable
(
this
,
ev
->
getTableName
()));
if
(
tab
==
0
)
{
DBUG_PRINT
(
"error"
,(
"unable to find table %s"
,
ev
->
getTableName
()));
delete
ev
;
DBUG_RETURN
(
NULL
);
}
if
((
info
->
m_table_impl
->
m_status
!=
NdbDictionary
::
Object
::
Retrieved
)
||
(
info
->
m_table_impl
->
m_id
!=
ev
->
m_table_id
)
||
(
table_version_major
(
info
->
m_table_impl
->
m_version
)
!=
if
((
tab
->
m_status
!=
NdbDictionary
::
Object
::
Retrieved
)
||
(
tab
->
m_id
!=
ev
->
m_table_id
)
||
(
table_version_major
(
tab
->
m_version
)
!=
table_version_major
(
ev
->
m_table_version
)))
{
removeCachedObject
(
*
info
->
m_table_impl
);
info
=
get_local_table_info
(
ev
->
getTableName
());
if
(
info
==
0
)
DBUG_PRINT
(
"info"
,
(
"mismatch on verison in cache"
));
releaseTableGlobal
(
*
tab
,
1
);
tab
=
fetchGlobalTableImplRef
(
InitTable
(
this
,
ev
->
getTableName
()));
if
(
tab
==
0
)
{
DBUG_PRINT
(
"error"
,(
"unable to find table %s"
,
ev
->
getTableName
()));
delete
ev
;
DBUG_RETURN
(
NULL
);
}
}
tab
=
info
->
m_table_impl
;
}
ev
->
setTable
(
tab
);
ev
->
setTable
(
m_ndb
.
externalizeTableName
(
ev
->
getTableName
()));
// get the columns from the attrListBitmask
NdbTableImpl
&
table
=
*
ev
->
m_tableImpl
;
...
...
storage/ndb/src/ndbapi/NdbDictionaryImpl.hpp
View file @
b96c4289
...
...
@@ -35,6 +35,9 @@ is_ndb_blob_table(const char* name, Uint32* ptab_id = 0, Uint32* pcol_no = 0);
bool
is_ndb_blob_table
(
const
class
NdbTableImpl
*
t
);
extern
int
ndb_dictionary_is_mysqld
;
#define ASSERT_NOT_MYSQLD assert(ndb_dictionary_is_mysqld == 0)
class
NdbDictObjectImpl
{
public:
int
m_id
;
...
...
@@ -253,6 +256,8 @@ public:
BaseString
m_internalName
;
BaseString
m_externalName
;
BaseString
m_tableName
;
Uint32
m_table_id
;
Uint32
m_table_version
;
Vector
<
NdbColumnImpl
*>
m_columns
;
Vector
<
int
>
m_key_ids
;
...
...
@@ -539,6 +544,21 @@ private:
UtilBuffer
m_buffer
;
};
class
NdbDictionaryImpl
;
class
GlobalCacheInitObject
{
public:
NdbDictionaryImpl
*
m_dict
;
const
BaseString
&
m_name
;
GlobalCacheInitObject
(
NdbDictionaryImpl
*
dict
,
const
BaseString
&
name
)
:
m_dict
(
dict
),
m_name
(
name
)
{}
virtual
~
GlobalCacheInitObject
()
{}
virtual
int
init
(
NdbTableImpl
&
tab
)
const
=
0
;
};
class
NdbDictionaryImpl
:
public
NdbDictionary
::
Dictionary
{
public:
NdbDictionaryImpl
(
Ndb
&
ndb
);
...
...
@@ -558,6 +578,7 @@ public:
int
removeCachedObject
(
NdbTableImpl
&
);
int
createIndex
(
NdbIndexImpl
&
ix
);
int
createIndex
(
NdbIndexImpl
&
ix
,
NdbTableImpl
&
tab
);
int
dropIndex
(
const
char
*
indexName
,
const
char
*
tableName
);
int
dropIndex
(
NdbIndexImpl
&
,
const
char
*
tableName
);
...
...
@@ -578,6 +599,15 @@ public:
int
listObjects
(
List
&
list
,
NdbDictionary
::
Object
::
Type
type
);
int
listIndexes
(
List
&
list
,
Uint32
indexId
);
NdbTableImpl
*
getTableGlobal
(
const
char
*
tableName
);
NdbIndexImpl
*
getIndexGlobal
(
const
char
*
indexName
,
NdbTableImpl
&
ndbtab
);
int
alterTableGlobal
(
NdbTableImpl
&
orig_impl
,
NdbTableImpl
&
impl
);
int
dropTableGlobal
(
NdbTableImpl
&
);
int
dropIndexGlobal
(
NdbIndexImpl
&
impl
);
int
releaseTableGlobal
(
NdbTableImpl
&
impl
,
int
invalidate
);
int
releaseIndexGlobal
(
NdbIndexImpl
&
impl
,
int
invalidate
);
NdbTableImpl
*
getTable
(
const
char
*
tableName
,
void
**
data
=
0
);
NdbTableImpl
*
getBlobTable
(
const
NdbTableImpl
&
,
uint
col_no
);
NdbTableImpl
*
getBlobTable
(
uint
tab_id
,
uint
col_no
);
...
...
@@ -616,10 +646,14 @@ public:
NdbDictInterface
m_receiver
;
Ndb
&
m_ndb
;
private:
NdbIndexImpl
*
getIndexImpl
(
const
char
*
externalName
,
const
BaseString
&
internalName
,
NdbTableImpl
&
tab
,
NdbTableImpl
&
prim
);
NdbIndexImpl
*
getIndexImpl
(
const
char
*
name
,
const
BaseString
&
internalName
);
Ndb_local_table_info
*
fetchGlobalTableImpl
(
const
BaseString
&
internalName
);
private:
NdbTableImpl
*
fetchGlobalTableImplRef
(
const
GlobalCacheInitObject
&
obj
);
};
inline
...
...
@@ -852,6 +886,27 @@ NdbDictionaryImpl::getImpl(const NdbDictionary::Dictionary & t){
* Inline:d getters
*/
class
InitTable
:
public
GlobalCacheInitObject
{
public:
InitTable
(
NdbDictionaryImpl
*
dict
,
const
BaseString
&
name
)
:
GlobalCacheInitObject
(
dict
,
name
)
{}
int
init
(
NdbTableImpl
&
tab
)
const
{
return
m_dict
->
getBlobTables
(
tab
);
}
};
inline
NdbTableImpl
*
NdbDictionaryImpl
::
getTableGlobal
(
const
char
*
table_name
)
{
const
BaseString
internal_tabname
(
m_ndb
.
internalize_table_name
(
table_name
));
return
fetchGlobalTableImplRef
(
InitTable
(
this
,
internal_tabname
));
}
inline
NdbTableImpl
*
NdbDictionaryImpl
::
getTable
(
const
char
*
table_name
,
void
**
data
)
...
...
@@ -885,21 +940,134 @@ NdbDictionaryImpl::get_local_table_info(const BaseString& internalTableName)
DBUG_PRINT
(
"enter"
,
(
"table: %s"
,
internalTableName
.
c_str
()));
Ndb_local_table_info
*
info
=
m_localHash
.
get
(
internalTableName
.
c_str
());
if
(
info
==
0
)
{
info
=
fetchGlobalTableImpl
(
internalTableName
);
if
(
info
==
0
)
{
DBUG_RETURN
(
0
);
if
(
info
==
0
)
{
NdbTableImpl
*
tab
=
fetchGlobalTableImplRef
(
InitTable
(
this
,
internalTableName
));
if
(
tab
)
{
info
=
Ndb_local_table_info
::
create
(
tab
,
m_local_table_data_size
);
if
(
info
)
{
m_localHash
.
put
(
internalTableName
.
c_str
(),
info
);
m_ndb
.
theFirstTupleId
[
tab
->
getTableId
()]
=
~
0
;
m_ndb
.
theLastTupleId
[
tab
->
getTableId
()]
=
~
0
;
}
}
}
DBUG_RETURN
(
info
);
// autoincrement already initialized
}
class
InitIndexGlobal
:
public
GlobalCacheInitObject
{
public:
const
char
*
m_index_name
;
NdbTableImpl
&
m_prim
;
InitIndexGlobal
(
NdbDictionaryImpl
*
dict
,
const
BaseString
&
internal_indexname
,
const
char
*
index_name
,
NdbTableImpl
&
prim
)
:
GlobalCacheInitObject
(
dict
,
internal_indexname
),
m_index_name
(
index_name
),
m_prim
(
prim
)
{}
int
init
(
NdbTableImpl
&
tab
)
const
{
tab
.
m_index
=
m_dict
->
getIndexImpl
(
m_index_name
,
m_name
,
tab
,
m_prim
);
if
(
tab
.
m_index
==
0
)
return
1
;
tab
.
m_index
->
m_table
=
&
tab
;
return
0
;
}
};
class
InitIndex
:
public
GlobalCacheInitObject
{
public:
const
char
*
m_index_name
;
InitIndex
(
NdbDictionaryImpl
*
dict
,
const
BaseString
&
internal_indexname
,
const
char
*
index_name
)
:
GlobalCacheInitObject
(
dict
,
internal_indexname
),
m_index_name
(
index_name
)
{}
int
init
(
NdbTableImpl
&
tab
)
const
{
DBUG_ASSERT
(
tab
.
m_index
==
0
);
tab
.
m_index
=
m_dict
->
getIndexImpl
(
m_index_name
,
m_name
);
if
(
tab
.
m_index
)
{
tab
.
m_index
->
m_table
=
&
tab
;
return
0
;
}
return
1
;
}
};
inline
NdbIndexImpl
*
NdbDictionaryImpl
::
getIndexGlobal
(
const
char
*
index_name
,
NdbTableImpl
&
ndbtab
)
{
DBUG_ENTER
(
"NdbDictionaryImpl::getIndexGlobal"
);
const
BaseString
internal_indexname
(
m_ndb
.
internalize_index_name
(
&
ndbtab
,
index_name
));
int
retry
=
2
;
while
(
retry
)
{
NdbTableImpl
*
tab
=
fetchGlobalTableImplRef
(
InitIndexGlobal
(
this
,
internal_indexname
,
index_name
,
ndbtab
));
if
(
tab
)
{
// tab->m_index sould be set. otherwise tab == 0
NdbIndexImpl
*
idx
=
tab
->
m_index
;
if
(
idx
->
m_table_id
!=
ndbtab
.
getObjectId
()
||
idx
->
m_table_version
!=
ndbtab
.
getObjectVersion
())
{
releaseIndexGlobal
(
*
idx
,
1
);
retry
--
;
continue
;
}
DBUG_RETURN
(
idx
);
}
break
;
}
m_error
.
code
=
4243
;
DBUG_RETURN
(
0
);
}
inline
int
NdbDictionaryImpl
::
releaseTableGlobal
(
NdbTableImpl
&
impl
,
int
invalidate
)
{
DBUG_ENTER
(
"NdbDictionaryImpl::releaseTableGlobal"
);
DBUG_PRINT
(
"enter"
,
(
"internal_name: %s"
,
impl
.
m_internalName
.
c_str
()));
m_globalHash
->
lock
();
m_globalHash
->
release
(
&
impl
,
invalidate
);
m_globalHash
->
unlock
();
DBUG_RETURN
(
0
);
}
inline
int
NdbDictionaryImpl
::
releaseIndexGlobal
(
NdbIndexImpl
&
impl
,
int
invalidate
)
{
DBUG_ENTER
(
"NdbDictionaryImpl::releaseIndexGlobal"
);
DBUG_PRINT
(
"enter"
,
(
"internal_name: %s"
,
impl
.
m_internalName
.
c_str
()));
m_globalHash
->
lock
();
m_globalHash
->
release
(
impl
.
m_table
,
invalidate
);
m_globalHash
->
unlock
();
DBUG_RETURN
(
0
);
}
inline
NdbIndexImpl
*
NdbDictionaryImpl
::
getIndex
(
const
char
*
index_name
,
const
char
*
table_name
)
{
if
(
table_name
||
m_ndb
.
usingFullyQualifiedNames
())
while
(
table_name
||
m_ndb
.
usingFullyQualifiedNames
())
{
const
BaseString
internal_indexname
(
(
table_name
)
...
...
@@ -910,18 +1078,28 @@ NdbDictionaryImpl::getIndex(const char * index_name,
if
(
internal_indexname
.
length
())
{
Ndb_local_table_info
*
info
=
get_local_table_info
(
internal_indexname
)
;
if
(
info
)
Ndb_local_table_info
*
info
=
m_localHash
.
get
(
internal_indexname
.
c_str
());
NdbTableImpl
*
tab
;
if
(
info
==
0
)
{
NdbTableImpl
*
tab
=
info
->
m_table_impl
;
if
(
tab
->
m_index
==
0
)
tab
->
m_index
=
getIndexImpl
(
index_name
,
internal_indexname
);
if
(
tab
->
m_index
!=
0
)
tab
->
m_index
->
m_table
=
tab
;
return
tab
->
m_index
;
tab
=
fetchGlobalTableImplRef
(
InitIndex
(
this
,
internal_indexname
,
index_name
));
if
(
tab
)
{
info
=
Ndb_local_table_info
::
create
(
tab
,
0
);
if
(
info
)
m_localHash
.
put
(
internal_indexname
.
c_str
(),
info
);
else
break
;
}
else
break
;
}
else
tab
=
info
->
m_table_impl
;
return
tab
->
m_index
;
}
break
;
}
m_error
.
code
=
4243
;
...
...
storage/ndb/src/ndbapi/NdbIndexStat.cpp
View file @
b96c4289
...
...
@@ -377,7 +377,7 @@ NdbIndexStat::stat_select(const Uint32* key1, Uint32 keylen1, const Uint32* key2
}
int
NdbIndexStat
::
records_in_range
(
NdbDictionary
::
Index
*
index
,
NdbIndexScanOperation
*
op
,
Uint64
table_rows
,
Uint64
*
count
,
int
flags
)
NdbIndexStat
::
records_in_range
(
const
NdbDictionary
::
Index
*
index
,
NdbIndexScanOperation
*
op
,
Uint64
table_rows
,
Uint64
*
count
,
int
flags
)
{
DBUG_ENTER
(
"NdbIndexStat::records_in_range"
);
Uint64
rows
;
...
...
storage/ndb/src/ndbapi/Ndbinit.cpp
View file @
b96c4289
...
...
@@ -99,6 +99,7 @@ void Ndb::setup(Ndb_cluster_connection *ndb_cluster_connection,
for
(
i
=
0
;
i
<
MAX_NDB_NODES
;
i
++
)
{
theConnectionArray
[
i
]
=
NULL
;
}
//forg
m_sys_tab_0
=
NULL
;
for
(
i
=
0
;
i
<
2048
;
i
++
)
{
theFirstTupleId
[
i
]
=
0
;
theLastTupleId
[
i
]
=
0
;
...
...
@@ -137,6 +138,9 @@ Ndb::~Ndb()
DBUG_ENTER
(
"Ndb::~Ndb()"
);
DBUG_PRINT
(
"enter"
,(
"this=0x%x"
,
this
));
if
(
m_sys_tab_0
)
getDictionary
()
->
removeTableGlobal
(
*
m_sys_tab_0
,
0
);
assert
(
theImpl
->
m_ev_op
==
0
);
// user should return NdbEventOperation's
for
(
NdbEventOperationImpl
*
op
=
theImpl
->
m_ev_op
;
op
;
op
=
op
->
m_next
)
{
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment