Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
M
MariaDB
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
nexedi
MariaDB
Commits
58703a43
Commit
58703a43
authored
Jul 06, 2005
by
mskold@mysql.com
Browse files
Options
Browse Files
Download
Plain Diff
Merge mskold@bk-internal.mysql.com:/home/bk/mysql-4.1
into mysql.com:/usr/local/home/marty/MySQL/mysql-4.1
parents
29653d5a
7063d85b
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
23 additions
and
42 deletions
+23
-42
sql/ha_ndbcluster.cc
sql/ha_ndbcluster.cc
+21
-39
sql/ha_ndbcluster.h
sql/ha_ndbcluster.h
+2
-3
No files found.
sql/ha_ndbcluster.cc
View file @
58703a43
...
@@ -1113,42 +1113,23 @@ int ha_ndbcluster::set_primary_key(NdbOperation *op, const byte *key)
...
@@ -1113,42 +1113,23 @@ int ha_ndbcluster::set_primary_key(NdbOperation *op, const byte *key)
}
}
int
ha_ndbcluster
::
set_primary_key_from_
old_data
(
NdbOperation
*
op
,
const
byte
*
old_data
)
int
ha_ndbcluster
::
set_primary_key_from_
record
(
NdbOperation
*
op
,
const
byte
*
record
)
{
{
KEY
*
key_info
=
table
->
key_info
+
table
->
primary_key
;
KEY
*
key_info
=
table
->
key_info
+
table
->
primary_key
;
KEY_PART_INFO
*
key_part
=
key_info
->
key_part
;
KEY_PART_INFO
*
key_part
=
key_info
->
key_part
;
KEY_PART_INFO
*
end
=
key_part
+
key_info
->
key_parts
;
KEY_PART_INFO
*
end
=
key_part
+
key_info
->
key_parts
;
DBUG_ENTER
(
"set_primary_key_from_
old_data
"
);
DBUG_ENTER
(
"set_primary_key_from_
record
"
);
for
(;
key_part
!=
end
;
key_part
++
)
for
(;
key_part
!=
end
;
key_part
++
)
{
{
Field
*
field
=
key_part
->
field
;
Field
*
field
=
key_part
->
field
;
if
(
set_ndb_key
(
op
,
field
,
if
(
set_ndb_key
(
op
,
field
,
key_part
->
fieldnr
-
1
,
old_data
+
key_part
->
offset
))
key_part
->
fieldnr
-
1
,
record
+
key_part
->
offset
))
ERR_RETURN
(
op
->
getNdbError
());
ERR_RETURN
(
op
->
getNdbError
());
}
}
DBUG_RETURN
(
0
);
DBUG_RETURN
(
0
);
}
}
int
ha_ndbcluster
::
set_primary_key
(
NdbOperation
*
op
)
{
DBUG_ENTER
(
"set_primary_key"
);
KEY
*
key_info
=
table
->
key_info
+
table
->
primary_key
;
KEY_PART_INFO
*
key_part
=
key_info
->
key_part
;
KEY_PART_INFO
*
end
=
key_part
+
key_info
->
key_parts
;
for
(;
key_part
!=
end
;
key_part
++
)
{
Field
*
field
=
key_part
->
field
;
if
(
set_ndb_key
(
op
,
field
,
key_part
->
fieldnr
-
1
,
field
->
ptr
))
ERR_RETURN
(
op
->
getNdbError
());
}
DBUG_RETURN
(
0
);
}
/*
/*
Read one record from NDB using primary key
Read one record from NDB using primary key
*/
*/
...
@@ -1241,9 +1222,9 @@ int ha_ndbcluster::complemented_pk_read(const byte *old_data, byte *new_data)
...
@@ -1241,9 +1222,9 @@ int ha_ndbcluster::complemented_pk_read(const byte *old_data, byte *new_data)
op
->
readTuple
(
lm
)
!=
0
)
op
->
readTuple
(
lm
)
!=
0
)
ERR_RETURN
(
trans
->
getNdbError
());
ERR_RETURN
(
trans
->
getNdbError
());
int
res
;
int
res
;
if
((
res
=
set_primary_key_from_old_data
(
op
,
old_data
)))
if
((
res
=
set_primary_key_from_record
(
op
,
old_data
)))
ERR_RETURN
(
trans
->
getNdbError
());
ERR_RETURN
(
trans
->
getNdbError
());
// Read all unreferenced non-key field(s)
// Read all unreferenced non-key field(s)
for
(
i
=
0
;
i
<
no_fields
;
i
++
)
for
(
i
=
0
;
i
<
no_fields
;
i
++
)
...
@@ -1273,7 +1254,7 @@ int ha_ndbcluster::complemented_pk_read(const byte *old_data, byte *new_data)
...
@@ -1273,7 +1254,7 @@ int ha_ndbcluster::complemented_pk_read(const byte *old_data, byte *new_data)
Peek to check if a particular row already exists
Peek to check if a particular row already exists
*/
*/
int
ha_ndbcluster
::
peek_row
()
int
ha_ndbcluster
::
peek_row
(
const
byte
*
record
)
{
{
NdbConnection
*
trans
=
m_active_trans
;
NdbConnection
*
trans
=
m_active_trans
;
NdbOperation
*
op
;
NdbOperation
*
op
;
...
@@ -1287,7 +1268,7 @@ int ha_ndbcluster::peek_row()
...
@@ -1287,7 +1268,7 @@ int ha_ndbcluster::peek_row()
ERR_RETURN
(
trans
->
getNdbError
());
ERR_RETURN
(
trans
->
getNdbError
());
int
res
;
int
res
;
if
((
res
=
set_primary_key
(
op
)))
if
((
res
=
set_primary_key
_from_record
(
op
,
record
)))
ERR_RETURN
(
trans
->
getNdbError
());
ERR_RETURN
(
trans
->
getNdbError
());
if
(
execute_no_commit_ie
(
this
,
trans
)
!=
0
)
if
(
execute_no_commit_ie
(
this
,
trans
)
!=
0
)
...
@@ -1841,7 +1822,7 @@ int ha_ndbcluster::write_row(byte *record)
...
@@ -1841,7 +1822,7 @@ int ha_ndbcluster::write_row(byte *record)
if
(
m_ignore_dup_key
&&
table
->
primary_key
!=
MAX_KEY
)
if
(
m_ignore_dup_key
&&
table
->
primary_key
!=
MAX_KEY
)
{
{
int
peek_res
=
peek_row
();
int
peek_res
=
peek_row
(
record
);
if
(
!
peek_res
)
if
(
!
peek_res
)
{
{
...
@@ -1891,9 +1872,7 @@ int ha_ndbcluster::write_row(byte *record)
...
@@ -1891,9 +1872,7 @@ int ha_ndbcluster::write_row(byte *record)
m_skip_auto_increment
=
!
auto_increment_column_changed
;
m_skip_auto_increment
=
!
auto_increment_column_changed
;
}
}
if
((
res
=
(
m_primary_key_update
?
if
((
res
=
set_primary_key_from_record
(
op
,
record
)))
set_primary_key_from_old_data
(
op
,
record
)
:
set_primary_key
(
op
))))
return
res
;
return
res
;
}
}
...
@@ -2030,7 +2009,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
...
@@ -2030,7 +2009,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
if
((
table
->
primary_key
!=
MAX_KEY
)
&&
if
((
table
->
primary_key
!=
MAX_KEY
)
&&
(
key_cmp
(
table
->
primary_key
,
old_data
,
new_data
)))
(
key_cmp
(
table
->
primary_key
,
old_data
,
new_data
)))
{
{
int
read_res
,
insert_res
,
delete_res
;
int
read_res
,
insert_res
,
delete_res
,
undo_res
;
DBUG_PRINT
(
"info"
,
(
"primary key update, doing pk read+delete+insert"
));
DBUG_PRINT
(
"info"
,
(
"primary key update, doing pk read+delete+insert"
));
// Get all old fields, since we optimize away fields not in query
// Get all old fields, since we optimize away fields not in query
...
@@ -2059,9 +2038,14 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
...
@@ -2059,9 +2038,14 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
DBUG_PRINT
(
"info"
,
(
"insert failed"
));
DBUG_PRINT
(
"info"
,
(
"insert failed"
));
if
(
trans
->
commitStatus
()
==
NdbConnection
::
Started
)
if
(
trans
->
commitStatus
()
==
NdbConnection
::
Started
)
{
{
// Undo write_row(new
_data)
// Undo delete_row(old
_data)
m_primary_key_update
=
TRUE
;
m_primary_key_update
=
TRUE
;
insert_res
=
write_row
((
byte
*
)
old_data
);
undo_res
=
write_row
((
byte
*
)
old_data
);
if
(
undo_res
)
push_warning
(
current_thd
,
MYSQL_ERROR
::
WARN_LEVEL_WARN
,
undo_res
,
"NDB failed undoing delete at primary key update"
);
m_primary_key_update
=
FALSE
;
m_primary_key_update
=
FALSE
;
}
}
DBUG_RETURN
(
insert_res
);
DBUG_RETURN
(
insert_res
);
...
@@ -2110,7 +2094,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
...
@@ -2110,7 +2094,7 @@ int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
else
else
{
{
int
res
;
int
res
;
if
((
res
=
set_primary_key_from_
old_data
(
op
,
old_data
)))
if
((
res
=
set_primary_key_from_
record
(
op
,
old_data
)))
DBUG_RETURN
(
res
);
DBUG_RETURN
(
res
);
}
}
}
}
...
@@ -2191,10 +2175,8 @@ int ha_ndbcluster::delete_row(const byte *record)
...
@@ -2191,10 +2175,8 @@ int ha_ndbcluster::delete_row(const byte *record)
else
else
{
{
int
res
;
int
res
;
if
((
res
=
(
m_primary_key_update
?
if
((
res
=
set_primary_key_from_record
(
op
,
record
)))
set_primary_key_from_old_data
(
op
,
record
)
return
res
;
:
set_primary_key
(
op
))))
return
res
;
}
}
}
}
...
...
sql/ha_ndbcluster.h
View file @
58703a43
...
@@ -168,7 +168,7 @@ class ha_ndbcluster: public handler
...
@@ -168,7 +168,7 @@ class ha_ndbcluster: public handler
int
pk_read
(
const
byte
*
key
,
uint
key_len
,
byte
*
buf
);
int
pk_read
(
const
byte
*
key
,
uint
key_len
,
byte
*
buf
);
int
complemented_pk_read
(
const
byte
*
old_data
,
byte
*
new_data
);
int
complemented_pk_read
(
const
byte
*
old_data
,
byte
*
new_data
);
int
peek_row
();
int
peek_row
(
const
byte
*
record
);
int
unique_index_read
(
const
byte
*
key
,
uint
key_len
,
int
unique_index_read
(
const
byte
*
key
,
uint
key_len
,
byte
*
buf
);
byte
*
buf
);
int
ordered_index_scan
(
const
key_range
*
start_key
,
int
ordered_index_scan
(
const
key_range
*
start_key
,
...
@@ -196,8 +196,7 @@ class ha_ndbcluster: public handler
...
@@ -196,8 +196,7 @@ class ha_ndbcluster: public handler
friend
int
g_get_ndb_blobs_value
(
NdbBlob
*
ndb_blob
,
void
*
arg
);
friend
int
g_get_ndb_blobs_value
(
NdbBlob
*
ndb_blob
,
void
*
arg
);
int
get_ndb_blobs_value
(
NdbBlob
*
last_ndb_blob
);
int
get_ndb_blobs_value
(
NdbBlob
*
last_ndb_blob
);
int
set_primary_key
(
NdbOperation
*
op
,
const
byte
*
key
);
int
set_primary_key
(
NdbOperation
*
op
,
const
byte
*
key
);
int
set_primary_key
(
NdbOperation
*
op
);
int
set_primary_key_from_record
(
NdbOperation
*
op
,
const
byte
*
record
);
int
set_primary_key_from_old_data
(
NdbOperation
*
op
,
const
byte
*
old_data
);
int
set_bounds
(
NdbIndexScanOperation
*
ndb_op
,
const
key_range
*
keys
[
2
]);
int
set_bounds
(
NdbIndexScanOperation
*
ndb_op
,
const
key_range
*
keys
[
2
]);
int
key_cmp
(
uint
keynr
,
const
byte
*
old_row
,
const
byte
*
new_row
);
int
key_cmp
(
uint
keynr
,
const
byte
*
old_row
,
const
byte
*
new_row
);
void
print_results
();
void
print_results
();
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment