- 02 Mar, 2005 3 commits
-
-
ingo@mysql.com authored
-
ingo@mysql.com authored
Added a check, if the table, which we are going to create, is open. This can happen if a MERGE mapped table is TRUNCATEd.
-
jan@hundin.mysql.fi authored
into hundin.mysql.fi:/home/jan/mysql-5.0
-
- 01 Mar, 2005 24 commits
-
-
heikki@hundin.mysql.fi authored
Correct indentation + some small style corrections
-
tomas@poseidon.ndb.mysql.com authored
updated ndb test makefile to make src dist complete
-
jani@ua141d10.elisa.omakaista.fi authored
into ua141d10.elisa.omakaista.fi:/home/my/bk/mysql-5.0
-
jani@ua141d10.elisa.omakaista.fi authored
-
marko@hundin.mysql.fi authored
delete_all_rows(): Revert previous change. The transaction has been marked started already in external_lock().
-
heikki@hundin.mysql.fi authored
into hundin.mysql.fi:/home/heikki/mysql-5.0
-
heikki@hundin.mysql.fi authored
In TRUNCATE TABLE, MySQL may call innobase_xa_prepare() even though there is no active transaction because it was committed in ha_innobase::delete_all_rows(); in innobase_xa_prepare() InnoDB starts then a new transaction; let us adjust the trx->active_trans flag accordingly
-
marko@hundin.mysql.fi authored
-
marko@hundin.mysql.fi authored
into hundin.mysql.fi:/home/marko/mysql-5.0
-
marko@hundin.mysql.fi authored
CREATE INDEX and DROP INDEX every 10,000 rows, similar to ALTER TABLE.
-
jan@hundin.mysql.fi authored
into hundin.mysql.fi:/home/jan/mysql-5.0
-
jani@ua141d10.elisa.omakaista.fi authored
-
marko@hundin.mysql.fi authored
into hundin.mysql.fi:/home/marko/mysql-4.1
-
jani@ua141d10.elisa.omakaista.fi authored
-
reggie@mdk10.(none) authored
Changed #warning pragma at the bottom to only occur when being used on non-Windows OS
-
marko@hundin.mysql.fi authored
remove old junk and to improve compression ratio. InnoDB: Make implicit type conversions explicit. (Bug #8826)
-
konstantin@mysql.com authored
with view-merge algorithm and prepared statements: in case when some Item_ref pointing to a view column was substituted with a reference pointing to the view expression for that column Item_ref::ref member of the original Item_ref was left pointing to not_found_item (0x1). As we currently perform expression substition part of the view-merge algorithm per each execution of a prepared statement or stored procedure, we need to preserve original Item_ref objects usable.
-
jan@hundin.mysql.fi authored
into hundin.mysql.fi:/home/jan/mysql-5.0
-
ramil@mysql.com authored
into mysql.com:/usr/home/ram/work/4.1.b6089
-
ramil@mysql.com authored
into mysql.com:/usr/home/ram/work/4.1.b6089
-
ramil@mysql.com authored
into mysql.com:/usr/home/ram/work/4.1.b5615
-
marko@hundin.mysql.fi authored
-
marko@hundin.mysql.fi authored
-
jan@hundin.mysql.fi authored
until it has been comprehensive tested.
-
- 28 Feb, 2005 13 commits
-
-
petr@mysql.com authored
into mysql.com:/home/cps/mysql/devel/im-fix-review
-
petr@mysql.com authored
-
monty@mysql.com authored
into mysql.com:/home/my/mysql-5.0
-
pem@mysql.com authored
into mysql.com:/home/pem/work/mysql-5.0
-
sergefp@mysql.com authored
the error to be sent to the client.
-
timour@mysql.com authored
into mysql.com:/home/timka/mysql/src/5.0-bug-8532
-
pem@mysql.com authored
into mysql.com:/home/pem/work/mysql-5.0
-
pem@mysql.comhem.se authored
a DECLARE ? HANDLER FOR stmt.
-
ramil@mysql.com authored
-
sergefp@mysql.com authored
-
pem@mysql.comhem.se authored
Return false from show_binlog_events() if successful, otherwise stored procedures will think it failed.
-
monty@mysql.com authored
into mysql.com:/home/my/mysql-5.0
-
timour@mysql.com authored
The problem was in that the code that analyses the applicability of the QUICK_GROUP_MIN_MAX access method for DISTINC queries assumed that there are no duplicate column references in the DISTINCT clause, and it added non-exiting key parts for the duplicate column references. The solution adds a test to check whether the select list already contained a field with the same name. If such field was already present, then it was already decided to use its key part for index access. In this such case we must skip the duplicate field instead of counting it as a new field.
-