- 02 Jul, 2007 1 commit
-
-
lars/lthalmann@dl145j.mysql.com authored
into mysql.com:/nfsdisk1/lars/MERGE/mysql-5.1-merge
-
- 01 Jul, 2007 4 commits
-
-
igor@olga.mysql.com authored
-
igor@olga.mysql.com authored
-
igor@olga.mysql.com authored
-
igor@olga.mysql.com authored
into olga.mysql.com:/home/igor/dev-opt/mysql-5.1-opt
-
- 30 Jun, 2007 9 commits
-
-
igor@olga.mysql.com authored
into olga.mysql.com:/home/igor/mysql-5.0-opt
-
gkodinov/kgeorge@magare.gmz authored
into magare.gmz:/home/kgeorge/mysql/autopush/B29157-5.1-opt
-
tomas@whalegate.ndb.mysql.com authored
into whalegate.ndb.mysql.com:/home/tomas/mysql-5.1-new-rpl
-
tomas@whalegate.ndb.mysql.com authored
-
tomas@whalegate.ndb.mysql.com authored
-
tomas@whalegate.ndb.mysql.com authored
into whalegate.ndb.mysql.com:/home/tomas/mysql-5.1-new-rpl
-
tomas@whalegate.ndb.mysql.com authored
-
stewart@flamingspork.com[stewart] authored
Test that we can start a MySQL Server with a default multibyte charset with NDB running. Test *really* basic functionality too. Index: ndb-work/mysql-test/r/rpl_ndb_ctype_ucs2_def.result ===================================================================
-
stewart@flamingspork.com[stewart] authored
NDB util thread calls mysql_parse internally with plain old c strings (7bit ascii) to create tables (e.g. mysql.ndb_schema). With mysqld default charset set to a multibyte one (e.g. ucs2) mysql_parse would try to interpret the 7bit string as UCS2 and promptly explode in a heap. Solution is to set the util thread to be using utf8 charset. Index: ndb-work/sql/ha_ndbcluster.cc ===================================================================
-
- 29 Jun, 2007 11 commits
-
-
gshchepa/uchum@gleb.loc authored
into gleb.loc:/home/uchum/work/bk/5.0-opt
-
cbell/Chuck@mysql_cab_desk. authored
into mysql_cab_desk.:C:/source/c++/mysql-5.1_BUG_28991
-
gshchepa/uchum@gleb.loc authored
When a UNION statement forced conversion of an UTF8 charset value to a binary charset value, the byte length of the result values was truncated to the CHAR_LENGTH of the original UTF8 value.
-
evgen@moonbone.local authored
into moonbone.local:/mnt/gentoo64/work/29261-bug-5.0-opt-mysql
-
evgen@moonbone.local authored
spaces. When the my_strnncollsp_simple function compares two strings and one is a prefix of another then this function compares characters in the rest of longer key with the space character to find whether the longer key is greater or less. But the sort order of the collation isn't used in this comparison. This may lead to a wrong comparison result, wrongly created index or wrong order of the result set of a query with the ORDER BY clause. Now the my_strnncollsp_simple function uses collation sort order to compare the characters in the rest of longer key with the space character.
-
gkodinov/kgeorge@magare.gmz authored
into magare.gmz:/home/kgeorge/mysql/autopush/B27333-gcov-5.0-opt
-
gkodinov/kgeorge@magare.gmz authored
query / no aggregate of subquery The optimizer counts the aggregate functions that appear as top level expressions (in all_fields) in the current subquery. Later it makes a list of these that it uses to actually execute the aggregates in end_send_group(). That count is used in several places as a flag whether there are aggregates functions. While collecting the above info it must not consider aggregates that are not aggregated in the current context. It must treat them as normal expressions instead. Not doing that leads to incorrect data about the query, e.g. running a query that actually has no aggregate functions as if it has some (and hence is expected to return only one row). Fixed by ignoring the aggregates that are not aggregated in the current context. One other smaller omission discovered and fixed in the process : the place of aggregation was not calculated for user defined functions. Fixed by calling Item_sum::init_sum_func_check() and Item_sum::check_sum_func() as it's done for the rest of the aggregate functions.
-
holyfoot/hf@hfmain.(none) authored
into mysql.com:/home/hf/work/29247/my51-29247
-
holyfoot/hf@hfmain.(none) authored
into mysql.com:/home/hf/work/29247/my50-29247
-
holyfoot/hf@hfmain.(none) authored
into mysql.com:/home/hf/work/29247/my51-29247
-
holyfoot/hf@hfmain.(none) authored
into mysql.com:/home/hf/work/29247/my51-29247
-
- 28 Jun, 2007 6 commits
-
-
gkodinov/kgeorge@magare.gmz authored
Sometimes the number of really updated rows (with changed column values) cannot be determined at the server level alone (e.g. if the storage engine does not return enough column values to verify that). So the only dependable way in such cases is to let the storage engine return that information if possible. Fixed the bug at server level by providing a way for the storage engine to return information about wether it actually updated the row or the old and the new column values are the same. It can do that by returning HA_ERR_RECORD_IS_THE_SAME in ha_update_row(). Note that each storage engine may choose not to try to return this status code, so this behaviour remains storage engine specific.
-
gkodinov/kgeorge@magare.gmz authored
into magare.gmz:/home/kgeorge/mysql/autopush/B26564-5.1-opt
-
holyfoot/hf@mysql.com/hfmain.(none) authored
what caused some consequitive tests failures
-
gkodinov/kgeorge@magare.gmz authored
into magare.gmz:/home/kgeorge/mysql/autopush/B26642-5.0-opt
-
tomas@poseidon.mysql.com authored
into poseidon.mysql.com:/home/tomas/mysql-5.1-new-rpl
-
tomas@poseidon.mysql.com authored
-
- 27 Jun, 2007 9 commits
-
-
tomas@poseidon.mysql.com authored
-
tomas@poseidon.mysql.com authored
- wrong results checked in
-
tomas@poseidon.mysql.com authored
- enabling test, duplicate bug, other bug fixed
-
tomas@poseidon.mysql.com authored
- only log statements locally (changes will not be logged on other servers)
-
tomas@poseidon.mysql.com authored
- binlog should always run if opt_bin_log is set
-
tomas@poseidon.mysql.com authored
- correct includes
-
holyfoot/hf@hfmain.(none) authored
into mysql.com:/home/hf/work/29156/my51-29156
-
mhansson@dl145s.mysql.com authored
into dl145s.mysql.com:/dev/shm/mhansson/my50-bug28677
-
gkodinov/kgeorge@magare.gmz authored
Thanks to Martin Friebe for finding and submitting a fix for this bug! A table with maximum number of key segments and maximum length key name would have a corrupted .frm file, due to an incorrect calculation of the complete key length. Now the key length is computed correctly (I hope) :-) MyISAM would reject a table with the maximum number of keys and the maximum number of key segments in all keys. It would allow one less than this total maximum. Now MyISAM accepts a table defined with the maximum. (This is a very minor issue.)
-