- 20 May, 2019 3 commits
-
-
Eugene Kosov authored
Disallow DATETIME for SYSTEM VERSIONING tables.
-
Marko Mäkelä authored
-
Eugene Kosov authored
MDEV-19486 Server crashes in row_upd or row_upd_del_mark_clust_rec on REPLACE into a versioned table row_insert_for_mysql(): InnoDB sets values for row_start and row_end. And this function used to return those values to server in ha_innobase::write_row(). This buggy behavior was removed. Also, a piece of code in this function was reformatted. upd_node_t::make_versioned_helper(): Assert that the preallocated size of the update vector is not exceeded.
-
- 19 May, 2019 1 commit
-
-
Varun Gupta authored
The issue in this case is that we take in account the estimates from quick keys instead of rec_per_key. The estimates for quick keys are better than rec_per_key only if we have ref(const), so we need to check that all keyparts in the ref key are of the type ref(const).
-
- 18 May, 2019 1 commit
-
-
Sergei Golubchik authored
-
- 17 May, 2019 8 commits
-
-
Sergei Golubchik authored
-
Sergei Golubchik authored
-
Sergei Golubchik authored
* Versioning tests support Closes #1043
-
Sergei Golubchik authored
handler::write_row() should not destroy the state of the running index_search/index_next/... or rnd_init/rnd_next/... scan
-
Eugene Kosov authored
Closes #677
-
Eugene Kosov authored
Improve diagnostics. Try to guess what type user tried to type.
-
Alexey Botchkov authored
JSON_MERGE_PATCH implemented. Added JSON_MERGE_PRESERVE as a synonim for the JSON_MERGE.
-
Varun Gupta authored
Fixed, now server can be configured with eq_range_index_dive_limit set in cnf file
-
- 16 May, 2019 12 commits
-
-
Sergey Vojtovich authored
-
Monty authored
The bug was that when using mysql_list_fields, then table_list->schema_table_name was not filled in. Fixed by using table_list->schema_table instead, which is always filled in.
-
Sergei Petrunia authored
Fix both code paths: - Change the test source code so it doesn't cause the "Unused variable" warning (which -Werror converted into error and caused CMake not to set HAVE_THREAD_LOCAL) - If the system doesn't seem to support HAVE_THREAD_LOCAL, refuse to compile (rather than producing a binary that crashes for some tests) Originally submitted at https://github.com/facebook/mysql-5.6/pull/905
-
Sergey Vojtovich authored
This test takes ~6 minutes, split it for better parallelism.
-
Sergey Vojtovich authored
Use thd_get_ha_data()/thd_set_ha_data() which protect against plugin removal until it has THD ha_data. Do not reset THD ha_data in rocksdb_close_connection(), cleaner approach is to let ha_close_connection() do it. Removed transaction objects cleanup from rocksdb_done_func(). As we lock plugin properly, there must be no transaction objects during RocksDB shutdown.
-
Sergey Vojtovich authored
-
Marko Mäkelä authored
-
Marko Mäkelä authored
The bug was introduced in MariaDB 10.4.0 by commit 0e5a4ac2 but it is good to have a regression test for this scenario in all applicable MariaDB versions. Cover the purge of an undo log record that was written before the completion of ADD SPATIAL INDEX.
-
Marko Mäkelä authored
-
Varun Gupta authored
we had the statistics tables in the FROM list of the select. The statistics for tables are not read in such cases, so we need to check this case separately.
-
Alexander Barkov authored
-
Varun Gupta authored
Statistics were not read for a table when we had a CREATE TABLE query. Enforce reading statistics for commands CREATE TABLE, SET and DO.
-
- 15 May, 2019 6 commits
-
-
Robert Bindar authored
-
Marko Mäkelä authored
-
Alexander Barkov authored
-
Eugene Kosov authored
log_buffer_extend(): Do not write to disk. Just allocate new bigger buffer and copy contents of old one to it. Do not acquire write_mutex. log_t::is_extending: Removed as unneeded now. LOG_BUFFER_SIZE: Removed to make the dependence on srv_log_buffer_size visible.
-
Alexander Barkov authored
Fixing problems revealed by buildbot: - Fixing compilation failure on Windows - Recoding correct engines/iuds/r/insert_decimal.result
-
Alexander Barkov authored
-
- 14 May, 2019 9 commits
-
-
Marko Mäkelä authored
Pass a correct string pointer to RETURN_IF_INNODB_NOT_STARTED. This was caught on Windows. The error was introduced in the merge commit 874f8f30.
-
Marko Mäkelä authored
-
Marko Mäkelä authored
-
Marko Mäkelä authored
-
Daniel Bartholomew authored
-
Marko Mäkelä authored
Add the test case. The parent commit, which cherry-picked the MDEV-17167 fix from 10.3 (commit bad2f156) fixed the bug.
-
Oleksandr Byelkin authored
-
Sergey Vojtovich authored
truncating a temporary table TRUNCATE expects only one TABLE instance (which is used by TRUNCATE itself) to be open. However this requirement wasn't enforced after "MDEV-5535: Cannot reopen temporary table". Fixed by closing unused table instances before performing TRUNCATE.
-
Sujatha authored
Problem: ======== We have a Master/Master Setup on two servers, but are only writing to one of those servers (so it is essentially Master/Slave) We upgraded from 10.1.* to 10.2.22 last week and starting with the upgrade, we are getting duplicate key errors on the slave. BINLOG=mixed. Analysis: ========= This issue happens with LOCK TABLES and binlog_format=MIXED combination. When an UNSAFE statement is encountered in 'MIXED' mode, it is logged in the form of 'ROW' format. For all the tables that are part of LOCK TABLES list their table maps are written into the binary log. For each table in the list a check is done to see if 'check_table_binlog_row_based_done' flag is set or not. If it is not set a check process is initiated to see if table qualifies for row based binary logging or not and 'check_table_binlog_row_based_done' is set. This flag will be cleared at the time of closing thread tables. But there can be special cases where the LOCK TABLES contains more number of tables but the unsafe query is actually using subset of tables from LOCK TABLES list. For example: LOCK TABLES locks t1,t2,t3 but the unsafe statement makes use of only two tables t1,t3. In this case the 'check_table_binlog_row_based_done' flag is enabled for table 't2' while writing table map, but 'close_thread_tables' function call will not reset this flag. Since the flag is not cleared for table 't2' even a safe statement which used t2 will be logged in the form of row based format. This leads to an assert on debug builds and causes duplicate entries in release builds. In release builds a statement is logged in the form of both ROW and STATEMENT format. This causes the slave to fail with duplicate key error. Fix: === During 'close_thread_tables' when LOCK TABLE modes are active "ha_reset" is done for all the tables which were part of current statement. As mentioned in the example 'ha_reset' is called for tables 't1' and 't3'. This will clear the 'check_table_binlog_row_based_done' flag. At this point add a check for the rest of the tables to see if 'check_table_binlog_row_based_done' is enabled or not. If enabled clear the flag.
-