- 13 Aug, 2017 1 commit
-
-
Igor Babaev authored
with window functions (mdev-10855). This patch just modified the function pushdown_cond_for_derived() to support this feature. Some test cases demonstrating this optimization were added to derived_cond_pushdown.test.
-
- 11 Aug, 2017 5 commits
-
-
halfspawn authored
-
Alexey Botchkov authored
Conflicts: sql/item_cmpfunc.cc storage/innobase/buf/buf0flu.cc storage/innobase/include/ut0stage.h storage/innobase/row/row0upd.cc
-
Alexey Botchkov authored
-
Alexey Botchkov authored
JSON_EXTRACT behaves specifically in the comparison, so we have to implement specific method for that in Arg_comparator. Conflicts: sql/item_cmpfunc.cc
-
Igor Babaev authored
-
- 10 Aug, 2017 6 commits
-
-
Igor Babaev authored
developed to cover the case of mdev-13389: "Optimization for equi-joins of derived tables with window functions".
-
Igor Babaev authored
"Optimization for equi-joins of derived tables with GROUP BY" should be considered rather as a 'proof of concept'. The task itself is targeted at an optimization that employs re-writing equi-joins with grouping derived tables / views into lateral derived tables. Here's an example of such transformation: select t1.a,t.max,t.min from t1 [left] join (select a, max(t2.b) max, min(t2.b) min from t2 group by t2.a) as t on t1.a=t.a; => select t1.a,tl.max,tl.min from t1 [left] join lateral (select a, max(t2.b) max, min(t2.b) min from t2 where t1.a=t2.a) as t on 1=1; The transformation pushes the equi-join condition t1.a=t.a into the derived table making it dependent on table t1. It means that for every row from t1 a new derived table must be filled out. However the size of any of these derived tables is just a fraction of the original derived table t. One could say that transformation 'splits' the rows used for the GROUP BY operation into separate groups performing aggregation for a group only in the case when there is a match for the current row of t1. Apparently the transformation may produce a query with a better performance only in the case when - the GROUP BY list refers only to fields returned by the derived table - there is an index I on one of the tables T used in FROM list of the specification of the derived table whose prefix covers the the fields from the proper beginning of the GROUP BY list or fields that are equal to those fields. Whether the result of the re-writing can be executed faster depends on many factors: - the size of the original derived table - the size of the table T - whether the index I is clustering for table T - whether the index I fully covers the GROUP BY list. This patch only tries to improve the chosen execution plan using this transformation. It tries to do it only when the chosen plan reaches the derived table by a key whose prefix covers all the fields of the derived table produced by the fields of the table T from the GROUP BY list. The code of the patch does not evaluates the cost of the improved plan. If certain conditions are met the transformation is applied.
-
Alexey Botchkov authored
JSON_EXTRACT behaves specifically in the comparison, so we have to implement specific method for that in Arg_comparator.
-
Marko Mäkelä authored
buf_page_io_complete(): Do not test bpage for NULL, because it is declared (and always passed) as nonnull. buf_flush_batch(): Remove the constant local variable count=0. fil_ibd_load(): Use magic comment to suppress -Wimplicit-fallthrough. ut_stage_alter_t::inc(ulint): Disable references to an unused parameter. lock_queue_validate(), sync_array_find_thread(), rbt_check_ordering(): Define only in debug builds.
-
Oleksandr Byelkin authored
The bug is result adding ability to have derived tables inside views. Fixed checks should be a switch between view/derived or select derived and information schema.
-
Marko Mäkelä authored
Only a relevant subset of the InnoDB changes was merged. In particular, two follow-up bug fixes for the bugs that were introduced in 5.7.18 but not MariaDB 10.2.7 were omitted. Because MariaDB 10.2.7 omitted the risky change Bug#23481444 OPTIMISER CALL ROW_SEARCH_MVCC() AND READ THE INDEX APPLIED BY UNCOMMITTED ROWS we do not need the follow-up fixes that were introduced in MySQL 5.6.37 and MySQL 5.7.19: Bug#25175249 ASSERTION: (TEMPL->IS_VIRTUAL && !FIELD) || ... Bug#25793677 INNODB: FAILING ASSERTION: CLUST_TEMPL_FOR_SEC || LEN
-
- 09 Aug, 2017 15 commits
-
-
Thirunarayanan Balathandayuthapani authored
Analysis: ========= During alter table rebuild, InnoDB fails to apply concurrent insert log. If the insert log record is present across the blocks then apply phase trying to access the next block without fetching it. Fix: ==== During virtual column parsing, check whether the record is present across the blocks before accessing the virtual column information. Reviewed-by: Jimmy Yang <jimmy.yang@oracle.com> RB: 16243
-
Thirunarayanan Balathandayuthapani authored
Analysis: ======== During alter table rebuild, InnoDB fails to apply concurrent delete log. Parsing and validation of merge record happens while applying the log operation on a table. Validation goes wrong for the virtual column. Validation assumes that virtual column information can't be the end of the merge record end. Fix: ==== Virtual column information in the merge record can be end of the merge record. Virtual column information is written at the end for row_log_table_delete(). Reviewed-by: Satya Bodapati<satya.bodapati@oracle.com> RB: 16155
-
Marko Mäkelä authored
The test is for a bug that was introduced in MySQL 5.7.18 but not MariaDB 10.2, because MariaDB did not merge the change that was considered incomplete and too risky for a GA release: Bug#23481444 OPTIMISER CALL ROW_SEARCH_MVCC() AND READ THE INDEX APPLIED BY UNCOMMITTED ROWS So, we are only merging the test changes from the bug fix in MySQL 5.7.19, not any code changes: commit 4f86aca37d551cc756d9187ec901f8c4a68a0543 Author: Thirunarayanan Balathandayuthapani <thirunarayanan.balathandayuth@oracle.com> Date: Wed Apr 26 11:10:41 2017 +0530 Bug #25793677 INNODB: FAILING ASSERTION: CLUST_TEMPL_FOR_SEC || LEN
-
Marko Mäkelä authored
MariaDB 10.2 never contained the Oracle change Bug#23481444 OPTIMISER CALL ROW_SEARCH_MVCC() AND READ THE INDEX APPLIED BY UNCOMMITTED ROWS because it was considered risky for a GA release and incomplete. Remove the references that were added when merging MySQL 5.6.36 to MariaDB 10.0.31, 10.1.24, and 10.2.7.
-
Thirunarayanan Balathandayuthapani authored
Analysis: ======== (1) During TRUNCATE of file_per_table tablespace, dict_operation_lock is released before eviction of dirty pages of a tablespace from the buffer pool. After eviction, we try to re-acquire dict_operation_lock (higher level latch) but we already hold lower level latch (index->lock). This causes latch order violation (2) Deadlock issue is present if child table is being truncated and it holds index lock. At the same time, cascade dml happens and it took dict_operation_lock and waiting for index lock. Fix: ==== 1) Release the indexes lock before releasing the dict operation lock. 2) Ignore the cascading dml operation on the parent table, for the cascading foreign key, if the child table is truncated or if it is in the process of being truncated. Reviewed-by: Jimmy Yang <jimmy.yang@oracle.com> Reviewed-by: Kevin Lewis <kevin.lewis@oracle.com> RB: 16122
-
Darshan M N authored
Issue ==== The issue is that the info message that InnoDB prints when a table is created with a reference which doesn't exist fills up the log as it's printed for every insert when foreign_key_checks is disabled. Fix === The fix is to display the message only if foreign_key_checks is enabled. Reviewed-by: Jimmy Yang <jimmy.yang@oracle.com>
-
Thirunarayanan Balathandayuthapani authored
Problem: ======= Offsets allocates memory from row_heap even for deleted row traversal during table rebuild. Solution: ========= Empty the row_heap even for deleted record. So that offsets don't allocate memory everytime. Reviewed-by: Jimmy Yang <jimmy.yang@oracle.com> RB: 15694
-
Marko Mäkelä authored
Cherry-pick the commit from MySQL 5.7.19, and adapt the test case: commit 45c933ac19c73a3e9c756a87ee1ba18ba1ac564c Author: Aakanksha Verma <aakanksha.verma@oracle.com> Date: Tue Mar 21 10:31:43 2017 +0530 Bug #25189192 ERRORS WHEN RESTARTING MYSQL AFTER RENAME TABLE. PROBLEM While renaming table innodb doesn't update the InnoDB Dictionary table INNODB_SYS_DATAFILES incase there is change in database while doing rename table. Hence on a restart the server log shows error that it couldnt find table with old path before rename which has actually been renamed. So the errors would only vanish if we update the system tablespace FIX Update the innodb dictionary table with new path in the case there is not a change in the table but the database holding the table as well. Reviewed-by: Jimmy Yang<Jimmy.Yang@oracle.com> RB: 15751
-
Oleksandr Byelkin authored
take into account new capabilty flag which has influence on result format.
-
Jan Lindström authored
This happens because on line 59 of /usr/bin/wsrep_sst_mysqldump we have a check : "if ! $MYSQL_CLIENT --version | grep 'Distrib 10.1' >/dev/null" but the client is 10.2 so it doesnt match the grep expression. Fixed check to be: "if ! $MYSQL_CLIENT --version | grep 'Distrib 10.' >/dev/null" so that every 10. version is accepted.
-
Marko Mäkelä authored
-
Marko Mäkelä authored
-
Sergei Petrunia authored
-
Marko Mäkelä authored
Disable change buffering, so that some data that was previously written to the encrypted redo log will not end up being copied to the unencrypted redo log due to change buffer merge.
-
Marko Mäkelä authored
The thd_destructor_proxy detects that no transactions are active and starts srv_shutdown_bg_undo_sources(), but fails to take into account that new transactions can still start, especially be slave but also by other threads. In addition there is no mutex when checking for active transaction so this is not safe. We relax the failing InnoDB debug assertion by allowing the execution of user transactions after the purge thread has been shut down. FIXME: If innodb_fast_shutdown=0, we should somehow guarantee that no new transactions can start after thd_destructor_proxy observed that trx_sys_any_active_transactions() did not hold.
-
- 08 Aug, 2017 11 commits
-
-
Marko Mäkelä authored
row_update_for_mysql(): Remove the wrapper function and rename the function from row_update_for_mysql_using_upd_graph(). Remove the unused parameter mysql_rec.
-
Marko Mäkelä authored
-
Marko Mäkelä authored
This is basically a duplicate or a reincarnation of MDEV-117. For some reason, the test innodb.mdev-117 started failing in 10.2. It is uncertain when this test started failing. The test is nondeterministic, because there is a race condition between the concurrently executing DELETE IGNORE and DELETE statements. When a deadlock is reported for DELETE IGNORE, the SQL layer would call handler::print_error() but then proceed to the next row, as if no error had happened (which is the purpose of DELETE IGNORE). So, when it proceeded to handler::ha_rnd_next(), InnoDB would hit an assertion failure, because the transaction no longer exists, and we are not executing at the start of a statement. handler::print_error(): If thd_mark_transaction_to_rollback(thd, true) was called, clear the ME_JUST_WARNING and ME_JUST_INFO errflags, so that a note or warning will be promoted to an error if the transaction was aborted by a storage engine.
-
Alexey Botchkov authored
Check for duplicating keys added.
-
Alexey Botchkov authored
outside. The result_limit variable wasn't always initialized in Item_func_json_array::fix_length_and_dec().
-
Marko Mäkelä authored
-
Marko Mäkelä authored
The file wait_innodb_all_purged.inc waited for InnoDB purge in a way that only worked in debug builds. The file wait_all_purged.inc provides a better mechanism.
-
Marko Mäkelä authored
-
Marko Mäkelä authored
If the latest InnoDB redo log checkpoint was stored in the first checkpoint slot and not the second one, InnoDB would incorrectly set log_sys->log.lsn to the previous checkpoint. It is possible that this logic error did not exist before commit 86927cc7, which removed traces of multiple InnoDB redo logs, to prepare for MDEV-12548 (Mariabackup for MariaDB 10.2). In the worst case, this error could mean that InnoDB unnecessarily fails to recover from redo log when the last-but-one checkpoint was overwritten, but the last checkpoint is intact. recv_find_max_checkpoint(), recv_find_max_checkpoint_0(): Do not overwrite the fields of log_sys->log with the information of an older checkpoint. recv_find_max_checkpoint(): Do not return DB_SUCCESS on an error. recv_recovery_from_checkpoint_start(): Return early if the log is in a version-tagged format but not in the latest format. (In this case, the log must be logically empty, and there is nothing to apply.)
-
Jan Lindström authored
Always read full page 0 to determine does tablespace contain encryption metadata. Tablespaces that are page compressed or page compressed and encrypted do not compare checksum as it does not exists. For encrypted tables use checksum verification written for encrypted tables and normal tables use normal method. buf_page_is_checksum_valid_crc32 buf_page_is_checksum_valid_innodb buf_page_is_checksum_valid_none Modify Innochecksum logging to file to avoid compilation warnings. fil0crypt.cc fil0crypt.h Modify to be able to use in innochecksum compilation and move fil_space_verify_crypt_checksum to end of the file. Add innochecksum logging to file. univ.i Add innochecksum strict_verify, log_file and cur_page_num variables as extern. page_zip_verify_checksum Add innochecksum logging to file and remove unnecessary code. innochecksum.cc Lot of changes most notable able to read encryption metadata from page 0 of the tablespace. Added test case where we corrupt intentionally FIL_PAGE_FILE_FLUSH_LSN_OR_KEY_VERSION (encryption key version) FIL_PAGE_FILE_FLUSH_LSN_OR_KEY_VERSION+4 (post encryption checksum) FIL_DATA+10 (data)
-
Alexey Botchkov authored
Comparison fixed to take the actual type of JSON value into account. Bug in escaping handling fixed.
-
- 07 Aug, 2017 2 commits
-
-
Monty authored
Added extra memcpy to get rid of valgrind warning for sequence tables with InnoDB. When reading a row from InnoDB, some of the bytes in the row are marked as not initialized. Needs to be investigated later, but this is a safe patch for now.
-
Monty authored
Problem was that SEQUENCE::table was shared among threads, which caused several threads to use the same object at the same time.
-