- 01 Oct, 2018 2 commits
-
-
Alexander Barkov authored
-
Alexander Barkov authored
A cleanup for MDEV-17317 Add THD* parameter into Item::get_date() and stricter data type control to "fuzzydate" Fixing C++ function check_date() to get the "fuzzydate" as date_mode_t rather than ulonglong, so conversion from date_time_t to ulonglong is now done inside C++ check_date(), and no conversion is needed in the callers' code. As an additional safety, modified the code not to pass TIME_FUZZY_DATE to the low level C functions: - check_date() - str_to_datetime() - str_to_time() - number_to_datetime() because TIME_FUZZY_DATE is known only on the C++ level, C functions do not know it. Soon we'll be adding more flags into the C++ level (i.e. to date_time_t), e.g. for rounding. It's a good idea to prevent passing C++ specific flags into pure C routines before this change. Asserts were added into the affected C functions to verify that the caller passed only known C level flags.
-
- 30 Sep, 2018 1 commit
-
-
Alexander Barkov authored
Also fixes: MDEV-17330 Wrong result for 0 + LEAST(TIME'-10:00:00',TIME'10:00:00') Problems: 1. These methods did not take into account the current session date flags and passed date_mode_t(0) to func->get_date(): Type_handler_temporal_result::Item_func_min_max_val_real Type_handler_temporal_result::Item_func_min_max_val_int Type_handler_temporal_result::Item_func_min_max_val_decimal Fixing to pass sql_mode_for_dates(thd) instead of date_mode_t(0). Note, sql_mode_for_dates(thd) is only needed for DATE/DATETIME data types. It is not needed for TIME. So splitting value methods Type_handler_temporal_result::Item_func_min_max_xxx into individual implementations for Type_handler_{time|date|datetime|timestamp}_common and, instead of calling get_date(), reusing inside classes Time(), Date(), Datetime() and their methods to_longlong(). sql_mode_for_dates(thd) is automatically passed to get_date() inside Date() and Datetime() constructors. The switch to classes also fixed the problem reported in MDEV-17330. Type_handler_temporal_result::Item_func_min_max_val_int() used to call TIME_to_ulonglong(), which was not correct for TIME. Changing the code to use Time().to_longlong() solved this. 2. Type_handler_temporal_result::Item_func_min_max_get_date also did not take into account the current session date flags in case of conversion from DATE/DATETIME to time and passed date_mode_t(0) to get_date_native(). Fixing to pass sql_mode_for_dates(thd) in case of conversion from DATE/DATETIME to TIME.
-
- 29 Sep, 2018 1 commit
-
-
Alexander Barkov authored
-
- 28 Sep, 2018 1 commit
-
-
Alexander Barkov authored
-
- 27 Sep, 2018 1 commit
-
-
Alexander Barkov authored
MDEV-15406 NO_ZERO_IN_DATE erroneously affects how CAST(AS DATE) warns about fractional digit truncation
-
- 26 Sep, 2018 1 commit
-
-
Alexander Barkov authored
MDEV-17219 Assertion `!t->fraction_remainder(decimals())' failed in Field_time::store_TIME_with_warning
-
- 25 Sep, 2018 1 commit
-
-
Alexander Barkov authored
-
- 22 Sep, 2018 1 commit
-
-
Alexander Barkov authored
-
- 21 Sep, 2018 2 commits
-
-
Marko Mäkelä authored
-
Marko Mäkelä authored
Clarify some comments about accessing an externally stored column on which a spatial index has been defined. Add a TODO comment that we should actually write the minimum bounding rectangle (MBR) to the undo log record, so that we can avoid fetching BLOBs and recomputing MBR. row_build_spatial_index_key(): Split from row_build_index_entry_low().
-
- 19 Sep, 2018 5 commits
-
-
Marko Mäkelä authored
-
Marko Mäkelä authored
Allow combination of non-instant, non-rebuilding operations with changes of table options that do not require a rebuild. For example, DROP INDEX or ADD INDEX can be performed with ALGORITHM=NOCOPY together with changing such table options. Changing the table options alone would be allowed with ALGORITHM=INSTANT. INNOBASE_ALTER_NOCREATE: A new set of flags, for operations that are refused for ALGORITHM=INSTANT and do not involve creating index trees. Move ALTER_RENAME_INDEX to the proper place (INNOBASE_ALTER_INSTANT). innobase_need_rebuild(): Do not require a rebuild if INNOBASE_ALTER_NOREBUILD operations are combined with ALTER_OPTIONS. ha_innobase::prepare_inplace_alter_table(), ha_innobase::inplace_alter_table(): Use the fast path if ALTER_OPTIONS is combined with INNOBASE_ALTER_NOCREATE. In this case, the actual changes would be deferred to ha_innobase::commit_inplace_alter_table().
-
Marko Mäkelä authored
This follows up to commit 755187c8. TRX_UNDO_INSERT_METADATA: Renamed from TRX_UNDO_INSERT_DEFAULT trx_undo_metadata: Renamed from trx_undo_default_rec
-
Marko Mäkelä authored
-
Marko Mäkelä authored
For instant ALTER TABLE, we store a hidden metadata record at the start of the clustered index, to indicate how the format of the records differs from the latest table definition. The term 'default row' is too specific, because it applies to instant ADD COLUMN only, and we will be supporting more classes of instant ALTER TABLE later on. For instant ADD COLUMN, we store the initial default values in the metadata record.
-
- 18 Sep, 2018 5 commits
-
-
Marko Mäkelä authored
Remove some code duplication and dead code. If no 'default row' record exists, the root page must be in the conventional format. Should the page type already be FIL_PAGE_TYPE_INSTANT, we would necessarily hit a debug assertion failure in page_set_instant().
-
Jacob Mathew authored
The crash occurs when the Spider node server attempts to create an error message stating that the temporary table is not found. The function to create the error message is called with incorrect parameters. I fixed the crash by correcting the incorrect parameter values. Author: Jacob Mathew. Reviewer: Kentoku Shiba. Cherry-Picked: Commit e3396161 on branch bb-10.3-mdev-17144
-
Jacob Mathew authored
The crash occurs when the Spider node server attempts to create an error message stating that the temporary table is not found. The function to create the error message is called with incorrect parameters. I fixed the crash by correcting the incorrect parameter values. Author: Jacob Mathew. Reviewer: Kentoku Shiba. Merged: Commit e3396161 branch bb-10.3-mdev-17144
-
Igor Babaev authored
The function JOIN_TAB::choose_best_splitting() did not take into account that for some tables whose fields were used in the GROUP BY list of the specification of a splittable materialized derived there might exist no elements in the array ext_keyuses_for_splitting.
-
Jacob Mathew authored
The crash occurs when the Spider node server attempts to create an error message stating that the temporary table is not found. The function to create the error message is called with incorrect parameters. I fixed the crash by correcting the incorrect parameter values. Author: Jacob Mathew. Reviewer: Kentoku Shiba.
-
- 17 Sep, 2018 4 commits
-
-
Marko Mäkelä authored
-
Marko Mäkelä authored
Now that ha_innobase::prepare_inplace_alter_table() is accessing ha_alter_info->create_info->option_struct, we must initialize it in the Mroonga wrapper for ALTER TABLE based on the parsed table options for the wrap_altered_table.
-
Marko Mäkelä authored
The table option page_compression_level is something that only affects future writes, not actually the data format. Therefore, we can allow instant changes of this option. Similarly, the table option page_compressed can be set on a previously uncompressed table without rebuilding the table, because an uncompressed page would be considered valid when reading a page_compressed table. Removing the page_compressed option will continue to require the table to be rebuilt. ha_innobase_inplace_ctx::page_compression_level: The requested page_compression_level at the start of ALTER TABLE, or 0 if page_compressed=OFF. alter_options_need_rebuild(): Renamed from create_option_need_rebuild(). Allow page_compression_level and page_compressed to be changed as above, without rebuilding the table. ha_innobase::check_if_supported_inplace_alter(): Allow ALGORITHM=INSTANT for ALTER_OPTIONS if the table is not to be rebuilt. If rebuild is needed, set ha_alter_info->unsupported_reason. innobase_page_compression_try(): Update SYS_TABLES.TYPE according to the table flags, for an instant change of page_compression_level or page_compressed. commit_cache_norebuild(): Adjust dict_table_t::flags, fil_space_t::flags and (if needed) FSP_SPACE_FLAGS if page_compression_level was specified.
-
Alexander Barkov authored
-
- 16 Sep, 2018 2 commits
-
-
Alexander Barkov authored
-
Alexander Barkov authored
MDEV-17203 Move fractional second truncation from Item_xxx_typecast::get_date() to Time and Datetime constructors
-
- 15 Sep, 2018 1 commit
-
-
Igor Babaev authored
The optimizer erroneously allowed to use join cache when joining a splittable materialized table together with splitting optimization. As a consequence in some rare cases the server returned wrong result sets for queries with materialized derived. This patch allows to use either join cache without usage of splitting technique for materialization of a splittable derived table or splitting without usage of join cache when joining such table. The costs the these alternatives are compared and the best variant is chosen.
-
- 14 Sep, 2018 2 commits
-
-
Marko Mäkelä authored
-
Marko Mäkelä authored
A debug assertion would fail if an instant ADD COLUMN operation involves splitting the leftmost leaf page and storing a default value off-page. Another debug assertion could fail if the default value does not fit in an undo log page. btr_cur_pessimistic_update(): Invoke rec_offs_make_valid() in order to prevent rec_offs_validate() assertion failure. innobase_add_instant_try(): Invoke btr_cur_pessimistic_update() with the BTR_KEEP_POS_FLAG, which is the correct course of action when BLOBs may need to be written. Whenever returning true, ensure that my_error() will have been called.
-
- 13 Sep, 2018 2 commits
-
-
Jacob Mathew authored
The problem occurs in 10.2 and earlier releases of MariaDB Server because the Partition Engine was not pushing the engine conditions to the underlying storage engine of each partition. This caused Spider to return the first 5 rows in the table with the data provided by the customer. 2 of the 5 rows did not qualify the WHERE clause, so they were removed from the result set by the server. To fix the problem, I have back-ported support for engine condition pushdown in the Partition Engine from MariaDB Server 10.3 to 10.2 and 10.1. In 10.3 and 10.4 I have merged the comments and the test case. Author: Jacob Mathew. Reviewer: Kentoku Shiba. Cherry-Picked: Commit ed49f9aa on branch 10.3
-
Jacob Mathew authored
The problem occurs in 10.2 and earlier releases of MariaDB Server because the Partition Engine was not pushing the engine conditions to the underlying storage engine of each partition. This caused Spider to return the first 5 rows in the table with the data provided by the customer. 2 of the 5 rows did not qualify the WHERE clause, so they were removed from the result set by the server. To fix the problem, I have back-ported support for engine condition pushdown in the Partition Engine from MariaDB Server 10.3 to 10.2 and 10.1. In 10.3 and 10.4 I have merged the comments and the test case. Author: Jacob Mathew. Reviewer: Kentoku Shiba. Merged: Commit eb2ca3d4 on branch bb-10.2-MDEV-16912
-
- 12 Sep, 2018 3 commits
-
-
Alexander Barkov authored
-
Ming Lin authored
The blob key length could be shorter than the length of the entire blob, for example, CREATE TABLE t1 (b BLOB, i INT, KEY(b(8))); INSERT INTO t1 VALUES (REPEAT('a',9),1); The key length is 8, while the blob length is 9. So we need to set the correct key length in Field_blob::sort_string().
-
Alexander Barkov authored
-
- 11 Sep, 2018 5 commits
-
-
Marko Mäkelä authored
Implement a 10.4 redo log format, which extends the 10.3 format by introducing the MLOG_MEMSET record. MLOG_MEMSET: A new redo log record type for filling an area with a byte. mlog_memset(): Write the MLOG_MEMSET record. mlog_parse_nbytes(): Handle MLOG_MEMSET as well. trx_rseg_header_create(): Reduce the redo log volume by making use of mlog_memset() and the zero-initialization that happens inside page allocation. fil_addr_null: Remove. flst_init(): Create a variant that takes a zero-initialized buf_block_t* as a parameter, and only writes the FIL_NULL using mlog_memset(). flst_zero_addr(): A variant of flst_write_addr() that writes a null address using mlog_memset() for the FIL_NULL. The following fixes are replacing some use of MLOG_WRITE_STRING with the more compact MLOG_MEMSET record, or eliminating redundant redo log writes: btr_store_big_rec_extern_fields(): Invoke mlog_memset() for zero-initializing the tail of the ROW_FORMAT=COMPRESSED BLOB page. trx_sysf_create(), trx_rseg_format_upgrade(): Invoke mlog_memset() for zero-initializing the page trailer. fsp_header_init(), trx_rseg_header_create(): Remove redundant zero-initializations.
-
Marko Mäkelä authored
Stop supporting the additional *trunc.log files that were introduced via MySQL 5.7 to MariaDB Server 10.2 and 10.3. DB_TABLESPACE_TRUNCATED: Remove. purge_sys.truncate: A new structure to track undo tablespace file truncation. srv_start(): Remove the call to buf_pool_invalidate(). It is no longer necessary, given that we no longer access things in ways that violate the ARIES protocol. This call was originally added for innodb_file_format, and it may later have been necessary for the proper function of the MySQL 5.7 TRUNCATE recovery, which we are now removing. trx_purge_cleanse_purge_queue(): Take the undo tablespace as a parameter. trx_purge_truncate_history(): Rewrite everything mostly in a single function, replacing references to undo::Truncate. recv_apply_hashed_log_recs(): If any redo log is to be applied, and if the log_sys.log.subformat indicates that separately logged truncate may have been used, refuse to proceed except if innodb_force_recovery is set. We will still refuse crash-upgrade if TRUNCATE TABLE was logged. Undo tablespace truncation would only be logged in undo*trunc.log files, which we are no longer checking for.
-
Marko Mäkelä authored
-
Marko Mäkelä authored
-
Jan Lindström authored
Return code from starting MariaDB.
-