1. 05 May, 2014 3 commits
  2. 04 May, 2014 1 commit
  3. 01 May, 2014 7 commits
  4. 23 Apr, 2014 1 commit
    • Sergey Vojtovich's avatar
      MDEV-5792 - Deadlock between SELECTs from METADATA_LOCK_INFO and another · b1232490
      Sergey Vojtovich authored
                  I_S table
      
      mdl_iterate() helper function (which is used by the plugin to iterate mdl
      locks) acquired mutexes in reverse order.
      
      Fixed by iterating MDL locks in two stages:
      1. Iterate locks hash under the protection of hash mutex, store all
         lock pointers in a thread local array and increment reference counter
         for the lock.
      2. Iterate local array without protection of hash mutex, handle destroyed
         locks.
      
      It somewhat echoes hack in MDL_map_partition::move_from_hash_to_lock_mutex.
      b1232490
  5. 05 May, 2014 1 commit
  6. 02 May, 2014 1 commit
  7. 29 Apr, 2014 1 commit
  8. 25 Apr, 2014 2 commits
  9. 28 Apr, 2014 1 commit
    • Sergey Petrunya's avatar
      Revert these two changes (wrong push) : · 182f7d76
      Sergey Petrunya authored
      MDEV-5980: EITS: if condition is used for REF access, its selectivity is still in filtered%
      MDEV-5985: EITS: selectivity estimates look illogical for join and non-key equalities
      MDEV-6003: EITS: ref access, keypart2=const vs keypart2=expr - inconsistent filtered% value
      - Made a number of fixes in table_cond_selectivity() so that it returns
        correct selectivity estimates.
      - Added comments in related code.
      Better comments
      182f7d76
  10. 25 Apr, 2014 1 commit
  11. 21 Apr, 2014 1 commit
  12. 27 Apr, 2014 1 commit
  13. 25 Apr, 2014 2 commits
    • Alexander Barkov's avatar
    • unknown's avatar
      MDEV-6156: Parallel replication incorrectly caches charset between worker threads · 64923bb6
      unknown authored
      The previous patch for this bug was unfortunately completely wrong.
      
      The purpose of cached_charset is to remember which character set we
      have installed currently in the THD, so that in the common case where
      charset does not change between queries, we do not need to update it
      in the THD. Thus, it is important that the cached_charset field is
      tightly coupled to the THD for which it handles caching.
      
      Thus the right place to put cached_charset seems to be in the THD.
      This patch introduces a field THD:system_thread_info where such info
      in general can be placed without further inflating the THD with unused
      data for other threads (THD is already far too big as it is). It then
      moves the cached_charset into this slot for the SQL driver thread and
      for the parallel replication worker threads.
      
      The THD::rpl_filter field is also moved inside system_thread_info, to
      keep the size of THD unchanged. Moving further fields in to reduce the
      size of THD is a separate task, filed as MDEV-6164.
      64923bb6
  14. 24 Apr, 2014 3 commits
  15. 23 Apr, 2014 2 commits
  16. 22 Apr, 2014 3 commits
    • Igor Babaev's avatar
      Merge. · bcf16fa6
      Igor Babaev authored
      bcf16fa6
    • Igor Babaev's avatar
      Fixed the problem of mdev-5947. · 3e0f63c1
      Igor Babaev authored
      Back-ported from the mysql 5.6 code line the patch with
      the following comment:
      
        Fix for Bug#11757108 CHANGE IN EXECUTION PLAN FOR COUNT_DISTINCT_GROUP_ON_KEY
                             CAUSES PEFORMANCE REGRESSION
      
        The cause for the performance regression is that the access strategy for the
        GROUP BY query is changed form using "index scan" in mysql-5.1 to use "loose
        index scan" in mysql-5.5. The index used for group by is unique and thus each
        "loose scan" group will only contain one record. Since loose scan needs to
        re-position on each "loose scan" group this query will do a re-position for
        each index entry. Compared to just reading the next index entry as a normal
        index scan does, the use of loose scan for this query becomes more expensive.
      
        The cause for selecting to use loose scan for this query is that in the current
        code when the size of the "loose scan" group is one, the formula for
        calculating the cost estimates becomes almost identical to the cost of using
        normal index scan. Differences in use of integer versus floating point arithmetic
        can cause one or the other access strategy to be selected.
      
        The main issue with the formula for estimating the cost of using loose scan is
        that it does not take into account that it is more costly to do a re-position
        for each "loose scan" group compared to just reading the next index entry.
        Both index scan and loose scan estimates the cpu cost as:
      
          "number of entries needed too read/scan" * ROW_EVALUATE_COST
      
        The results from testing with the query in this bug indicates that the real
        cost for doing re-position four to eight times higher than just reading the
        next index entry. Thus, the cpu cost estimate for loose scan should be increased.
        To account for the extra work to re-position in the index we increase the
        cost for loose index scan to include the cost of navigating the index.
        This is modelled as a function of the height of the b-tree:
      
          navigation cost= ceil(log(records in table)/log(indexes per block))
                         * ROWID_COMPARE_COST;
      
        This will avoid loose index scan being used for indexes where the "loose scan"
        group contains very few index entries.
      3e0f63c1
    • Alexander Barkov's avatar
  17. 21 Apr, 2014 2 commits
  18. 19 Apr, 2014 1 commit
    • Olivier Bertrand's avatar
      - Implement "remote" index (similar to FEDERATED ones) for MYSQL tables. · 187e4169
      Olivier Bertrand authored
        Not yet done for ODBC tables.
      modified:
        storage/connect/connect.cc
        storage/connect/ha_connect.cc
        storage/connect/ha_connect.h
        storage/connect/mycat.cc
        storage/connect/plgdbsem.h
        storage/connect/reldef.h
        storage/connect/tabdos.h
        storage/connect/tabmysql.cpp
        storage/connect/tabmysql.h
        storage/connect/tabodbc.cpp
        storage/connect/tabodbc.h
        storage/connect/xindex.cpp
        storage/connect/xtable.h
      
      - Return error in "info" on Cardinality error.
      modified:
        storage/connect/ha_connect.cc
      187e4169
  19. 18 Apr, 2014 1 commit
  20. 17 Apr, 2014 1 commit
    • Igor Babaev's avatar
      Fixed the problem of mdev-5970: · 12eb6d88
      Igor Babaev authored
      back-ported the patch for bug #13256831 from mysql-5.6 code line.
      
        Here's the comment this patch was provided with:
      
        Fixed bug#13256831 - ERROR 1032 (HY000): CAN'T FIND RECORD.
      
        This bug only occurs if a user tries to update a base table using
        an updatable view and this view was created as a join for which
        the clause 'WITH CHECK OPTION' was specified.
      
        The reason for the bug was that when such an update was
        executed, row positions were not properly handled for tables
        that were not updated but had constraints that had to be
        checked due to the 'WITH CHECK OPTION' clause.
      
        The reason for the bug was that when such update is executed
        then for tables specified in the view definition and
        also listed in the 'WITH CHECK OPTION' clause the positioning to
        row being updated is not performed.
      12eb6d88
  21. 16 Apr, 2014 1 commit
  22. 15 Apr, 2014 1 commit
  23. 14 Apr, 2014 1 commit
    • Olivier Bertrand's avatar
      - In info, the file length sometimes could not be caculated because the · 213ecbbb
      Olivier Bertrand authored
        catalog data path had not been set. This was added into ha_connect::info.
      modified:
        storage/connect/ha_connect.cc
      
      - All the functions querying table options could return information from the wrong
        table when several CONNECT tables were used in the same query (for instance joined
        together) This was because they belonged to the catalog class that is shared between
        all tables in the same query. They have been moved from the catalog class to the
        TABDEF/RELDEF class that is attached to each table. This was a major potential bug.
      modified:
        storage/connect/catalog.h
        storage/connect/filamvct.cpp
        storage/connect/filamzip.cpp
        storage/connect/mycat.cc
        storage/connect/mycat.h
        storage/connect/reldef.cpp
        storage/connect/reldef.h
        storage/connect/tabdos.cpp
        storage/connect/tabfmt.cpp
        storage/connect/tabmul.cpp
        storage/connect/tabmysql.cpp
        storage/connect/taboccur.cpp
        storage/connect/tabodbc.cpp
        storage/connect/tabpivot.cpp
        storage/connect/tabsys.cpp
        storage/connect/tabtbl.cpp
        storage/connect/tabutil.cpp
        storage/connect/tabvct.cpp
        storage/connect/tabwmi.cpp
        storage/connect/tabxcl.cpp
        storage/connect/tabxml.cpp
        storage/connect/xindex.cpp
      
      - Prepare indexing of MYSQL/ODBC tables (as does FEDERATED) (Not implemented yet)
      modified:
        storage/connect/ha_connect.cc
        storage/connect/ha_connect.h
        storage/connect/mycat.cc
        storage/connect/mycat.h
      
      - Typo
      modified:
        storage/connect/plgdbutl.cpp
      213ecbbb
  24. 11 Apr, 2014 1 commit
    • Sergey Petrunya's avatar
      MDEV-6081: ORDER BY+ref(const): selectivity is very incorrect (MySQL Bug#14338686) · 244d4b53
      Sergey Petrunya authored
      Add a testcase and backport this fix:
      
      Bug#14338686: MYSQL IS GENERATING DIFFERENT AND SLOWER
                    (IN NEWER VERSIONS) EXECUTION PLAN
      PROBLEM:
      While checking for an index to sort for the order by clause
      in this query
      "SELECT datestamp FROM contractStatusHistory WHERE
      contract_id = contracts.id ORDER BY datestamp asc limit 1;"
      
      we do not calculate the number of rows to be examined correctly.
      As a result we choose index 'idx_contractStatusHistory_datestamp'
      defined on the 'datestamp' field, rather than choosing index
      'contract_id'. And hence the lower performance.
      
      ANALYSIS:
      While checking if an index is present to give the records in
      sorted order(datestamp), we consider the selectivity of the
      'ref_key'(contract_id here) using 'table->quick_condition_rows'.
      'ref_key' here can be an index from 'REF_ACCESS' or from 'RANGE'.
      
      As this is a 'REF_ACCESS', 'table->quick_condition_rows' is not
      set to the actual value which is 2. Instead is set to the number
      of tuples present in the table indicating that every row that
      is selected would be satisfying the condition present in the query.
      
      Hence, the selectivity becomes 1 even when we choose the index
      on the order by column instead of the join_condition.
      
      But, in reality as only 2 rows satisy the condition, we need to
      examine half of the entire data set to get one tuple when we
      choose index on the order by column.
      Had we chosen the 'REF_ACCESS' we would have examined only 2 tuples.
      Hence the delay in executing the query specified.
        
      FIX:
      While calculating the selectivity of the ref_key:
      For REF_ACCESS consider quick_rows[ref_key] if range 
      optimizer has an estimate for this key. Else consider 
      'rec_per_key' statistic.
      For RANGE ACCESS consider 'table->quick_condition_rows'.
      244d4b53