- 06 May, 2014 3 commits
-
-
Sergei Golubchik authored
-
Sergei Golubchik authored
XtraDB 5.6.16-64.2
-
Sergei Golubchik authored
InnoDB 5.6.16
-
- 05 May, 2014 5 commits
-
-
Sergei Golubchik authored
revno: 5716 committer: Praveenkumar Hulakund <praveenkumar.hulakund@oracle.com> branch nick: mysql_5_6 timestamp: Sat 2013-12-28 22:08:40 +0530 message: Bug#17862905: MYSQLDUMP CREATES USELESS METADATA LOCKS
-
Sergei Golubchik authored
-
Sergei Golubchik authored
-
Sergei Golubchik authored
-
Sergei Golubchik authored
MDEV-6095 replicate- filters for slaves with a connection name that contain an underscore are ignored Revert revision-id: guilhem@mysql.com-20091118203617-3jfudef971r7bf2y Implement a non-destructive fix instead.
-
- 04 May, 2014 1 commit
-
-
Sergei Golubchik authored
-
- 01 May, 2014 7 commits
-
-
Sergei Golubchik authored
-
Sergei Golubchik authored
-
Sergei Golubchik authored
various places in the code.
-
Sergei Golubchik authored
-
Sergei Golubchik authored
the linker script.
-
Sergei Golubchik authored
-
Sergei Golubchik authored
correct the length encoding for the extra2 frm segment. add the comment.
-
- 23 Apr, 2014 1 commit
-
-
Sergey Vojtovich authored
I_S table mdl_iterate() helper function (which is used by the plugin to iterate mdl locks) acquired mutexes in reverse order. Fixed by iterating MDL locks in two stages: 1. Iterate locks hash under the protection of hash mutex, store all lock pointers in a thread local array and increment reference counter for the lock. 2. Iterate local array without protection of hash mutex, handle destroyed locks. It somewhat echoes hack in MDL_map_partition::move_from_hash_to_lock_mutex.
-
- 05 May, 2014 1 commit
-
-
Sergey Petrunya authored
- Use floating-point division in selectivity calculations.
-
- 02 May, 2014 1 commit
-
-
Igor Babaev authored
The problem appeared when a loose scan used a key prefix whose last component called uint3korr in the implementation of the key_cmp virtual function.
-
- 29 Apr, 2014 1 commit
-
-
Sergey Petrunya authored
-
- 25 Apr, 2014 2 commits
-
-
Sergey Petrunya authored
- Testcase. The bug is fixed by commit for MDEV-6003
-
Sergey Petrunya authored
- Fix table_cond_selectivity() to work correctly for ref access and "keypart2=const" case.
-
- 28 Apr, 2014 1 commit
-
-
Sergey Petrunya authored
MDEV-5980: EITS: if condition is used for REF access, its selectivity is still in filtered% MDEV-5985: EITS: selectivity estimates look illogical for join and non-key equalities MDEV-6003: EITS: ref access, keypart2=const vs keypart2=expr - inconsistent filtered% value - Made a number of fixes in table_cond_selectivity() so that it returns correct selectivity estimates. - Added comments in related code. Better comments
-
- 25 Apr, 2014 1 commit
-
-
Sergey Petrunya authored
-
- 21 Apr, 2014 1 commit
-
-
Sergey Petrunya authored
MDEV-5985: EITS: selectivity estimates look illogical for join and non-key equalities MDEV-6003: EITS: ref access, keypart2=const vs keypart2=expr - inconsistent filtered% value - Made a number of fixes in table_cond_selectivity() so that it returns correct selectivity estimates. - Added comments in related code.
-
- 27 Apr, 2014 1 commit
-
-
Elena Stepanova authored
Added a test, the problem itself was fixed in MDEV-6068
-
- 25 Apr, 2014 2 commits
-
-
Alexander Barkov authored
-
unknown authored
The previous patch for this bug was unfortunately completely wrong. The purpose of cached_charset is to remember which character set we have installed currently in the THD, so that in the common case where charset does not change between queries, we do not need to update it in the THD. Thus, it is important that the cached_charset field is tightly coupled to the THD for which it handles caching. Thus the right place to put cached_charset seems to be in the THD. This patch introduces a field THD:system_thread_info where such info in general can be placed without further inflating the THD with unused data for other threads (THD is already far too big as it is). It then moves the cached_charset into this slot for the SQL driver thread and for the parallel replication worker threads. The THD::rpl_filter field is also moved inside system_thread_info, to keep the size of THD unchanged. Moving further fields in to reduce the size of THD is a separate task, filed as MDEV-6164.
-
- 24 Apr, 2014 3 commits
-
-
Alexander Barkov authored
-
unknown authored
The problem is the async binlog checkpointing; this could on rare occasions occur too late, causing SHOW BINLOG EVENTS to show the wrong events and cause .result file difference.
-
Alexander Barkov authored
-
- 23 Apr, 2014 2 commits
-
-
unknown authored
Replication caches the character sets used in a query, to be able to quickly reuse them for the next query in the common case of them not having changed. In parallel replication, this caching needs to be per-worker-thread. The code was not modified to handle this correctly, so the caching in one worker could cause another worker to run a query using the wrong character set, causing replication corruption.
-
Alexander Barkov authored
Added a new system variable: default_regex_flags='DOTALL,DUPNAMES,EXTENDED,EXTRA,MULTILINE,UNGREEDY'
-
- 22 Apr, 2014 3 commits
-
-
Igor Babaev authored
-
Igor Babaev authored
Back-ported from the mysql 5.6 code line the patch with the following comment: Fix for Bug#11757108 CHANGE IN EXECUTION PLAN FOR COUNT_DISTINCT_GROUP_ON_KEY CAUSES PEFORMANCE REGRESSION The cause for the performance regression is that the access strategy for the GROUP BY query is changed form using "index scan" in mysql-5.1 to use "loose index scan" in mysql-5.5. The index used for group by is unique and thus each "loose scan" group will only contain one record. Since loose scan needs to re-position on each "loose scan" group this query will do a re-position for each index entry. Compared to just reading the next index entry as a normal index scan does, the use of loose scan for this query becomes more expensive. The cause for selecting to use loose scan for this query is that in the current code when the size of the "loose scan" group is one, the formula for calculating the cost estimates becomes almost identical to the cost of using normal index scan. Differences in use of integer versus floating point arithmetic can cause one or the other access strategy to be selected. The main issue with the formula for estimating the cost of using loose scan is that it does not take into account that it is more costly to do a re-position for each "loose scan" group compared to just reading the next index entry. Both index scan and loose scan estimates the cpu cost as: "number of entries needed too read/scan" * ROW_EVALUATE_COST The results from testing with the query in this bug indicates that the real cost for doing re-position four to eight times higher than just reading the next index entry. Thus, the cpu cost estimate for loose scan should be increased. To account for the extra work to re-position in the index we increase the cost for loose index scan to include the cost of navigating the index. This is modelled as a function of the height of the b-tree: navigation cost= ceil(log(records in table)/log(indexes per block)) * ROWID_COMPARE_COST; This will avoid loose index scan being used for indexes where the "loose scan" group contains very few index entries.
-
Alexander Barkov authored
-
- 21 Apr, 2014 2 commits
-
-
Sergei Golubchik authored
-
Alexander Barkov authored
-
- 19 Apr, 2014 1 commit
-
-
Olivier Bertrand authored
Not yet done for ODBC tables. modified: storage/connect/connect.cc storage/connect/ha_connect.cc storage/connect/ha_connect.h storage/connect/mycat.cc storage/connect/plgdbsem.h storage/connect/reldef.h storage/connect/tabdos.h storage/connect/tabmysql.cpp storage/connect/tabmysql.h storage/connect/tabodbc.cpp storage/connect/tabodbc.h storage/connect/xindex.cpp storage/connect/xtable.h - Return error in "info" on Cardinality error. modified: storage/connect/ha_connect.cc
-
- 18 Apr, 2014 1 commit
-
-
Sergey Petrunya authored
- Backport the fix for MySQL Bug#13947868 - Add our testcase (they don't publish theirs)
-