1) Ensure "init_db.sql" and "test_db-sql" really get built.
2) Ensure the "*.def" files with NetWare linker options get distributed to the proper directories.
netware/BUILD/compile-netware-END:
Ensure the "*.def" files are built for NetWare.
This is a backport of a 5.1 fix which may not be needed in 5.0 but cannot do any harm:
the general "link_sources" step might fall victim to a cleanup which would be fatal
just for NetWare, because of problems in the ordering of SUBDIR entries.
netware/Makefile.am:
1) The scripts "init_db.sql" and "test_db.sql" must be built in the NetWare phase.
2) Use "basename", not sed.
- Move increment of "i" to "increment section" of for loop
- Protect against writing after end of "buff"(backport from 5.1)
sql/sql_show.cc:
- Move the increment of i to "increment" section of for loop. Since "i"
was initially 0 the for loop exited immediately
- Add protection for writing after end of "buff"
since this flag was explicitly removed in pushbuild for GCOV builds.
BUILD_CMD => ['sh', '-c', 'perl -i.bak -pe "s/ \\\\\$static_link//" ' .
'BUILD/compile-pentium-gcov; BUILD/compile-pentium-gcov'],
Moving $static_link to SETUP.sh broke this, and is now fixed.
Should this flag be needed on some platforms,
the proper location is compile-<platform>-gcov
Tested the amd64 and pentium64 build fine without it, and can run NDB tests.
BUILD/SETUP.sh:
Removed $static_link from GCOV builds.
1) We do not provide the "isam" table handler in 5.0 and up (different from "myisam" !),
so we do not need the ".def" files for the "isam"-specific tools.
2) Use "basename" to get the base name of a file, not a harder-to-read sed expression.
BitKeeper/deleted/.del-isamchk.def:
Delete: netware/isamchk.def
BitKeeper/deleted/.del-isamlog.def:
Delete: netware/isamlog.def
BitKeeper/deleted/.del-pack_isam.def:
Delete: netware/pack_isam.def
netware/Makefile.am:
Use a plain "basename" showing the purpose, not a sed command which is harder to read.
This is a performance bug, related to the parsing or 'OR' and 'AND' boolean
expressions.
Let N be the number of expressions involved in a OR (respectively AND).
When N=1
For example, "select 1" involve only 1 term: there is no OR operator.
In 4.0 and 4.1, parsing expressions not involving OR had no overhead.
In 5.0, parsing adds some overhead, with Select->expr_list.
With this patch, the overhead introduced in 5.0 has been removed,
so that performances for N=1 should be identical to the 4.0 performances,
which are optimal (there is no code executed at all)
The overhead in 5.0 was in fact affecting significantly some operations.
For example, loading 1 Million rows into a table with INSERTs,
for a table that has 100 columns, leads to parsing 100 Millions of
expressions, which means that the overhead related to Select->expr_list
is executed 100 Million times ...
Considering that N=1 is by far the most probable expression,
this case should be optimal.
When N=2
For example, "select a OR b" involves 2 terms in the OR operator.
In 4.0 and 4.1, parsing expressions involving 2 terms created 1 Item_cond_or
node, which is the expected result.
In 5.0, parsing these expression also produced 1 node, but with some extra
overhead related to Select->expr_list : creating 1 list in Select->expr_list
and another in Item_cond::list is inefficient.
With this patch, the overhead introduced in 5.0 has been removed
so that performances for N=2 should be identical to the 4.0 performances.
Note that the memory allocation uses the new (thd->mem_root) syntax
directly.
The cost of "is_cond_or" is estimated to be neglectable: the real problem
of the performance degradation comes from unneeded memory allocations.
When N>=3
For example, "select a OR b OR c ...", which involves 3 or more terms.
In 4.0 and 4.1, the parser had no significant cost overhead, but produced
an Item tree which is difficult to evaluate / optimize during runtime.
In 5.0, the parser produces a better Item tree, using the Item_cond
constructor that accepts a list of children directly, but at an extra cost
related to Select->expr_list.
With this patch, the code is implemented to take the best of the two
implementations:
- there is no overhead with Select->expr_list
- the Item tree generated is optimized and flattened.
This is achieved by adding children nodes into the Item tree directly,
with Item_cond::add(), which avoids the need for temporary lists and memory
allocation
Note that this patch also provide an extra optimization, that the previous
code in 5.0 did not provide: expressions are flattened in the Item tree,
based on what the expression already parsed is, and not based on the order
in which rules are reduced.
For example : "(a OR b) OR c", "a OR (b OR c)" would both be represented
with 2 Item_cond_or nodes before this patch, and with 1 node only with this
patch. The logic used is based on the mathematical properties of the OR
operator (it's associative), and produces a simpler tree.
sql/item_cmpfunc.h:
Improved performances for parsing boolean expressions
sql/sql_yacc.yy:
Improved performances for parsing boolean expressions
mysql-test/r/parser_precedence.result:
Added test cases to cover boolean operator precedence
mysql-test/t/parser_precedence.test:
Added test cases to cover boolean operator precedence
Killing a SELECT query with KILL QUERY or KILL CONNECTION
causes a server crash if the query cache is enabled.
Normal evaluation of a query may be interrupted by the
KILL QUERY/CONNECTION statement, in this case the mysql_execute_command
function returns TRUE, and the thd->killed flag has true value.
In this case the result of the query may
be cached incompletely (omitting call to query_cache_insert inside
the net_real_write function), and next call to query_cache_end_of_result
may lead to server crash.
Thus, the query_cache_end_of_result function has been modified to abort
query cache in the case of killed thread.
sql/sql_cache.cc:
Fixed bug #30201.
The query_cache_end_of_result function has been modified to abort query
cache in the case of query execution failure. Also this function has been
modified to remove incomplete query block.
into adventure.(none):/home/thek/Development/cpp/mysql-5.0-runtime
mysql-test/r/query_cache.result:
Auto merged
mysql-test/t/query_cache.test:
Auto merged
Although the query cache doesn't support retrieval of statements containing
column level access control, it was still possible to cache such statements
thus wasting memory.
This patch extends the access control check on the target tables to avoid
caching a statement with column level restrictions.
Views are excepted and can be cached but only retrieved by super user account.
mysql-test/t/query_cache_with_views.test:
Rename: mysql-test/t/view_query_cache.test -> mysql-test/t/query_cache_with_views.test
mysql-test/r/query_cache_with_views.result:
Rename: mysql-test/r/view_query_cache.result -> mysql-test/r/query_cache_with_views.result
mysql-test/r/query_cache.result:
Modified test case to allow caching of views
mysql-test/t/query_cache.test:
Modified test case to allow caching of views
sql/sql_cache.cc:
Allow caching of views
The server created temporary tables for filesort in the working directory
instead of the specified tmpdir directory.
sql/item.cc:
Fixed bug #30287.
The Item_field::set_field method has been modified to reset the any_privileges
flag to false in case of system temporary table. This modification prevents the
server from unnecessary checking of user privileges to access system temporary
tables.
sql/sql_select.cc:
Fixed bug #30287.
Bugfix for #29015 has been removed: TABLE_SHARE::table_name of system
temporary tables contains full path to table file basename again.
sql/sql_view.cc:
Fixed bug #30287.
Commentary has been added.
Use "SHOW TABLES FROM `db`" instead of $dbh->tables() in the
get_list_of_tables() routine.
The symptom is that, when used with recent versions of DBD::mysql,
mysqlhotcopy uses a double-qualified table name, for example:
Invalid db.table name 'test.test`.`x' at /usr/bin/mysqlhotcopy line 855.
This is caused by a change in DBD::mysql. See this diff:
http://svn.perl.org/viewcvs/modules/DBD-mysql/trunk/lib/DBD/mysql.pm?r1=9183&r2=9188
Basically, older DBD::mysql implemented a limited ->table_info method;
now the full method is implemented, and as a result DBI's ->tables()
method has access to the schema value, so it uses it.
scripts/mysqlhotcopy.sh:
Use "SHOW TABLES FROM `db`" instead of $dbh->tables() in
get_list_of_tables(). DBI's ->tables() method calls
"SHOW TABLES" under the hood, and then quotes the
names and (depending on DBD::mysql version) adds the schema
name, too. mysqlhotcopy doesn't want the schema name
or the quotes, so it's all around better to just call
"SHOW TABLES" itself.
to find "ar" but if it cannot be found sets AR=false. This
leads to confusing failures during the build rather than at
configure time.
We have our own checks for ar, but as AR was already set
earlier by the libtool tests they were never exectuted.
Therefore, update the tests so that we catch any libtool
failures, and run AC_CHECK_PROG explicitly to ensure that we
see sensible output from configure prior to any potential
failure.
config/ac-macros/misc.m4:
AC_CHECK_PROG does nothing if the first argument (variable) is
already set, so remove redundant check for the empty string.
This also ensures that a 'checking for ar' message is printed
prior to any failure, which is far less confusing.