1
0
mirror of https://github.com/MariaDB/server.git synced 2025-07-29 05:21:33 +03:00

Fixed bug when joining with caching.

Fixed race condition when using the binary log and INSERT DELAYED which could cause the binary log to have rows that was not yet written to MyISAM tables.


Docs/manual.texi:
  Changelog
mysql-test/r/null_key.result:
  Fix of testcase after changing optimizer to only use range keys if it would use a long part of the SAME key.
sql/sql_insert.cc:
  Fixed race condition with binary log and INSERT DELAYED
sql/sql_select.cc:
  Fixed bug when joining with caching
This commit is contained in:
unknown
2001-11-27 02:50:20 +02:00
parent 0c9d051c16
commit 312e5e82c0
4 changed files with 65 additions and 44 deletions

View File

@ -46785,7 +46785,7 @@ users use this code as the rest of the code and because of this we are
not yet 100% confident in this code. not yet 100% confident in this code.
@menu @menu
* News-3.23.46:: * News-3.23.46:: Changes in release 3.23.46
* News-3.23.45:: Changes in release 3.23.45 * News-3.23.45:: Changes in release 3.23.45
* News-3.23.44:: Changes in release 3.23.44 * News-3.23.44:: Changes in release 3.23.44
* News-3.23.43:: Changes in release 3.23.43 * News-3.23.43:: Changes in release 3.23.43
@ -46846,6 +46846,12 @@ One can now kill @code{ANALYZE},@code{REPAIR} and @code{OPTIMIZE TABLE} when
the thread is waiting to get a lock on the table. the thread is waiting to get a lock on the table.
@item @item
Fixed race condition in @code{ANALYZE TABLE}. Fixed race condition in @code{ANALYZE TABLE}.
@item
Fixed bug when joining with caching (unlikely to happen).
@item
Fixed race condition when using the binary log and @code{INSERT DELAYED}
which could cause the binary log to have rows that was not yet written
to MyISAM tables.
@end itemize @end itemize
@node News-3.23.45, News-3.23.44, News-3.23.46, News-3.23.x @node News-3.23.45, News-3.23.44, News-3.23.46, News-3.23.x

View File

@ -11,7 +11,7 @@ t1 index NULL a 9 NULL 12 where used; Using index
table type possible_keys key key_len ref rows Extra table type possible_keys key key_len ref rows Extra
t1 range a,b a 9 NULL 3 where used; Using index t1 range a,b a 9 NULL 3 where used; Using index
table type possible_keys key key_len ref rows Extra table type possible_keys key key_len ref rows Extra
t1 range a,b a 9 NULL 2 where used; Using index t1 ref a,b b 4 const 2 where used
table type possible_keys key key_len ref rows Extra table type possible_keys key key_len ref rows Extra
t1 ref a,b a 5 const 3 where used; Using index t1 ref a,b a 5 const 3 where used; Using index
table type possible_keys key key_len ref rows Extra table type possible_keys key key_len ref rows Extra

View File

@ -1099,7 +1099,7 @@ bool delayed_insert::handle_inserts(void)
{ {
int error; int error;
uint max_rows; uint max_rows;
bool using_ignore=0; bool using_ignore=0, using_bin_log=mysql_bin_log.is_open();
delayed_row *row; delayed_row *row;
DBUG_ENTER("handle_inserts"); DBUG_ENTER("handle_inserts");
@ -1124,7 +1124,13 @@ bool delayed_insert::handle_inserts(void)
max_rows= ~0; // Do as much as possible max_rows= ~0; // Do as much as possible
} }
table->file->extra(HA_EXTRA_WRITE_CACHE); /*
We can't use row caching when using the binary log because if
we get a crash, then binary log will contain rows that are not yet
written to disk, which will cause problems in replication.
*/
if (!using_bin_log)
table->file->extra(HA_EXTRA_WRITE_CACHE);
pthread_mutex_lock(&mutex); pthread_mutex_lock(&mutex);
while ((row=rows.get())) while ((row=rows.get()))
{ {
@ -1161,7 +1167,7 @@ bool delayed_insert::handle_inserts(void)
if (row->query && row->log_query) if (row->query && row->log_query)
{ {
mysql_update_log.write(&thd,row->query, row->query_length); mysql_update_log.write(&thd,row->query, row->query_length);
if (mysql_bin_log.is_open()) if (using_bin_log)
{ {
thd.query_length = row->query_length; thd.query_length = row->query_length;
Query_log_event qinfo(&thd, row->query); Query_log_event qinfo(&thd, row->query);
@ -1197,7 +1203,8 @@ bool delayed_insert::handle_inserts(void)
/* This should never happen */ /* This should never happen */
sql_print_error(ER(ER_DELAYED_CANT_CHANGE_LOCK),table->real_name); sql_print_error(ER(ER_DELAYED_CANT_CHANGE_LOCK),table->real_name);
} }
table->file->extra(HA_EXTRA_WRITE_CACHE); if (!using_bin_log)
table->file->extra(HA_EXTRA_WRITE_CACHE);
pthread_mutex_lock(&mutex); pthread_mutex_lock(&mutex);
thd.proc_info="insert"; thd.proc_info="insert";
} }

View File

@ -1881,52 +1881,55 @@ find_best(JOIN *join,table_map rest_tables,uint idx,double record_count,
** Find how much space the prevous read not const tables takes in cache ** Find how much space the prevous read not const tables takes in cache
*/ */
static void calc_used_field_length(THD *thd, JOIN_TAB *join_tab)
{
uint null_fields,blobs,fields,rec_length;
null_fields=blobs=fields=rec_length=0;
Field **f_ptr,*field;
for (f_ptr=join_tab->table->field ; (field= *f_ptr) ; f_ptr++)
{
if (field->query_id == thd->query_id)
{
uint flags=field->flags;
fields++;
rec_length+=field->pack_length();
if (flags & BLOB_FLAG)
blobs++;
if (!(flags & NOT_NULL_FLAG))
null_fields++;
}
}
if (null_fields)
rec_length+=(join_tab->table->null_fields+7)/8;
if (join_tab->table->maybe_null)
rec_length+=sizeof(my_bool);
if (blobs)
{
uint blob_length=(uint) (join_tab->table->file->mean_rec_length-
(join_tab->table->reclength- rec_length));
rec_length+=(uint) max(4,blob_length);
}
join_tab->used_fields=fields;
join_tab->used_fieldlength=rec_length;
join_tab->used_blobs=blobs;
}
static uint static uint
cache_record_length(JOIN *join,uint idx) cache_record_length(JOIN *join,uint idx)
{ {
uint length; uint length=0;
JOIN_TAB **pos,**end; JOIN_TAB **pos,**end;
THD *thd=join->thd; THD *thd=join->thd;
length=0;
for (pos=join->best_ref+join->const_tables,end=join->best_ref+idx ; for (pos=join->best_ref+join->const_tables,end=join->best_ref+idx ;
pos != end ; pos != end ;
pos++) pos++)
{ {
JOIN_TAB *join_tab= *pos; JOIN_TAB *join_tab= *pos;
if (!join_tab->used_fieldlength) if (!join_tab->used_fieldlength) /* Not calced yet */
{ /* Not calced yet */ calc_used_field_length(thd, join_tab);
uint null_fields,blobs,fields,rec_length;
null_fields=blobs=fields=rec_length=0;
Field **f_ptr,*field;
for (f_ptr=join_tab->table->field ; (field= *f_ptr) ; f_ptr++)
{
if (field->query_id == thd->query_id)
{
uint flags=field->flags;
fields++;
rec_length+=field->pack_length();
if (flags & BLOB_FLAG)
blobs++;
if (!(flags & NOT_NULL_FLAG))
null_fields++;
}
}
if (null_fields)
rec_length+=(join_tab->table->null_fields+7)/8;
if (join_tab->table->maybe_null)
rec_length+=sizeof(my_bool);
if (blobs)
{
uint blob_length=(uint) (join_tab->table->file->mean_rec_length-
(join_tab->table->reclength- rec_length));
rec_length+=(uint) max(4,blob_length);
}
join_tab->used_fields=fields;
join_tab->used_fieldlength=rec_length;
join_tab->used_blobs=blobs;
}
length+=join_tab->used_fieldlength; length+=join_tab->used_fieldlength;
} }
return length; return length;
@ -2248,6 +2251,7 @@ make_join_select(JOIN *join,SQL_SELECT *select,COND *cond)
used_tables|=current_map; used_tables|=current_map;
if (tab->type == JT_REF && tab->quick && if (tab->type == JT_REF && tab->quick &&
tab->ref.key == tab->quick->index &&
tab->ref.key_length < tab->quick->max_used_key_length) tab->ref.key_length < tab->quick->max_used_key_length)
{ {
/* Range uses longer key; Use this instead of ref on key */ /* Range uses longer key; Use this instead of ref on key */
@ -5631,15 +5635,19 @@ join_init_cache(THD *thd,JOIN_TAB *tables,uint table_count)
uint length,blobs,size; uint length,blobs,size;
CACHE_FIELD *copy,**blob_ptr; CACHE_FIELD *copy,**blob_ptr;
JOIN_CACHE *cache; JOIN_CACHE *cache;
JOIN_TAB *join_tab;
DBUG_ENTER("join_init_cache"); DBUG_ENTER("join_init_cache");
cache= &tables[table_count].cache; cache= &tables[table_count].cache;
cache->fields=blobs=0; cache->fields=blobs=0;
for (i=0 ; i < table_count ; i++) join_tab=tables;
for (i=0 ; i < table_count ; i++,join_tab++)
{ {
cache->fields+=tables[i].used_fields; if (!join_tab->used_fieldlength) /* Not calced yet */
blobs+=tables[i].used_blobs; calc_used_field_length(thd, join_tab);
cache->fields+=join_tab->used_fields;
blobs+=join_tab->used_blobs;
} }
if (!(cache->field=(CACHE_FIELD*) if (!(cache->field=(CACHE_FIELD*)
sql_alloc(sizeof(CACHE_FIELD)*(cache->fields+table_count*2)+(blobs+1)* sql_alloc(sizeof(CACHE_FIELD)*(cache->fields+table_count*2)+(blobs+1)*