mirror of
				https://github.com/MariaDB/server.git
				synced 2025-10-28 17:15:19 +03:00 
			
		
		
		
	- Changed to still use bcmp() in certain cases becasue - Faster for short unaligneed strings than memcmp() - Bettern when using valgrind - Changed to use my_sprintf() instead of sprintf() to get higher portability for old systems - Changed code to use MariaDB version of select->skip_record() - Removed -%::SCCS/s.% from Makefile.am:s to remove automake warnings
		
			
				
	
	
		
			1943 lines
		
	
	
		
			58 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1943 lines
		
	
	
		
			58 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* Copyright (C) 2000 MySQL AB
 | |
| 
 | |
|    This program is free software; you can redistribute it and/or modify
 | |
|    it under the terms of the GNU General Public License as published by
 | |
|    the Free Software Foundation; version 2 of the License.
 | |
| 
 | |
|    This program is distributed in the hope that it will be useful,
 | |
|    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|    GNU General Public License for more details.
 | |
| 
 | |
|    You should have received a copy of the GNU General Public License
 | |
|    along with this program; if not, write to the Free Software
 | |
|    Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA */
 | |
| 
 | |
| /*
 | |
|   Cashing of files with only does (sequential) read or writes of fixed-
 | |
|   length records. A read isn't allowed to go over file-length. A read is ok
 | |
|   if it ends at file-length and next read can try to read after file-length
 | |
|   (and get a EOF-error).
 | |
|   Possibly use of asyncronic io.
 | |
|   macros for read and writes for faster io.
 | |
|   Used instead of FILE when reading or writing whole files.
 | |
|   This code makes mf_rec_cache obsolete (currently only used by ISAM)
 | |
|   One can change info->pos_in_file to a higher value to skip bytes in file if
 | |
|   also info->read_pos is set to info->read_end.
 | |
|   If called through open_cached_file(), then the temporary file will
 | |
|   only be created if a write exeeds the file buffer or if one calls
 | |
|   my_b_flush_io_cache().
 | |
| 
 | |
|   If one uses SEQ_READ_APPEND, then two buffers are allocated, one for
 | |
|   reading and another for writing.  Reads are first done from disk and
 | |
|   then done from the write buffer.  This is an efficient way to read
 | |
|   from a log file when one is writing to it at the same time.
 | |
|   For this to work, the file has to be opened in append mode!
 | |
|   Note that when one uses SEQ_READ_APPEND, one MUST write using
 | |
|   my_b_append !  This is needed because we need to lock the mutex
 | |
|   every time we access the write buffer.
 | |
| 
 | |
| TODO:
 | |
|   When one SEQ_READ_APPEND and we are reading and writing at the same time,
 | |
|   each time the write buffer gets full and it's written to disk, we will
 | |
|   always do a disk read to read a part of the buffer from disk to the
 | |
|   read buffer.
 | |
|   This should be fixed so that when we do a my_b_flush_io_cache() and
 | |
|   we have been reading the write buffer, we should transfer the rest of the
 | |
|   write buffer to the read buffer before we start to reuse it.
 | |
| */
 | |
| 
 | |
| #define MAP_TO_USE_RAID
 | |
| #include "mysys_priv.h"
 | |
| #include <m_string.h>
 | |
| #ifdef HAVE_AIOWAIT
 | |
| #include "mysys_err.h"
 | |
| static void my_aiowait(my_aio_result *result);
 | |
| #endif
 | |
| #include <errno.h>
 | |
| 
 | |
| #ifdef THREAD
 | |
| #define lock_append_buffer(info) \
 | |
|  pthread_mutex_lock(&(info)->append_buffer_lock)
 | |
| #define unlock_append_buffer(info) \
 | |
|  pthread_mutex_unlock(&(info)->append_buffer_lock)
 | |
| #else
 | |
| #define lock_append_buffer(info)
 | |
| #define unlock_append_buffer(info)
 | |
| #endif
 | |
| 
 | |
| #define IO_ROUND_UP(X) (((X)+IO_SIZE-1) & ~(IO_SIZE-1))
 | |
| #define IO_ROUND_DN(X) ( (X)            & ~(IO_SIZE-1))
 | |
| 
 | |
| /*
 | |
|   Setup internal pointers inside IO_CACHE
 | |
| 
 | |
|   SYNOPSIS
 | |
|     setup_io_cache()
 | |
|     info		IO_CACHE handler
 | |
| 
 | |
|   NOTES
 | |
|     This is called on automaticly on init or reinit of IO_CACHE
 | |
|     It must be called externally if one moves or copies an IO_CACHE
 | |
|     object.
 | |
| */
 | |
| 
 | |
| void setup_io_cache(IO_CACHE* info)
 | |
| {
 | |
|   /* Ensure that my_b_tell() and my_b_bytes_in_cache works */
 | |
|   if (info->type == WRITE_CACHE)
 | |
|   {
 | |
|     info->current_pos= &info->write_pos;
 | |
|     info->current_end= &info->write_end;
 | |
|   }
 | |
|   else
 | |
|   {
 | |
|     info->current_pos= &info->read_pos;
 | |
|     info->current_end= &info->read_end;
 | |
|   }
 | |
| }
 | |
| 
 | |
| 
 | |
| static void
 | |
| init_functions(IO_CACHE* info)
 | |
| {
 | |
|   enum cache_type type= info->type;
 | |
|   switch (type) {
 | |
|   case READ_NET:
 | |
|     /*
 | |
|       Must be initialized by the caller. The problem is that
 | |
|       _my_b_net_read has to be defined in sql directory because of
 | |
|       the dependency on THD, and therefore cannot be visible to
 | |
|       programs that link against mysys but know nothing about THD, such
 | |
|       as myisamchk
 | |
|     */
 | |
|     break;
 | |
|   case SEQ_READ_APPEND:
 | |
|     info->read_function = _my_b_seq_read;
 | |
|     info->write_function = 0;			/* Force a core if used */
 | |
|     break;
 | |
|   default:
 | |
|     info->read_function =
 | |
| #ifdef THREAD
 | |
|                           info->share ? _my_b_read_r :
 | |
| #endif
 | |
|                                         _my_b_read;
 | |
|     info->write_function = _my_b_write;
 | |
|   }
 | |
| 
 | |
|   setup_io_cache(info);
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Initialize an IO_CACHE object
 | |
| 
 | |
|   SYNOPSOS
 | |
|     init_io_cache()
 | |
|     info		cache handler to initialize
 | |
|     file		File that should be associated to to the handler
 | |
| 			If == -1 then real_open_cached_file()
 | |
| 			will be called when it's time to open file.
 | |
|     cachesize		Size of buffer to allocate for read/write
 | |
| 			If == 0 then use my_default_record_cache_size
 | |
|     type		Type of cache
 | |
|     seek_offset		Where cache should start reading/writing
 | |
|     use_async_io	Set to 1 of we should use async_io (if avaiable)
 | |
|     cache_myflags	Bitmap of differnt flags
 | |
| 			MY_WME | MY_FAE | MY_NABP | MY_FNABP |
 | |
| 			MY_DONT_CHECK_FILESIZE
 | |
| 
 | |
|   RETURN
 | |
|     0  ok
 | |
|     #  error
 | |
| */
 | |
| 
 | |
| int init_io_cache(IO_CACHE *info, File file, size_t cachesize,
 | |
| 		  enum cache_type type, my_off_t seek_offset,
 | |
| 		  pbool use_async_io, myf cache_myflags)
 | |
| {
 | |
|   size_t min_cache;
 | |
|   my_off_t pos;
 | |
|   my_off_t end_of_file= ~(my_off_t) 0;
 | |
|   DBUG_ENTER("init_io_cache");
 | |
|   DBUG_PRINT("enter",("cache: 0x%lx  type: %d  pos: %ld",
 | |
| 		      (ulong) info, (int) type, (ulong) seek_offset));
 | |
| 
 | |
|   info->file= file;
 | |
|   info->type= TYPE_NOT_SET;	    /* Don't set it until mutex are created */
 | |
|   info->pos_in_file= seek_offset;
 | |
|   info->pre_close = info->pre_read = info->post_read = 0;
 | |
|   info->arg = 0;
 | |
|   info->alloced_buffer = 0;
 | |
|   info->buffer=0;
 | |
|   info->seek_not_done= 0;
 | |
| 
 | |
|   if (file >= 0)
 | |
|   {
 | |
|     pos= my_tell(file, MYF(0));
 | |
|     if ((pos == (my_off_t) -1) && (my_errno == ESPIPE))
 | |
|     {
 | |
|       /*
 | |
|          This kind of object doesn't support seek() or tell(). Don't set a
 | |
|          flag that will make us again try to seek() later and fail.
 | |
|       */
 | |
|       info->seek_not_done= 0;
 | |
|       /*
 | |
|         Additionally, if we're supposed to start somewhere other than the
 | |
|         the beginning of whatever this file is, then somebody made a bad
 | |
|         assumption.
 | |
|       */
 | |
|       DBUG_ASSERT(seek_offset == 0);
 | |
|     }
 | |
|     else
 | |
|       info->seek_not_done= test(seek_offset != pos);
 | |
|   }
 | |
| 
 | |
|   info->disk_writes= 0;
 | |
| #ifdef THREAD
 | |
|   info->share=0;
 | |
| #endif
 | |
| 
 | |
|   if (!cachesize && !(cachesize= my_default_record_cache_size))
 | |
|     DBUG_RETURN(1);				/* No cache requested */
 | |
|   min_cache=use_async_io ? IO_SIZE*4 : IO_SIZE*2;
 | |
|   if (type == READ_CACHE || type == SEQ_READ_APPEND)
 | |
|   {						/* Assume file isn't growing */
 | |
|     if (!(cache_myflags & MY_DONT_CHECK_FILESIZE))
 | |
|     {
 | |
|       /* Calculate end of file to avoid allocating oversized buffers */
 | |
|       end_of_file=my_seek(file,0L,MY_SEEK_END,MYF(0));
 | |
|       /* Need to reset seek_not_done now that we just did a seek. */
 | |
|       info->seek_not_done= end_of_file == seek_offset ? 0 : 1;
 | |
|       if (end_of_file < seek_offset)
 | |
| 	end_of_file=seek_offset;
 | |
|       /* Trim cache size if the file is very small */
 | |
|       if ((my_off_t) cachesize > end_of_file-seek_offset+IO_SIZE*2-1)
 | |
|       {
 | |
| 	cachesize= (size_t) (end_of_file-seek_offset)+IO_SIZE*2-1;
 | |
| 	use_async_io=0;				/* No need to use async */
 | |
|       }
 | |
|     }
 | |
|   }
 | |
|   cache_myflags &= ~MY_DONT_CHECK_FILESIZE;
 | |
|   if (type != READ_NET && type != WRITE_NET)
 | |
|   {
 | |
|     /* Retry allocating memory in smaller blocks until we get one */
 | |
|     cachesize= ((cachesize + min_cache-1) & ~(min_cache-1));
 | |
|     for (;;)
 | |
|     {
 | |
|       size_t buffer_block;
 | |
|       /*
 | |
|         Unset MY_WAIT_IF_FULL bit if it is set, to prevent conflict with
 | |
|         MY_ZEROFILL.
 | |
|       */
 | |
|       myf flags= (myf) (cache_myflags & ~(MY_WME | MY_WAIT_IF_FULL));
 | |
| 
 | |
|       if (cachesize < min_cache)
 | |
| 	cachesize = min_cache;
 | |
|       buffer_block= cachesize;
 | |
|       if (type == SEQ_READ_APPEND)
 | |
| 	buffer_block *= 2;
 | |
|       if (cachesize == min_cache)
 | |
|         flags|= (myf) MY_WME;
 | |
| 
 | |
|       if ((info->buffer= (uchar*) my_malloc(buffer_block, flags)) != 0)
 | |
|       {
 | |
| 	info->write_buffer=info->buffer;
 | |
| 	if (type == SEQ_READ_APPEND)
 | |
| 	  info->write_buffer = info->buffer + cachesize;
 | |
| 	info->alloced_buffer=1;
 | |
| 	break;					/* Enough memory found */
 | |
|       }
 | |
|       if (cachesize == min_cache)
 | |
| 	DBUG_RETURN(2);				/* Can't alloc cache */
 | |
|       /* Try with less memory */
 | |
|       cachesize= (cachesize*3/4 & ~(min_cache-1));
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   DBUG_PRINT("info",("init_io_cache: cachesize = %lu", (ulong) cachesize));
 | |
|   info->read_length=info->buffer_length=cachesize;
 | |
|   info->myflags=cache_myflags & ~(MY_NABP | MY_FNABP);
 | |
|   info->request_pos= info->read_pos= info->write_pos = info->buffer;
 | |
|   if (type == SEQ_READ_APPEND)
 | |
|   {
 | |
|     info->append_read_pos = info->write_pos = info->write_buffer;
 | |
|     info->write_end = info->write_buffer + info->buffer_length;
 | |
| #ifdef THREAD
 | |
|     pthread_mutex_init(&info->append_buffer_lock,MY_MUTEX_INIT_FAST);
 | |
| #endif
 | |
|   }
 | |
| #if defined(SAFE_MUTEX) && defined(THREAD)
 | |
|   else
 | |
|   {
 | |
|     /* Clear mutex so that safe_mutex will notice that it's not initialized */
 | |
|     bzero((char*) &info->append_buffer_lock, sizeof(info));
 | |
|   }
 | |
| #endif
 | |
| 
 | |
|   if (type == WRITE_CACHE)
 | |
|     info->write_end=
 | |
|       info->buffer+info->buffer_length- (seek_offset & (IO_SIZE-1));
 | |
|   else
 | |
|     info->read_end=info->buffer;		/* Nothing in cache */
 | |
| 
 | |
|   /* End_of_file may be changed by user later */
 | |
|   info->end_of_file= end_of_file;
 | |
|   info->error=0;
 | |
|   info->type= type;
 | |
|   init_functions(info);
 | |
| #ifdef HAVE_AIOWAIT
 | |
|   if (use_async_io && ! my_disable_async_io)
 | |
|   {
 | |
|     DBUG_PRINT("info",("Using async io"));
 | |
|     info->read_length/=2;
 | |
|     info->read_function=_my_b_async_read;
 | |
|   }
 | |
|   info->inited=info->aio_result.pending=0;
 | |
| #endif
 | |
|   DBUG_RETURN(0);
 | |
| }						/* init_io_cache */
 | |
| 
 | |
| 	/* Wait until current request is ready */
 | |
| 
 | |
| #ifdef HAVE_AIOWAIT
 | |
| static void my_aiowait(my_aio_result *result)
 | |
| {
 | |
|   if (result->pending)
 | |
|   {
 | |
|     struct aio_result_t *tmp;
 | |
|     for (;;)
 | |
|     {
 | |
|       if ((int) (tmp=aiowait((struct timeval *) 0)) == -1)
 | |
|       {
 | |
| 	if (errno == EINTR)
 | |
| 	  continue;
 | |
| 	DBUG_PRINT("error",("No aio request, error: %d",errno));
 | |
| 	result->pending=0;			/* Assume everythings is ok */
 | |
| 	break;
 | |
|       }
 | |
|       ((my_aio_result*) tmp)->pending=0;
 | |
|       if ((my_aio_result*) tmp == result)
 | |
| 	break;
 | |
|     }
 | |
|   }
 | |
|   return;
 | |
| }
 | |
| #endif
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Use this to reset cache to re-start reading or to change the type
 | |
|   between READ_CACHE <-> WRITE_CACHE
 | |
|   If we are doing a reinit of a cache where we have the start of the file
 | |
|   in the cache, we are reusing this memory without flushing it to disk.
 | |
| */
 | |
| 
 | |
| my_bool reinit_io_cache(IO_CACHE *info, enum cache_type type,
 | |
| 			my_off_t seek_offset,
 | |
| 			pbool use_async_io __attribute__((unused)),
 | |
| 			pbool clear_cache)
 | |
| {
 | |
|   DBUG_ENTER("reinit_io_cache");
 | |
|   DBUG_PRINT("enter",("cache: 0x%lx type: %d  seek_offset: %lu  clear_cache: %d",
 | |
| 		      (ulong) info, type, (ulong) seek_offset,
 | |
| 		      (int) clear_cache));
 | |
| 
 | |
|   /* One can't do reinit with the following types */
 | |
|   DBUG_ASSERT(type != READ_NET && info->type != READ_NET &&
 | |
| 	      type != WRITE_NET && info->type != WRITE_NET &&
 | |
| 	      type != SEQ_READ_APPEND && info->type != SEQ_READ_APPEND);
 | |
| 
 | |
|   /* If the whole file is in memory, avoid flushing to disk */
 | |
|   if (! clear_cache &&
 | |
|       seek_offset >= info->pos_in_file &&
 | |
|       seek_offset <= my_b_tell(info))
 | |
|   {
 | |
|     /* Reuse current buffer without flushing it to disk */
 | |
|     uchar *pos;
 | |
|     if (info->type == WRITE_CACHE && type == READ_CACHE)
 | |
|     {
 | |
|       info->read_end=info->write_pos;
 | |
|       info->end_of_file=my_b_tell(info);
 | |
|       /*
 | |
|         Trigger a new seek only if we have a valid
 | |
|         file handle.
 | |
|       */
 | |
|       info->seek_not_done= (info->file != -1);
 | |
|     }
 | |
|     else if (type == WRITE_CACHE)
 | |
|     {
 | |
|       if (info->type == READ_CACHE)
 | |
|       {
 | |
| 	info->write_end=info->write_buffer+info->buffer_length;
 | |
| 	info->seek_not_done=1;
 | |
|       }
 | |
|       info->end_of_file = ~(my_off_t) 0;
 | |
|     }
 | |
|     pos=info->request_pos+(seek_offset-info->pos_in_file);
 | |
|     if (type == WRITE_CACHE)
 | |
|       info->write_pos=pos;
 | |
|     else
 | |
|       info->read_pos= pos;
 | |
| #ifdef HAVE_AIOWAIT
 | |
|     my_aiowait(&info->aio_result);		/* Wait for outstanding req */
 | |
| #endif
 | |
|   }
 | |
|   else
 | |
|   {
 | |
|     /*
 | |
|       If we change from WRITE_CACHE to READ_CACHE, assume that everything
 | |
|       after the current positions should be ignored
 | |
|     */
 | |
|     if (info->type == WRITE_CACHE && type == READ_CACHE)
 | |
|       info->end_of_file=my_b_tell(info);
 | |
|     /* flush cache if we want to reuse it */
 | |
|     if (!clear_cache && my_b_flush_io_cache(info,1))
 | |
|       DBUG_RETURN(1);
 | |
|     info->pos_in_file=seek_offset;
 | |
|     /* Better to do always do a seek */
 | |
|     info->seek_not_done=1;
 | |
|     info->request_pos=info->read_pos=info->write_pos=info->buffer;
 | |
|     if (type == READ_CACHE)
 | |
|     {
 | |
|       info->read_end=info->buffer;		/* Nothing in cache */
 | |
|     }
 | |
|     else
 | |
|     {
 | |
|       info->write_end=(info->buffer + info->buffer_length -
 | |
| 		       (seek_offset & (IO_SIZE-1)));
 | |
|       info->end_of_file= ~(my_off_t) 0;
 | |
|     }
 | |
|   }
 | |
|   info->type=type;
 | |
|   info->error=0;
 | |
|   init_functions(info);
 | |
| 
 | |
| #ifdef HAVE_AIOWAIT
 | |
|   if (use_async_io && ! my_disable_async_io &&
 | |
|       ((ulong) info->buffer_length <
 | |
|        (ulong) (info->end_of_file - seek_offset)))
 | |
|   {
 | |
|     info->read_length=info->buffer_length/2;
 | |
|     info->read_function=_my_b_async_read;
 | |
|   }
 | |
|   info->inited=0;
 | |
| #endif
 | |
|   DBUG_RETURN(0);
 | |
| } /* reinit_io_cache */
 | |
| 
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Read buffered.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     _my_b_read()
 | |
|       info                      IO_CACHE pointer
 | |
|       Buffer                    Buffer to retrieve count bytes from file
 | |
|       Count                     Number of bytes to read into Buffer
 | |
| 
 | |
|   NOTE
 | |
|     This function is only called from the my_b_read() macro when there
 | |
|     isn't enough characters in the buffer to satisfy the request.
 | |
| 
 | |
|   WARNING
 | |
| 
 | |
|     When changing this function, be careful with handling file offsets
 | |
|     (end-of_file, pos_in_file). Do not cast them to possibly smaller
 | |
|     types than my_off_t unless you can be sure that their value fits.
 | |
|     Same applies to differences of file offsets.
 | |
| 
 | |
|     When changing this function, check _my_b_read_r(). It might need the
 | |
|     same change.
 | |
| 
 | |
|   RETURN
 | |
|     0      we succeeded in reading all data
 | |
|     1      Error: can't read requested characters
 | |
| */
 | |
| 
 | |
| int _my_b_read(register IO_CACHE *info, uchar *Buffer, size_t Count)
 | |
| {
 | |
|   size_t length,diff_length,left_length, max_length;
 | |
|   my_off_t pos_in_file;
 | |
|   DBUG_ENTER("_my_b_read");
 | |
| 
 | |
|   if ((left_length= (size_t) (info->read_end-info->read_pos)))
 | |
|   {
 | |
|     DBUG_ASSERT(Count >= left_length);	/* User is not using my_b_read() */
 | |
|     memcpy(Buffer,info->read_pos, left_length);
 | |
|     Buffer+=left_length;
 | |
|     Count-=left_length;
 | |
|   }
 | |
| 
 | |
|   /* pos_in_file always point on where info->buffer was read */
 | |
|   pos_in_file=info->pos_in_file+ (size_t) (info->read_end - info->buffer);
 | |
| 
 | |
|   /* 
 | |
|     Whenever a function which operates on IO_CACHE flushes/writes
 | |
|     some part of the IO_CACHE to disk it will set the property
 | |
|     "seek_not_done" to indicate this to other functions operating
 | |
|     on the IO_CACHE.
 | |
|   */
 | |
|   if (info->seek_not_done)
 | |
|   {
 | |
|     if ((my_seek(info->file,pos_in_file,MY_SEEK_SET,MYF(0)) 
 | |
|         != MY_FILEPOS_ERROR))
 | |
|     {
 | |
|       /* No error, reset seek_not_done flag. */
 | |
|       info->seek_not_done= 0;
 | |
|     }
 | |
|     else
 | |
|     {
 | |
|       /*
 | |
|         If the seek failed and the error number is ESPIPE, it is because
 | |
|         info->file is a pipe or socket or FIFO.  We never should have tried
 | |
|         to seek on that.  See Bugs#25807 and #22828 for more info.
 | |
|       */
 | |
|       DBUG_ASSERT(my_errno != ESPIPE);
 | |
|       info->error= -1;
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   diff_length= (size_t) (pos_in_file & (IO_SIZE-1));
 | |
|   if (Count >= (size_t) (IO_SIZE+(IO_SIZE-diff_length)))
 | |
|   {					/* Fill first intern buffer */
 | |
|     size_t read_length;
 | |
|     if (info->end_of_file <= pos_in_file)
 | |
|     {					/* End of file */
 | |
|       info->error= (int) left_length;
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
|     length=(Count & (size_t) ~(IO_SIZE-1))-diff_length;
 | |
|     if ((read_length= my_read(info->file,Buffer, length, info->myflags))
 | |
| 	!= length)
 | |
|     {
 | |
|       info->error= (read_length == (size_t) -1 ? -1 :
 | |
| 		    (int) (read_length+left_length));
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
|     Count-=length;
 | |
|     Buffer+=length;
 | |
|     pos_in_file+=length;
 | |
|     left_length+=length;
 | |
|     diff_length=0;
 | |
|   }
 | |
| 
 | |
|   max_length= info->read_length-diff_length;
 | |
|   if (info->type != READ_FIFO &&
 | |
|       max_length > (info->end_of_file - pos_in_file))
 | |
|     max_length= (size_t) (info->end_of_file - pos_in_file);
 | |
|   if (!max_length)
 | |
|   {
 | |
|     if (Count)
 | |
|     {
 | |
|       info->error= (int) left_length; /* We only got this many char */
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
|     length=0;				/* Didn't read any chars */
 | |
|   }
 | |
|   else if ((length= my_read(info->file,info->buffer, max_length,
 | |
|                             info->myflags)) < Count ||
 | |
| 	   length == (size_t) -1)
 | |
|   {
 | |
|     if (length != (size_t) -1)
 | |
|       memcpy(Buffer, info->buffer, length);
 | |
|     info->pos_in_file= pos_in_file;
 | |
|     info->error= length == (size_t) -1 ? -1 : (int) (length+left_length);
 | |
|     info->read_pos=info->read_end=info->buffer;
 | |
|     DBUG_RETURN(1);
 | |
|   }
 | |
|   info->read_pos=info->buffer+Count;
 | |
|   info->read_end=info->buffer+length;
 | |
|   info->pos_in_file=pos_in_file;
 | |
|   memcpy(Buffer, info->buffer, Count);
 | |
|   DBUG_RETURN(0);
 | |
| }
 | |
| 
 | |
| 
 | |
| #ifdef THREAD
 | |
| /*
 | |
|   Prepare IO_CACHE for shared use.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     init_io_cache_share()
 | |
|       read_cache                A read cache. This will be copied for
 | |
|                                 every thread after setup.
 | |
|       cshare                    The share.
 | |
|       write_cache               If non-NULL a write cache that is to be
 | |
|                                 synchronized with the read caches.
 | |
|       num_threads               Number of threads sharing the cache
 | |
|                                 including the write thread if any.
 | |
| 
 | |
|   DESCRIPTION
 | |
| 
 | |
|     The shared cache is used so: One IO_CACHE is initialized with
 | |
|     init_io_cache(). This includes the allocation of a buffer. Then a
 | |
|     share is allocated and init_io_cache_share() is called with the io
 | |
|     cache and the share. Then the io cache is copied for each thread. So
 | |
|     every thread has its own copy of IO_CACHE. But the allocated buffer
 | |
|     is shared because cache->buffer is the same for all caches.
 | |
| 
 | |
|     One thread reads data from the file into the buffer. All threads
 | |
|     read from the buffer, but every thread maintains its own set of
 | |
|     pointers into the buffer. When all threads have used up the buffer
 | |
|     contents, one of the threads reads the next block of data into the
 | |
|     buffer. To accomplish this, each thread enters the cache lock before
 | |
|     accessing the buffer. They wait in lock_io_cache() until all threads
 | |
|     joined the lock. The last thread entering the lock is in charge of
 | |
|     reading from file to buffer. It wakes all threads when done.
 | |
| 
 | |
|     Synchronizing a write cache to the read caches works so: Whenever
 | |
|     the write buffer needs a flush, the write thread enters the lock and
 | |
|     waits for all other threads to enter the lock too. They do this when
 | |
|     they have used up the read buffer. When all threads are in the lock,
 | |
|     the write thread copies the write buffer to the read buffer and
 | |
|     wakes all threads.
 | |
| 
 | |
|     share->running_threads is the number of threads not being in the
 | |
|     cache lock. When entering lock_io_cache() the number is decreased.
 | |
|     When the thread that fills the buffer enters unlock_io_cache() the
 | |
|     number is reset to the number of threads. The condition
 | |
|     running_threads == 0 means that all threads are in the lock. Bumping
 | |
|     up the number to the full count is non-intuitive. But increasing the
 | |
|     number by one for each thread that leaves the lock could lead to a
 | |
|     solo run of one thread. The last thread to join a lock reads from
 | |
|     file to buffer, wakes the other threads, processes the data in the
 | |
|     cache and enters the lock again. If no other thread left the lock
 | |
|     meanwhile, it would think it's the last one again and read the next
 | |
|     block...
 | |
| 
 | |
|     The share has copies of 'error', 'buffer', 'read_end', and
 | |
|     'pos_in_file' from the thread that filled the buffer. We may not be
 | |
|     able to access this information directly from its cache because the
 | |
|     thread may be removed from the share before the variables could be
 | |
|     copied by all other threads. Or, if a write buffer is synchronized,
 | |
|     it would change its 'pos_in_file' after waking the other threads,
 | |
|     possibly before they could copy its value.
 | |
| 
 | |
|     However, the 'buffer' variable in the share is for a synchronized
 | |
|     write cache. It needs to know where to put the data. Otherwise it
 | |
|     would need access to the read cache of one of the threads that is
 | |
|     not yet removed from the share.
 | |
| 
 | |
|   RETURN
 | |
|     void
 | |
| */
 | |
| 
 | |
| void init_io_cache_share(IO_CACHE *read_cache, IO_CACHE_SHARE *cshare,
 | |
|                          IO_CACHE *write_cache, uint num_threads)
 | |
| {
 | |
|   DBUG_ENTER("init_io_cache_share");
 | |
|   DBUG_PRINT("io_cache_share", ("read_cache: 0x%lx  share: 0x%lx  "
 | |
|                                 "write_cache: 0x%lx  threads: %u",
 | |
|                                 (long) read_cache, (long) cshare,
 | |
|                                 (long) write_cache, num_threads));
 | |
| 
 | |
|   DBUG_ASSERT(num_threads > 1);
 | |
|   DBUG_ASSERT(read_cache->type == READ_CACHE);
 | |
|   DBUG_ASSERT(!write_cache || (write_cache->type == WRITE_CACHE));
 | |
| 
 | |
|   pthread_mutex_init(&cshare->mutex, MY_MUTEX_INIT_FAST);
 | |
|   pthread_cond_init(&cshare->cond, 0);
 | |
|   pthread_cond_init(&cshare->cond_writer, 0);
 | |
| 
 | |
|   cshare->running_threads= num_threads;
 | |
|   cshare->total_threads=   num_threads;
 | |
|   cshare->error=           0;    /* Initialize. */
 | |
|   cshare->buffer=          read_cache->buffer;
 | |
|   cshare->read_end=        NULL; /* See function comment of lock_io_cache(). */
 | |
|   cshare->pos_in_file=     0;    /* See function comment of lock_io_cache(). */
 | |
|   cshare->source_cache=    write_cache; /* Can be NULL. */
 | |
| 
 | |
|   read_cache->share=         cshare;
 | |
|   read_cache->read_function= _my_b_read_r;
 | |
|   read_cache->current_pos=   NULL;
 | |
|   read_cache->current_end=   NULL;
 | |
| 
 | |
|   if (write_cache)
 | |
|     write_cache->share= cshare;
 | |
| 
 | |
|   DBUG_VOID_RETURN;
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Remove a thread from shared access to IO_CACHE.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     remove_io_thread()
 | |
|       cache                     The IO_CACHE to be removed from the share.
 | |
| 
 | |
|   NOTE
 | |
| 
 | |
|     Every thread must do that on exit for not to deadlock other threads.
 | |
| 
 | |
|     The last thread destroys the pthread resources.
 | |
| 
 | |
|     A writer flushes its cache first.
 | |
| 
 | |
|   RETURN
 | |
|     void
 | |
| */
 | |
| 
 | |
| void remove_io_thread(IO_CACHE *cache)
 | |
| {
 | |
|   IO_CACHE_SHARE *cshare= cache->share;
 | |
|   uint total;
 | |
|   DBUG_ENTER("remove_io_thread");
 | |
| 
 | |
|   /* If the writer goes, it needs to flush the write cache. */
 | |
|   if (cache == cshare->source_cache)
 | |
|     flush_io_cache(cache);
 | |
| 
 | |
|   pthread_mutex_lock(&cshare->mutex);
 | |
|   DBUG_PRINT("io_cache_share", ("%s: 0x%lx",
 | |
|                                 (cache == cshare->source_cache) ?
 | |
|                                 "writer" : "reader", (long) cache));
 | |
| 
 | |
|   /* Remove from share. */
 | |
|   total= --cshare->total_threads;
 | |
|   DBUG_PRINT("io_cache_share", ("remaining threads: %u", total));
 | |
| 
 | |
|   /* Detach from share. */
 | |
|   cache->share= NULL;
 | |
| 
 | |
|   /* If the writer goes, let the readers know. */
 | |
|   if (cache == cshare->source_cache)
 | |
|   {
 | |
|     DBUG_PRINT("io_cache_share", ("writer leaves"));
 | |
|     cshare->source_cache= NULL;
 | |
|   }
 | |
| 
 | |
|   /* If all threads are waiting for me to join the lock, wake them. */
 | |
|   if (!--cshare->running_threads)
 | |
|   {
 | |
|     DBUG_PRINT("io_cache_share", ("the last running thread leaves, wake all"));
 | |
|     pthread_cond_signal(&cshare->cond_writer);
 | |
|     pthread_cond_broadcast(&cshare->cond);
 | |
|   }
 | |
| 
 | |
|   pthread_mutex_unlock(&cshare->mutex);
 | |
| 
 | |
|   if (!total)
 | |
|   {
 | |
|     DBUG_PRINT("io_cache_share", ("last thread removed, destroy share"));
 | |
|     pthread_cond_destroy (&cshare->cond_writer);
 | |
|     pthread_cond_destroy (&cshare->cond);
 | |
|     pthread_mutex_destroy(&cshare->mutex);
 | |
|   }
 | |
| 
 | |
|   DBUG_VOID_RETURN;
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Lock IO cache and wait for all other threads to join.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     lock_io_cache()
 | |
|       cache                     The cache of the thread entering the lock.
 | |
|       pos                       File position of the block to read.
 | |
|                                 Unused for the write thread.
 | |
| 
 | |
|   DESCRIPTION
 | |
| 
 | |
|     Wait for all threads to finish with the current buffer. We want
 | |
|     all threads to proceed in concert. The last thread to join
 | |
|     lock_io_cache() will read the block from file and all threads start
 | |
|     to use it. Then they will join again for reading the next block.
 | |
| 
 | |
|     The waiting threads detect a fresh buffer by comparing
 | |
|     cshare->pos_in_file with the position they want to process next.
 | |
|     Since the first block may start at position 0, we take
 | |
|     cshare->read_end as an additional condition. This variable is
 | |
|     initialized to NULL and will be set after a block of data is written
 | |
|     to the buffer.
 | |
| 
 | |
|   RETURN
 | |
|     1           OK, lock in place, go ahead and read.
 | |
|     0           OK, unlocked, another thread did the read.
 | |
| */
 | |
| 
 | |
| static int lock_io_cache(IO_CACHE *cache, my_off_t pos)
 | |
| {
 | |
|   IO_CACHE_SHARE *cshare= cache->share;
 | |
|   DBUG_ENTER("lock_io_cache");
 | |
| 
 | |
|   /* Enter the lock. */
 | |
|   pthread_mutex_lock(&cshare->mutex);
 | |
|   cshare->running_threads--;
 | |
|   DBUG_PRINT("io_cache_share", ("%s: 0x%lx  pos: %lu  running: %u",
 | |
|                                 (cache == cshare->source_cache) ?
 | |
|                                 "writer" : "reader", (long) cache, (ulong) pos,
 | |
|                                 cshare->running_threads));
 | |
| 
 | |
|   if (cshare->source_cache)
 | |
|   {
 | |
|     /* A write cache is synchronized to the read caches. */
 | |
| 
 | |
|     if (cache == cshare->source_cache)
 | |
|     {
 | |
|       /* The writer waits until all readers are here. */
 | |
|       while (cshare->running_threads)
 | |
|       {
 | |
|         DBUG_PRINT("io_cache_share", ("writer waits in lock"));
 | |
|         pthread_cond_wait(&cshare->cond_writer, &cshare->mutex);
 | |
|       }
 | |
|       DBUG_PRINT("io_cache_share", ("writer awoke, going to copy"));
 | |
| 
 | |
|       /* Stay locked. Leave the lock later by unlock_io_cache(). */
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
| 
 | |
|     /* The last thread wakes the writer. */
 | |
|     if (!cshare->running_threads)
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("waking writer"));
 | |
|       pthread_cond_signal(&cshare->cond_writer);
 | |
|     }
 | |
| 
 | |
|     /*
 | |
|       Readers wait until the data is copied from the writer. Another
 | |
|       reason to stop waiting is the removal of the write thread. If this
 | |
|       happens, we leave the lock with old data in the buffer.
 | |
|     */
 | |
|     while ((!cshare->read_end || (cshare->pos_in_file < pos)) &&
 | |
|            cshare->source_cache)
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("reader waits in lock"));
 | |
|       pthread_cond_wait(&cshare->cond, &cshare->mutex);
 | |
|     }
 | |
| 
 | |
|     /*
 | |
|       If the writer was removed from the share while this thread was
 | |
|       asleep, we need to simulate an EOF condition. The writer cannot
 | |
|       reset the share variables as they might still be in use by readers
 | |
|       of the last block. When we awake here then because the last
 | |
|       joining thread signalled us. If the writer is not the last, it
 | |
|       will not signal. So it is safe to clear the buffer here.
 | |
|     */
 | |
|     if (!cshare->read_end || (cshare->pos_in_file < pos))
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("reader found writer removed. EOF"));
 | |
|       cshare->read_end= cshare->buffer; /* Empty buffer. */
 | |
|       cshare->error= 0; /* EOF is not an error. */
 | |
|     }
 | |
|   }
 | |
|   else
 | |
|   {
 | |
|     /*
 | |
|       There are read caches only. The last thread arriving in
 | |
|       lock_io_cache() continues with a locked cache and reads the block.
 | |
|     */
 | |
|     if (!cshare->running_threads)
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("last thread joined, going to read"));
 | |
|       /* Stay locked. Leave the lock later by unlock_io_cache(). */
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
| 
 | |
|     /*
 | |
|       All other threads wait until the requested block is read by the
 | |
|       last thread arriving. Another reason to stop waiting is the
 | |
|       removal of a thread. If this leads to all threads being in the
 | |
|       lock, we have to continue also. The first of the awaken threads
 | |
|       will then do the read.
 | |
|     */
 | |
|     while ((!cshare->read_end || (cshare->pos_in_file < pos)) &&
 | |
|            cshare->running_threads)
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("reader waits in lock"));
 | |
|       pthread_cond_wait(&cshare->cond, &cshare->mutex);
 | |
|     }
 | |
| 
 | |
|     /* If the block is not yet read, continue with a locked cache and read. */
 | |
|     if (!cshare->read_end || (cshare->pos_in_file < pos))
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("reader awoke, going to read"));
 | |
|       /* Stay locked. Leave the lock later by unlock_io_cache(). */
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
| 
 | |
|     /* Another thread did read the block already. */
 | |
|   }
 | |
|   DBUG_PRINT("io_cache_share", ("reader awoke, going to process %u bytes",
 | |
|                                 (uint) (cshare->read_end ? (size_t)
 | |
|                                         (cshare->read_end - cshare->buffer) :
 | |
|                                         0)));
 | |
| 
 | |
|   /*
 | |
|     Leave the lock. Do not call unlock_io_cache() later. The thread that
 | |
|     filled the buffer did this and marked all threads as running.
 | |
|   */
 | |
|   pthread_mutex_unlock(&cshare->mutex);
 | |
|   DBUG_RETURN(0);
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Unlock IO cache.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     unlock_io_cache()
 | |
|       cache                     The cache of the thread leaving the lock.
 | |
| 
 | |
|   NOTE
 | |
|     This is called by the thread that filled the buffer. It marks all
 | |
|     threads as running and awakes them. This must not be done by any
 | |
|     other thread.
 | |
| 
 | |
|     Do not signal cond_writer. Either there is no writer or the writer
 | |
|     is the only one who can call this function.
 | |
| 
 | |
|     The reason for resetting running_threads to total_threads before
 | |
|     waking all other threads is that it could be possible that this
 | |
|     thread is so fast with processing the buffer that it enters the lock
 | |
|     before even one other thread has left it. If every awoken thread
 | |
|     would increase running_threads by one, this thread could think that
 | |
|     he is again the last to join and would not wait for the other
 | |
|     threads to process the data.
 | |
| 
 | |
|   RETURN
 | |
|     void
 | |
| */
 | |
| 
 | |
| static void unlock_io_cache(IO_CACHE *cache)
 | |
| {
 | |
|   IO_CACHE_SHARE *cshare= cache->share;
 | |
|   DBUG_ENTER("unlock_io_cache");
 | |
|   DBUG_PRINT("io_cache_share", ("%s: 0x%lx  pos: %lu  running: %u",
 | |
|                                 (cache == cshare->source_cache) ?
 | |
|                                 "writer" : "reader",
 | |
|                                 (long) cache, (ulong) cshare->pos_in_file,
 | |
|                                 cshare->total_threads));
 | |
| 
 | |
|   cshare->running_threads= cshare->total_threads;
 | |
|   pthread_cond_broadcast(&cshare->cond);
 | |
|   pthread_mutex_unlock(&cshare->mutex);
 | |
|   DBUG_VOID_RETURN;
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Read from IO_CACHE when it is shared between several threads.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     _my_b_read_r()
 | |
|       cache                     IO_CACHE pointer
 | |
|       Buffer                    Buffer to retrieve count bytes from file
 | |
|       Count                     Number of bytes to read into Buffer
 | |
| 
 | |
|   NOTE
 | |
|     This function is only called from the my_b_read() macro when there
 | |
|     isn't enough characters in the buffer to satisfy the request.
 | |
| 
 | |
|   IMPLEMENTATION
 | |
| 
 | |
|     It works as follows: when a thread tries to read from a file (that
 | |
|     is, after using all the data from the (shared) buffer), it just
 | |
|     hangs on lock_io_cache(), waiting for other threads. When the very
 | |
|     last thread attempts a read, lock_io_cache() returns 1, the thread
 | |
|     does actual IO and unlock_io_cache(), which signals all the waiting
 | |
|     threads that data is in the buffer.
 | |
| 
 | |
|   WARNING
 | |
| 
 | |
|     When changing this function, be careful with handling file offsets
 | |
|     (end-of_file, pos_in_file). Do not cast them to possibly smaller
 | |
|     types than my_off_t unless you can be sure that their value fits.
 | |
|     Same applies to differences of file offsets. (Bug #11527)
 | |
| 
 | |
|     When changing this function, check _my_b_read(). It might need the
 | |
|     same change.
 | |
| 
 | |
|   RETURN
 | |
|     0      we succeeded in reading all data
 | |
|     1      Error: can't read requested characters
 | |
| */
 | |
| 
 | |
| int _my_b_read_r(register IO_CACHE *cache, uchar *Buffer, size_t Count)
 | |
| {
 | |
|   my_off_t pos_in_file;
 | |
|   size_t length, diff_length, left_length;
 | |
|   IO_CACHE_SHARE *cshare= cache->share;
 | |
|   DBUG_ENTER("_my_b_read_r");
 | |
| 
 | |
|   if ((left_length= (size_t) (cache->read_end - cache->read_pos)))
 | |
|   {
 | |
|     DBUG_ASSERT(Count >= left_length);	/* User is not using my_b_read() */
 | |
|     memcpy(Buffer, cache->read_pos, left_length);
 | |
|     Buffer+= left_length;
 | |
|     Count-= left_length;
 | |
|   }
 | |
|   while (Count)
 | |
|   {
 | |
|     size_t cnt, len;
 | |
| 
 | |
|     pos_in_file= cache->pos_in_file + (cache->read_end - cache->buffer);
 | |
|     diff_length= (size_t) (pos_in_file & (IO_SIZE-1));
 | |
|     length=IO_ROUND_UP(Count+diff_length)-diff_length;
 | |
|     length= ((length <= cache->read_length) ?
 | |
|              length + IO_ROUND_DN(cache->read_length - length) :
 | |
|              length - IO_ROUND_UP(length - cache->read_length));
 | |
|     if (cache->type != READ_FIFO &&
 | |
| 	(length > (cache->end_of_file - pos_in_file)))
 | |
|       length= (size_t) (cache->end_of_file - pos_in_file);
 | |
|     if (length == 0)
 | |
|     {
 | |
|       cache->error= (int) left_length;
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
|     if (lock_io_cache(cache, pos_in_file))
 | |
|     {
 | |
|       /* With a synchronized write/read cache we won't come here... */
 | |
|       DBUG_ASSERT(!cshare->source_cache);
 | |
|       /*
 | |
|         ... unless the writer has gone before this thread entered the
 | |
|         lock. Simulate EOF in this case. It can be distinguished by
 | |
|         cache->file.
 | |
|       */
 | |
|       if (cache->file < 0)
 | |
|         len= 0;
 | |
|       else
 | |
|       {
 | |
|         /*
 | |
|           Whenever a function which operates on IO_CACHE flushes/writes
 | |
|           some part of the IO_CACHE to disk it will set the property
 | |
|           "seek_not_done" to indicate this to other functions operating
 | |
|           on the IO_CACHE.
 | |
|         */
 | |
|         if (cache->seek_not_done)
 | |
|         {
 | |
|           if (my_seek(cache->file, pos_in_file, MY_SEEK_SET, MYF(0))
 | |
|               == MY_FILEPOS_ERROR)
 | |
|           {
 | |
|             cache->error= -1;
 | |
|             unlock_io_cache(cache);
 | |
|             DBUG_RETURN(1);
 | |
|           }
 | |
|         }
 | |
|         len= my_read(cache->file, cache->buffer, length, cache->myflags);
 | |
|       }
 | |
|       DBUG_PRINT("io_cache_share", ("read %lu bytes", (ulong) len));
 | |
| 
 | |
|       cache->read_end=    cache->buffer + (len == (size_t) -1 ? 0 : len);
 | |
|       cache->error=       (len == length ? 0 : (int) len);
 | |
|       cache->pos_in_file= pos_in_file;
 | |
| 
 | |
|       /* Copy important values to the share. */
 | |
|       cshare->error=       cache->error;
 | |
|       cshare->read_end=    cache->read_end;
 | |
|       cshare->pos_in_file= pos_in_file;
 | |
| 
 | |
|       /* Mark all threads as running and wake them. */
 | |
|       unlock_io_cache(cache);
 | |
|     }
 | |
|     else
 | |
|     {
 | |
|       /*
 | |
|         With a synchronized write/read cache readers always come here.
 | |
|         Copy important values from the share.
 | |
|       */
 | |
|       cache->error=       cshare->error;
 | |
|       cache->read_end=    cshare->read_end;
 | |
|       cache->pos_in_file= cshare->pos_in_file;
 | |
| 
 | |
|       len= ((cache->error == -1) ? (size_t) -1 :
 | |
|             (size_t) (cache->read_end - cache->buffer));
 | |
|     }
 | |
|     cache->read_pos=      cache->buffer;
 | |
|     cache->seek_not_done= 0;
 | |
|     if (len == 0 || len == (size_t) -1)
 | |
|     {
 | |
|       DBUG_PRINT("io_cache_share", ("reader error. len %lu  left %lu",
 | |
|                                     (ulong) len, (ulong) left_length));
 | |
|       cache->error= (int) left_length;
 | |
|       DBUG_RETURN(1);
 | |
|     }
 | |
|     cnt= (len > Count) ? Count : len;
 | |
|     memcpy(Buffer, cache->read_pos, cnt);
 | |
|     Count -= cnt;
 | |
|     Buffer+= cnt;
 | |
|     left_length+= cnt;
 | |
|     cache->read_pos+= cnt;
 | |
|   }
 | |
|   DBUG_RETURN(0);
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Copy data from write cache to read cache.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     copy_to_read_buffer()
 | |
|       write_cache               The write cache.
 | |
|       write_buffer              The source of data, mostly the cache buffer.
 | |
|       write_length              The number of bytes to copy.
 | |
| 
 | |
|   NOTE
 | |
|     The write thread will wait for all read threads to join the cache
 | |
|     lock. Then it copies the data over and wakes the read threads.
 | |
| 
 | |
|   RETURN
 | |
|     void
 | |
| */
 | |
| 
 | |
| static void copy_to_read_buffer(IO_CACHE *write_cache,
 | |
|                                 const uchar *write_buffer, size_t write_length)
 | |
| {
 | |
|   IO_CACHE_SHARE *cshare= write_cache->share;
 | |
| 
 | |
|   DBUG_ASSERT(cshare->source_cache == write_cache);
 | |
|   /*
 | |
|     write_length is usually less or equal to buffer_length.
 | |
|     It can be bigger if _my_b_write() is called with a big length.
 | |
|   */
 | |
|   while (write_length)
 | |
|   {
 | |
|     size_t copy_length= min(write_length, write_cache->buffer_length);
 | |
|     int  __attribute__((unused)) rc;
 | |
| 
 | |
|     rc= lock_io_cache(write_cache, write_cache->pos_in_file);
 | |
|     /* The writing thread does always have the lock when it awakes. */
 | |
|     DBUG_ASSERT(rc);
 | |
| 
 | |
|     memcpy(cshare->buffer, write_buffer, copy_length);
 | |
| 
 | |
|     cshare->error=       0;
 | |
|     cshare->read_end=    cshare->buffer + copy_length;
 | |
|     cshare->pos_in_file= write_cache->pos_in_file;
 | |
| 
 | |
|     /* Mark all threads as running and wake them. */
 | |
|     unlock_io_cache(write_cache);
 | |
| 
 | |
|     write_buffer+= copy_length;
 | |
|     write_length-= copy_length;
 | |
|   }
 | |
| }
 | |
| #endif /*THREAD*/
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Do sequential read from the SEQ_READ_APPEND cache.
 | |
|   
 | |
|   We do this in three stages:
 | |
|    - first read from info->buffer
 | |
|    - then if there are still data to read, try the file descriptor
 | |
|    - afterwards, if there are still data to read, try append buffer
 | |
| 
 | |
|   RETURNS
 | |
|     0  Success
 | |
|     1  Failed to read
 | |
| */
 | |
| 
 | |
| int _my_b_seq_read(register IO_CACHE *info, uchar *Buffer, size_t Count)
 | |
| {
 | |
|   size_t length, diff_length, left_length, save_count, max_length;
 | |
|   my_off_t pos_in_file;
 | |
|   save_count=Count;
 | |
| 
 | |
|   /* first, read the regular buffer */
 | |
|   if ((left_length=(size_t) (info->read_end-info->read_pos)))
 | |
|   {
 | |
|     DBUG_ASSERT(Count > left_length);	/* User is not using my_b_read() */
 | |
|     memcpy(Buffer,info->read_pos, left_length);
 | |
|     Buffer+=left_length;
 | |
|     Count-=left_length;
 | |
|   }
 | |
|   lock_append_buffer(info);
 | |
| 
 | |
|   /* pos_in_file always point on where info->buffer was read */
 | |
|   if ((pos_in_file=info->pos_in_file +
 | |
|        (size_t) (info->read_end - info->buffer)) >= info->end_of_file)
 | |
|     goto read_append_buffer;
 | |
| 
 | |
|   /*
 | |
|     With read-append cache we must always do a seek before we read,
 | |
|     because the write could have moved the file pointer astray
 | |
|   */
 | |
|   if (my_seek(info->file,pos_in_file,MY_SEEK_SET,MYF(0)) == MY_FILEPOS_ERROR)
 | |
|   {
 | |
|    info->error= -1;
 | |
|    unlock_append_buffer(info);
 | |
|    return (1);
 | |
|   }
 | |
|   info->seek_not_done=0;
 | |
| 
 | |
|   diff_length= (size_t) (pos_in_file & (IO_SIZE-1));
 | |
| 
 | |
|   /* now the second stage begins - read from file descriptor */
 | |
|   if (Count >= (size_t) (IO_SIZE+(IO_SIZE-diff_length)))
 | |
|   {
 | |
|     /* Fill first intern buffer */
 | |
|     size_t read_length;
 | |
| 
 | |
|     length=(Count & (size_t) ~(IO_SIZE-1))-diff_length;
 | |
|     if ((read_length= my_read(info->file,Buffer, length,
 | |
|                               info->myflags)) == (size_t) -1)
 | |
|     {
 | |
|       info->error= -1;
 | |
|       unlock_append_buffer(info);
 | |
|       return 1;
 | |
|     }
 | |
|     Count-=read_length;
 | |
|     Buffer+=read_length;
 | |
|     pos_in_file+=read_length;
 | |
| 
 | |
|     if (read_length != length)
 | |
|     {
 | |
|       /*
 | |
| 	We only got part of data;  Read the rest of the data from the
 | |
| 	write buffer
 | |
|       */
 | |
|       goto read_append_buffer;
 | |
|     }
 | |
|     left_length+=length;
 | |
|     diff_length=0;
 | |
|   }
 | |
| 
 | |
|   max_length= info->read_length-diff_length;
 | |
|   if (max_length > (info->end_of_file - pos_in_file))
 | |
|     max_length= (size_t) (info->end_of_file - pos_in_file);
 | |
|   if (!max_length)
 | |
|   {
 | |
|     if (Count)
 | |
|       goto read_append_buffer;
 | |
|     length=0;				/* Didn't read any more chars */
 | |
|   }
 | |
|   else
 | |
|   {
 | |
|     length= my_read(info->file,info->buffer, max_length, info->myflags);
 | |
|     if (length == (size_t) -1)
 | |
|     {
 | |
|       info->error= -1;
 | |
|       unlock_append_buffer(info);
 | |
|       return 1;
 | |
|     }
 | |
|     if (length < Count)
 | |
|     {
 | |
|       memcpy(Buffer, info->buffer, length);
 | |
|       Count -= length;
 | |
|       Buffer += length;
 | |
| 
 | |
|       /*
 | |
| 	 added the line below to make
 | |
| 	 DBUG_ASSERT(pos_in_file==info->end_of_file) pass.
 | |
| 	 otherwise this does not appear to be needed
 | |
|       */
 | |
|       pos_in_file += length;
 | |
|       goto read_append_buffer;
 | |
|     }
 | |
|   }
 | |
|   unlock_append_buffer(info);
 | |
|   info->read_pos=info->buffer+Count;
 | |
|   info->read_end=info->buffer+length;
 | |
|   info->pos_in_file=pos_in_file;
 | |
|   memcpy(Buffer,info->buffer,(size_t) Count);
 | |
|   return 0;
 | |
| 
 | |
| read_append_buffer:
 | |
| 
 | |
|   /*
 | |
|      Read data from the current write buffer.
 | |
|      Count should never be == 0 here (The code will work even if count is 0)
 | |
|   */
 | |
| 
 | |
|   {
 | |
|     /* First copy the data to Count */
 | |
|     size_t len_in_buff = (size_t) (info->write_pos - info->append_read_pos);
 | |
|     size_t copy_len;
 | |
|     size_t transfer_len;
 | |
| 
 | |
|     DBUG_ASSERT(info->append_read_pos <= info->write_pos);
 | |
|     /*
 | |
|       TODO: figure out if the assert below is needed or correct.
 | |
|     */
 | |
|     DBUG_ASSERT(pos_in_file == info->end_of_file);
 | |
|     copy_len=min(Count, len_in_buff);
 | |
|     memcpy(Buffer, info->append_read_pos, copy_len);
 | |
|     info->append_read_pos += copy_len;
 | |
|     Count -= copy_len;
 | |
|     if (Count)
 | |
|       info->error= (int) (save_count - Count);
 | |
| 
 | |
|     /* Fill read buffer with data from write buffer */
 | |
|     memcpy(info->buffer, info->append_read_pos,
 | |
| 	   (size_t) (transfer_len=len_in_buff - copy_len));
 | |
|     info->read_pos= info->buffer;
 | |
|     info->read_end= info->buffer+transfer_len;
 | |
|     info->append_read_pos=info->write_pos;
 | |
|     info->pos_in_file=pos_in_file+copy_len;
 | |
|     info->end_of_file+=len_in_buff;
 | |
|   }
 | |
|   unlock_append_buffer(info);
 | |
|   return Count ? 1 : 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| #ifdef HAVE_AIOWAIT
 | |
| 
 | |
| /*
 | |
|   Read from the IO_CACHE into a buffer and feed asynchronously
 | |
|   from disk when needed.
 | |
| 
 | |
|   SYNOPSIS
 | |
|     _my_b_async_read()
 | |
|       info                      IO_CACHE pointer
 | |
|       Buffer                    Buffer to retrieve count bytes from file
 | |
|       Count                     Number of bytes to read into Buffer
 | |
| 
 | |
|   RETURN VALUE
 | |
|     -1          An error has occurred; my_errno is set.
 | |
|      0          Success
 | |
|      1          An error has occurred; IO_CACHE to error state.
 | |
| */
 | |
| 
 | |
| int _my_b_async_read(register IO_CACHE *info, uchar *Buffer, size_t Count)
 | |
| {
 | |
|   size_t length,read_length,diff_length,left_length,use_length,org_Count;
 | |
|   size_t max_length;
 | |
|   my_off_t next_pos_in_file;
 | |
|   uchar *read_buffer;
 | |
| 
 | |
|   memcpy(Buffer,info->read_pos,
 | |
| 	 (left_length= (size_t) (info->read_end-info->read_pos)));
 | |
|   Buffer+=left_length;
 | |
|   org_Count=Count;
 | |
|   Count-=left_length;
 | |
| 
 | |
|   if (info->inited)
 | |
|   {						/* wait for read block */
 | |
|     info->inited=0;				/* No more block to read */
 | |
|     my_aiowait(&info->aio_result);		/* Wait for outstanding req */
 | |
|     if (info->aio_result.result.aio_errno)
 | |
|     {
 | |
|       if (info->myflags & MY_WME)
 | |
| 	my_error(EE_READ, MYF(ME_BELL+ME_WAITTANG),
 | |
| 		 my_filename(info->file),
 | |
| 		 info->aio_result.result.aio_errno);
 | |
|       my_errno=info->aio_result.result.aio_errno;
 | |
|       info->error= -1;
 | |
|       return(1);
 | |
|     }
 | |
|     if (! (read_length= (size_t) info->aio_result.result.aio_return) ||
 | |
| 	read_length == (size_t) -1)
 | |
|     {
 | |
|       my_errno=0;				/* For testing */
 | |
|       info->error= (read_length == (size_t) -1 ? -1 :
 | |
| 		    (int) (read_length+left_length));
 | |
|       return(1);
 | |
|     }
 | |
|     info->pos_in_file+= (size_t) (info->read_end - info->request_pos);
 | |
| 
 | |
|     if (info->request_pos != info->buffer)
 | |
|       info->request_pos=info->buffer;
 | |
|     else
 | |
|       info->request_pos=info->buffer+info->read_length;
 | |
|     info->read_pos=info->request_pos;
 | |
|     next_pos_in_file=info->aio_read_pos+read_length;
 | |
| 
 | |
| 	/* Check if pos_in_file is changed
 | |
| 	   (_ni_read_cache may have skipped some bytes) */
 | |
| 
 | |
|     if (info->aio_read_pos < info->pos_in_file)
 | |
|     {						/* Fix if skipped bytes */
 | |
|       if (info->aio_read_pos + read_length < info->pos_in_file)
 | |
|       {
 | |
| 	read_length=0;				/* Skip block */
 | |
| 	next_pos_in_file=info->pos_in_file;
 | |
|       }
 | |
|       else
 | |
|       {
 | |
| 	my_off_t offset= (info->pos_in_file - info->aio_read_pos);
 | |
| 	info->pos_in_file=info->aio_read_pos; /* Whe are here */
 | |
| 	info->read_pos=info->request_pos+offset;
 | |
| 	read_length-=offset;			/* Bytes left from read_pos */
 | |
|       }
 | |
|     }
 | |
| #ifndef DBUG_OFF
 | |
|     if (info->aio_read_pos > info->pos_in_file)
 | |
|     {
 | |
|       my_errno=EINVAL;
 | |
|       return(info->read_length= (size_t) -1);
 | |
|     }
 | |
| #endif
 | |
| 	/* Copy found bytes to buffer */
 | |
|     length=min(Count,read_length);
 | |
|     memcpy(Buffer,info->read_pos,(size_t) length);
 | |
|     Buffer+=length;
 | |
|     Count-=length;
 | |
|     left_length+=length;
 | |
|     info->read_end=info->rc_pos+read_length;
 | |
|     info->read_pos+=length;
 | |
|   }
 | |
|   else
 | |
|     next_pos_in_file=(info->pos_in_file+ (size_t)
 | |
| 		      (info->read_end - info->request_pos));
 | |
| 
 | |
| 	/* If reading large blocks, or first read or read with skip */
 | |
|   if (Count)
 | |
|   {
 | |
|     if (next_pos_in_file == info->end_of_file)
 | |
|     {
 | |
|       info->error=(int) (read_length+left_length);
 | |
|       return 1;
 | |
|     }
 | |
|     
 | |
|     if (my_seek(info->file,next_pos_in_file,MY_SEEK_SET,MYF(0))
 | |
|         == MY_FILEPOS_ERROR)
 | |
|     {
 | |
|       info->error= -1;
 | |
|       return (1);
 | |
|     }
 | |
| 
 | |
|     read_length=IO_SIZE*2- (size_t) (next_pos_in_file & (IO_SIZE-1));
 | |
|     if (Count < read_length)
 | |
|     {					/* Small block, read to cache */
 | |
|       if ((read_length=my_read(info->file,info->request_pos,
 | |
| 			       read_length, info->myflags)) == (size_t) -1)
 | |
|         return info->error= -1;
 | |
|       use_length=min(Count,read_length);
 | |
|       memcpy(Buffer,info->request_pos,(size_t) use_length);
 | |
|       info->read_pos=info->request_pos+Count;
 | |
|       info->read_end=info->request_pos+read_length;
 | |
|       info->pos_in_file=next_pos_in_file;	/* Start of block in cache */
 | |
|       next_pos_in_file+=read_length;
 | |
| 
 | |
|       if (Count != use_length)
 | |
|       {					/* Didn't find hole block */
 | |
| 	if (info->myflags & (MY_WME | MY_FAE | MY_FNABP) && Count != org_Count)
 | |
| 	  my_error(EE_EOFERR, MYF(ME_BELL+ME_WAITTANG),
 | |
| 		   my_filename(info->file),my_errno);
 | |
| 	info->error=(int) (read_length+left_length);
 | |
| 	return 1;
 | |
|       }
 | |
|     }
 | |
|     else
 | |
|     {						/* Big block, don't cache it */
 | |
|       if ((read_length= my_read(info->file,Buffer, Count,info->myflags))
 | |
| 	  != Count)
 | |
|       {
 | |
| 	info->error= read_length == (size_t) -1 ? -1 : read_length+left_length;
 | |
| 	return 1;
 | |
|       }
 | |
|       info->read_pos=info->read_end=info->request_pos;
 | |
|       info->pos_in_file=(next_pos_in_file+=Count);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   /* Read next block with asyncronic io */
 | |
|   diff_length=(next_pos_in_file & (IO_SIZE-1));
 | |
|   max_length= info->read_length - diff_length;
 | |
|   if (max_length > info->end_of_file - next_pos_in_file)
 | |
|     max_length= (size_t) (info->end_of_file - next_pos_in_file);
 | |
| 
 | |
|   if (info->request_pos != info->buffer)
 | |
|     read_buffer=info->buffer;
 | |
|   else
 | |
|     read_buffer=info->buffer+info->read_length;
 | |
|   info->aio_read_pos=next_pos_in_file;
 | |
|   if (max_length)
 | |
|   {
 | |
|     info->aio_result.result.aio_errno=AIO_INPROGRESS;	/* Marker for test */
 | |
|     DBUG_PRINT("aioread",("filepos: %ld  length: %lu",
 | |
| 			  (ulong) next_pos_in_file, (ulong) max_length));
 | |
|     if (aioread(info->file,read_buffer, max_length,
 | |
| 		(my_off_t) next_pos_in_file,MY_SEEK_SET,
 | |
| 		&info->aio_result.result))
 | |
|     {						/* Skip async io */
 | |
|       my_errno=errno;
 | |
|       DBUG_PRINT("error",("got error: %d, aio_result: %d from aioread, async skipped",
 | |
| 			  errno, info->aio_result.result.aio_errno));
 | |
|       if (info->request_pos != info->buffer)
 | |
|       {
 | |
| 	bmove(info->buffer,info->request_pos,
 | |
| 	      (size_t) (info->read_end - info->read_pos));
 | |
| 	info->request_pos=info->buffer;
 | |
| 	info->read_pos-=info->read_length;
 | |
| 	info->read_end-=info->read_length;
 | |
|       }
 | |
|       info->read_length=info->buffer_length;	/* Use hole buffer */
 | |
|       info->read_function=_my_b_read;		/* Use normal IO_READ next */
 | |
|     }
 | |
|     else
 | |
|       info->inited=info->aio_result.pending=1;
 | |
|   }
 | |
|   return 0;					/* Block read, async in use */
 | |
| } /* _my_b_async_read */
 | |
| #endif
 | |
| 
 | |
| 
 | |
| /* Read one byte when buffer is empty */
 | |
| 
 | |
| int _my_b_get(IO_CACHE *info)
 | |
| {
 | |
|   uchar buff;
 | |
|   IO_CACHE_CALLBACK pre_read,post_read;
 | |
|   if ((pre_read = info->pre_read))
 | |
|     (*pre_read)(info);
 | |
|   if ((*(info)->read_function)(info,&buff,1))
 | |
|     return my_b_EOF;
 | |
|   if ((post_read = info->post_read))
 | |
|     (*post_read)(info);
 | |
|   return (int) (uchar) buff;
 | |
| }
 | |
| 
 | |
| /* 
 | |
|    Write a byte buffer to IO_CACHE and flush to disk
 | |
|    if IO_CACHE is full.
 | |
| 
 | |
|    RETURN VALUE
 | |
|     1 On error on write
 | |
|     0 On success
 | |
|    -1 On error; my_errno contains error code.
 | |
| */
 | |
| 
 | |
| int _my_b_write(register IO_CACHE *info, const uchar *Buffer, size_t Count)
 | |
| {
 | |
|   size_t rest_length,length;
 | |
| 
 | |
|   if (info->pos_in_file+info->buffer_length > info->end_of_file)
 | |
|   {
 | |
|     my_errno=errno=EFBIG;
 | |
|     return info->error = -1;
 | |
|   }
 | |
| 
 | |
|   rest_length= (size_t) (info->write_end - info->write_pos);
 | |
|   memcpy(info->write_pos,Buffer,(size_t) rest_length);
 | |
|   Buffer+=rest_length;
 | |
|   Count-=rest_length;
 | |
|   info->write_pos+=rest_length;
 | |
| 
 | |
|   if (my_b_flush_io_cache(info,1))
 | |
|     return 1;
 | |
|   if (Count >= IO_SIZE)
 | |
|   {					/* Fill first intern buffer */
 | |
|     length=Count & (size_t) ~(IO_SIZE-1);
 | |
|     if (info->seek_not_done)
 | |
|     {
 | |
|       /*
 | |
|         Whenever a function which operates on IO_CACHE flushes/writes
 | |
|         some part of the IO_CACHE to disk it will set the property
 | |
|         "seek_not_done" to indicate this to other functions operating
 | |
|         on the IO_CACHE.
 | |
|       */
 | |
|       if (my_seek(info->file,info->pos_in_file,MY_SEEK_SET,MYF(0)))
 | |
|       {
 | |
|         info->error= -1;
 | |
|         return (1);
 | |
|       }
 | |
|       info->seek_not_done=0;
 | |
|     }
 | |
|     if (my_write(info->file, Buffer, length, info->myflags | MY_NABP))
 | |
|       return info->error= -1;
 | |
| 
 | |
| #ifdef THREAD
 | |
|     /*
 | |
|       In case of a shared I/O cache with a writer we normally do direct
 | |
|       write cache to read cache copy. Simulate this here by direct
 | |
|       caller buffer to read cache copy. Do it after the write so that
 | |
|       the cache readers actions on the flushed part can go in parallel
 | |
|       with the write of the extra stuff. copy_to_read_buffer()
 | |
|       synchronizes writer and readers so that after this call the
 | |
|       readers can act on the extra stuff while the writer can go ahead
 | |
|       and prepare the next output. copy_to_read_buffer() relies on
 | |
|       info->pos_in_file.
 | |
|     */
 | |
|     if (info->share)
 | |
|       copy_to_read_buffer(info, Buffer, length);
 | |
| #endif
 | |
| 
 | |
|     Count-=length;
 | |
|     Buffer+=length;
 | |
|     info->pos_in_file+=length;
 | |
|   }
 | |
|   memcpy(info->write_pos,Buffer,(size_t) Count);
 | |
|   info->write_pos+=Count;
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Append a block to the write buffer.
 | |
|   This is done with the buffer locked to ensure that we don't read from
 | |
|   the write buffer before we are ready with it.
 | |
| */
 | |
| 
 | |
| int my_b_append(register IO_CACHE *info, const uchar *Buffer, size_t Count)
 | |
| {
 | |
|   size_t rest_length,length;
 | |
| 
 | |
| #ifdef THREAD
 | |
|   /*
 | |
|     Assert that we cannot come here with a shared cache. If we do one
 | |
|     day, we might need to add a call to copy_to_read_buffer().
 | |
|   */
 | |
|   DBUG_ASSERT(!info->share);
 | |
| #endif
 | |
| 
 | |
|   lock_append_buffer(info);
 | |
|   rest_length= (size_t) (info->write_end - info->write_pos);
 | |
|   if (Count <= rest_length)
 | |
|     goto end;
 | |
|   memcpy(info->write_pos, Buffer, rest_length);
 | |
|   Buffer+=rest_length;
 | |
|   Count-=rest_length;
 | |
|   info->write_pos+=rest_length;
 | |
|   if (my_b_flush_io_cache(info,0))
 | |
|   {
 | |
|     unlock_append_buffer(info);
 | |
|     return 1;
 | |
|   }
 | |
|   if (Count >= IO_SIZE)
 | |
|   {					/* Fill first intern buffer */
 | |
|     length=Count & (size_t) ~(IO_SIZE-1);
 | |
|     if (my_write(info->file,Buffer, length, info->myflags | MY_NABP))
 | |
|     {
 | |
|       unlock_append_buffer(info);
 | |
|       return info->error= -1;
 | |
|     }
 | |
|     Count-=length;
 | |
|     Buffer+=length;
 | |
|     info->end_of_file+=length;
 | |
|   }
 | |
| 
 | |
| end:
 | |
|   memcpy(info->write_pos,Buffer,(size_t) Count);
 | |
|   info->write_pos+=Count;
 | |
|   unlock_append_buffer(info);
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| int my_b_safe_write(IO_CACHE *info, const uchar *Buffer, size_t Count)
 | |
| {
 | |
|   /*
 | |
|     Sasha: We are not writing this with the ? operator to avoid hitting
 | |
|     a possible compiler bug. At least gcc 2.95 cannot deal with 
 | |
|     several layers of ternary operators that evaluated comma(,) operator
 | |
|     expressions inside - I do have a test case if somebody wants it
 | |
|   */
 | |
|   if (info->type == SEQ_READ_APPEND)
 | |
|     return my_b_append(info, Buffer, Count);
 | |
|   return my_b_write(info, Buffer, Count);
 | |
| }
 | |
| 
 | |
| 
 | |
| /*
 | |
|   Write a block to disk where part of the data may be inside the record
 | |
|   buffer.  As all write calls to the data goes through the cache,
 | |
|   we will never get a seek over the end of the buffer
 | |
| */
 | |
| 
 | |
| int my_block_write(register IO_CACHE *info, const uchar *Buffer, size_t Count,
 | |
| 		   my_off_t pos)
 | |
| {
 | |
|   size_t length;
 | |
|   int error=0;
 | |
| 
 | |
| #ifdef THREAD
 | |
|   /*
 | |
|     Assert that we cannot come here with a shared cache. If we do one
 | |
|     day, we might need to add a call to copy_to_read_buffer().
 | |
|   */
 | |
|   DBUG_ASSERT(!info->share);
 | |
| #endif
 | |
| 
 | |
|   if (pos < info->pos_in_file)
 | |
|   {
 | |
|     /* Of no overlap, write everything without buffering */
 | |
|     if (pos + Count <= info->pos_in_file)
 | |
|       return (int) my_pwrite(info->file, Buffer, Count, pos,
 | |
|                              info->myflags | MY_NABP);
 | |
|     /* Write the part of the block that is before buffer */
 | |
|     length= (uint) (info->pos_in_file - pos);
 | |
|     if (my_pwrite(info->file, Buffer, length, pos, info->myflags | MY_NABP))
 | |
|       info->error= error= -1;
 | |
|     Buffer+=length;
 | |
|     pos+=  length;
 | |
|     Count-= length;
 | |
| #ifndef HAVE_PREAD
 | |
|     info->seek_not_done=1;
 | |
| #endif
 | |
|   }
 | |
| 
 | |
|   /* Check if we want to write inside the used part of the buffer.*/
 | |
|   length= (size_t) (info->write_end - info->buffer);
 | |
|   if (pos < info->pos_in_file + length)
 | |
|   {
 | |
|     size_t offset= (size_t) (pos - info->pos_in_file);
 | |
|     length-=offset;
 | |
|     if (length > Count)
 | |
|       length=Count;
 | |
|     memcpy(info->buffer+offset, Buffer, length);
 | |
|     Buffer+=length;
 | |
|     Count-= length;
 | |
|     /* Fix length of buffer if the new data was larger */
 | |
|     if (info->buffer+length > info->write_pos)
 | |
|       info->write_pos=info->buffer+length;
 | |
|     if (!Count)
 | |
|       return (error);
 | |
|   }
 | |
|   /* Write at the end of the current buffer; This is the normal case */
 | |
|   if (_my_b_write(info, Buffer, Count))
 | |
|     error= -1;
 | |
|   return error;
 | |
| }
 | |
| 
 | |
| 
 | |
| 	/* Flush write cache */
 | |
| 
 | |
| #ifdef THREAD
 | |
| #define LOCK_APPEND_BUFFER if (need_append_buffer_lock) \
 | |
|   lock_append_buffer(info);
 | |
| #define UNLOCK_APPEND_BUFFER if (need_append_buffer_lock) \
 | |
|   unlock_append_buffer(info);
 | |
| #else
 | |
| #define LOCK_APPEND_BUFFER
 | |
| #define UNLOCK_APPEND_BUFFER
 | |
| #endif
 | |
| 
 | |
| 
 | |
| int my_b_flush_io_cache(IO_CACHE *info,
 | |
|                         int need_append_buffer_lock __attribute__((unused)))
 | |
| {
 | |
|   size_t length;
 | |
|   my_off_t pos_in_file;
 | |
|   my_bool append_cache= (info->type == SEQ_READ_APPEND);
 | |
|   DBUG_ENTER("my_b_flush_io_cache");
 | |
|   DBUG_PRINT("enter", ("cache: 0x%lx", (long) info));
 | |
| 
 | |
| #ifdef THREAD
 | |
|   if (!append_cache)
 | |
|     need_append_buffer_lock= 0;
 | |
| #endif
 | |
| 
 | |
|   if (info->type == WRITE_CACHE || append_cache)
 | |
|   {
 | |
|     if (info->file == -1)
 | |
|     {
 | |
|       if (real_open_cached_file(info))
 | |
| 	DBUG_RETURN((info->error= -1));
 | |
|     }
 | |
|     LOCK_APPEND_BUFFER;
 | |
| 
 | |
|     if ((length=(size_t) (info->write_pos - info->write_buffer)))
 | |
|     {
 | |
| #ifdef THREAD
 | |
|       /*
 | |
|         In case of a shared I/O cache with a writer we do direct write
 | |
|         cache to read cache copy. Do it before the write here so that
 | |
|         the readers can work in parallel with the write.
 | |
|         copy_to_read_buffer() relies on info->pos_in_file.
 | |
|       */
 | |
|       if (info->share)
 | |
|         copy_to_read_buffer(info, info->write_buffer, length);
 | |
| #endif
 | |
| 
 | |
|       pos_in_file=info->pos_in_file;
 | |
|       /*
 | |
| 	If we have append cache, we always open the file with
 | |
| 	O_APPEND which moves the pos to EOF automatically on every write
 | |
|       */
 | |
|       if (!append_cache && info->seek_not_done)
 | |
|       {					/* File touched, do seek */
 | |
| 	if (my_seek(info->file,pos_in_file,MY_SEEK_SET,MYF(0)) ==
 | |
| 	    MY_FILEPOS_ERROR)
 | |
| 	{
 | |
| 	  UNLOCK_APPEND_BUFFER;
 | |
| 	  DBUG_RETURN((info->error= -1));
 | |
| 	}
 | |
| 	if (!append_cache)
 | |
| 	  info->seek_not_done=0;
 | |
|       }
 | |
|       if (!append_cache)
 | |
| 	info->pos_in_file+=length;
 | |
|       info->write_end= (info->write_buffer+info->buffer_length-
 | |
| 			((pos_in_file+length) & (IO_SIZE-1)));
 | |
| 
 | |
|       if (my_write(info->file,info->write_buffer,length,
 | |
| 		   info->myflags | MY_NABP))
 | |
| 	info->error= -1;
 | |
|       else
 | |
| 	info->error= 0;
 | |
|       if (!append_cache)
 | |
|       {
 | |
|         set_if_bigger(info->end_of_file,(pos_in_file+length));
 | |
|       }
 | |
|       else
 | |
|       {
 | |
| 	info->end_of_file+=(info->write_pos-info->append_read_pos);
 | |
| 	DBUG_ASSERT(info->end_of_file == my_tell(info->file,MYF(0)));
 | |
|       }
 | |
| 
 | |
|       info->append_read_pos=info->write_pos=info->write_buffer;
 | |
|       ++info->disk_writes;
 | |
|       UNLOCK_APPEND_BUFFER;
 | |
|       DBUG_RETURN(info->error);
 | |
|     }
 | |
|   }
 | |
| #ifdef HAVE_AIOWAIT
 | |
|   else if (info->type != READ_NET)
 | |
|   {
 | |
|     my_aiowait(&info->aio_result);		/* Wait for outstanding req */
 | |
|     info->inited=0;
 | |
|   }
 | |
| #endif
 | |
|   UNLOCK_APPEND_BUFFER;
 | |
|   DBUG_RETURN(0);
 | |
| }
 | |
| 
 | |
| /*
 | |
|   Free an IO_CACHE object
 | |
| 
 | |
|   SYNOPSOS
 | |
|     end_io_cache()
 | |
|     info		IO_CACHE Handle to free
 | |
| 
 | |
|   NOTES
 | |
|     It's currently safe to call this if one has called init_io_cache()
 | |
|     on the 'info' object, even if init_io_cache() failed.
 | |
|     This function is also safe to call twice with the same handle.
 | |
| 
 | |
|   RETURN
 | |
|    0  ok
 | |
|    #  Error
 | |
| */
 | |
| 
 | |
| int end_io_cache(IO_CACHE *info)
 | |
| {
 | |
|   int error=0;
 | |
|   IO_CACHE_CALLBACK pre_close;
 | |
|   DBUG_ENTER("end_io_cache");
 | |
|   DBUG_PRINT("enter",("cache: 0x%lx", (ulong) info));
 | |
| 
 | |
| #ifdef THREAD
 | |
|   /*
 | |
|     Every thread must call remove_io_thread(). The last one destroys
 | |
|     the share elements.
 | |
|   */
 | |
|   DBUG_ASSERT(!info->share || !info->share->total_threads);
 | |
| #endif
 | |
| 
 | |
|   if ((pre_close=info->pre_close))
 | |
|   {
 | |
|     (*pre_close)(info);
 | |
|     info->pre_close= 0;
 | |
|   }
 | |
|   if (info->alloced_buffer)
 | |
|   {
 | |
|     info->alloced_buffer=0;
 | |
|     if (info->file != -1)			/* File doesn't exist */
 | |
|       error= my_b_flush_io_cache(info,1);
 | |
|     my_free((uchar*) info->buffer,MYF(MY_WME));
 | |
|     info->buffer=info->read_pos=(uchar*) 0;
 | |
|   }
 | |
|   if (info->type == SEQ_READ_APPEND)
 | |
|   {
 | |
|     /* Destroy allocated mutex */
 | |
|     info->type= TYPE_NOT_SET;
 | |
| #ifdef THREAD
 | |
|     pthread_mutex_destroy(&info->append_buffer_lock);
 | |
| #endif
 | |
|   }
 | |
| #ifdef THREAD
 | |
|   info->share= 0;
 | |
| #endif
 | |
|   DBUG_RETURN(error);
 | |
| } /* end_io_cache */
 | |
| 
 | |
| 
 | |
| /**********************************************************************
 | |
|  Testing of MF_IOCACHE
 | |
| **********************************************************************/
 | |
| 
 | |
| #ifdef MAIN
 | |
| 
 | |
| #include <my_dir.h>
 | |
| 
 | |
| void die(const char* fmt, ...)
 | |
| {
 | |
|   va_list va_args;
 | |
|   va_start(va_args,fmt);
 | |
|   fprintf(stderr,"Error:");
 | |
|   vfprintf(stderr, fmt,va_args);
 | |
|   fprintf(stderr,", errno=%d\n", errno);
 | |
|   exit(1);
 | |
| }
 | |
| 
 | |
| int open_file(const char* fname, IO_CACHE* info, int cache_size)
 | |
| {
 | |
|   int fd;
 | |
|   if ((fd=my_open(fname,O_CREAT | O_RDWR,MYF(MY_WME))) < 0)
 | |
|     die("Could not open %s", fname);
 | |
|   if (init_io_cache(info, fd, cache_size, SEQ_READ_APPEND, 0,0,MYF(MY_WME)))
 | |
|     die("failed in init_io_cache()");
 | |
|   return fd;
 | |
| }
 | |
| 
 | |
| void close_file(IO_CACHE* info)
 | |
| {
 | |
|   end_io_cache(info);
 | |
|   my_close(info->file, MYF(MY_WME));
 | |
| }
 | |
| 
 | |
| int main(int argc, char** argv)
 | |
| {
 | |
|   IO_CACHE sra_cache; /* SEQ_READ_APPEND */
 | |
|   MY_STAT status;
 | |
|   const char* fname="/tmp/iocache.test";
 | |
|   int cache_size=16384;
 | |
|   char llstr_buf[22];
 | |
|   int max_block,total_bytes=0;
 | |
|   int i,num_loops=100,error=0;
 | |
|   char *p;
 | |
|   char* block, *block_end;
 | |
|   MY_INIT(argv[0]);
 | |
|   max_block = cache_size*3;
 | |
|   if (!(block=(char*)my_malloc(max_block,MYF(MY_WME))))
 | |
|     die("Not enough memory to allocate test block");
 | |
|   block_end = block + max_block;
 | |
|   for (p = block,i=0; p < block_end;i++)
 | |
|   {
 | |
|     *p++ = (char)i;
 | |
|   }
 | |
|   if (my_stat(fname,&status, MYF(0)) &&
 | |
|       my_delete(fname,MYF(MY_WME)))
 | |
|     {
 | |
|       die("Delete of %s failed, aborting", fname);
 | |
|     }
 | |
|   open_file(fname,&sra_cache, cache_size);
 | |
|   for (i = 0; i < num_loops; i++)
 | |
|   {
 | |
|     char buf[4];
 | |
|     int block_size = abs(rand() % max_block);
 | |
|     int4store(buf, block_size);
 | |
|     if (my_b_append(&sra_cache,buf,4) ||
 | |
| 	my_b_append(&sra_cache, block, block_size))
 | |
|       die("write failed");
 | |
|     total_bytes += 4+block_size;
 | |
|   }
 | |
|   close_file(&sra_cache);
 | |
|   my_free(block,MYF(MY_WME));
 | |
|   if (!my_stat(fname,&status,MYF(MY_WME)))
 | |
|     die("%s failed to stat, but I had just closed it,\
 | |
|  wonder how that happened");
 | |
|   printf("Final size of %s is %s, wrote %d bytes\n",fname,
 | |
| 	 llstr(status.st_size,llstr_buf),
 | |
| 	 total_bytes);
 | |
|   my_delete(fname, MYF(MY_WME));
 | |
|   /* check correctness of tests */
 | |
|   if (total_bytes != status.st_size)
 | |
|   {
 | |
|     fprintf(stderr,"Not the same number of bytes acutally  in file as bytes \
 | |
| supposedly written\n");
 | |
|     error=1;
 | |
|   }
 | |
|   exit(error);
 | |
|   return 0;
 | |
| }
 | |
| #endif
 |