477
489
info->file is a pipe or socket or FIFO. We never should have tried
478
490
to seek on that. See Bugs#25807 and #22828 for more info.
480
assert(errno != ESPIPE);
492
DBUG_ASSERT(my_errno != ESPIPE);
486
diff_length= (size_t) (pos_in_file_local & (IO_SIZE-1));
498
diff_length= (size_t) (pos_in_file & (IO_SIZE-1));
487
499
if (Count >= (size_t) (IO_SIZE+(IO_SIZE-diff_length)))
488
500
{ /* Fill first intern buffer */
489
501
size_t read_length;
490
if (info->end_of_file <= pos_in_file_local)
502
if (info->end_of_file <= pos_in_file)
491
503
{ /* End of file */
492
504
info->error= (int) left_length;
495
length_local=(Count & (size_t) ~(IO_SIZE-1))-diff_length;
496
if ((read_length= my_read(info->file,Buffer, length_local, info->myflags)) != length_local)
507
length=(Count & (size_t) ~(IO_SIZE-1))-diff_length;
508
if ((read_length= my_read(info->file,Buffer, length, info->myflags))
498
511
info->error= (read_length == (size_t) -1 ? -1 :
499
512
(int) (read_length+left_length));
502
Count-= length_local;
503
Buffer+= length_local;
504
pos_in_file_local+= length_local;
505
left_length+= length_local;
509
522
max_length= info->read_length-diff_length;
510
523
if (info->type != READ_FIFO &&
511
max_length > (info->end_of_file - pos_in_file_local))
512
max_length= (size_t) (info->end_of_file - pos_in_file_local);
524
max_length > (info->end_of_file - pos_in_file))
525
max_length= (size_t) (info->end_of_file - pos_in_file);
517
info->error= static_cast<int>(left_length); /* We only got this many char */
530
info->error= left_length; /* We only got this many char */
520
length_local=0; /* Didn't read any chars */
533
length=0; /* Didn't read any chars */
522
else if (( length_local= my_read(info->file,info->buffer, max_length,
535
else if ((length= my_read(info->file,info->buffer, max_length,
523
536
info->myflags)) < Count ||
524
length_local == (size_t) -1)
537
length == (size_t) -1)
526
if ( length_local != (size_t) -1)
527
memcpy(Buffer, info->buffer, length_local);
528
info->pos_in_file= pos_in_file_local;
529
info->error= length_local == (size_t) -1 ? -1 : (int) ( length_local+left_length);
539
if (length != (size_t) -1)
540
memcpy(Buffer, info->buffer, length);
541
info->pos_in_file= pos_in_file;
542
info->error= length == (size_t) -1 ? -1 : (int) (length+left_length);
530
543
info->read_pos=info->read_end=info->buffer;
533
546
info->read_pos=info->buffer+Count;
534
info->read_end=info->buffer+ length_local;
535
info->pos_in_file=pos_in_file_local;
547
info->read_end=info->buffer+length;
548
info->pos_in_file=pos_in_file;
536
549
memcpy(Buffer, info->buffer, Count);
556
Prepare IO_CACHE for shared use.
559
init_io_cache_share()
560
read_cache A read cache. This will be copied for
561
every thread after setup.
563
write_cache If non-NULL a write cache that is to be
564
synchronized with the read caches.
565
num_threads Number of threads sharing the cache
566
including the write thread if any.
570
The shared cache is used so: One IO_CACHE is initialized with
571
init_io_cache(). This includes the allocation of a buffer. Then a
572
share is allocated and init_io_cache_share() is called with the io
573
cache and the share. Then the io cache is copied for each thread. So
574
every thread has its own copy of IO_CACHE. But the allocated buffer
575
is shared because cache->buffer is the same for all caches.
577
One thread reads data from the file into the buffer. All threads
578
read from the buffer, but every thread maintains its own set of
579
pointers into the buffer. When all threads have used up the buffer
580
contents, one of the threads reads the next block of data into the
581
buffer. To accomplish this, each thread enters the cache lock before
582
accessing the buffer. They wait in lock_io_cache() until all threads
583
joined the lock. The last thread entering the lock is in charge of
584
reading from file to buffer. It wakes all threads when done.
586
Synchronizing a write cache to the read caches works so: Whenever
587
the write buffer needs a flush, the write thread enters the lock and
588
waits for all other threads to enter the lock too. They do this when
589
they have used up the read buffer. When all threads are in the lock,
590
the write thread copies the write buffer to the read buffer and
593
share->running_threads is the number of threads not being in the
594
cache lock. When entering lock_io_cache() the number is decreased.
595
When the thread that fills the buffer enters unlock_io_cache() the
596
number is reset to the number of threads. The condition
597
running_threads == 0 means that all threads are in the lock. Bumping
598
up the number to the full count is non-intuitive. But increasing the
599
number by one for each thread that leaves the lock could lead to a
600
solo run of one thread. The last thread to join a lock reads from
601
file to buffer, wakes the other threads, processes the data in the
602
cache and enters the lock again. If no other thread left the lock
603
meanwhile, it would think it's the last one again and read the next
606
The share has copies of 'error', 'buffer', 'read_end', and
607
'pos_in_file' from the thread that filled the buffer. We may not be
608
able to access this information directly from its cache because the
609
thread may be removed from the share before the variables could be
610
copied by all other threads. Or, if a write buffer is synchronized,
611
it would change its 'pos_in_file' after waking the other threads,
612
possibly before they could copy its value.
614
However, the 'buffer' variable in the share is for a synchronized
615
write cache. It needs to know where to put the data. Otherwise it
616
would need access to the read cache of one of the threads that is
617
not yet removed from the share.
623
void init_io_cache_share(IO_CACHE *read_cache, IO_CACHE_SHARE *cshare,
624
IO_CACHE *write_cache, uint num_threads)
626
DBUG_ENTER("init_io_cache_share");
627
DBUG_PRINT("io_cache_share", ("read_cache: 0x%lx share: 0x%lx "
628
"write_cache: 0x%lx threads: %u",
629
(long) read_cache, (long) cshare,
630
(long) write_cache, num_threads));
632
DBUG_ASSERT(num_threads > 1);
633
DBUG_ASSERT(read_cache->type == READ_CACHE);
634
DBUG_ASSERT(!write_cache || (write_cache->type == WRITE_CACHE));
636
pthread_mutex_init(&cshare->mutex, MY_MUTEX_INIT_FAST);
637
pthread_cond_init(&cshare->cond, 0);
638
pthread_cond_init(&cshare->cond_writer, 0);
640
cshare->running_threads= num_threads;
641
cshare->total_threads= num_threads;
642
cshare->error= 0; /* Initialize. */
643
cshare->buffer= read_cache->buffer;
644
cshare->read_end= NULL; /* See function comment of lock_io_cache(). */
645
cshare->pos_in_file= 0; /* See function comment of lock_io_cache(). */
646
cshare->source_cache= write_cache; /* Can be NULL. */
648
read_cache->share= cshare;
649
read_cache->read_function= _my_b_read_r;
650
read_cache->current_pos= NULL;
651
read_cache->current_end= NULL;
654
write_cache->share= cshare;
661
Remove a thread from shared access to IO_CACHE.
665
cache The IO_CACHE to be removed from the share.
669
Every thread must do that on exit for not to deadlock other threads.
671
The last thread destroys the pthread resources.
673
A writer flushes its cache first.
679
void remove_io_thread(IO_CACHE *cache)
681
IO_CACHE_SHARE *cshare= cache->share;
683
DBUG_ENTER("remove_io_thread");
685
/* If the writer goes, it needs to flush the write cache. */
686
if (cache == cshare->source_cache)
687
flush_io_cache(cache);
689
pthread_mutex_lock(&cshare->mutex);
690
DBUG_PRINT("io_cache_share", ("%s: 0x%lx",
691
(cache == cshare->source_cache) ?
692
"writer" : "reader", (long) cache));
694
/* Remove from share. */
695
total= --cshare->total_threads;
696
DBUG_PRINT("io_cache_share", ("remaining threads: %u", total));
698
/* Detach from share. */
701
/* If the writer goes, let the readers know. */
702
if (cache == cshare->source_cache)
704
DBUG_PRINT("io_cache_share", ("writer leaves"));
705
cshare->source_cache= NULL;
708
/* If all threads are waiting for me to join the lock, wake them. */
709
if (!--cshare->running_threads)
711
DBUG_PRINT("io_cache_share", ("the last running thread leaves, wake all"));
712
pthread_cond_signal(&cshare->cond_writer);
713
pthread_cond_broadcast(&cshare->cond);
716
pthread_mutex_unlock(&cshare->mutex);
720
DBUG_PRINT("io_cache_share", ("last thread removed, destroy share"));
721
pthread_cond_destroy (&cshare->cond_writer);
722
pthread_cond_destroy (&cshare->cond);
723
pthread_mutex_destroy(&cshare->mutex);
731
Lock IO cache and wait for all other threads to join.
735
cache The cache of the thread entering the lock.
736
pos File position of the block to read.
737
Unused for the write thread.
741
Wait for all threads to finish with the current buffer. We want
742
all threads to proceed in concert. The last thread to join
743
lock_io_cache() will read the block from file and all threads start
744
to use it. Then they will join again for reading the next block.
746
The waiting threads detect a fresh buffer by comparing
747
cshare->pos_in_file with the position they want to process next.
748
Since the first block may start at position 0, we take
749
cshare->read_end as an additional condition. This variable is
750
initialized to NULL and will be set after a block of data is written
754
1 OK, lock in place, go ahead and read.
755
0 OK, unlocked, another thread did the read.
758
static int lock_io_cache(IO_CACHE *cache, my_off_t pos)
760
IO_CACHE_SHARE *cshare= cache->share;
761
DBUG_ENTER("lock_io_cache");
763
/* Enter the lock. */
764
pthread_mutex_lock(&cshare->mutex);
765
cshare->running_threads--;
766
DBUG_PRINT("io_cache_share", ("%s: 0x%lx pos: %lu running: %u",
767
(cache == cshare->source_cache) ?
768
"writer" : "reader", (long) cache, (ulong) pos,
769
cshare->running_threads));
771
if (cshare->source_cache)
773
/* A write cache is synchronized to the read caches. */
775
if (cache == cshare->source_cache)
777
/* The writer waits until all readers are here. */
778
while (cshare->running_threads)
780
DBUG_PRINT("io_cache_share", ("writer waits in lock"));
781
pthread_cond_wait(&cshare->cond_writer, &cshare->mutex);
783
DBUG_PRINT("io_cache_share", ("writer awoke, going to copy"));
785
/* Stay locked. Leave the lock later by unlock_io_cache(). */
789
/* The last thread wakes the writer. */
790
if (!cshare->running_threads)
792
DBUG_PRINT("io_cache_share", ("waking writer"));
793
pthread_cond_signal(&cshare->cond_writer);
797
Readers wait until the data is copied from the writer. Another
798
reason to stop waiting is the removal of the write thread. If this
799
happens, we leave the lock with old data in the buffer.
801
while ((!cshare->read_end || (cshare->pos_in_file < pos)) &&
802
cshare->source_cache)
804
DBUG_PRINT("io_cache_share", ("reader waits in lock"));
805
pthread_cond_wait(&cshare->cond, &cshare->mutex);
809
If the writer was removed from the share while this thread was
810
asleep, we need to simulate an EOF condition. The writer cannot
811
reset the share variables as they might still be in use by readers
812
of the last block. When we awake here then because the last
813
joining thread signalled us. If the writer is not the last, it
814
will not signal. So it is safe to clear the buffer here.
816
if (!cshare->read_end || (cshare->pos_in_file < pos))
818
DBUG_PRINT("io_cache_share", ("reader found writer removed. EOF"));
819
cshare->read_end= cshare->buffer; /* Empty buffer. */
820
cshare->error= 0; /* EOF is not an error. */
826
There are read caches only. The last thread arriving in
827
lock_io_cache() continues with a locked cache and reads the block.
829
if (!cshare->running_threads)
831
DBUG_PRINT("io_cache_share", ("last thread joined, going to read"));
832
/* Stay locked. Leave the lock later by unlock_io_cache(). */
837
All other threads wait until the requested block is read by the
838
last thread arriving. Another reason to stop waiting is the
839
removal of a thread. If this leads to all threads being in the
840
lock, we have to continue also. The first of the awaken threads
841
will then do the read.
843
while ((!cshare->read_end || (cshare->pos_in_file < pos)) &&
844
cshare->running_threads)
846
DBUG_PRINT("io_cache_share", ("reader waits in lock"));
847
pthread_cond_wait(&cshare->cond, &cshare->mutex);
850
/* If the block is not yet read, continue with a locked cache and read. */
851
if (!cshare->read_end || (cshare->pos_in_file < pos))
853
DBUG_PRINT("io_cache_share", ("reader awoke, going to read"));
854
/* Stay locked. Leave the lock later by unlock_io_cache(). */
858
/* Another thread did read the block already. */
860
DBUG_PRINT("io_cache_share", ("reader awoke, going to process %u bytes",
861
(uint) (cshare->read_end ? (size_t)
862
(cshare->read_end - cshare->buffer) :
866
Leave the lock. Do not call unlock_io_cache() later. The thread that
867
filled the buffer did this and marked all threads as running.
869
pthread_mutex_unlock(&cshare->mutex);
879
cache The cache of the thread leaving the lock.
882
This is called by the thread that filled the buffer. It marks all
883
threads as running and awakes them. This must not be done by any
886
Do not signal cond_writer. Either there is no writer or the writer
887
is the only one who can call this function.
889
The reason for resetting running_threads to total_threads before
890
waking all other threads is that it could be possible that this
891
thread is so fast with processing the buffer that it enters the lock
892
before even one other thread has left it. If every awoken thread
893
would increase running_threads by one, this thread could think that
894
he is again the last to join and would not wait for the other
895
threads to process the data.
901
static void unlock_io_cache(IO_CACHE *cache)
903
IO_CACHE_SHARE *cshare= cache->share;
904
DBUG_ENTER("unlock_io_cache");
905
DBUG_PRINT("io_cache_share", ("%s: 0x%lx pos: %lu running: %u",
906
(cache == cshare->source_cache) ?
908
(long) cache, (ulong) cshare->pos_in_file,
909
cshare->total_threads));
911
cshare->running_threads= cshare->total_threads;
912
pthread_cond_broadcast(&cshare->cond);
913
pthread_mutex_unlock(&cshare->mutex);
919
Read from IO_CACHE when it is shared between several threads.
923
cache IO_CACHE pointer
924
Buffer Buffer to retrieve count bytes from file
925
Count Number of bytes to read into Buffer
928
This function is only called from the my_b_read() macro when there
929
isn't enough characters in the buffer to satisfy the request.
933
It works as follows: when a thread tries to read from a file (that
934
is, after using all the data from the (shared) buffer), it just
935
hangs on lock_io_cache(), waiting for other threads. When the very
936
last thread attempts a read, lock_io_cache() returns 1, the thread
937
does actual IO and unlock_io_cache(), which signals all the waiting
938
threads that data is in the buffer.
942
When changing this function, be careful with handling file offsets
943
(end-of_file, pos_in_file). Do not cast them to possibly smaller
944
types than my_off_t unless you can be sure that their value fits.
945
Same applies to differences of file offsets. (Bug #11527)
947
When changing this function, check _my_b_read(). It might need the
951
0 we succeeded in reading all data
952
1 Error: can't read requested characters
955
int _my_b_read_r(register IO_CACHE *cache, uchar *Buffer, size_t Count)
957
my_off_t pos_in_file;
958
size_t length, diff_length, left_length;
959
IO_CACHE_SHARE *cshare= cache->share;
960
DBUG_ENTER("_my_b_read_r");
962
if ((left_length= (size_t) (cache->read_end - cache->read_pos)))
964
DBUG_ASSERT(Count >= left_length); /* User is not using my_b_read() */
965
memcpy(Buffer, cache->read_pos, left_length);
966
Buffer+= left_length;
973
pos_in_file= cache->pos_in_file + (cache->read_end - cache->buffer);
974
diff_length= (size_t) (pos_in_file & (IO_SIZE-1));
975
length=IO_ROUND_UP(Count+diff_length)-diff_length;
976
length= ((length <= cache->read_length) ?
977
length + IO_ROUND_DN(cache->read_length - length) :
978
length - IO_ROUND_UP(length - cache->read_length));
979
if (cache->type != READ_FIFO &&
980
(length > (cache->end_of_file - pos_in_file)))
981
length= (size_t) (cache->end_of_file - pos_in_file);
984
cache->error= (int) left_length;
987
if (lock_io_cache(cache, pos_in_file))
989
/* With a synchronized write/read cache we won't come here... */
990
DBUG_ASSERT(!cshare->source_cache);
992
... unless the writer has gone before this thread entered the
993
lock. Simulate EOF in this case. It can be distinguished by
1001
Whenever a function which operates on IO_CACHE flushes/writes
1002
some part of the IO_CACHE to disk it will set the property
1003
"seek_not_done" to indicate this to other functions operating
1006
if (cache->seek_not_done)
1008
if (my_seek(cache->file, pos_in_file, MY_SEEK_SET, MYF(0))
1009
== MY_FILEPOS_ERROR)
1012
unlock_io_cache(cache);
1016
len= my_read(cache->file, cache->buffer, length, cache->myflags);
1018
DBUG_PRINT("io_cache_share", ("read %lu bytes", (ulong) len));
1020
cache->read_end= cache->buffer + (len == (size_t) -1 ? 0 : len);
1021
cache->error= (len == length ? 0 : (int) len);
1022
cache->pos_in_file= pos_in_file;
1024
/* Copy important values to the share. */
1025
cshare->error= cache->error;
1026
cshare->read_end= cache->read_end;
1027
cshare->pos_in_file= pos_in_file;
1029
/* Mark all threads as running and wake them. */
1030
unlock_io_cache(cache);
1035
With a synchronized write/read cache readers always come here.
1036
Copy important values from the share.
1038
cache->error= cshare->error;
1039
cache->read_end= cshare->read_end;
1040
cache->pos_in_file= cshare->pos_in_file;
1042
len= ((cache->error == -1) ? (size_t) -1 :
1043
(size_t) (cache->read_end - cache->buffer));
1045
cache->read_pos= cache->buffer;
1046
cache->seek_not_done= 0;
1047
if (len == 0 || len == (size_t) -1)
1049
DBUG_PRINT("io_cache_share", ("reader error. len %lu left %lu",
1050
(ulong) len, (ulong) left_length));
1051
cache->error= (int) left_length;
1054
cnt= (len > Count) ? Count : len;
1055
memcpy(Buffer, cache->read_pos, cnt);
1059
cache->read_pos+= cnt;
1066
Copy data from write cache to read cache.
1069
copy_to_read_buffer()
1070
write_cache The write cache.
1071
write_buffer The source of data, mostly the cache buffer.
1072
write_length The number of bytes to copy.
1075
The write thread will wait for all read threads to join the cache
1076
lock. Then it copies the data over and wakes the read threads.
1082
static void copy_to_read_buffer(IO_CACHE *write_cache,
1083
const uchar *write_buffer, size_t write_length)
1085
IO_CACHE_SHARE *cshare= write_cache->share;
1087
DBUG_ASSERT(cshare->source_cache == write_cache);
1089
write_length is usually less or equal to buffer_length.
1090
It can be bigger if _my_b_write() is called with a big length.
1092
while (write_length)
1094
size_t copy_length= min(write_length, write_cache->buffer_length);
1095
int __attribute__((unused)) rc;
1097
rc= lock_io_cache(write_cache, write_cache->pos_in_file);
1098
/* The writing thread does always have the lock when it awakes. */
1101
memcpy(cshare->buffer, write_buffer, copy_length);
1104
cshare->read_end= cshare->buffer + copy_length;
1105
cshare->pos_in_file= write_cache->pos_in_file;
1107
/* Mark all threads as running and wake them. */
1108
unlock_io_cache(write_cache);
1110
write_buffer+= copy_length;
1111
write_length-= copy_length;
1118
Do sequential read from the SEQ_READ_APPEND cache.
1120
We do this in three stages:
1121
- first read from info->buffer
1122
- then if there are still data to read, try the file descriptor
1123
- afterwards, if there are still data to read, try append buffer
1130
int _my_b_seq_read(register IO_CACHE *info, uchar *Buffer, size_t Count)
1132
size_t length, diff_length, left_length, save_count, max_length;
1133
my_off_t pos_in_file;
1136
/* first, read the regular buffer */
1137
if ((left_length=(size_t) (info->read_end-info->read_pos)))
1139
DBUG_ASSERT(Count > left_length); /* User is not using my_b_read() */
1140
memcpy(Buffer,info->read_pos, left_length);
1141
Buffer+=left_length;
1144
lock_append_buffer(info);
1146
/* pos_in_file always point on where info->buffer was read */
1147
if ((pos_in_file=info->pos_in_file +
1148
(size_t) (info->read_end - info->buffer)) >= info->end_of_file)
1149
goto read_append_buffer;
1152
With read-append cache we must always do a seek before we read,
1153
because the write could have moved the file pointer astray
1155
if (my_seek(info->file,pos_in_file,MY_SEEK_SET,MYF(0)) == MY_FILEPOS_ERROR)
1158
unlock_append_buffer(info);
1161
info->seek_not_done=0;
1163
diff_length= (size_t) (pos_in_file & (IO_SIZE-1));
1165
/* now the second stage begins - read from file descriptor */
1166
if (Count >= (size_t) (IO_SIZE+(IO_SIZE-diff_length)))
1168
/* Fill first intern buffer */
1171
length=(Count & (size_t) ~(IO_SIZE-1))-diff_length;
1172
if ((read_length= my_read(info->file,Buffer, length,
1173
info->myflags)) == (size_t) -1)
1176
unlock_append_buffer(info);
1180
Buffer+=read_length;
1181
pos_in_file+=read_length;
1183
if (read_length != length)
1186
We only got part of data; Read the rest of the data from the
1189
goto read_append_buffer;
1191
left_length+=length;
1195
max_length= info->read_length-diff_length;
1196
if (max_length > (info->end_of_file - pos_in_file))
1197
max_length= (size_t) (info->end_of_file - pos_in_file);
1201
goto read_append_buffer;
1202
length=0; /* Didn't read any more chars */
1206
length= my_read(info->file,info->buffer, max_length, info->myflags);
1207
if (length == (size_t) -1)
1210
unlock_append_buffer(info);
1215
memcpy(Buffer, info->buffer, length);
1220
added the line below to make
1221
DBUG_ASSERT(pos_in_file==info->end_of_file) pass.
1222
otherwise this does not appear to be needed
1224
pos_in_file += length;
1225
goto read_append_buffer;
1228
unlock_append_buffer(info);
1229
info->read_pos=info->buffer+Count;
1230
info->read_end=info->buffer+length;
1231
info->pos_in_file=pos_in_file;
1232
memcpy(Buffer,info->buffer,(size_t) Count);
1238
Read data from the current write buffer.
1239
Count should never be == 0 here (The code will work even if count is 0)
1243
/* First copy the data to Count */
1244
size_t len_in_buff = (size_t) (info->write_pos - info->append_read_pos);
1246
size_t transfer_len;
1248
DBUG_ASSERT(info->append_read_pos <= info->write_pos);
1250
TODO: figure out if the assert below is needed or correct.
1252
DBUG_ASSERT(pos_in_file == info->end_of_file);
1253
copy_len=min(Count, len_in_buff);
1254
memcpy(Buffer, info->append_read_pos, copy_len);
1255
info->append_read_pos += copy_len;
1258
info->error = save_count - Count;
1260
/* Fill read buffer with data from write buffer */
1261
memcpy(info->buffer, info->append_read_pos,
1262
(size_t) (transfer_len=len_in_buff - copy_len));
1263
info->read_pos= info->buffer;
1264
info->read_end= info->buffer+transfer_len;
1265
info->append_read_pos=info->write_pos;
1266
info->pos_in_file=pos_in_file+copy_len;
1267
info->end_of_file+=len_in_buff;
1269
unlock_append_buffer(info);
1270
return Count ? 1 : 0;
541
1274
#ifdef HAVE_AIOWAIT
545
* Read from the st_io_cache into a buffer and feed asynchronously from disk when needed.
547
* @param info st_io_cache pointer
548
* @param Buffer Buffer to retrieve count bytes from file
549
* @param Count Number of bytes to read into Buffer
551
* @retval -1 An error has occurred; errno is set.
553
* @retval 1 An error has occurred; st_io_cache to error state.
555
int _my_b_async_read(register st_io_cache *info, unsigned char *Buffer, size_t Count)
1277
Read from the IO_CACHE into a buffer and feed asynchronously
1278
from disk when needed.
1282
info IO_CACHE pointer
1283
Buffer Buffer to retrieve count bytes from file
1284
Count Number of bytes to read into Buffer
1287
-1 An error has occurred; my_errno is set.
1289
1 An error has occurred; IO_CACHE to error state.
1292
int _my_b_async_read(register IO_CACHE *info, uchar *Buffer, size_t Count)
557
size_t length_local,read_length,diff_length,left_length,use_length,org_Count;
1294
size_t length,read_length,diff_length,left_length,use_length,org_Count;
558
1295
size_t max_length;
559
1296
my_off_t next_pos_in_file;
560
unsigned char *read_buffer;
562
1299
memcpy(Buffer,info->read_pos,
563
1300
(left_length= (size_t) (info->read_end-info->read_pos)));
761
1510
if (Count >= IO_SIZE)
762
1511
{ /* Fill first intern buffer */
763
length_local=Count & (size_t) ~(IO_SIZE-1);
1512
length=Count & (size_t) ~(IO_SIZE-1);
764
1513
if (info->seek_not_done)
767
Whenever a function which operates on st_io_cache flushes/writes
768
some part of the st_io_cache to disk it will set the property
1516
Whenever a function which operates on IO_CACHE flushes/writes
1517
some part of the IO_CACHE to disk it will set the property
769
1518
"seek_not_done" to indicate this to other functions operating
772
if (lseek(info->file,info->pos_in_file,SEEK_SET))
1521
if (my_seek(info->file,info->pos_in_file,MY_SEEK_SET,MYF(0)))
774
1523
info->error= -1;
777
1526
info->seek_not_done=0;
779
if (my_write(info->file, Buffer, length_local, info->myflags | MY_NABP))
780
return info->error= -1;
783
Buffer+=length_local;
784
info->pos_in_file+=length_local;
786
memcpy(info->write_pos,Buffer,(size_t) Count);
787
info->write_pos+=Count;
793
* Write a block to disk where part of the data may be inside the record buffer.
794
* As all write calls to the data goes through the cache,
795
* we will never get a seek over the end of the buffer.
797
int my_block_write(register st_io_cache *info, const unsigned char *Buffer, size_t Count,
1528
if (my_write(info->file, Buffer, length, info->myflags | MY_NABP))
1529
return info->error= -1;
1533
In case of a shared I/O cache with a writer we normally do direct
1534
write cache to read cache copy. Simulate this here by direct
1535
caller buffer to read cache copy. Do it after the write so that
1536
the cache readers actions on the flushed part can go in parallel
1537
with the write of the extra stuff. copy_to_read_buffer()
1538
synchronizes writer and readers so that after this call the
1539
readers can act on the extra stuff while the writer can go ahead
1540
and prepare the next output. copy_to_read_buffer() relies on
1544
copy_to_read_buffer(info, Buffer, length);
1549
info->pos_in_file+=length;
1551
memcpy(info->write_pos,Buffer,(size_t) Count);
1552
info->write_pos+=Count;
1558
Append a block to the write buffer.
1559
This is done with the buffer locked to ensure that we don't read from
1560
the write buffer before we are ready with it.
1563
int my_b_append(register IO_CACHE *info, const uchar *Buffer, size_t Count)
1565
size_t rest_length,length;
1569
Assert that we cannot come here with a shared cache. If we do one
1570
day, we might need to add a call to copy_to_read_buffer().
1572
DBUG_ASSERT(!info->share);
1575
lock_append_buffer(info);
1576
rest_length= (size_t) (info->write_end - info->write_pos);
1577
if (Count <= rest_length)
1579
memcpy(info->write_pos, Buffer, rest_length);
1580
Buffer+=rest_length;
1582
info->write_pos+=rest_length;
1583
if (my_b_flush_io_cache(info,0))
1585
unlock_append_buffer(info);
1588
if (Count >= IO_SIZE)
1589
{ /* Fill first intern buffer */
1590
length=Count & (size_t) ~(IO_SIZE-1);
1591
if (my_write(info->file,Buffer, length, info->myflags | MY_NABP))
1593
unlock_append_buffer(info);
1594
return info->error= -1;
1598
info->end_of_file+=length;
1602
memcpy(info->write_pos,Buffer,(size_t) Count);
1603
info->write_pos+=Count;
1604
unlock_append_buffer(info);
1609
int my_b_safe_write(IO_CACHE *info, const uchar *Buffer, size_t Count)
1612
Sasha: We are not writing this with the ? operator to avoid hitting
1613
a possible compiler bug. At least gcc 2.95 cannot deal with
1614
several layers of ternary operators that evaluated comma(,) operator
1615
expressions inside - I do have a test case if somebody wants it
1617
if (info->type == SEQ_READ_APPEND)
1618
return my_b_append(info, Buffer, Count);
1619
return my_b_write(info, Buffer, Count);
1624
Write a block to disk where part of the data may be inside the record
1625
buffer. As all write calls to the data goes through the cache,
1626
we will never get a seek over the end of the buffer
1629
int my_block_write(register IO_CACHE *info, const uchar *Buffer, size_t Count,
1637
Assert that we cannot come here with a shared cache. If we do one
1638
day, we might need to add a call to copy_to_read_buffer().
1640
DBUG_ASSERT(!info->share);
803
1643
if (pos < info->pos_in_file)
805
1645
/* Of no overlap, write everything without buffering */
806
1646
if (pos + Count <= info->pos_in_file)
807
return (pwrite(info->file, Buffer, Count, pos) == 0);
1647
return my_pwrite(info->file, Buffer, Count, pos,
1648
info->myflags | MY_NABP);
808
1649
/* Write the part of the block that is before buffer */
809
length_local= (uint32_t) (info->pos_in_file - pos);
810
if (pwrite(info->file, Buffer, length_local, pos) == 0)
1650
length= (uint) (info->pos_in_file - pos);
1651
if (my_pwrite(info->file, Buffer, length, pos, info->myflags | MY_NABP))
811
1652
info->error= error= -1;
812
Buffer+=length_local;
814
Count-= length_local;
1657
info->seek_not_done=1;
817
1661
/* Check if we want to write inside the used part of the buffer.*/
818
length_local= (size_t) (info->write_end - info->buffer);
819
if (pos < info->pos_in_file + length_local)
1662
length= (size_t) (info->write_end - info->buffer);
1663
if (pos < info->pos_in_file + length)
821
1665
size_t offset= (size_t) (pos - info->pos_in_file);
822
length_local-=offset;
823
if (length_local > Count)
825
memcpy(info->buffer+offset, Buffer, length_local);
826
Buffer+=length_local;
827
Count-= length_local;
828
/* Fix length_local of buffer if the new data was larger */
829
if (info->buffer+length_local > info->write_pos)
830
info->write_pos=info->buffer+length_local;
1669
memcpy(info->buffer+offset, Buffer, length);
1672
/* Fix length of buffer if the new data was larger */
1673
if (info->buffer+length > info->write_pos)
1674
info->write_pos=info->buffer+length;
909
unlock_append_buffer(info, need_append_buffer_lock);
1780
UNLOCK_APPEND_BUFFER;
915
* Free an st_io_cache object
918
* It's currently safe to call this if one has called init_io_cache()
919
* on the 'info' object, even if init_io_cache() failed.
920
* This function is also safe to call twice with the same handle.
922
* @param info st_io_cache Handle to free
927
int st_io_cache::end_io_cache()
1785
Free an IO_CACHE object
1789
info IO_CACHE Handle to free
1792
It's currently safe to call this if one has called init_io_cache()
1793
on the 'info' object, even if init_io_cache() failed.
1794
This function is also safe to call twice with the same handle.
1801
int end_io_cache(IO_CACHE *info)
938
if (type == READ_CACHE)
939
global_read_buffer.sub(buffer_length);
941
if (file != -1) /* File doesn't exist */
942
_error= my_b_flush_io_cache(this, 1);
943
free((unsigned char*) buffer);
944
buffer=read_pos=(unsigned char*) 0;
1804
IO_CACHE_CALLBACK pre_close;
1805
DBUG_ENTER("end_io_cache");
1806
DBUG_PRINT("enter",("cache: 0x%lx", (ulong) info));
1810
Every thread must call remove_io_thread(). The last one destroys
1813
DBUG_ASSERT(!info->share || !info->share->total_threads);
1816
if ((pre_close=info->pre_close))
1821
if (info->alloced_buffer)
1823
info->alloced_buffer=0;
1824
if (info->file != -1) /* File doesn't exist */
1825
error= my_b_flush_io_cache(info,1);
1826
my_free((uchar*) info->buffer,MYF(MY_WME));
1827
info->buffer=info->read_pos=(uchar*) 0;
1829
if (info->type == SEQ_READ_APPEND)
1831
/* Destroy allocated mutex */
1832
info->type= TYPE_NOT_SET;
1834
pthread_mutex_destroy(&info->append_buffer_lock);
948
1838
} /* end_io_cache */
950
} /* namespace internal */
951
} /* namespace drizzled */
1841
/**********************************************************************
1842
Testing of MF_IOCACHE
1843
**********************************************************************/
1849
void die(const char* fmt, ...)
1852
va_start(va_args,fmt);
1853
fprintf(stderr,"Error:");
1854
vfprintf(stderr, fmt,va_args);
1855
fprintf(stderr,", errno=%d\n", errno);
1859
int open_file(const char* fname, IO_CACHE* info, int cache_size)
1862
if ((fd=my_open(fname,O_CREAT | O_RDWR,MYF(MY_WME))) < 0)
1863
die("Could not open %s", fname);
1864
if (init_io_cache(info, fd, cache_size, SEQ_READ_APPEND, 0,0,MYF(MY_WME)))
1865
die("failed in init_io_cache()");
1869
void close_file(IO_CACHE* info)
1872
my_close(info->file, MYF(MY_WME));
1875
int main(int argc, char** argv)
1877
IO_CACHE sra_cache; /* SEQ_READ_APPEND */
1879
const char* fname="/tmp/iocache.test";
1880
int cache_size=16384;
1882
int max_block,total_bytes=0;
1883
int i,num_loops=100,error=0;
1885
char* block, *block_end;
1887
max_block = cache_size*3;
1888
if (!(block=(char*)my_malloc(max_block,MYF(MY_WME))))
1889
die("Not enough memory to allocate test block");
1890
block_end = block + max_block;
1891
for (p = block,i=0; p < block_end;i++)
1895
if (my_stat(fname,&status, MYF(0)) &&
1896
my_delete(fname,MYF(MY_WME)))
1898
die("Delete of %s failed, aborting", fname);
1900
open_file(fname,&sra_cache, cache_size);
1901
for (i = 0; i < num_loops; i++)
1904
int block_size = abs(rand() % max_block);
1905
int4store(buf, block_size);
1906
if (my_b_append(&sra_cache,buf,4) ||
1907
my_b_append(&sra_cache, block, block_size))
1908
die("write failed");
1909
total_bytes += 4+block_size;
1911
close_file(&sra_cache);
1912
my_free(block,MYF(MY_WME));
1913
if (!my_stat(fname,&status,MYF(MY_WME)))
1914
die("%s failed to stat, but I had just closed it,\
1915
wonder how that happened");
1916
printf("Final size of %s is %s, wrote %d bytes\n",fname,
1917
llstr(status.st_size,llstr_buf),
1919
my_delete(fname, MYF(MY_WME));
1920
/* check correctness of tests */
1921
if (total_bytes != status.st_size)
1923
fprintf(stderr,"Not the same number of bytes acutally in file as bytes \
1924
supposedly written\n");