mirror of
https://codeberg.org/redict/redict.git
synced 2025-01-22 08:08:53 -05:00
IO performances greatly improved under high writes load
This commit is contained in:
parent
6eaad66373
commit
f771dc23a0
195
src/dscache.c
195
src/dscache.c
@ -107,7 +107,8 @@
|
||||
*/
|
||||
|
||||
void spawnIOThread(void);
|
||||
int cacheScheduleIOPushJobs(int onlyloads);
|
||||
int cacheScheduleIOPushJobs(int flags);
|
||||
int processActiveIOJobs(int max);
|
||||
|
||||
/* =================== Virtual Memory - Blocking Side ====================== */
|
||||
|
||||
@ -216,9 +217,9 @@ int cacheFreeOneEntry(void) {
|
||||
* otherwise we'll use an infinite amount of memory if changes to
|
||||
* the dataset are faster than I/O */
|
||||
if (listLength(server.cache_io_queue) > 0) {
|
||||
cacheScheduleIOPushJobs(0);
|
||||
waitEmptyIOJobsQueue();
|
||||
processAllPendingIOJobs();
|
||||
redisLog(REDIS_DEBUG,"--- Busy waiting IO to reclaim memory");
|
||||
cacheScheduleIOPushJobs(REDIS_IO_ASAP);
|
||||
processActiveIOJobs(1);
|
||||
return REDIS_OK;
|
||||
}
|
||||
/* Nothing to free at all... */
|
||||
@ -484,62 +485,86 @@ void spawnIOThread(void) {
|
||||
server.io_active_threads++;
|
||||
}
|
||||
|
||||
/* Wait that all the pending IO Jobs are processed */
|
||||
void waitEmptyIOJobsQueue(void) {
|
||||
while(1) {
|
||||
/* Wait that up to 'max' pending IO Jobs are processed by the I/O thread.
|
||||
* From our point of view an IO job processed means that the count of
|
||||
* server.io_processed must increase by one.
|
||||
*
|
||||
* If max is -1, all the pending IO jobs will be processed.
|
||||
*
|
||||
* Returns the number of IO jobs processed.
|
||||
*
|
||||
* NOTE: while this may appear like a busy loop, we are actually blocked
|
||||
* by IO since we continuously acquire/release the IO lock. */
|
||||
int processActiveIOJobs(int max) {
|
||||
int processed = 0;
|
||||
|
||||
while(max == -1 || max > 0) {
|
||||
int io_processed_len;
|
||||
|
||||
lockThreadedIO();
|
||||
if (listLength(server.io_newjobs) == 0 &&
|
||||
listLength(server.io_processing) == 0)
|
||||
{
|
||||
/* There is nothing more to process */
|
||||
unlockThreadedIO();
|
||||
return;
|
||||
break;
|
||||
}
|
||||
|
||||
#if 0
|
||||
/* If there are new jobs we need to signal the thread to
|
||||
* process the next one. */
|
||||
redisLog(REDIS_DEBUG,"waitEmptyIOJobsQueue: new %d, processing %d",
|
||||
listLength(server.io_newjobs),
|
||||
listLength(server.io_processing));
|
||||
|
||||
/* FIXME: signal or not?
|
||||
if (listLength(server.io_newjobs)) {
|
||||
pthread_cond_signal(&server.io_condvar);
|
||||
}
|
||||
*/
|
||||
/* While waiting for empty jobs queue condition we post-process some
|
||||
* finshed job, as I/O threads may be hanging trying to write against
|
||||
* the io_ready_pipe_write FD but there are so much pending jobs that
|
||||
* it's blocking. */
|
||||
#endif
|
||||
|
||||
/* Check if we can process some finished job */
|
||||
io_processed_len = listLength(server.io_processed);
|
||||
unlockThreadedIO();
|
||||
if (io_processed_len) {
|
||||
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
|
||||
(void*)0xdeadbeef,0);
|
||||
/* FIXME: probably wiser to drop this sleeps. Anyway
|
||||
* the contention on the IO thread will avoid we to loop
|
||||
* too fast here. */
|
||||
usleep(1000); /* 1 millisecond */
|
||||
} else {
|
||||
/* FIXME: same as fixme above. */
|
||||
usleep(10000); /* 10 milliseconds */
|
||||
processed++;
|
||||
if (max != -1) max--;
|
||||
}
|
||||
}
|
||||
return processed;
|
||||
}
|
||||
|
||||
/* Process all the IO Jobs already completed by threads but still waiting
|
||||
* processing from the main thread. */
|
||||
void processAllPendingIOJobs(void) {
|
||||
while(1) {
|
||||
void waitEmptyIOJobsQueue(void) {
|
||||
processActiveIOJobs(-1);
|
||||
}
|
||||
|
||||
/* Process up to 'max' IO Jobs already completed by threads but still waiting
|
||||
* processing from the main thread.
|
||||
*
|
||||
* If max == -1 all the pending jobs are processed.
|
||||
*
|
||||
* The number of processed jobs is returned. */
|
||||
int processPendingIOJobs(int max) {
|
||||
int processed = 0;
|
||||
|
||||
while(max == -1 || max > 0) {
|
||||
int io_processed_len;
|
||||
|
||||
lockThreadedIO();
|
||||
io_processed_len = listLength(server.io_processed);
|
||||
unlockThreadedIO();
|
||||
if (io_processed_len == 0) return;
|
||||
if (io_processed_len == 0) break;
|
||||
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
|
||||
(void*)0xdeadbeef,0);
|
||||
if (max != -1) max--;
|
||||
processed++;
|
||||
}
|
||||
return processed;
|
||||
}
|
||||
|
||||
void processAllPendingIOJobs(void) {
|
||||
processPendingIOJobs(-1);
|
||||
}
|
||||
|
||||
/* This function must be called while with threaded IO locked */
|
||||
@ -665,7 +690,7 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
|
||||
* in queue for the same key. */
|
||||
if (type == REDIS_IO_LOAD && !(flags & REDIS_IO_SAVE)) {
|
||||
listAddNodeHead(server.cache_io_queue, op);
|
||||
cacheScheduleIOPushJobs(1);
|
||||
cacheScheduleIOPushJobs(REDIS_IO_ONLYLOADS);
|
||||
} else {
|
||||
/* FIXME: probably when this happens we want to at least move
|
||||
* the write job about this queue on top, and set the creation time
|
||||
@ -675,13 +700,19 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
|
||||
}
|
||||
|
||||
/* Push scheduled IO operations into IO Jobs that the IO thread can process.
|
||||
* If 'onlyloads' is true only IO_LOAD jobs are processed: this is useful
|
||||
* since it's save to push LOAD IO jobs from any place of the code, while
|
||||
*
|
||||
* If flags include REDIS_IO_ONLYLOADS only load jobs are processed:this is
|
||||
* useful since it's safe to push LOAD IO jobs from any place of the code, while
|
||||
* SAVE io jobs should never be pushed while we are processing a command
|
||||
* (not protected by lookupKey() that will block on keys in IO_SAVEINPROG
|
||||
* state. */
|
||||
* state.
|
||||
*
|
||||
* The REDIS_IO_ASAP flag tells the function to don't wait for the IO job
|
||||
* scheduled completion time, but just do the operation ASAP. This is useful
|
||||
* when we need to reclaim memory from the IO queue.
|
||||
*/
|
||||
#define MAX_IO_JOBS_QUEUE 100
|
||||
int cacheScheduleIOPushJobs(int onlyloads) {
|
||||
int cacheScheduleIOPushJobs(int flags) {
|
||||
time_t now = time(NULL);
|
||||
listNode *ln;
|
||||
int jobs, topush = 0, pushed = 0;
|
||||
@ -699,66 +730,64 @@ int cacheScheduleIOPushJobs(int onlyloads) {
|
||||
|
||||
while((ln = listFirst(server.cache_io_queue)) != NULL) {
|
||||
ioop *op = ln->value;
|
||||
struct dictEntry *de;
|
||||
robj *val;
|
||||
|
||||
if (!topush) break;
|
||||
topush--;
|
||||
|
||||
if (op->type == REDIS_IO_LOAD ||
|
||||
(!onlyloads && (now - op->ctime) >= server.cache_flush_delay))
|
||||
if (op->type != REDIS_IO_LOAD && flags & REDIS_IO_ONLYLOADS) break;
|
||||
|
||||
if (!(flags & REDIS_IO_ASAP) &&
|
||||
(now - op->ctime) < server.cache_flush_delay) break;
|
||||
|
||||
/* Don't add a SAVE job in the IO thread queue if there is already
|
||||
* a save in progress for the same key. */
|
||||
if (op->type == REDIS_IO_SAVE &&
|
||||
cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
|
||||
{
|
||||
struct dictEntry *de;
|
||||
robj *val;
|
||||
|
||||
/* Don't add a SAVE job in queue if there is already
|
||||
* a save in progress for the same key. */
|
||||
if (op->type == REDIS_IO_SAVE &&
|
||||
cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
|
||||
{
|
||||
/* Move the operation at the end of the list of there
|
||||
* are other operations. Otherwise break, nothing to do
|
||||
* here. */
|
||||
if (listLength(server.cache_io_queue) > 1) {
|
||||
listDelNode(server.cache_io_queue,ln);
|
||||
listAddNodeTail(server.cache_io_queue,op);
|
||||
continue;
|
||||
} else {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
redisLog(REDIS_DEBUG,"Creating IO %s Job for key %s",
|
||||
op->type == REDIS_IO_LOAD ? "load" : "save", op->key->ptr);
|
||||
|
||||
if (op->type == REDIS_IO_LOAD) {
|
||||
dsCreateIOJob(REDIS_IOJOB_LOAD,op->db,op->key,NULL);
|
||||
/* Move the operation at the end of the list if there
|
||||
* are other operations, so we can try to process the next one.
|
||||
* Otherwise break, nothing to do here. */
|
||||
if (listLength(server.cache_io_queue) > 1) {
|
||||
listDelNode(server.cache_io_queue,ln);
|
||||
listAddNodeTail(server.cache_io_queue,op);
|
||||
continue;
|
||||
} else {
|
||||
/* Lookup the key, in order to put the current value in the IO
|
||||
* Job. Otherwise if the key does not exists we schedule a disk
|
||||
* store delete operation, setting the value to NULL. */
|
||||
de = dictFind(op->db->dict,op->key->ptr);
|
||||
if (de) {
|
||||
val = dictGetEntryVal(de);
|
||||
} else {
|
||||
/* Setting the value to NULL tells the IO thread to delete
|
||||
* the key on disk. */
|
||||
val = NULL;
|
||||
}
|
||||
dsCreateIOJob(REDIS_IOJOB_SAVE,op->db,op->key,val);
|
||||
break;
|
||||
}
|
||||
/* Mark the operation as in progress. */
|
||||
cacheScheduleIODelFlag(op->db,op->key,op->type);
|
||||
cacheScheduleIOAddFlag(op->db,op->key,
|
||||
(op->type == REDIS_IO_LOAD) ? REDIS_IO_LOADINPROG :
|
||||
REDIS_IO_SAVEINPROG);
|
||||
/* Finally remove the operation from the queue.
|
||||
* But we'll have trace of it in the hash table. */
|
||||
listDelNode(server.cache_io_queue,ln);
|
||||
decrRefCount(op->key);
|
||||
zfree(op);
|
||||
pushed++;
|
||||
} else {
|
||||
break; /* too early */
|
||||
}
|
||||
|
||||
redisLog(REDIS_DEBUG,"Creating IO %s Job for key %s",
|
||||
op->type == REDIS_IO_LOAD ? "load" : "save", op->key->ptr);
|
||||
|
||||
if (op->type == REDIS_IO_LOAD) {
|
||||
dsCreateIOJob(REDIS_IOJOB_LOAD,op->db,op->key,NULL);
|
||||
} else {
|
||||
/* Lookup the key, in order to put the current value in the IO
|
||||
* Job. Otherwise if the key does not exists we schedule a disk
|
||||
* store delete operation, setting the value to NULL. */
|
||||
de = dictFind(op->db->dict,op->key->ptr);
|
||||
if (de) {
|
||||
val = dictGetEntryVal(de);
|
||||
} else {
|
||||
/* Setting the value to NULL tells the IO thread to delete
|
||||
* the key on disk. */
|
||||
val = NULL;
|
||||
}
|
||||
dsCreateIOJob(REDIS_IOJOB_SAVE,op->db,op->key,val);
|
||||
}
|
||||
/* Mark the operation as in progress. */
|
||||
cacheScheduleIODelFlag(op->db,op->key,op->type);
|
||||
cacheScheduleIOAddFlag(op->db,op->key,
|
||||
(op->type == REDIS_IO_LOAD) ? REDIS_IO_LOADINPROG :
|
||||
REDIS_IO_SAVEINPROG);
|
||||
/* Finally remove the operation from the queue.
|
||||
* But we'll have trace of it in the hash table. */
|
||||
listDelNode(server.cache_io_queue,ln);
|
||||
decrRefCount(op->key);
|
||||
zfree(op);
|
||||
pushed++;
|
||||
}
|
||||
return pushed;
|
||||
}
|
||||
|
@ -125,6 +125,10 @@
|
||||
#define REDIS_IO_LOADINPROG 4
|
||||
#define REDIS_IO_SAVEINPROG 8
|
||||
|
||||
/* Generic IO flags */
|
||||
#define REDIS_IO_ONLYLOADS 1
|
||||
#define REDIS_IO_ASAP 2
|
||||
|
||||
#define REDIS_MAX_COMPLETED_JOBS_PROCESSED 1
|
||||
#define REDIS_THREAD_STACK_SIZE (1024*1024*4)
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user