Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756795AbZGBUKT (ORCPT ); Thu, 2 Jul 2009 16:10:19 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1755997AbZGBUD7 (ORCPT ); Thu, 2 Jul 2009 16:03:59 -0400 Received: from mx2.redhat.com ([66.187.237.31]:53126 "EHLO mx2.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755665AbZGBUDv (ORCPT ); Thu, 2 Jul 2009 16:03:51 -0400 From: Vivek Goyal To: linux-kernel@vger.kernel.org, containers@lists.linux-foundation.org, dm-devel@redhat.com, jens.axboe@oracle.com, nauman@google.com, dpshah@google.com, lizf@cn.fujitsu.com, mikew@google.com, fchecconi@gmail.com, paolo.valente@unimore.it, ryov@valinux.co.jp, fernando@oss.ntt.co.jp, s-uchida@ap.jp.nec.com, taka@valinux.co.jp, guijianfeng@cn.fujitsu.com, jmoyer@redhat.com, dhaval@linux.vnet.ibm.com, balbir@linux.vnet.ibm.com, righi.andrea@gmail.com, m-ikeda@ds.jp.nec.com, jbaron@redhat.com Cc: agk@redhat.com, snitzer@redhat.com, vgoyal@redhat.com, akpm@linux-foundation.org, peterz@infradead.org Subject: [PATCH 12/25] io-controller: idle for sometime on sync queue before expiring it Date: Thu, 2 Jul 2009 16:01:44 -0400 Message-Id: <1246564917-19603-13-git-send-email-vgoyal@redhat.com> In-Reply-To: <1246564917-19603-1-git-send-email-vgoyal@redhat.com> References: <1246564917-19603-1-git-send-email-vgoyal@redhat.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 14224 Lines: 399 o When a sync queue expires, in many cases it might be empty and then it will be deleted from the active tree. This will lead to a scenario where out of two competing queues, only one is on the tree and when a new queue is selected, vtime jump takes place and we don't see services provided in proportion to weight. o In general this is a fundamental problem with fairness of sync queues where queues are not continuously backlogged. Looks like idling is only solution to make sure such kind of queues can get some decent amount of disk bandwidth in the face of competion from continusouly backlogged queues. But excessive idling has potential to reduce performance on SSD and disks with commnad queuing. o This patch experiments with waiting for next request to come before a queue is expired after it has consumed its time slice. This can ensure more accurate fairness numbers in some cases. o Introduced a tunable "fairness". If set, io-controller will put more focus on getting fairness right than getting throughput right. o When writes are being done on a file opened with O_SYNC, ioscheduler sees synchronous write requests with noidle flag set. But the fact is we are seeing a continuous stream of writes with-in 1ms or so. Hence it makes sense to wait on these writes. For the time being to achieve fairness for O_SYNC writes, continue to idle even if last request was sync write and noidle flag was set. (Only done if "fairness" is set). Probably right fix is to make sure in O_SYNC path, requests are not marked with noidle flag. Signed-off-by: Vivek Goyal --- block/cfq-iosched.c | 1 + block/elevator-fq.c | 133 ++++++++++++++++++++++++++++++++++++++++++++------- block/elevator-fq.h | 15 ++++++ 3 files changed, 132 insertions(+), 17 deletions(-) diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c index 02b5cd5..98a35fd 100644 --- a/block/cfq-iosched.c +++ b/block/cfq-iosched.c @@ -2004,6 +2004,7 @@ static struct elv_fs_entry cfq_attrs[] = { ELV_ATTR(slice_idle), ELV_ATTR(slice_sync), ELV_ATTR(slice_async), + ELV_ATTR(fairness), __ATTR_NULL }; diff --git a/block/elevator-fq.c b/block/elevator-fq.c index bab01b5..68be1dc 100644 --- a/block/elevator-fq.c +++ b/block/elevator-fq.c @@ -424,6 +424,7 @@ static void bfq_active_insert(struct io_service_tree *st, struct rb_node *node = &entity->rb_node; bfq_insert(&st->active, entity); + entity->sched_data->nr_active++; if (node->rb_left != NULL) node = node->rb_left; @@ -483,6 +484,7 @@ static void bfq_active_remove(struct io_service_tree *st, node = bfq_find_deepest(&entity->rb_node); bfq_remove(&st->active, entity); + entity->sched_data->nr_active--; if (node != NULL) bfq_update_active_tree(node); @@ -569,6 +571,21 @@ static void bfq_forget_idle(struct io_service_tree *st) bfq_put_idle_entity(st, first_idle); } +/* + * Returns the number of active entities a particular io group has. This + * includes number of active entities on service tree as well as the active + * entity which is being served currently, if any. + */ + +static inline int elv_iog_nr_active(struct io_group *iog) +{ + struct io_sched_data *sd = &iog->sched_data; + + if (sd->active_entity) + return sd->nr_active + 1; + else + return sd->nr_active; +} static struct io_service_tree * __bfq_entity_update_prio(struct io_service_tree *old_st, @@ -1995,6 +2012,8 @@ SHOW_FUNCTION(elv_slice_sync_show, efqd->elv_slice[1], 1); EXPORT_SYMBOL(elv_slice_sync_show); SHOW_FUNCTION(elv_slice_async_show, efqd->elv_slice[0], 1); EXPORT_SYMBOL(elv_slice_async_show); +SHOW_FUNCTION(elv_fairness_show, efqd->fairness, 0); +EXPORT_SYMBOL(elv_fairness_show); #undef SHOW_FUNCTION #define STORE_FUNCTION(__FUNC, __PTR, MIN, MAX, __CONV) \ @@ -2019,6 +2038,8 @@ STORE_FUNCTION(elv_slice_sync_store, &efqd->elv_slice[1], 1, UINT_MAX, 1); EXPORT_SYMBOL(elv_slice_sync_store); STORE_FUNCTION(elv_slice_async_store, &efqd->elv_slice[0], 1, UINT_MAX, 1); EXPORT_SYMBOL(elv_slice_async_store); +STORE_FUNCTION(elv_fairness_store, &efqd->fairness, 0, 1, 0); +EXPORT_SYMBOL(elv_fairness_store); #undef STORE_FUNCTION void elv_schedule_dispatch(struct request_queue *q) @@ -2142,7 +2163,7 @@ static void elv_ioq_update_idle_window(struct elevator_queue *eq, * io scheduler if it wants to disable idling based on additional * considrations like seek pattern. */ - if (enable_idle) { + if (enable_idle && !efqd->fairness) { if (eq->ops->elevator_update_idle_window_fn) enable_idle = eq->ops->elevator_update_idle_window_fn( eq, ioq->sched_queue, rq); @@ -2328,6 +2349,7 @@ static void __elv_set_active_ioq(struct elv_fq_data *efqd, struct io_queue *ioq, elv_clear_ioq_wait_request(ioq); elv_clear_ioq_must_dispatch(ioq); + elv_clear_ioq_wait_busy_done(ioq); elv_mark_ioq_slice_new(ioq); del_timer(&efqd->idle_slice_timer); @@ -2483,10 +2505,12 @@ void __elv_ioq_slice_expired(struct request_queue *q, struct io_queue *ioq) assert_spin_locked(q->queue_lock); elv_log_ioq(efqd, ioq, "slice expired"); - if (elv_ioq_wait_request(ioq)) + if (elv_ioq_wait_request(ioq) || elv_ioq_wait_busy(ioq)) del_timer(&efqd->idle_slice_timer); elv_clear_ioq_wait_request(ioq); + elv_clear_ioq_wait_busy(ioq); + elv_clear_ioq_wait_busy_done(ioq); /* * if ioq->slice_end = 0, that means a queue was expired before first @@ -2659,7 +2683,7 @@ void elv_ioq_request_add(struct request_queue *q, struct request *rq) * has other work pending, don't risk delaying until the * idle timer unplug to continue working. */ - if (elv_ioq_wait_request(ioq)) { + if (elv_ioq_wait_request(ioq) && !elv_ioq_wait_busy(ioq)) { if (blk_rq_bytes(rq) > PAGE_CACHE_SIZE || efqd->busy_queues > 1) { del_timer(&efqd->idle_slice_timer); @@ -2667,6 +2691,18 @@ void elv_ioq_request_add(struct request_queue *q, struct request *rq) } elv_mark_ioq_must_dispatch(ioq); } + + /* + * If we were waiting for a request on this queue, wait is + * done. Schedule the next dispatch + */ + if (elv_ioq_wait_busy(ioq)) { + del_timer(&efqd->idle_slice_timer); + elv_clear_ioq_wait_busy(ioq); + elv_mark_ioq_wait_busy_done(ioq); + elv_clear_ioq_must_dispatch(ioq); + elv_schedule_dispatch(q); + } } else if (elv_should_preempt(q, ioq, rq)) { /* * not the active queue - expire current slice if it is @@ -2694,6 +2730,9 @@ static void elv_idle_slice_timer(unsigned long data) if (ioq) { + if (elv_ioq_wait_busy(ioq)) + goto expire; + /* * We saw a request before the queue expired, let it through */ @@ -2727,7 +2766,7 @@ out_cont: spin_unlock_irqrestore(q->queue_lock, flags); } -static void elv_ioq_arm_slice_timer(struct request_queue *q) +static void elv_ioq_arm_slice_timer(struct request_queue *q, int wait_for_busy) { struct elv_fq_data *efqd = &q->elevator->efqd; struct io_queue *ioq = elv_active_ioq(q->elevator); @@ -2740,26 +2779,38 @@ static void elv_ioq_arm_slice_timer(struct request_queue *q) * for devices that support queuing, otherwise we still have a problem * with sync vs async workloads. */ - if (blk_queue_nonrot(q) && efqd->hw_tag) + if (blk_queue_nonrot(q) && efqd->hw_tag && !efqd->fairness) return; /* - * still requests with the driver, don't idle + * idle is disabled, either manually or by past process history */ - if (efqd->rq_in_driver) + if (!efqd->elv_slice_idle || !elv_ioq_idle_window(ioq)) return; /* - * idle is disabled, either manually or by past process history + * This queue has consumed its time slice. We are waiting only for + * it to become busy before we select next queue for dispatch. */ - if (!efqd->elv_slice_idle || !elv_ioq_idle_window(ioq)) + if (wait_for_busy) { + elv_mark_ioq_wait_busy(ioq); + sl = efqd->elv_slice_idle; + mod_timer(&efqd->idle_slice_timer, jiffies + sl); + elv_log_ioq(efqd, ioq, "arm idle: %lu wait busy=1", sl); + return; + } + + /* + * still requests with the driver, don't idle + */ + if (efqd->rq_in_driver && !efqd->fairness) return; /* * may be iosched got its own idling logic. In that case io * schduler will take care of arming the timer, if need be. */ - if (q->elevator->ops->elevator_arm_slice_timer_fn) { + if (q->elevator->ops->elevator_arm_slice_timer_fn && !efqd->fairness) { q->elevator->ops->elevator_arm_slice_timer_fn(q, ioq->sched_queue); } else { @@ -2822,11 +2873,38 @@ void *elv_fq_select_ioq(struct request_queue *q, int force) goto expire; } + /* We are waiting for this queue to become busy before it expires.*/ + if (efqd->fairness && elv_ioq_wait_busy(ioq)) { + ioq = NULL; + goto keep_queue; + } + /* * The active queue has run out of time, expire it and select new. */ - if (elv_ioq_slice_used(ioq) && !elv_ioq_must_dispatch(ioq)) - goto expire; + if (elv_ioq_slice_used(ioq) && !elv_ioq_must_dispatch(ioq)) { + /* + * Queue has used up its slice. Wait busy is not on otherwise + * we wouldn't have been here. There is a chance that after + * slice expiry no request from the queue completed hence + * wait busy timer could not be turned on. If that's the case + * don't expire the queue yet. Next request completion from + * the queue will arm the wait busy timer. + * + * Don't wait if this group has other active queues. This + * will make sure that we don't loose fairness at group level + * at the same time in root group we will not see cfq + * regressions. + */ + if (elv_ioq_sync(ioq) && !ioq->nr_queued + && elv_ioq_nr_dispatched(ioq) + && (elv_iog_nr_active(ioq_to_io_group(ioq)) <= 1) + && !elv_ioq_wait_busy_done(ioq)) { + ioq = NULL; + goto keep_queue; + } else + goto expire; + } /* * If we have a RT cfqq waiting, then we pre-empt the current non-rt @@ -2977,11 +3055,13 @@ void elv_ioq_completed_request(struct request_queue *q, struct request *rq) const int sync = rq_is_sync(rq); struct io_queue *ioq; struct elv_fq_data *efqd = &q->elevator->efqd; + struct io_group *iog; if (!elv_iosched_fair_queuing_enabled(q->elevator)) return; ioq = rq->ioq; + iog = ioq_to_io_group(ioq); elv_log_ioq(efqd, ioq, "complete"); @@ -3007,6 +3087,12 @@ void elv_ioq_completed_request(struct request_queue *q, struct request *rq) elv_ioq_set_prio_slice(q, ioq); elv_clear_ioq_slice_new(ioq); } + + if (elv_ioq_class_idle(ioq)) { + elv_ioq_slice_expired(q); + goto done; + } + /* * If there are no requests waiting in this queue, and * there are other queues ready to issue requests, AND @@ -3014,13 +3100,24 @@ void elv_ioq_completed_request(struct request_queue *q, struct request *rq) * mean seek distance, give them a chance to run instead * of idling. */ - if (elv_ioq_slice_used(ioq) || elv_ioq_class_idle(ioq)) - elv_ioq_slice_expired(q); - else if (!ioq->nr_queued && !elv_close_cooperator(q, ioq, 1) - && sync && !rq_noidle(rq)) - elv_ioq_arm_slice_timer(q); + if (elv_ioq_slice_used(ioq)) { + if (sync && !ioq->nr_queued + && (elv_iog_nr_active(iog) <= 1)) { + /* + * Idle for one extra period in hierarchical + * setup + */ + elv_ioq_arm_slice_timer(q, 1); + } else { + /* Expire the queue */ + elv_ioq_slice_expired(q); + } + } else if (!ioq->nr_queued && !elv_close_cooperator(q, ioq, 1) + && sync && (!rq_noidle(rq) || efqd->fairness)) + elv_ioq_arm_slice_timer(q, 0); } +done: if (!efqd->rq_in_driver) elv_schedule_dispatch(q); } @@ -3125,6 +3222,8 @@ int elv_init_fq_data(struct request_queue *q, struct elevator_queue *e) efqd->elv_slice_idle = elv_slice_idle; efqd->hw_tag = 1; + /* For the time being keep fairness enabled by default */ + efqd->fairness = 1; return 0; } diff --git a/block/elevator-fq.h b/block/elevator-fq.h index d76bd96..a414309 100644 --- a/block/elevator-fq.h +++ b/block/elevator-fq.h @@ -75,6 +75,7 @@ struct io_service_tree { struct io_sched_data { struct io_entity *active_entity; struct io_entity *next_active; + int nr_active; struct io_service_tree service_tree[IO_IOPRIO_CLASSES]; }; @@ -337,6 +338,13 @@ struct elv_fq_data { unsigned long long rate_sampling_start; /*sampling window start jifies*/ /* number of sectors finished io during current sampling window */ unsigned long rate_sectors_current; + + /* + * If set to 1, will disable many optimizations done for boost + * throughput and focus more on providing fairness for sync + * queues. + */ + unsigned int fairness; }; /* Logging facilities. */ @@ -358,6 +366,8 @@ enum elv_queue_state_flags { ELV_QUEUE_FLAG_wait_request, /* waiting for a request */ ELV_QUEUE_FLAG_must_dispatch, /* must be allowed a dispatch */ ELV_QUEUE_FLAG_slice_new, /* no requests dispatched in slice */ + ELV_QUEUE_FLAG_wait_busy, /* wait for this queue to get busy */ + ELV_QUEUE_FLAG_wait_busy_done, /* Have already waited on this queue*/ }; #define ELV_IO_QUEUE_FLAG_FNS(name) \ @@ -380,6 +390,8 @@ ELV_IO_QUEUE_FLAG_FNS(wait_request) ELV_IO_QUEUE_FLAG_FNS(must_dispatch) ELV_IO_QUEUE_FLAG_FNS(idle_window) ELV_IO_QUEUE_FLAG_FNS(slice_new) +ELV_IO_QUEUE_FLAG_FNS(wait_busy) +ELV_IO_QUEUE_FLAG_FNS(wait_busy_done) static inline struct io_service_tree * io_entity_service_tree(struct io_entity *entity) @@ -532,6 +544,9 @@ extern ssize_t elv_slice_sync_store(struct elevator_queue *q, const char *name, extern ssize_t elv_slice_async_show(struct elevator_queue *q, char *name); extern ssize_t elv_slice_async_store(struct elevator_queue *q, const char *name, size_t count); +extern ssize_t elv_fairness_show(struct elevator_queue *q, char *name); +extern ssize_t elv_fairness_store(struct elevator_queue *q, const char *name, + size_t count); /* Functions used by elevator.c */ extern int elv_init_fq_data(struct request_queue *q, struct elevator_queue *e); -- 1.6.0.6 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/