Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754543AbZGXUar (ORCPT ); Fri, 24 Jul 2009 16:30:47 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1754490AbZGXUan (ORCPT ); Fri, 24 Jul 2009 16:30:43 -0400 Received: from mx2.redhat.com ([66.187.237.31]:35515 "EHLO mx2.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754281AbZGXUab (ORCPT ); Fri, 24 Jul 2009 16:30:31 -0400 From: Vivek Goyal To: linux-kernel@vger.kernel.org, containers@lists.linux-foundation.org, dm-devel@redhat.com, jens.axboe@oracle.com, nauman@google.com, dpshah@google.com, ryov@valinux.co.jp, guijianfeng@cn.fujitsu.com, balbir@linux.vnet.ibm.com, righi.andrea@gmail.com Cc: lizf@cn.fujitsu.com, mikew@google.com, fchecconi@gmail.com, paolo.valente@unimore.it, fernando@oss.ntt.co.jp, s-uchida@ap.jp.nec.com, taka@valinux.co.jp, jmoyer@redhat.com, dhaval@linux.vnet.ibm.com, m-ikeda@ds.jp.nec.com, agk@redhat.com, vgoyal@redhat.com, akpm@linux-foundation.org, peterz@infradead.org Subject: [PATCH 12/24] io-controller: Introduce group idling Date: Fri, 24 Jul 2009 16:27:42 -0400 Message-Id: <1248467274-32073-13-git-send-email-vgoyal@redhat.com> In-Reply-To: <1248467274-32073-1-git-send-email-vgoyal@redhat.com> References: <1248467274-32073-1-git-send-email-vgoyal@redhat.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 16194 Lines: 461 o It is not always that IO from a process or group is continuous. There are cases of dependent reads where next read is not issued till previous read has finished. For such cases, CFQ introduced the notion of slice_idle, where we idle on the queue for sometime hoping next request will come and that's how fairness is provided otherwise queue will be deleted immediately from the service tree and this process will not get the fair share. o This patch introduces the similar concept at group level. Idle on the group for a period of "group_idle" which is tunable through sysfs interface. So if a group is empty and about to be deleted, we idle for the next request. o This patch also introduces the notion of wait busy where we wait for one extra group_idle period even if queue has consumed its time slice. The reason being that group will loose its share upon removal from service tree as some other entity will be picked for dispatch and vtime jump will take place. Signed-off-by: Vivek Goyal --- block/cfq-iosched.c | 3 + block/elevator-fq.c | 177 ++++++++++++++++++++++++++++++++++++++++++++++++--- block/elevator-fq.h | 42 ++++++++++++ 3 files changed, 213 insertions(+), 9 deletions(-) diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c index a362ce1..6238567 100644 --- a/block/cfq-iosched.c +++ b/block/cfq-iosched.c @@ -2124,6 +2124,9 @@ static struct elv_fs_entry cfq_attrs[] = { CFQ_ATTR(slice_idle), ELV_ATTR(slice_sync), ELV_ATTR(slice_async), +#ifdef CONFIG_GROUP_IOSCHED + ELV_ATTR(group_idle), +#endif __ATTR_NULL }; diff --git a/block/elevator-fq.c b/block/elevator-fq.c index ca26960..396bdcd 100644 --- a/block/elevator-fq.c +++ b/block/elevator-fq.c @@ -21,6 +21,7 @@ const int elv_slice_sync = HZ / 10; int elv_slice_async = HZ / 25; const int elv_slice_async_rq = 2; +int elv_group_idle = HZ / 125; static struct kmem_cache *elv_ioq_pool; #define ELV_SLICE_SCALE (5) @@ -52,7 +53,6 @@ elv_release_ioq(struct elevator_queue *eq, struct io_queue **ioq_ptr); #define for_each_entity_safe(entity, parent) \ for (; entity && ({ parent = entity->parent; 1; }); entity = parent) - static struct io_entity *bfq_lookup_next_entity(struct io_sched_data *sd, int extract); @@ -1579,6 +1579,7 @@ io_group_chain_alloc(struct request_queue *q, void *key, struct cgroup *cgroup) atomic_set(&iog->ref, 0); iog->deleting = 0; + elv_mark_iog_idle_window(iog); /* * Take the initial reference that will be released on destroy * This can be thought of a joint reference by cgroup and @@ -2217,6 +2218,8 @@ ssize_t __FUNC(struct elevator_queue *e, char *page) \ __data = jiffies_to_msecs(__data); \ return elv_var_show(__data, (page)); \ } +SHOW_FUNCTION(elv_group_idle_show, efqd->elv_group_idle, 1); +EXPORT_SYMBOL(elv_group_idle_show); SHOW_FUNCTION(elv_slice_sync_show, efqd->elv_slice[1], 1); EXPORT_SYMBOL(elv_slice_sync_show); SHOW_FUNCTION(elv_slice_async_show, efqd->elv_slice[0], 1); @@ -2239,6 +2242,8 @@ ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count)\ *(__PTR) = __data; \ return ret; \ } +STORE_FUNCTION(elv_group_idle_store, &efqd->elv_group_idle, 0, UINT_MAX, 1); +EXPORT_SYMBOL(elv_group_idle_store); STORE_FUNCTION(elv_slice_sync_store, &efqd->elv_slice[1], 1, UINT_MAX, 1); EXPORT_SYMBOL(elv_slice_sync_store); STORE_FUNCTION(elv_slice_async_store, &efqd->elv_slice[0], 1, UINT_MAX, 1); @@ -2506,7 +2511,9 @@ static void __elv_set_active_ioq(struct elv_fq_data *efqd, struct io_queue *ioq, ioq->slice_start = jiffies; elv_clear_ioq_wait_request(ioq); + elv_clear_iog_wait_request(iog); elv_clear_ioq_must_dispatch(ioq); + elv_clear_iog_wait_busy_done(iog); elv_mark_ioq_slice_new(ioq); del_timer(&efqd->idle_slice_timer); @@ -2634,14 +2641,19 @@ void __elv_ioq_slice_expired(struct request_queue *q, struct io_queue *ioq) struct elv_fq_data *efqd = &q->elevator->efqd; struct io_entity *entity = &ioq->entity; long slice_unused = 0, slice_used = 0, slice_overshoot = 0; + struct io_group *iog = ioq_to_io_group(ioq); assert_spin_locked(q->queue_lock); elv_log_ioq(efqd, ioq, "slice expired"); - if (elv_ioq_wait_request(ioq)) + if (elv_ioq_wait_request(ioq) || elv_iog_wait_request(iog) + || elv_iog_wait_busy(iog)) del_timer(&efqd->idle_slice_timer); elv_clear_ioq_wait_request(ioq); + elv_clear_iog_wait_request(iog); + elv_clear_iog_wait_busy(iog); + elv_clear_iog_wait_busy_done(iog); slice_used = jiffies - ioq->slice_start; if (!slice_used) { @@ -2783,6 +2795,8 @@ void elv_ioq_request_add(struct request_queue *q, struct request *rq) { struct elv_fq_data *efqd = &q->elevator->efqd; struct io_queue *ioq = rq->ioq; + struct io_group *iog = ioq_to_io_group(ioq); + int group_wait = 0; if (!elv_iosched_fair_queuing_enabled(q->elevator)) return; @@ -2795,6 +2809,24 @@ void elv_ioq_request_add(struct request_queue *q, struct request *rq) if (!elv_ioq_busy(ioq)) elv_add_ioq_busy(efqd, ioq); + if (elv_iog_wait_request(iog)) { + del_timer(&efqd->idle_slice_timer); + elv_clear_iog_wait_request(iog); + group_wait = 1; + } + + /* + * If we were waiting for a request on this group, wait is + * done. Schedule the next dispatch + */ + if (elv_iog_wait_busy(iog)) { + del_timer(&efqd->idle_slice_timer); + elv_clear_iog_wait_busy(iog); + elv_mark_iog_wait_busy_done(iog); + elv_schedule_dispatch(q); + return; + } + if (ioq == elv_active_ioq(q->elevator)) { /* * Remember that we saw a request from this process, but @@ -2806,7 +2838,7 @@ void elv_ioq_request_add(struct request_queue *q, struct request *rq) * has other work pending, don't risk delaying until the * idle timer unplug to continue working. */ - if (elv_ioq_wait_request(ioq)) { + if (group_wait || elv_ioq_wait_request(ioq)) { del_timer(&efqd->idle_slice_timer); if (blk_rq_bytes(rq) > PAGE_CACHE_SIZE || efqd->busy_queues > 1 || !blk_queue_plugged(q)) @@ -2823,6 +2855,13 @@ void elv_ioq_request_add(struct request_queue *q, struct request *rq) */ elv_preempt_queue(q, ioq); __blk_run_queue(q); + } else if (group_wait) { + /* + * Got a request in the group we were waiting for. Request + * does not belong to active queue and we have not decided + * to preempt the current active queue. Schedule the dispatch. + */ + elv_schedule_dispatch(q); } } @@ -2840,6 +2879,14 @@ static void elv_idle_slice_timer(unsigned long data) ioq = efqd->active_queue; if (ioq) { + struct io_group *iog = ioq_to_io_group(ioq); + + elv_clear_iog_wait_request(iog); + + if (elv_iog_wait_busy(iog)) { + elv_clear_iog_wait_busy(iog); + goto expire; + } /* * We saw a request before the queue expired, let it through @@ -2889,6 +2936,32 @@ static void elv_ioq_arm_slice_timer(struct request_queue *q) eq->ops->elevator_arm_slice_timer_fn(q, ioq->sched_queue); } +static void elv_iog_arm_slice_timer(struct request_queue *q, + struct io_group *iog, int wait_for_busy) +{ + struct elv_fq_data *efqd = &q->elevator->efqd; + unsigned long sl; + + if (!efqd->elv_group_idle || !elv_iog_idle_window(iog)) + return; + /* + * This queue has consumed its time slice. We are waiting only for + * it to become busy before we select next queue for dispatch. + */ + if (wait_for_busy) { + elv_mark_iog_wait_busy(iog); + sl = efqd->elv_group_idle; + mod_timer(&efqd->idle_slice_timer, jiffies + sl); + elv_log_iog(efqd, iog, "arm idle group: %lu wait busy=1", sl); + return; + } + + elv_mark_iog_wait_request(iog); + sl = efqd->elv_group_idle; + mod_timer(&efqd->idle_slice_timer, jiffies + sl); + elv_log_iog(efqd, iog, "arm_idle group: %lu", sl); +} + /* * If io scheduler has functionality of keeping track of close cooperator, check * with it if it has got a closely co-operating queue. @@ -2926,6 +2999,8 @@ void *elv_fq_select_ioq(struct request_queue *q, int force) if (ioq == NULL) goto new_queue; + iog = ioq_to_io_group(ioq); + /* * Force dispatch. Continue to dispatch from current queue as long * as it has requests. @@ -2937,11 +3012,48 @@ void *elv_fq_select_ioq(struct request_queue *q, int force) goto expire; } + /* We are waiting for this group to become busy before it expires.*/ + if (elv_iog_wait_busy(iog)) { + ioq = NULL; + goto keep_queue; + } + /* * The active queue has run out of time, expire it and select new. */ - if (elv_ioq_slice_used(ioq) && !elv_ioq_must_dispatch(ioq)) - goto expire; + if (elv_ioq_slice_used(ioq) && !elv_ioq_must_dispatch(ioq)) { + /* + * Queue has used up its slice. Wait busy is not on otherwise + * we wouldn't have been here. If this group will be deleted + * after the queue expiry, then make sure we have onece + * done wait busy on the group in an attempt to make it + * backlogged. + * + * Following check helps in two conditions. + * - If there are requests dispatched from the queue and + * select_ioq() comes before a request completed from the + * queue and got a chance to arm any of the idle timers. + * + * - If at request completion time slice had not expired and + * we armed either a ioq timer or group timer but when + * select_ioq() hits, slice has expired and it will expire + * the queue without doing busy wait on group. + * + * In similar situations cfq lets delte the queue even if + * idle timer is armed. That does not impact fairness in non + * hierarhical setup due to weighted slice lengths. But in + * hierarchical setup where group slice lengths are derived + * from queue and is not proportional to group's weight, it + * harms the fairness of the group. + */ + if ((elv_iog_nr_active(iog) <= 1) && !ioq->nr_queued + && !elv_iog_wait_busy_done(iog) && efqd->elv_group_idle + && elv_iog_idle_window(iog)) { + ioq = NULL; + goto keep_queue; + } else + goto expire; + } /* * The active queue has requests and isn't expired, allow it to @@ -2973,6 +3085,13 @@ void *elv_fq_select_ioq(struct request_queue *q, int force) goto keep_queue; } + /* Check for group idling */ + if (elv_iog_idle_window(iog) && (elv_iog_nr_active(iog) <= 1) + && elv_ioq_nr_dispatched(ioq)) { + ioq = NULL; + goto keep_queue; + } + expire: elv_ioq_slice_expired(q); new_queue: @@ -3047,11 +3166,13 @@ void elv_ioq_completed_request(struct request_queue *q, struct request *rq) const int sync = rq_is_sync(rq); struct io_queue *ioq; struct elv_fq_data *efqd = &q->elevator->efqd; + struct io_group *iog; if (!elv_iosched_fair_queuing_enabled(q->elevator)) return; ioq = rq->ioq; + iog = ioq_to_io_group(ioq); WARN_ON(!efqd->rq_in_driver); WARN_ON(!ioq->dispatched); efqd->rq_in_driver--; @@ -3070,6 +3191,12 @@ void elv_ioq_completed_request(struct request_queue *q, struct request *rq) elv_ioq_set_prio_slice(q, ioq); elv_clear_ioq_slice_new(ioq); } + + if (elv_ioq_class_idle(ioq)) { + elv_ioq_slice_expired(q); + goto done; + } + /* * If there are no requests waiting in this queue, and * there are other queues ready to issue requests, AND @@ -3077,13 +3204,44 @@ void elv_ioq_completed_request(struct request_queue *q, struct request *rq) * mean seek distance, give them a chance to run instead * of idling. */ - if (elv_ioq_slice_used(ioq) || elv_ioq_class_idle(ioq)) + if (elv_ioq_slice_used(ioq)) { + /* This is the last empty queue in the group and it + * has consumed its slice. If we expire it right away + * group might loose its share. Wait for an extra + * group_idle period for a request before queue + * expires. + */ + if ((elv_iog_nr_active(iog) <= 1) && !ioq->nr_queued) { + elv_iog_arm_slice_timer(q, iog, 1); + goto done; + } + + /* Expire the queue */ elv_ioq_slice_expired(q); - else if (!ioq->nr_queued && !elv_close_cooperator(q, ioq) - && sync && !rq_noidle(rq)) + } else if (!ioq->nr_queued && !elv_close_cooperator(q, ioq) + && sync && !rq_noidle(rq)) elv_ioq_arm_slice_timer(q); - } + /* + * If this is the last queue in the group and we did not + * decide to idle on queue, idle on group. + */ + if (elv_active_ioq(q->elevator) && !ioq->nr_queued && + !ioq->dispatched && !timer_pending(&efqd->idle_slice_timer) + && (elv_iog_nr_active(iog) <= 1)) { + /* + * If queue has used up its slice, wait for the + * one extra group_idle period to let the group + * backlogged again. This is to avoid a group loosing + * its fair share. + */ + if (elv_ioq_slice_used(ioq)) + elv_iog_arm_slice_timer(q, iog, 1); + else + elv_iog_arm_slice_timer(q, iog, 0); + } + } +done: if (!efqd->rq_in_driver) elv_schedule_dispatch(q); } @@ -3196,6 +3354,7 @@ int elv_init_fq_data(struct request_queue *q, struct elevator_queue *e) efqd->elv_slice[0] = elv_slice_async; efqd->elv_slice[1] = elv_slice_sync; + efqd->elv_group_idle = elv_group_idle; return 0; } diff --git a/block/elevator-fq.h b/block/elevator-fq.h index 1da7ecc..95c1d94 100644 --- a/block/elevator-fq.h +++ b/block/elevator-fq.h @@ -218,6 +218,7 @@ struct io_queue { */ struct io_group { struct io_entity entity; + unsigned int flags; struct hlist_node elv_data_node; struct hlist_node group_node; struct io_sched_data sched_data; @@ -311,6 +312,8 @@ struct elv_fq_data { struct timer_list idle_slice_timer; struct work_struct unplug_work; + unsigned int elv_group_idle; + /* Base slice length for sync and async queues */ unsigned int elv_slice[2]; @@ -381,6 +384,42 @@ ELV_IO_QUEUE_FLAG_FNS(must_dispatch) ELV_IO_QUEUE_FLAG_FNS(idle_window) ELV_IO_QUEUE_FLAG_FNS(slice_new) +#ifdef CONFIG_GROUP_IOSCHED + +enum elv_group_state_flags { + ELV_GROUP_FLAG_idle_window, /* elevator group idling enabled */ + ELV_GROUP_FLAG_wait_request, /* waiting for a request */ + ELV_GROUP_FLAG_wait_busy, /* wait for this queue to get busy */ + ELV_GROUP_FLAG_wait_busy_done, /* Have already waited on this group*/ +}; + +#define ELV_IO_GROUP_FLAG_FNS(name) \ +static inline void elv_mark_iog_##name(struct io_group *iog) \ +{ \ + (iog)->flags |= (1 << ELV_GROUP_FLAG_##name); \ +} \ +static inline void elv_clear_iog_##name(struct io_group *iog) \ +{ \ + (iog)->flags &= ~(1 << ELV_GROUP_FLAG_##name); \ +} \ +static inline int elv_iog_##name(struct io_group *iog) \ +{ \ + return ((iog)->flags & (1 << ELV_GROUP_FLAG_##name)) != 0; \ +} + +#else /* GROUP_IOSCHED */ + +#define ELV_IO_GROUP_FLAG_FNS(name) \ +static inline void elv_mark_iog_##name(struct io_group *iog) {} \ +static inline void elv_clear_iog_##name(struct io_group *iog) {} \ +static inline int elv_iog_##name(struct io_group *iog) { return 0; } +#endif /* GROUP_IOSCHED */ + +ELV_IO_GROUP_FLAG_FNS(idle_window) +ELV_IO_GROUP_FLAG_FNS(wait_request) +ELV_IO_GROUP_FLAG_FNS(wait_busy) +ELV_IO_GROUP_FLAG_FNS(wait_busy_done) + static inline struct io_service_tree * io_entity_service_tree(struct io_entity *entity) { @@ -480,6 +519,9 @@ static inline struct io_group *ioq_to_io_group(struct io_queue *ioq) #ifdef CONFIG_GROUP_IOSCHED extern int io_group_allow_merge(struct request *rq, struct bio *bio); extern void elv_put_iog(struct io_group *iog); +extern ssize_t elv_group_idle_show(struct elevator_queue *q, char *name); +extern ssize_t elv_group_idle_store(struct elevator_queue *q, const char *name, + size_t count); static inline void elv_get_iog(struct io_group *iog) { atomic_inc(&iog->ref); -- 1.6.0.6 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/