For most implementations of reference count, atomic_t is preferred
for their natural-born atomic ops capability.
Change the reference count of bfq_group/bfq_queue, both data structures
and related ops, into atomic.
Signed-off-by: Dawei Li <[email protected]>
---
block/bfq-cgroup.c | 8 +++----
block/bfq-iosched.c | 54 +++++++++++++++++++++++----------------------
block/bfq-iosched.h | 6 ++---
block/bfq-wf2q.c | 6 ++---
4 files changed, 37 insertions(+), 37 deletions(-)
diff --git a/block/bfq-cgroup.c b/block/bfq-cgroup.c
index 144bca006463..714126ba21b6 100644
--- a/block/bfq-cgroup.c
+++ b/block/bfq-cgroup.c
@@ -316,14 +316,12 @@ struct bfq_group *bfqq_group(struct bfq_queue *bfqq)
static void bfqg_get(struct bfq_group *bfqg)
{
- bfqg->ref++;
+ atomic_inc(&bfqg->ref);
}
static void bfqg_put(struct bfq_group *bfqg)
{
- bfqg->ref--;
-
- if (bfqg->ref == 0)
+ if (atomic_dec_and_test(&bfqg->ref))
kfree(bfqg);
}
@@ -659,7 +657,7 @@ void bfq_bfqq_move(struct bfq_data *bfqd, struct bfq_queue *bfqq,
* Get extra reference to prevent bfqq from being freed in
* next possible expire or deactivate.
*/
- bfqq->ref++;
+ atomic_inc(&bfqq->ref);
/* If bfqq is empty, then bfq_bfqq_expire also invokes
* bfq_del_bfqq_busy, thereby removing bfqq and its entity
diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
index 7ea427817f7f..fbe5624be71f 100644
--- a/block/bfq-iosched.c
+++ b/block/bfq-iosched.c
@@ -935,7 +935,7 @@ void bfq_weights_tree_add(struct bfq_data *bfqd, struct bfq_queue *bfqq,
inc_counter:
bfqq->weight_counter->num_active++;
- bfqq->ref++;
+ atomic_inc(&bfqq->ref);
}
/*
@@ -1224,9 +1224,10 @@ bfq_bfqq_resume_state(struct bfq_queue *bfqq, struct bfq_data *bfqd,
static int bfqq_process_refs(struct bfq_queue *bfqq)
{
- return bfqq->ref - bfqq->entity.allocated -
+ return atomic_read(&bfqq->ref) - bfqq->entity.allocated -
bfqq->entity.on_st_or_in_serv -
- (bfqq->weight_counter != NULL) - bfqq->stable_ref;
+ (bfqq->weight_counter != NULL) -
+ atomic_read(&bfqq->stable_ref);
}
/* Empty burst list and add just bfqq (see comments on bfq_handle_burst) */
@@ -2818,7 +2819,7 @@ bfq_setup_merge(struct bfq_queue *bfqq, struct bfq_queue *new_bfqq)
* expected to be associated with new_bfqq as they happen to
* issue I/O.
*/
- new_bfqq->ref += process_refs;
+ atomic_add(process_refs, &new_bfqq->ref);
return new_bfqq;
}
@@ -5255,10 +5256,10 @@ void bfq_put_queue(struct bfq_queue *bfqq)
struct hlist_node *n;
struct bfq_group *bfqg = bfqq_group(bfqq);
- bfq_log_bfqq(bfqq->bfqd, bfqq, "put_queue: %p %d", bfqq, bfqq->ref);
+ bfq_log_bfqq(bfqq->bfqd, bfqq, "put_queue: %p %d", bfqq,
+ atomic_read(&bfqq->ref));
- bfqq->ref--;
- if (bfqq->ref)
+ if (!atomic_dec_and_test(&bfqq->ref))
return;
if (!hlist_unhashed(&bfqq->burst_list_node)) {
@@ -5328,7 +5329,7 @@ void bfq_put_queue(struct bfq_queue *bfqq)
static void bfq_put_stable_ref(struct bfq_queue *bfqq)
{
- bfqq->stable_ref--;
+ atomic_dec(&bfqq->stable_ref);
bfq_put_queue(bfqq);
}
@@ -5358,7 +5359,7 @@ static void bfq_exit_bfqq(struct bfq_data *bfqd, struct bfq_queue *bfqq)
bfq_schedule_dispatch(bfqd);
}
- bfq_log_bfqq(bfqd, bfqq, "exit_bfqq: %p, %d", bfqq, bfqq->ref);
+ bfq_log_bfqq(bfqd, bfqq, "exit_bfqq: %p, %d", bfqq, atomic_read(&bfqq->ref));
bfq_put_cooperator(bfqq);
@@ -5507,7 +5508,7 @@ static void bfq_init_bfqq(struct bfq_data *bfqd, struct bfq_queue *bfqq,
INIT_HLIST_NODE(&bfqq->woken_list_node);
INIT_HLIST_HEAD(&bfqq->woken_list);
- bfqq->ref = 0;
+ atomic_set(&bfqq->ref, 0);
bfqq->bfqd = bfqd;
if (bic)
@@ -5710,12 +5711,12 @@ static struct bfq_queue *bfq_do_or_sched_stable_merge(struct bfq_data *bfqd,
* to prevent it from being freed,
* until we decide whether to merge
*/
- last_bfqq_created->ref++;
+ atomic_inc(&last_bfqq_created->ref);
/*
* need to keep track of stable refs, to
* compute process refs correctly
*/
- last_bfqq_created->stable_ref++;
+ atomic_inc(&last_bfqq_created->stable_ref);
/*
* Record the bfqq to merge to.
*/
@@ -5767,20 +5768,21 @@ static struct bfq_queue *bfq_get_queue(struct bfq_data *bfqd,
* prune it.
*/
if (async_bfqq) {
- bfqq->ref++; /*
- * Extra group reference, w.r.t. sync
- * queue. This extra reference is removed
- * only if bfqq->bfqg disappears, to
- * guarantee that this queue is not freed
- * until its group goes away.
- */
+ atomic_inc(&bfqq->ref);
+ /*
+ * Extra group reference, w.r.t. sync
+ * queue. This extra reference is removed
+ * only if bfqq->bfqg disappears, to
+ * guarantee that this queue is not freed
+ * until its group goes away.
+ */
bfq_log_bfqq(bfqd, bfqq, "get_queue, bfqq not in async: %p, %d",
- bfqq, bfqq->ref);
+ bfqq, atomic_read(&bfqq->ref));
*async_bfqq = bfqq;
}
out:
- bfqq->ref++; /* get a process reference to this queue */
+ atomic_inc(&bfqq->ref); /* get a process reference to this queue */
if (bfqq != &bfqd->oom_bfqq && is_sync && !respawn)
bfqq = bfq_do_or_sched_stable_merge(bfqd, bfqq, bic);
@@ -6059,7 +6061,7 @@ static bool __bfq_insert_request(struct bfq_data *bfqd, struct request *rq)
*/
bfqq_request_allocated(new_bfqq);
bfqq_request_freed(bfqq);
- new_bfqq->ref++;
+ atomic_inc(&new_bfqq->ref);
/*
* If the bic associated with the process
* issuing this request still points to bfqq
@@ -6803,10 +6805,10 @@ static struct bfq_queue *bfq_init_rq(struct request *rq)
}
bfqq_request_allocated(bfqq);
- bfqq->ref++;
+ atomic_inc(&bfqq->ref);
bic->requests++;
bfq_log_bfqq(bfqd, bfqq, "get_request %p: bfqq %p, %d",
- rq, bfqq, bfqq->ref);
+ rq, bfqq, atomic_read(&bfqq->ref));
rq->elv.priv[0] = bic;
rq->elv.priv[1] = bfqq;
@@ -6939,7 +6941,7 @@ static void __bfq_put_async_bfqq(struct bfq_data *bfqd,
bfq_bfqq_move(bfqd, bfqq, bfqd->root_group);
bfq_log_bfqq(bfqd, bfqq, "put_async_bfqq: putting %p, %d",
- bfqq, bfqq->ref);
+ bfqq, atomic_read(&bfqq->ref));
bfq_put_queue(bfqq);
*bfqq_ptr = NULL;
}
@@ -7092,7 +7094,7 @@ static int bfq_init_queue(struct request_queue *q, struct elevator_type *e)
* will not attempt to free it.
*/
bfq_init_bfqq(bfqd, &bfqd->oom_bfqq, NULL, 1, 0);
- bfqd->oom_bfqq.ref++;
+ atomic_inc(&bfqd->oom_bfqq.ref);
bfqd->oom_bfqq.new_ioprio = BFQ_DEFAULT_QUEUE_IOPRIO;
bfqd->oom_bfqq.new_ioprio_class = IOPRIO_CLASS_BE;
bfqd->oom_bfqq.entity.new_weight =
diff --git a/block/bfq-iosched.h b/block/bfq-iosched.h
index 64ee618064ba..71ac0de80bb0 100644
--- a/block/bfq-iosched.h
+++ b/block/bfq-iosched.h
@@ -234,9 +234,9 @@ struct bfq_ttime {
*/
struct bfq_queue {
/* reference counter */
- int ref;
+ atomic_t ref;
/* counter of references from other queues for delayed stable merge */
- int stable_ref;
+ atomic_t stable_ref;
/* parent bfq_data */
struct bfq_data *bfqd;
@@ -928,7 +928,7 @@ struct bfq_group {
char blkg_path[128];
/* reference counter (see comments in bfq_bic_update_cgroup) */
- int ref;
+ atomic_t ref;
/* Is bfq_group still online? */
bool online;
diff --git a/block/bfq-wf2q.c b/block/bfq-wf2q.c
index 8fc3da4c23bb..60a9a2c1fc8d 100644
--- a/block/bfq-wf2q.c
+++ b/block/bfq-wf2q.c
@@ -512,9 +512,9 @@ static void bfq_get_entity(struct bfq_entity *entity)
struct bfq_queue *bfqq = bfq_entity_to_bfqq(entity);
if (bfqq) {
- bfqq->ref++;
+ atomic_inc(&bfqq->ref);
bfq_log_bfqq(bfqq->bfqd, bfqq, "get_entity: %p %d",
- bfqq, bfqq->ref);
+ bfqq, atomic_read(&bfqq->ref));
}
}
@@ -1611,7 +1611,7 @@ bool __bfq_bfqd_reset_in_service(struct bfq_data *bfqd)
* reference to the queue. If this is the case, then
* bfqq gets freed here.
*/
- int ref = in_serv_bfqq->ref;
+ int ref = atomic_read(&in_serv_bfqq->ref);
bfq_put_queue(in_serv_bfqq);
if (ref == 1)
return true;
--
2.25.1
Hi,
?? 2022/10/11 22:52, Dawei Li ะด??:
> For most implementations of reference count, atomic_t is preferred
> for their natural-born atomic ops capability.
> Change the reference count of bfq_group/bfq_queue, both data structures
> and related ops, into atomic.
I'm afraid that this is unnecessary, the modifications of reference
count are inside spin_lock() in bfq.
Thanks,
Kuai
>
> Signed-off-by: Dawei Li <[email protected]>
> ---
> block/bfq-cgroup.c | 8 +++----
> block/bfq-iosched.c | 54 +++++++++++++++++++++++----------------------
> block/bfq-iosched.h | 6 ++---
> block/bfq-wf2q.c | 6 ++---
> 4 files changed, 37 insertions(+), 37 deletions(-)
>
> diff --git a/block/bfq-cgroup.c b/block/bfq-cgroup.c
> index 144bca006463..714126ba21b6 100644
> --- a/block/bfq-cgroup.c
> +++ b/block/bfq-cgroup.c
> @@ -316,14 +316,12 @@ struct bfq_group *bfqq_group(struct bfq_queue *bfqq)
>
> static void bfqg_get(struct bfq_group *bfqg)
> {
> - bfqg->ref++;
> + atomic_inc(&bfqg->ref);
> }
>
> static void bfqg_put(struct bfq_group *bfqg)
> {
> - bfqg->ref--;
> -
> - if (bfqg->ref == 0)
> + if (atomic_dec_and_test(&bfqg->ref))
> kfree(bfqg);
> }
>
> @@ -659,7 +657,7 @@ void bfq_bfqq_move(struct bfq_data *bfqd, struct bfq_queue *bfqq,
> * Get extra reference to prevent bfqq from being freed in
> * next possible expire or deactivate.
> */
> - bfqq->ref++;
> + atomic_inc(&bfqq->ref);
>
> /* If bfqq is empty, then bfq_bfqq_expire also invokes
> * bfq_del_bfqq_busy, thereby removing bfqq and its entity
> diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
> index 7ea427817f7f..fbe5624be71f 100644
> --- a/block/bfq-iosched.c
> +++ b/block/bfq-iosched.c
> @@ -935,7 +935,7 @@ void bfq_weights_tree_add(struct bfq_data *bfqd, struct bfq_queue *bfqq,
>
> inc_counter:
> bfqq->weight_counter->num_active++;
> - bfqq->ref++;
> + atomic_inc(&bfqq->ref);
> }
>
> /*
> @@ -1224,9 +1224,10 @@ bfq_bfqq_resume_state(struct bfq_queue *bfqq, struct bfq_data *bfqd,
>
> static int bfqq_process_refs(struct bfq_queue *bfqq)
> {
> - return bfqq->ref - bfqq->entity.allocated -
> + return atomic_read(&bfqq->ref) - bfqq->entity.allocated -
> bfqq->entity.on_st_or_in_serv -
> - (bfqq->weight_counter != NULL) - bfqq->stable_ref;
> + (bfqq->weight_counter != NULL) -
> + atomic_read(&bfqq->stable_ref);
> }
>
> /* Empty burst list and add just bfqq (see comments on bfq_handle_burst) */
> @@ -2818,7 +2819,7 @@ bfq_setup_merge(struct bfq_queue *bfqq, struct bfq_queue *new_bfqq)
> * expected to be associated with new_bfqq as they happen to
> * issue I/O.
> */
> - new_bfqq->ref += process_refs;
> + atomic_add(process_refs, &new_bfqq->ref);
> return new_bfqq;
> }
>
> @@ -5255,10 +5256,10 @@ void bfq_put_queue(struct bfq_queue *bfqq)
> struct hlist_node *n;
> struct bfq_group *bfqg = bfqq_group(bfqq);
>
> - bfq_log_bfqq(bfqq->bfqd, bfqq, "put_queue: %p %d", bfqq, bfqq->ref);
> + bfq_log_bfqq(bfqq->bfqd, bfqq, "put_queue: %p %d", bfqq,
> + atomic_read(&bfqq->ref));
>
> - bfqq->ref--;
> - if (bfqq->ref)
> + if (!atomic_dec_and_test(&bfqq->ref))
> return;
>
> if (!hlist_unhashed(&bfqq->burst_list_node)) {
> @@ -5328,7 +5329,7 @@ void bfq_put_queue(struct bfq_queue *bfqq)
>
> static void bfq_put_stable_ref(struct bfq_queue *bfqq)
> {
> - bfqq->stable_ref--;
> + atomic_dec(&bfqq->stable_ref);
> bfq_put_queue(bfqq);
> }
>
> @@ -5358,7 +5359,7 @@ static void bfq_exit_bfqq(struct bfq_data *bfqd, struct bfq_queue *bfqq)
> bfq_schedule_dispatch(bfqd);
> }
>
> - bfq_log_bfqq(bfqd, bfqq, "exit_bfqq: %p, %d", bfqq, bfqq->ref);
> + bfq_log_bfqq(bfqd, bfqq, "exit_bfqq: %p, %d", bfqq, atomic_read(&bfqq->ref));
>
> bfq_put_cooperator(bfqq);
>
> @@ -5507,7 +5508,7 @@ static void bfq_init_bfqq(struct bfq_data *bfqd, struct bfq_queue *bfqq,
> INIT_HLIST_NODE(&bfqq->woken_list_node);
> INIT_HLIST_HEAD(&bfqq->woken_list);
>
> - bfqq->ref = 0;
> + atomic_set(&bfqq->ref, 0);
> bfqq->bfqd = bfqd;
>
> if (bic)
> @@ -5710,12 +5711,12 @@ static struct bfq_queue *bfq_do_or_sched_stable_merge(struct bfq_data *bfqd,
> * to prevent it from being freed,
> * until we decide whether to merge
> */
> - last_bfqq_created->ref++;
> + atomic_inc(&last_bfqq_created->ref);
> /*
> * need to keep track of stable refs, to
> * compute process refs correctly
> */
> - last_bfqq_created->stable_ref++;
> + atomic_inc(&last_bfqq_created->stable_ref);
> /*
> * Record the bfqq to merge to.
> */
> @@ -5767,20 +5768,21 @@ static struct bfq_queue *bfq_get_queue(struct bfq_data *bfqd,
> * prune it.
> */
> if (async_bfqq) {
> - bfqq->ref++; /*
> - * Extra group reference, w.r.t. sync
> - * queue. This extra reference is removed
> - * only if bfqq->bfqg disappears, to
> - * guarantee that this queue is not freed
> - * until its group goes away.
> - */
> + atomic_inc(&bfqq->ref);
> + /*
> + * Extra group reference, w.r.t. sync
> + * queue. This extra reference is removed
> + * only if bfqq->bfqg disappears, to
> + * guarantee that this queue is not freed
> + * until its group goes away.
> + */
> bfq_log_bfqq(bfqd, bfqq, "get_queue, bfqq not in async: %p, %d",
> - bfqq, bfqq->ref);
> + bfqq, atomic_read(&bfqq->ref));
> *async_bfqq = bfqq;
> }
>
> out:
> - bfqq->ref++; /* get a process reference to this queue */
> + atomic_inc(&bfqq->ref); /* get a process reference to this queue */
>
> if (bfqq != &bfqd->oom_bfqq && is_sync && !respawn)
> bfqq = bfq_do_or_sched_stable_merge(bfqd, bfqq, bic);
> @@ -6059,7 +6061,7 @@ static bool __bfq_insert_request(struct bfq_data *bfqd, struct request *rq)
> */
> bfqq_request_allocated(new_bfqq);
> bfqq_request_freed(bfqq);
> - new_bfqq->ref++;
> + atomic_inc(&new_bfqq->ref);
> /*
> * If the bic associated with the process
> * issuing this request still points to bfqq
> @@ -6803,10 +6805,10 @@ static struct bfq_queue *bfq_init_rq(struct request *rq)
> }
>
> bfqq_request_allocated(bfqq);
> - bfqq->ref++;
> + atomic_inc(&bfqq->ref);
> bic->requests++;
> bfq_log_bfqq(bfqd, bfqq, "get_request %p: bfqq %p, %d",
> - rq, bfqq, bfqq->ref);
> + rq, bfqq, atomic_read(&bfqq->ref));
>
> rq->elv.priv[0] = bic;
> rq->elv.priv[1] = bfqq;
> @@ -6939,7 +6941,7 @@ static void __bfq_put_async_bfqq(struct bfq_data *bfqd,
> bfq_bfqq_move(bfqd, bfqq, bfqd->root_group);
>
> bfq_log_bfqq(bfqd, bfqq, "put_async_bfqq: putting %p, %d",
> - bfqq, bfqq->ref);
> + bfqq, atomic_read(&bfqq->ref));
> bfq_put_queue(bfqq);
> *bfqq_ptr = NULL;
> }
> @@ -7092,7 +7094,7 @@ static int bfq_init_queue(struct request_queue *q, struct elevator_type *e)
> * will not attempt to free it.
> */
> bfq_init_bfqq(bfqd, &bfqd->oom_bfqq, NULL, 1, 0);
> - bfqd->oom_bfqq.ref++;
> + atomic_inc(&bfqd->oom_bfqq.ref);
> bfqd->oom_bfqq.new_ioprio = BFQ_DEFAULT_QUEUE_IOPRIO;
> bfqd->oom_bfqq.new_ioprio_class = IOPRIO_CLASS_BE;
> bfqd->oom_bfqq.entity.new_weight =
> diff --git a/block/bfq-iosched.h b/block/bfq-iosched.h
> index 64ee618064ba..71ac0de80bb0 100644
> --- a/block/bfq-iosched.h
> +++ b/block/bfq-iosched.h
> @@ -234,9 +234,9 @@ struct bfq_ttime {
> */
> struct bfq_queue {
> /* reference counter */
> - int ref;
> + atomic_t ref;
> /* counter of references from other queues for delayed stable merge */
> - int stable_ref;
> + atomic_t stable_ref;
> /* parent bfq_data */
> struct bfq_data *bfqd;
>
> @@ -928,7 +928,7 @@ struct bfq_group {
> char blkg_path[128];
>
> /* reference counter (see comments in bfq_bic_update_cgroup) */
> - int ref;
> + atomic_t ref;
> /* Is bfq_group still online? */
> bool online;
>
> diff --git a/block/bfq-wf2q.c b/block/bfq-wf2q.c
> index 8fc3da4c23bb..60a9a2c1fc8d 100644
> --- a/block/bfq-wf2q.c
> +++ b/block/bfq-wf2q.c
> @@ -512,9 +512,9 @@ static void bfq_get_entity(struct bfq_entity *entity)
> struct bfq_queue *bfqq = bfq_entity_to_bfqq(entity);
>
> if (bfqq) {
> - bfqq->ref++;
> + atomic_inc(&bfqq->ref);
> bfq_log_bfqq(bfqq->bfqd, bfqq, "get_entity: %p %d",
> - bfqq, bfqq->ref);
> + bfqq, atomic_read(&bfqq->ref));
> }
> }
>
> @@ -1611,7 +1611,7 @@ bool __bfq_bfqd_reset_in_service(struct bfq_data *bfqd)
> * reference to the queue. If this is the case, then
> * bfqq gets freed here.
> */
> - int ref = in_serv_bfqq->ref;
> + int ref = atomic_read(&in_serv_bfqq->ref);
> bfq_put_queue(in_serv_bfqq);
> if (ref == 1)
> return true;
>