Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp3510194imm; Mon, 6 Aug 2018 06:09:19 -0700 (PDT) X-Google-Smtp-Source: AAOMgpfNSN7wARzXwLE9FmjfGXvlmMlxXbMzHLWAAy8aVGZ3pjm32NNc+fCdorikGHLsqSfJxRZN X-Received: by 2002:a17:902:7446:: with SMTP id e6-v6mr14067006plt.161.1533560959630; Mon, 06 Aug 2018 06:09:19 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1533560959; cv=none; d=google.com; s=arc-20160816; b=jXMapSrgEfS5UJ28mAh5z23OLO8robSdX2sMtnA1GP6i04KjBBTKYPx3pjCb4f4nko 6OGOhLhZiEIy78TW6KNsSKU875klM0ju9Tun7FLzvdJYzpNnMDKiomc9t0FexqRAwfeY pgAlllrlYifNi5N2KXMKFiWVEAWX8TH4hhywBNHeGZhX/S5r08+cyzORZa2ZuheWBM0U sY+ibDE8ZGEYnCuMNlmVehPXxnLzrE+C84gBZ3FFDhPeg/YHxrloRFANAT0GjXBM5uhw g1wDlxuDIYl6vMKGtlTFgi9XMLayfhiNR+xe1G/pKk7hEyE5n5Z8AL1Dnl9yjnwgU0Vd Z/sg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:arc-authentication-results; bh=yWhfm9weQOLCMFxWi8RWy9MUmJldmRg9ny5D1AbXWqM=; b=yoppQ5Hs3SER07y+J5fmeoudtuVk5Xi7Y5JNPoV6qR59wSzS1Yek5T/w4n7x8fIhzS IxF6Fz9sSZn2Beuk5Q+VSWcFq7Z1wqD2EuOqudwOGPbvFiPIE8crFZDibbhbJSv8/nge fADNLglsRJhInuQRs/nFBuy8UOf4oTLSb3Pvn1PAFWwhMSk1d1ucT/ySM8LWjX19D5tf x5sGtbsolMh5oJez+a29kUu1us5dm3SUfsRfUa/Tvr+7Yi2n17Hyrgzg2CnDBQJ42/av vcdraC9inCEed+ZhHv+n4EiKZGq/hfSJKVJjjva3mxwK+IlZzn08RmAUf+mX860pJKI1 ROhw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j13-v6si13907811pfj.230.2018.08.06.06.09.04; Mon, 06 Aug 2018 06:09:19 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730456AbeHFNnA (ORCPT + 99 others); Mon, 6 Aug 2018 09:43:00 -0400 Received: from mx2.suse.de ([195.135.220.15]:33658 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1727340AbeHFNnA (ORCPT ); Mon, 6 Aug 2018 09:43:00 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay1.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id D4EADAE4C; Mon, 6 Aug 2018 11:34:18 +0000 (UTC) From: Juergen Gross To: linux-kernel@vger.kernel.org, xen-devel@lists.xenproject.org, linux-block@vger.kernel.org Cc: konrad.wilk@oracle.com, roger.pau@citrix.com, axboe@kernel.dk, boris.ostrovsky@oracle.com, Juergen Gross Subject: [PATCH 2/4] xen/blkfront: cleanup stale persistent grants Date: Mon, 6 Aug 2018 13:34:01 +0200 Message-Id: <20180806113403.24728-4-jgross@suse.com> X-Mailer: git-send-email 2.13.7 In-Reply-To: <20180806113403.24728-1-jgross@suse.com> References: <20180806113403.24728-1-jgross@suse.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Add a periodic cleanup function to remove old persistent grants which are no longer in use on the backend side. This avoids starvation in case there are lots of persistent grants for a device which no longer is involved in I/O business. Signed-off-by: Juergen Gross --- drivers/block/xen-blkfront.c | 99 ++++++++++++++++++++++++++++++++++++++++++-- 1 file changed, 95 insertions(+), 4 deletions(-) diff --git a/drivers/block/xen-blkfront.c b/drivers/block/xen-blkfront.c index b5cedccb5d7d..19feb8835fc4 100644 --- a/drivers/block/xen-blkfront.c +++ b/drivers/block/xen-blkfront.c @@ -46,6 +46,7 @@ #include #include #include +#include #include #include @@ -121,6 +122,9 @@ static inline struct blkif_req *blkif_req(struct request *rq) static DEFINE_MUTEX(blkfront_mutex); static const struct block_device_operations xlvbd_block_fops; +static struct delayed_work blkfront_work; +static LIST_HEAD(info_list); +static bool blkfront_work_active; /* * Maximum number of segments in indirect requests, the actual value used by @@ -216,6 +220,7 @@ struct blkfront_info /* Save uncomplete reqs and bios for migration. */ struct list_head requests; struct bio_list bio_list; + struct list_head info_list; }; static unsigned int nr_minors; @@ -1764,6 +1769,12 @@ static int write_per_ring_nodes(struct xenbus_transaction xbt, return err; } +static void free_info(struct blkfront_info *info) +{ + list_del(&info->info_list); + kfree(info); +} + /* Common code used when first setting up, and when resuming. */ static int talk_to_blkback(struct xenbus_device *dev, struct blkfront_info *info) @@ -1885,7 +1896,10 @@ static int talk_to_blkback(struct xenbus_device *dev, destroy_blkring: blkif_free(info, 0); - kfree(info); + mutex_lock(&blkfront_mutex); + free_info(info); + mutex_unlock(&blkfront_mutex); + dev_set_drvdata(&dev->dev, NULL); return err; @@ -1996,6 +2010,10 @@ static int blkfront_probe(struct xenbus_device *dev, info->handle = simple_strtoul(strrchr(dev->nodename, '/')+1, NULL, 0); dev_set_drvdata(&dev->dev, info); + mutex_lock(&blkfront_mutex); + list_add(&info->info_list, &info_list); + mutex_unlock(&blkfront_mutex); + return 0; } @@ -2306,6 +2324,15 @@ static void blkfront_gather_backend_features(struct blkfront_info *info) if (indirect_segments <= BLKIF_MAX_SEGMENTS_PER_REQUEST) indirect_segments = 0; info->max_indirect_segments = indirect_segments; + + if (info->feature_persistent) { + mutex_lock(&blkfront_mutex); + if (!blkfront_work_active) { + blkfront_work_active = true; + schedule_delayed_work(&blkfront_work, HZ * 10); + } + mutex_unlock(&blkfront_mutex); + } } /* @@ -2487,7 +2514,9 @@ static int blkfront_remove(struct xenbus_device *xbdev) mutex_unlock(&info->mutex); if (!bdev) { - kfree(info); + mutex_lock(&blkfront_mutex); + free_info(info); + mutex_unlock(&blkfront_mutex); return 0; } @@ -2507,7 +2536,9 @@ static int blkfront_remove(struct xenbus_device *xbdev) if (info && !bdev->bd_openers) { xlvbd_release_gendisk(info); disk->private_data = NULL; - kfree(info); + mutex_lock(&blkfront_mutex); + free_info(info); + mutex_unlock(&blkfront_mutex); } mutex_unlock(&bdev->bd_mutex); @@ -2590,7 +2621,7 @@ static void blkif_release(struct gendisk *disk, fmode_t mode) dev_info(disk_to_dev(bdev->bd_disk), "releasing disk\n"); xlvbd_release_gendisk(info); disk->private_data = NULL; - kfree(info); + free_info(info); } out: @@ -2623,6 +2654,62 @@ static struct xenbus_driver blkfront_driver = { .is_ready = blkfront_is_ready, }; +static void purge_persistent_grants(struct blkfront_info *info) +{ + unsigned int i; + unsigned long flags; + + for (i = 0; i < info->nr_rings; i++) { + struct blkfront_ring_info *rinfo = &info->rinfo[i]; + struct grant *gnt_list_entry, *tmp; + + spin_lock_irqsave(&rinfo->ring_lock, flags); + + if (rinfo->persistent_gnts_c == 0) { + spin_unlock_irqrestore(&rinfo->ring_lock, flags); + continue; + } + + list_for_each_entry_safe(gnt_list_entry, tmp, &rinfo->grants, + node) { + if (gnt_list_entry->gref == GRANT_INVALID_REF || + gnttab_query_foreign_access(gnt_list_entry->gref)) + continue; + + list_del(&gnt_list_entry->node); + gnttab_end_foreign_access(gnt_list_entry->gref, 0, 0UL); + rinfo->persistent_gnts_c--; + __free_page(gnt_list_entry->page); + kfree(gnt_list_entry); + } + + spin_unlock_irqrestore(&rinfo->ring_lock, flags); + } +} + +static void blkfront_delay_work(struct work_struct *work) +{ + struct blkfront_info *info; + + mutex_lock(&blkfront_mutex); + + blkfront_work_active = false; + + list_for_each_entry(info, &info_list, info_list) { + if (info->feature_persistent) { + blkfront_work_active = true; + mutex_lock(&info->mutex); + purge_persistent_grants(info); + mutex_unlock(&info->mutex); + } + } + + if (blkfront_work_active) + schedule_delayed_work(&blkfront_work, HZ * 10); + + mutex_unlock(&blkfront_mutex); +} + static int __init xlblk_init(void) { int ret; @@ -2655,6 +2742,8 @@ static int __init xlblk_init(void) return -ENODEV; } + INIT_DELAYED_WORK(&blkfront_work, blkfront_delay_work); + ret = xenbus_register_frontend(&blkfront_driver); if (ret) { unregister_blkdev(XENVBD_MAJOR, DEV_NAME); @@ -2668,6 +2757,8 @@ module_init(xlblk_init); static void __exit xlblk_exit(void) { + cancel_delayed_work_sync(&blkfront_work); + xenbus_unregister_driver(&blkfront_driver); unregister_blkdev(XENVBD_MAJOR, DEV_NAME); kfree(minors); -- 2.13.7