Received: by 2002:a05:6a10:1d13:0:0:0:0 with SMTP id pp19csp592476pxb; Wed, 25 Aug 2021 10:10:25 -0700 (PDT) X-Google-Smtp-Source: ABdhPJweZvpwjPZ1YuGlk32FP1fwGpzVskvRbXn5Q1QrVXduCWwL/XiCxaTbPDGGVvlhqYzbBOcm X-Received: by 2002:a92:444e:: with SMTP id a14mr24518534ilm.152.1629911425666; Wed, 25 Aug 2021 10:10:25 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1629911425; cv=none; d=google.com; s=arc-20160816; b=F26HrNEE4Dg7vVEMM/AeoK6SKgdBNVVMmdekRi9OSErVGZDYnOvm+WFXBgkWLwfXEX 0b6/v5dnyz+Y0704qL2fkllhRw8MxozS8LLS00lXVI3r/+bxzUEjOABjJuQRL0QlSG8g K4RcqX+l2RuA6123soqXXI0Mn3b296RBFApiOWJ6IxZ2hHZMNHGNs48jNIZWEerNzURp lCzBIU2oJwcs2L8xRG3nOY6s08sfWx5YNxWJGOGjFW7dT+qI49UXPz1nzE+YHDF2jAls AF3AaOi0EsLobKPl0bkuKwvBCwH1jH8fwC6Xrk12W/8t3MRGww5NEOisF0Kbi/9JEFNn pUPg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=9UW6pSpJ4FLqQ4wBMuO+kq8Bqfp9+2FRMMWJSAkncAI=; b=ZS8lljaiM6cUEipmfYm6bbkQ1nYJt60M+Ad6v6Xx3reFjp+qu5kxl+hrXmHzZky9xw wSAnA6OGLRz/mUNgp5Gs26rVuZnGRq6TwMi5tL6iUOql4bgNKLPf+lkVH6HYPD3OuL69 NQsV0jKQ8GXnUfJ9fD1to6AiJF7JqgUXQbOorjQHtqvKD0bONoU0zDCxwwG2ZolfJ1TY 82Po1maOGpSeVm9/I6NKDrngiEWA9VZc12nmL3URDcsOcmF/HLpkV0jWAFfYcm3fXuo8 6V+NQ/UtRX5POn7hYTP/LzEVoI2rxXc8e/6KkswGIiGmaXaR+RKIgrtGPYrltuBJB+Og O7CA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@intel-com.20150623.gappssmtp.com header.s=20150623 header.b=I9y2FdMr; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id t1si252011ioj.70.2021.08.25.10.10.12; Wed, 25 Aug 2021 10:10:25 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@intel-com.20150623.gappssmtp.com header.s=20150623 header.b=I9y2FdMr; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235699AbhHYRIb (ORCPT + 99 others); Wed, 25 Aug 2021 13:08:31 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:50562 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229711AbhHYRIa (ORCPT ); Wed, 25 Aug 2021 13:08:30 -0400 Received: from mail-pf1-x42c.google.com (mail-pf1-x42c.google.com [IPv6:2607:f8b0:4864:20::42c]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 56981C061757 for ; Wed, 25 Aug 2021 10:07:44 -0700 (PDT) Received: by mail-pf1-x42c.google.com with SMTP id y190so281602pfg.7 for ; Wed, 25 Aug 2021 10:07:44 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=intel-com.20150623.gappssmtp.com; s=20150623; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=9UW6pSpJ4FLqQ4wBMuO+kq8Bqfp9+2FRMMWJSAkncAI=; b=I9y2FdMrrmYyf94isIAJ1/5++twpxEF/rRiEBCII1BWzsHJnYRoj4J7VE0NAm3JrRd HZYC/xbPm39a37i+d8l6sLgmTRhu5G5LlK1sPZII33IlQRyvuu6GS1Has8ou6QawOcwp rDofaFIRlOHM8MxbvIgoV84p1SC8mN/fFj6OcEhcnECQBx404gZNxvVSuGolzSCYHHna 34LC/7CRu0SYTVFQmimM8xjMTOJ9K4mqNBCk+k7Xua5TUH0k3mWEcv5Km/r003ZRf2GP GlBoPGiXMeNGfSQBDSEPlj1yTudVNpllYYqGkEfTQvpsy/sLQpzuYaDRdgxkZrHtGBz5 omoA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=9UW6pSpJ4FLqQ4wBMuO+kq8Bqfp9+2FRMMWJSAkncAI=; b=dfY+6/NY6DLsidmsbAcRUPzwaTnGDyIJkFCoGjED71l4adiGIbQ4Y0gg0Wc1i529G2 F6bKWWXezt5rHhLL/8hOFdmLGMC8MTDhfdxSiIwWOXdOuVyYtc0P8/EiYvuRxFjKrsAa pIYlCllSt/bQ4hJEfu4sfcaJDIcqcsb4eFG9QluWAwHLChmZZgXb/724jeRQ3jjV/tZ8 0Czs1ZMFLaGxQR39wQfzMeYafAmS0rv+HHaa7V5pmBEN2cYwsAB96vD0IiN6Ow5sFIuK DQI+Svqp7EjzVax6HPz+TpAeX9n3D5Cn+tubWR6PEd6YPR22IxU8znqgylTAB7AoMYCA kXRg== X-Gm-Message-State: AOAM533orH6gtoRhebTcjHqcGoaM2gtx88dexCa4JFP2PAldu2ZLB2OC nWy27+5MpAJJcKGFmTvpIjJZdnD3StzPTyD4Eoj67A== X-Received: by 2002:a65:6642:: with SMTP id z2mr30617563pgv.240.1629911263764; Wed, 25 Aug 2021 10:07:43 -0700 (PDT) MIME-Version: 1.0 References: <20210816142505.28359-1-david@redhat.com> <20210816142505.28359-2-david@redhat.com> In-Reply-To: From: Dan Williams Date: Wed, 25 Aug 2021 10:07:33 -0700 Message-ID: Subject: Re: [PATCH v2 1/3] /dev/mem: disallow access to explicitly excluded system RAM regions To: David Hildenbrand Cc: Linux Kernel Mailing List , Arnd Bergmann , Greg Kroah-Hartman , "Michael S. Tsirkin" , Jason Wang , "Rafael J. Wysocki" , Andrew Morton , Hanjun Guo , Andy Shevchenko , virtualization@lists.linux-foundation.org, Linux MM Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Aug 25, 2021 at 12:23 AM David Hildenbrand wrote: > > On 25.08.21 02:58, Dan Williams wrote: > > On Mon, Aug 16, 2021 at 7:25 AM David Hildenbrand wrote: > >> > >> virtio-mem dynamically exposes memory inside a device memory region as > >> system RAM to Linux, coordinating with the hypervisor which parts are > >> actually "plugged" and consequently usable/accessible. On the one hand, the > >> virtio-mem driver adds/removes whole memory blocks, creating/removing busy > >> IORESOURCE_SYSTEM_RAM resources, on the other hand, it logically (un)plugs > >> memory inside added memory blocks, dynamically either exposing them to > >> the buddy or hiding them from the buddy and marking them PG_offline. > >> > >> virtio-mem wants to make sure that in a sane environment, nobody > >> "accidentially" accesses unplugged memory inside the device managed > >> region. After /proc/kcore has been sanitized and /dev/kmem has been > >> removed, /dev/mem is the remaining interface that still allows uncontrolled > >> access to the device-managed region of virtio-mem devices from user > >> space. > >> > >> There is no known sane use case for mapping virtio-mem device memory > >> via /dev/mem while virtio-mem driver concurrently (un)plugs memory inside > >> that region. So once the driver was loaded and detected the device > >> along the device-managed region, we just want to disallow any access via > >> /dev/mem to it. > >> > >> Let's add the basic infrastructure to exclude some physical memory > >> regions completely from /dev/mem access, on any architecture and under > >> any system configuration (independent of CONFIG_STRICT_DEVMEM and > >> independent of "iomem="). > > > > I'm certainly on team "/dev/mem considered harmful", but this approach > > feels awkward. It feels wrong for being non-committal about whether > > CONFIG_STRICT_DEVMEM is in wide enough use that the safety can be > > turned on all the time, and the configuration option dropped, or there > > are users clinging onto /dev/mem where they expect to be able to build > > a debug kernel to turn all of these restrictions off, even the > > virtio-mem ones. This splits the difference and says some /dev/mem > > accesses are always disallowed for "reasons", but I could say the same > > thing about pmem, there's no sane reason to allow /dev/mem which has > > no idea about the responsibilities of properly touching pmem to get > > access to it. > > For virtio-mem, there is no use case *and* access could be harmful; I > don't even want to allow if for debugging purposes. If you want to > inspect virtio-mem device memory content, use /proc/kcore, which > performs proper synchronized access checks. Modifying random virtio-mem > memory via /dev/mem in a debug kernel will not be possible: if you > really have to play with fire, use kdb or better don't load the > virtio-mem driver during boot, such that the kernel won't even be making > use of device memory. > > I don't want people disabling CONFIG_STRICT_DEVMEM, or booting with > "iomem=relaxed", and "accidentally" accessing any of virtio-mem memory > via /dev/mem, while it gets concurrently plugged/unplugged by the > virtio-mem driver. Not even for debugging purposes. That sounds more an argument that all of the existing "kernel is using this region" cases should become mandatory exclusions. If unloading the driver removes the exclusion then that's precisely CONFIG_IO_STRICT_DEVMEM. Why is the virtio-mem driver more special than any other driver that expects this integrity guarantee? > We disallow mapping to some other regions independent of > CONFIG_STRICT_DEVMEM already, so the idea to ignore CONFIG_STRICT_DEVMEM > is not completely new: > > "Note that with PAT support enabled, even in this case there are > restrictions on /dev/mem use due to the cache aliasing requirements." > > Maybe you even want to do something similar with PMEM now that there is > infrastructure for it and just avoid having to deal with revoking > /dev/mem mappings later. That would be like blocking writes to /dev/sda just because a filesytem might later be mounted on it. If the /dev/mem access is not actively colliding with other kernel operations what business does the kernel have saying no? I'm pushing on this topic because I am also considering an exclusion on PCI configuration access to the "DOE mailbox" since it can disrupt the kernel's operation, at the same time, root can go change PCI BARs to nonsensical values whenever it wants which is also in the category of "has no use case && could be harmful". > I think there are weird debugging/educational setups [1] that still > require CONFIG_STRICT_DEVMEM=n even with iomem=relaxed. Take a look at > lib/devmem_is_allowed.c:devmem_is_allowed(), it disallows any access to > (what's currently added as) System RAM. It might just do what people > want when dealing with system RAM that doesn't suddenly vanish , so I > don't ultimately see why we should remove CONFIG_STRICT_DEVMEM=n. Yes, I wanted to tease out more of your rationale on where the line should be drawn, I think a mostly unfettered /dev/mem mode is here to stay.