Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753353AbdIRKN6 (ORCPT ); Mon, 18 Sep 2017 06:13:58 -0400 Received: from mx2.suse.de ([195.135.220.15]:38785 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1753171AbdIRKN5 (ORCPT ); Mon, 18 Sep 2017 06:13:57 -0400 Date: Mon, 18 Sep 2017 12:13:50 +0200 From: Jan Kara To: Steven Whitehouse Cc: Al Viro , Linus Torvalds , Mimi Zohar , Christoph Hellwig , LSM List , Christoph Hellwig , linux-ima-devel@lists.sourceforge.net, James Morris , Linux Kernel Mailing List , Matthew Garrett , Jan Kara , "Theodore Ts'o" , Andreas Dilger , Jaegeuk Kim , Chao Yu , Bob Peterson , David Woodhouse , Dave Kleikamp , Ryusuke Konishi , Mark Fasheh , Joel Becker , Richard Weinberger , "Darrick J. Wong" , Hugh Dickins , Chris Mason Subject: Re: [PATCH 3/3] ima: use fs method to read integrity data (updated patch description) Message-ID: <20170918101350.GI32516@quack2.suse.cz> References: <1505451494-30228-1-git-send-email-zohar@linux.vnet.ibm.com> <1505451494-30228-4-git-send-email-zohar@linux.vnet.ibm.com> <1505507142.4200.103.camel@linux.vnet.ibm.com> <20170917151757.GA14262@infradead.org> <1505664935.4200.191.camel@linux.vnet.ibm.com> <20170917163828.GE5426@ZenIV.linux.org.uk> <517c83a6-d7c5-9638-ebaa-52800ca0962c@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <517c83a6-d7c5-9638-ebaa-52800ca0962c@redhat.com> User-Agent: Mutt/1.5.24 (2015-08-30) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2433 Lines: 49 On Mon 18-09-17 10:19:25, Steven Whitehouse wrote: > On 17/09/17 17:38, Al Viro wrote: > >On Sun, Sep 17, 2017 at 09:34:01AM -0700, Linus Torvalds wrote: > >>Now, I suspect most (all?) do, but that's a historical artifact rather > >>than "design". In particular, the VFS layer used to do the locking for > >>the filesystems, to guarantee the POSIX requirements (POSIX requires > >>that writes be seen atomically). > >> > >>But that lock was pushed down into the filesystems, since some > >>filesystems really wanted to have parallel writes (particularly for > >>direct IO, where that POSIX serialization requirement doesn't exist). > >> > >>That's all many years ago, though. New filesystems are likely to have > >>copied the pattern from old ones, but even then.. > >> > >>Also, it's worth noting that "inode->i_rwlock" isn't even well-defined > >>as a lock. You can have the question of *which* inode gets talked > >>about when you have things like eoverlayfs etc. Normally it would be > >>obvious, but sometimes you'd use "file->f_mapping->host" (which is the > >>same thing in the simple cases), and sometimes it really wouldn't be > >>obvious at all.. > >> > >>So... I'm really not at all convinced that i_rwsem is sensible. It's > >>one of those things that are "mostly right for the simple cases", > >>but... > >The thing pretty much common to all of them is that write() might need > >to modify permissions (suid removal), which brings ->i_rwsem in one > >way or another - notify_change() needs that held... > > For GFS2, if we are to hold the inode info constant while it is checked, we > would need to take a glock (read lock in this case) across the relevant > operations. The glock will be happy under i_rwlock, since we have a lock > ordering that takes local locks ahead of cluster locks. I've not dug into > this enough to figure out whether the current proposal will allow this to > work with GFS2 though. Does IMA cache the results from the > ->read_integrity() operation? So I have asked Mimi about clustered filesystems before. And for now the answer was that IMA for clustered filesystems is not supported (it will return some error since ->integrity_read is NULL). If we would ever want to support those it would require larger overhaul of the IMA architecture to give filesystem more control over the locking (which is essentially what Linus wants). Honza -- Jan Kara SUSE Labs, CR