Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751865AbaDYIip (ORCPT ); Fri, 25 Apr 2014 04:38:45 -0400 Received: from cantor2.suse.de ([195.135.220.15]:38290 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751522AbaDYIif (ORCPT ); Fri, 25 Apr 2014 04:38:35 -0400 Date: Fri, 25 Apr 2014 09:38:30 +0100 From: Mel Gorman To: Dan Streetman Cc: Hugh Dickins , Andrew Morton , Michal Hocko , Christian Ehrhardt , Shaohua Li , Weijie Yang , Linux-MM , linux-kernel Subject: Re: [PATCH 1/2] swap: change swap_info singly-linked list to list_head Message-ID: <20140425083830.GY23991@suse.de> References: <1397336454-13855-1-git-send-email-ddstreet@ieee.org> <1397336454-13855-2-git-send-email-ddstreet@ieee.org> <20140423103400.GH23991@suse.de> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-15 Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Apr 24, 2014 at 02:48:43PM -0400, Dan Streetman wrote: > >> > >> - } > >> - > >> + list_for_each(tmp, &swap_list_head) { > >> + si = list_entry(tmp, typeof(*si), list); > >> spin_lock(&si->lock); > >> - if (!si->highest_bit) { > >> - spin_unlock(&si->lock); > >> - continue; > >> - } > >> - if (!(si->flags & SWP_WRITEOK)) { > >> + if (!si->highest_bit || !(si->flags & SWP_WRITEOK)) { > >> spin_unlock(&si->lock); > >> continue; > >> } > >> > >> - swap_list.next = next; > >> + /* > >> + * rotate the current swap_info that we're going to use > >> + * to after any other swap_info that have the same prio, > >> + * so that all equal-priority swap_info get used equally > >> + */ > >> + next = si; > >> + list_for_each_entry_continue(next, &swap_list_head, list) { > >> + if (si->prio != next->prio) > >> + break; > >> + list_rotate_left(&si->list); > >> + next = si; > >> + } > >> > > > > The list manipulations will be a lot of cache writes as the list is shuffled > > around. On slow storage I do not think this will be noticable but it may > > be noticable on faster swap devices that are SSD based. I've added Shaohua > > Li to the cc as he has been concerned with the performance of swap in the > > past. Shaohua, can you run this patchset through any of your test cases > > with the addition that multiple swap files are used to see if the cache > > writes are noticable? You'll need multiple swap files, some of which are > > at equal priority so the list shuffling logic is triggered. > > One performance improvement could be instead of rotating the current > entry past each following same-prio entry, just scan to the end of the > same-prio entries and move the current entry there; that would skip > the extra writes. Especially since this code will run for each > get_swap_page(), no need for any unnecessary writes. > Shaohua is the person that would be most sensitive to performance problems in this area and his tests are in the clear. If he's happy then I don't think there is justification for changing the patch as-is. > > > >> spin_unlock(&swap_lock); > >> /* This is called for allocating swap entry for cache */ > >> offset = scan_swap_map(si, SWAP_HAS_CACHE); > >> spin_unlock(&si->lock); > >> if (offset) > >> - return swp_entry(type, offset); > >> + return swp_entry(si->type, offset); > >> spin_lock(&swap_lock); > >> - next = swap_list.next; > >> + /* > >> + * shouldn't really have got here, but for some reason the > >> + * scan_swap_map came back empty for this swap_info. > >> + * Since we dropped the swap_lock, there may now be > >> + * non-full higher prio swap_infos; let's start over. > >> + */ > >> + tmp = &swap_list_head; > >> } > > > > Has this ever triggered? The number of swap pages was examined under the > > swap lock so no other process should have been iterating through the > > swap files. Once a candidate was found, the si lock was acquired for the > > swap scan map so nothing else should have raced with it. > > Well scan_swap_map() does drop the si->lock if it has any trouble at > all finding an offset to use, so I think it's possible that for a > nearly-full si multiple concurrent get_swap_page() calls could enter > scan_swap_map() with the same si, only some of them actually get pages > from the si and then the si becomes full, and the other threads in > scan_swap_map() see it's full and exit in failure. I can update the > code comment there to better indicate why it was reached, instead of > just saying "we shouldn't have got here" :-) > With the updates to some comments then feel free to add Acked-by: Mel Gorman > It may also be worth trying to get a better indicator of "available" > swap_info_structs for use in get_swap_page(), either by looking at > something other than si->highest_bit and/or keeping the si out of the > prio_list until it's actually available for use, not just has a single > entry free. However, that probably won't be simple and might be > better as a separate patch to the rest of these changes. > I agree that it is likely outside the scope of what this series is meant to accomplish. -- Mel Gorman SUSE Labs -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/