Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S934141AbdC3Vyb (ORCPT ); Thu, 30 Mar 2017 17:54:31 -0400 Received: from mail-io0-f176.google.com ([209.85.223.176]:34843 "EHLO mail-io0-f176.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755124AbdC3Vya (ORCPT ); Thu, 30 Mar 2017 17:54:30 -0400 MIME-Version: 1.0 In-Reply-To: References: <20170317231636.142311-1-timmurray@google.com> <20170330155123.GA3929@cmpxchg.org> From: Tim Murray Date: Thu, 30 Mar 2017 14:54:28 -0700 Message-ID: Subject: Re: [RFC 0/1] add support for reclaiming priorities per mem cgroup To: Johannes Weiner Cc: Michal Hocko , Vladimir Davydov , LKML , cgroups@vger.kernel.org, Linux-MM , Suren Baghdasaryan , Patrik Torstensson , Android Kernel Team Content-Type: text/plain; charset=UTF-8 Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 676 Lines: 13 On Thu, Mar 30, 2017 at 12:40 PM, Tim Murray wrote: > The current critical vmpressure event > hasn't been that successful in avoiding oom-killer (on 3.18, at > least)--I've been able to get oom-killer to trigger without a > vmpressure event. Looked at this some more, and this is almost certainly because vmpressure relies on workqueues. Scheduling delay from CFS workqueues would explain vmpressure latency that results in oom-killer running long before the critical vmpressure notification is received in userspace, even if userspace is running as FIFO. We regularly see 10ms+ latency on workqueues, even when an Android device isn't heavily loaded.