Mel Gorman ([email protected])

Number of posts: 2306 (2.77 per day)
First post: 2015-07-23 11:05:19
Last post: 2017-11-02 13:12:33

Date List Subject
2022-04-22 19:51:00 linux-kernel [PATCH 2/6] mm/page_alloc: Use only one PCP list for THP-sized allocations
2022-04-22 14:23:05 linux-kernel [PATCH 3/6] mm/page_alloc: Split out buddy removal code from rmqueue into separate helper
2022-04-22 10:36:38 linux-kernel [RFC PATCH 0/6] Drain remote per-cpu directly
2022-04-21 09:09:16 linux-kernel [PATCH 5/6] mm/page_alloc: Protect PCP lists with a spinlock
2022-04-21 06:54:44 linux-kernel [PATCH 1/6] mm/page_alloc: Add page->buddy_list and page->pcp_list
2022-04-20 20:31:04 linux-kernel [PATCH 6/6] mm/page_alloc: Remotely drain per-cpu lists
2022-04-07 20:42:34 linux-kernel Re: [PATCH] mm/vmscan.c: no need to double-check if free pages are under high-watermark
2022-04-07 15:59:59 linux-kernel Re: [PATCH] mm, page_alloc: fix build_zonerefs_node()
2022-03-16 12:25:27 linux-kernel Re: [PATCH] mm/page_alloc: call check_pcp_refill() while zone spinlock is not held
2022-03-10 11:39:40 linux-kernel [PATCH] mm/page_alloc: check high-order pages for corruption during PCP operations
2022-03-09 14:28:57 linux-kernel Re: [PATCH v2] mm/page_alloc: call check_new_pages() while zone spinlock is not held
2022-03-09 11:46:19 linux-kernel Re: [LKP] Re: [sched/numa] 0fb3978b0a: stress-ng.fstat.ops_per_sec -18.9% regression
2022-03-07 09:51:50 linux-kernel Re: [PATCH v2] mm/page_alloc: call check_new_pages() while zone spinlock is not held
2022-02-24 01:57:58 linux-kernel Re: [PATCH 5/6] mm/page_alloc: Free pages in a single pass during bulk free
2022-02-24 00:30:46 linux-kernel Re: [PATCH v6 00/71] Introducing the Maple Tree
2022-02-23 23:40:10 linux-kernel Re: [PATCH v2 00/35] Speculative page faults
2022-02-23 15:41:13 linux-kernel Re: [PATCH v4 1/1] mm: vmscan: Reduce throttling due to a failure to make progress'
2022-02-22 17:03:04 linux-kernel Re: [PATCH v5] sched/fair: Consider cpu affinity when allowing NUMA imbalance in find_idlest_group
2022-02-22 09:54:18 linux-kernel Re: [PATCH v4] sched/fair: Consider cpu affinity when allowing NUMA imbalance in find_idlest_group
2022-02-21 17:45:28 linux-kernel Re: [PATCH] mm/pages_alloc.c: Don't create ZONE_MOVABLE beyond the end of a node
2022-02-21 15:04:42 linux-kernel [PATCH v2 0/6] More follow-up on high-order PCP caching
2022-02-21 11:41:24 linux-kernel [PATCH 1/1] mm/page_alloc: Do not prefetch buddies during bulk free
2022-02-18 12:58:05 linux-kernel Re: [PATCH 5/6] mm/page_alloc: Free pages in a single pass during bulk free
2022-02-18 11:00:40 linux-kernel Re: [PATCH 5/6] mm/page_alloc: Free pages in a single pass during bulk free
2022-02-17 19:50:09 linux-kernel Re: [PATCH 5/6] mm/page_alloc: Free pages in a single pass during bulk free
2022-02-17 16:06:40 linux-kernel Re: [PATCH v4] sched/fair: Consider cpu affinity when allowing NUMA imbalance in find_idlest_group
2022-02-17 13:56:16 linux-kernel [PATCH 2/6] mm/page_alloc: Track range of active PCP lists during bulk free
2022-02-17 13:42:56 linux-kernel [PATCH 3/6] mm/page_alloc: Simplify how many pages are selected per pcp list during bulk free
2022-02-17 12:52:23 linux-kernel Re: [PATCH v4] sched/fair: Consider cpu affinity when allowing NUMA imbalance in find_idlest_group
2022-02-17 10:59:43 linux-kernel [PATCH 6/6] mm/page_alloc: Limit number of high-order pages on PCP during bulk free
2022-02-17 09:53:59 linux-kernel [PATCH 1/6] mm/page_alloc: Fetch the correct pcp buddy during bulk free
2022-02-17 08:00:16 linux-kernel [PATCH 4/6] mm/page_alloc: Drain the requested list first during bulk free
2022-02-17 06:09:41 linux-kernel [PATCH v2 0/6] Follow-up on high-order PCP caching
2022-02-17 02:33:29 linux-kernel [PATCH 5/6] mm/page_alloc: Free pages in a single pass during bulk free
2022-02-16 13:19:41 linux-kernel Re: [PATCH 2/5] mm/page_alloc: Track range of active PCP lists during bulk free
2022-02-16 06:30:00 linux-kernel Re: [PATCH v4 1/1] mm: vmscan: Reduce throttling due to a failure to make progress
2022-02-15 21:50:25 linux-kernel [PATCH 2/5] mm/page_alloc: Track range of active PCP lists during bulk free
2022-02-15 18:33:45 linux-kernel [PATCH 1/5] mm/page_alloc: Fetch the correct pcp buddy during bulk free
2022-02-15 17:37:33 linux-kernel [PATCH 5/5] mm/page_alloc: Limit number of high-order pages on PCP during bulk free
2022-02-15 15:54:21 linux-kernel [PATCH 4/5] mm/page_alloc: Free pages in a single pass during bulk free
2022-02-15 15:37:24 linux-kernel [PATCH 3/5] mm/page_alloc: Simplify how many pages are selected per pcp list during bulk free
2022-02-15 15:33:56 linux-kernel [PATCH 0/5] Follow-up on high-order PCP caching
2022-02-09 11:58:03 linux-kernel Re: [PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs
2022-02-09 10:04:54 linux-kernel [PATCH 1/2] sched/fair: Improve consistency of allowed NUMA balance calculations
2022-02-09 09:38:43 linux-kernel Re: [PATCH] sched/fair: Consider cpu affinity when allowing NUMA imbalance in find_idlest_group
2022-02-09 05:41:14 linux-kernel [PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs
2022-02-09 00:34:39 linux-kernel [PATCH v6 0/2] Adjust NUMA imbalance for multiple LLCs
2022-02-07 13:13:38 linux-kernel Re: [PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs
2022-02-05 08:18:10 linux-kernel Re: [PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs
2022-02-04 19:30:56 linux-kernel [PATCH 1/2] sched/fair: Improve consistency of allowed NUMA balance calculations
2022-02-03 20:33:49 linux-kernel [PATCH v5 0/2] Adjust NUMA imbalance for multiple LLCs