2022-04-07 19:59:38

by Yixuan Cao

[permalink] [raw]
Subject: [PATCH] mm/slab.c: fix comments

While reading the source code,
I noticed some language errors in the comments, so I fixed them.

Signed-off-by: Yixuan Cao <[email protected]>
---
mm/slab.c | 12 ++++++------
1 file changed, 6 insertions(+), 6 deletions(-)

diff --git a/mm/slab.c b/mm/slab.c
index 90b16c7ae01a..e882657c1494 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -781,7 +781,7 @@ static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
int slab_node = slab_nid(virt_to_slab(objp));
int node = numa_mem_id();
/*
- * Make sure we are not freeing a object from another node to the array
+ * Make sure we are not freeing an object from another node to the array
* cache on this cpu.
*/
if (likely(node == slab_node))
@@ -832,7 +832,7 @@ static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp)

/*
* The kmem_cache_nodes don't come and go as CPUs
- * come and go. slab_mutex is sufficient
+ * come and go. slab_mutex provides sufficient
* protection here.
*/
cachep->node[node] = n;
@@ -845,7 +845,7 @@ static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp)
* Allocates and initializes node for a node on each slab cache, used for
* either memory or cpu hotplug. If memory is being hot-added, the kmem_cache_node
* will be allocated off-node since memory is not yet online for the new node.
- * When hotplugging memory or a cpu, existing node are not replaced if
+ * When hotplugging memory or a cpu, existing nodes are not replaced if
* already in use.
*
* Must hold slab_mutex.
@@ -1046,7 +1046,7 @@ int slab_prepare_cpu(unsigned int cpu)
* offline.
*
* Even if all the cpus of a node are down, we don't free the
- * kmem_cache_node of any cache. This to avoid a race between cpu_down, and
+ * kmem_cache_node of any cache. This is to avoid a race between cpu_down, and
* a kmalloc allocation from another cpu for memory from the node of
* the cpu going down. The kmem_cache_node structure is usually allocated from
* kmem_cache_create() and gets destroyed at kmem_cache_destroy().
@@ -1890,7 +1890,7 @@ static bool set_on_slab_cache(struct kmem_cache *cachep,
* @flags: SLAB flags
*
* Returns a ptr to the cache on success, NULL on failure.
- * Cannot be called within a int, but can be interrupted.
+ * Cannot be called within an int, but can be interrupted.
* The @ctor is run when new pages are allocated by the cache.
*
* The flags are
@@ -3138,7 +3138,7 @@ static void *fallback_alloc(struct kmem_cache *cache, gfp_t flags)
}

/*
- * A interface to enable slab creation on nodeid
+ * An interface to enable slab creation on nodeid
*/
static void *____cache_alloc_node(struct kmem_cache *cachep, gfp_t flags,
int nodeid)
--
2.17.1




2022-04-07 20:59:16

by Hyeonggon Yoo

[permalink] [raw]
Subject: Re: [PATCH] mm/slab.c: fix comments

On Thu, Apr 07, 2022 at 04:09:58PM +0800, Yixuan Cao wrote:
> While reading the source code,
> I noticed some language errors in the comments, so I fixed them.
>
> Signed-off-by: Yixuan Cao <[email protected]>
> ---
> mm/slab.c | 12 ++++++------
> 1 file changed, 6 insertions(+), 6 deletions(-)
>
> diff --git a/mm/slab.c b/mm/slab.c
> index 90b16c7ae01a..e882657c1494 100644
> --- a/mm/slab.c
> +++ b/mm/slab.c
> @@ -781,7 +781,7 @@ static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
> int slab_node = slab_nid(virt_to_slab(objp));
> int node = numa_mem_id();
> /*
> - * Make sure we are not freeing a object from another node to the array
> + * Make sure we are not freeing an object from another node to the array
> * cache on this cpu.
> */
> if (likely(node == slab_node))
> @@ -832,7 +832,7 @@ static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp)
>
> /*
> * The kmem_cache_nodes don't come and go as CPUs
> - * come and go. slab_mutex is sufficient
> + * come and go. slab_mutex provides sufficient
> * protection here.
> */
> cachep->node[node] = n;
> @@ -845,7 +845,7 @@ static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp)
> * Allocates and initializes node for a node on each slab cache, used for
> * either memory or cpu hotplug. If memory is being hot-added, the kmem_cache_node
> * will be allocated off-node since memory is not yet online for the new node.
> - * When hotplugging memory or a cpu, existing node are not replaced if
> + * When hotplugging memory or a cpu, existing nodes are not replaced if
> * already in use.
> *
> * Must hold slab_mutex.
> @@ -1046,7 +1046,7 @@ int slab_prepare_cpu(unsigned int cpu)
> * offline.
> *
> * Even if all the cpus of a node are down, we don't free the
> - * kmem_cache_node of any cache. This to avoid a race between cpu_down, and
> + * kmem_cache_node of any cache. This is to avoid a race between cpu_down, and
> * a kmalloc allocation from another cpu for memory from the node of
> * the cpu going down. The kmem_cache_node structure is usually allocated from
> * kmem_cache_create() and gets destroyed at kmem_cache_destroy().
> @@ -1890,7 +1890,7 @@ static bool set_on_slab_cache(struct kmem_cache *cachep,
> * @flags: SLAB flags
> *
> * Returns a ptr to the cache on success, NULL on failure.
> - * Cannot be called within a int, but can be interrupted.
> + * Cannot be called within an int, but can be interrupted.
> * The @ctor is run when new pages are allocated by the cache.
> *
> * The flags are
> @@ -3138,7 +3138,7 @@ static void *fallback_alloc(struct kmem_cache *cache, gfp_t flags)
> }
>
> /*
> - * A interface to enable slab creation on nodeid
> + * An interface to enable slab creation on nodeid
> */
> static void *____cache_alloc_node(struct kmem_cache *cachep, gfp_t flags,
> int nodeid)
> --
> 2.17.1
>
>
>
>

Acked-by: Hyeonggon Yoo <[email protected]>

Thanks!

--
Thanks,
Hyeonggon