mm, slab: remove the corner case of inc_slabs_node()

Message ID 20240222130233.2880176-1-chengming.zhou@linux.dev
State New
Headers
Series mm, slab: remove the corner case of inc_slabs_node() |

Commit Message

Chengming Zhou Feb. 22, 2024, 1:02 p.m. UTC
  From: Chengming Zhou <chengming.zhou@linux.dev>

We already have the inc_slabs_node() after kmem_cache_node->node[node]
initialized in early_kmem_cache_node_alloc(), this special case of
inc_slabs_node() can be removed. Then we don't need to consider the
existence of kmem_cache_node in inc_slabs_node() anymore.

Signed-off-by: Chengming Zhou <chengming.zhou@linux.dev>
---
 mm/slub.c | 13 ++-----------
 1 file changed, 2 insertions(+), 11 deletions(-)
  

Comments

Vlastimil Babka March 1, 2024, 4:03 p.m. UTC | #1
On 2/22/24 14:02, chengming.zhou@linux.dev wrote:
> From: Chengming Zhou <chengming.zhou@linux.dev>
> 
> We already have the inc_slabs_node() after kmem_cache_node->node[node]
> initialized in early_kmem_cache_node_alloc(), this special case of
> inc_slabs_node() can be removed. Then we don't need to consider the
> existence of kmem_cache_node in inc_slabs_node() anymore.
> 
> Signed-off-by: Chengming Zhou <chengming.zhou@linux.dev>

Well spotted, thank. Added to slab/for-next.

> ---
>  mm/slub.c | 13 ++-----------
>  1 file changed, 2 insertions(+), 11 deletions(-)
> 
> diff --git a/mm/slub.c b/mm/slub.c
> index 284b751b3b64..3f413e5e1415 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -1500,16 +1500,8 @@ static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects)
>  {
>  	struct kmem_cache_node *n = get_node(s, node);
>  
> -	/*
> -	 * May be called early in order to allocate a slab for the
> -	 * kmem_cache_node structure. Solve the chicken-egg
> -	 * dilemma by deferring the increment of the count during
> -	 * bootstrap (see early_kmem_cache_node_alloc).
> -	 */
> -	if (likely(n)) {
> -		atomic_long_inc(&n->nr_slabs);
> -		atomic_long_add(objects, &n->total_objects);
> -	}
> +	atomic_long_inc(&n->nr_slabs);
> +	atomic_long_add(objects, &n->total_objects);
>  }
>  static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects)
>  {
> @@ -4877,7 +4869,6 @@ static void early_kmem_cache_node_alloc(int node)
>  	slab = new_slab(kmem_cache_node, GFP_NOWAIT, node);
>  
>  	BUG_ON(!slab);
> -	inc_slabs_node(kmem_cache_node, slab_nid(slab), slab->objects);
>  	if (slab_nid(slab) != node) {
>  		pr_err("SLUB: Unable to allocate memory from node %d\n", node);
>  		pr_err("SLUB: Allocating a useless per node structure in order to be able to continue\n");
  

Patch

diff --git a/mm/slub.c b/mm/slub.c
index 284b751b3b64..3f413e5e1415 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -1500,16 +1500,8 @@  static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects)
 {
 	struct kmem_cache_node *n = get_node(s, node);
 
-	/*
-	 * May be called early in order to allocate a slab for the
-	 * kmem_cache_node structure. Solve the chicken-egg
-	 * dilemma by deferring the increment of the count during
-	 * bootstrap (see early_kmem_cache_node_alloc).
-	 */
-	if (likely(n)) {
-		atomic_long_inc(&n->nr_slabs);
-		atomic_long_add(objects, &n->total_objects);
-	}
+	atomic_long_inc(&n->nr_slabs);
+	atomic_long_add(objects, &n->total_objects);
 }
 static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects)
 {
@@ -4877,7 +4869,6 @@  static void early_kmem_cache_node_alloc(int node)
 	slab = new_slab(kmem_cache_node, GFP_NOWAIT, node);
 
 	BUG_ON(!slab);
-	inc_slabs_node(kmem_cache_node, slab_nid(slab), slab->objects);
 	if (slab_nid(slab) != node) {
 		pr_err("SLUB: Unable to allocate memory from node %d\n", node);
 		pr_err("SLUB: Allocating a useless per node structure in order to be able to continue\n");