All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH 1/5] mm/slab_common: move kmem_cache definition to internal header
@ 2014-08-21  8:09 ` Joonsoo Kim
  0 siblings, 0 replies; 18+ messages in thread
From: Joonsoo Kim @ 2014-08-21  8:09 UTC (permalink / raw)
  To: Andrew Morton
  Cc: Christoph Lameter, Pekka Enberg, David Rientjes, linux-mm,
	linux-kernel, Joonsoo Kim

We don't need to keep kmem_cache definition in include/linux/slab.h
if we don't need to inline kmem_cache_size(). According to my
code inspection, this function is only called at lc_create() in
lib/lru_cache.c which may be called at initialization phase of something,
so we don't need to inline it. Therfore, move it to slab_common.c and
move kmem_cache definition to internal header.

After this change, we can change kmem_cache definition easily without
full kernel build. For instance, we can turn on/off CONFIG_SLUB_STATS
without full kernel build.

Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
---
 include/linux/slab.h |   42 +-----------------------------------------
 mm/slab.h            |   33 +++++++++++++++++++++++++++++++++
 mm/slab_common.c     |    8 ++++++++
 3 files changed, 42 insertions(+), 41 deletions(-)

diff --git a/include/linux/slab.h b/include/linux/slab.h
index 1d9abb7..9062e4a 100644
--- a/include/linux/slab.h
+++ b/include/linux/slab.h
@@ -158,31 +158,6 @@ size_t ksize(const void *);
 #define ARCH_KMALLOC_MINALIGN __alignof__(unsigned long long)
 #endif
 
-#ifdef CONFIG_SLOB
-/*
- * Common fields provided in kmem_cache by all slab allocators
- * This struct is either used directly by the allocator (SLOB)
- * or the allocator must include definitions for all fields
- * provided in kmem_cache_common in their definition of kmem_cache.
- *
- * Once we can do anonymous structs (C11 standard) we could put a
- * anonymous struct definition in these allocators so that the
- * separate allocations in the kmem_cache structure of SLAB and
- * SLUB is no longer needed.
- */
-struct kmem_cache {
-	unsigned int object_size;/* The original size of the object */
-	unsigned int size;	/* The aligned/padded/added on size  */
-	unsigned int align;	/* Alignment as calculated */
-	unsigned long flags;	/* Active flags on the slab */
-	const char *name;	/* Slab name for sysfs */
-	int refcount;		/* Use counter */
-	void (*ctor)(void *);	/* Called on object slot creation */
-	struct list_head list;	/* List of all slab caches on the system */
-};
-
-#endif /* CONFIG_SLOB */
-
 /*
  * Kmalloc array related definitions
  */
@@ -363,14 +338,6 @@ kmem_cache_alloc_node_trace(struct kmem_cache *s,
 }
 #endif /* CONFIG_TRACING */
 
-#ifdef CONFIG_SLAB
-#include <linux/slab_def.h>
-#endif
-
-#ifdef CONFIG_SLUB
-#include <linux/slub_def.h>
-#endif
-
 extern void *kmalloc_order(size_t size, gfp_t flags, unsigned int order);
 
 #ifdef CONFIG_TRACING
@@ -650,14 +617,7 @@ static inline void *kzalloc_node(size_t size, gfp_t flags, int node)
 	return kmalloc_node(size, flags | __GFP_ZERO, node);
 }
 
-/*
- * Determine the size of a slab object
- */
-static inline unsigned int kmem_cache_size(struct kmem_cache *s)
-{
-	return s->object_size;
-}
-
+unsigned int kmem_cache_size(struct kmem_cache *s);
 void __init kmem_cache_init_late(void);
 
 #endif	/* _LINUX_SLAB_H */
diff --git a/mm/slab.h b/mm/slab.h
index 0e0fdd3..bd1c54a 100644
--- a/mm/slab.h
+++ b/mm/slab.h
@@ -4,6 +4,39 @@
  * Internal slab definitions
  */
 
+#ifdef CONFIG_SLOB
+/*
+ * Common fields provided in kmem_cache by all slab allocators
+ * This struct is either used directly by the allocator (SLOB)
+ * or the allocator must include definitions for all fields
+ * provided in kmem_cache_common in their definition of kmem_cache.
+ *
+ * Once we can do anonymous structs (C11 standard) we could put a
+ * anonymous struct definition in these allocators so that the
+ * separate allocations in the kmem_cache structure of SLAB and
+ * SLUB is no longer needed.
+ */
+struct kmem_cache {
+	unsigned int object_size;/* The original size of the object */
+	unsigned int size;	/* The aligned/padded/added on size  */
+	unsigned int align;	/* Alignment as calculated */
+	unsigned long flags;	/* Active flags on the slab */
+	const char *name;	/* Slab name for sysfs */
+	int refcount;		/* Use counter */
+	void (*ctor)(void *);	/* Called on object slot creation */
+	struct list_head list;	/* List of all slab caches on the system */
+};
+
+#endif /* CONFIG_SLOB */
+
+#ifdef CONFIG_SLAB
+#include <linux/slab_def.h>
+#endif
+
+#ifdef CONFIG_SLUB
+#include <linux/slub_def.h>
+#endif
+
 /*
  * State of the slab allocator.
  *
diff --git a/mm/slab_common.c b/mm/slab_common.c
index d319502..2088904 100644
--- a/mm/slab_common.c
+++ b/mm/slab_common.c
@@ -30,6 +30,14 @@ LIST_HEAD(slab_caches);
 DEFINE_MUTEX(slab_mutex);
 struct kmem_cache *kmem_cache;
 
+/*
+ * Determine the size of a slab object
+ */
+unsigned int kmem_cache_size(struct kmem_cache *s)
+{
+	return s->object_size;
+}
+
 #ifdef CONFIG_DEBUG_VM
 static int kmem_cache_sanity_check(const char *name, size_t size)
 {
-- 
1.7.9.5


^ permalink raw reply related	[flat|nested] 18+ messages in thread

end of thread, other threads:[~2014-08-25  8:29 UTC | newest]

Thread overview: 18+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2014-08-21  8:09 [PATCH 1/5] mm/slab_common: move kmem_cache definition to internal header Joonsoo Kim
2014-08-21  8:09 ` Joonsoo Kim
2014-08-21  8:09 ` [PATCH 2/5] mm/sl[ao]b: always track caller in kmalloc_(node_)track_caller() Joonsoo Kim
2014-08-21  8:09   ` Joonsoo Kim
2014-08-21 14:15   ` Christoph Lameter
2014-08-21 14:15     ` Christoph Lameter
2014-08-21  8:09 ` [PATCH 3/5] mm/slab: move cache_flusharray() out of unlikely.text section Joonsoo Kim
2014-08-21  8:09   ` Joonsoo Kim
2014-08-21  8:09 ` [PATCH 4/5] mm/slab: noinline __ac_put_obj() Joonsoo Kim
2014-08-21  8:09   ` Joonsoo Kim
2014-08-21  8:09 ` [PATCH 5/5] mm/slab: factor out unlikely part of cache_free_alien() Joonsoo Kim
2014-08-21  8:09   ` Joonsoo Kim
2014-08-21  8:27 ` [PATCH 1/5] mm/slab_common: move kmem_cache definition to internal header Zhang Yanfei
2014-08-21  8:27   ` Zhang Yanfei
2014-08-25  8:29   ` Joonsoo Kim
2014-08-25  8:29     ` Joonsoo Kim
2014-08-21 14:14 ` Christoph Lameter
2014-08-21 14:14   ` Christoph Lameter

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.