Commit aa137f9d authored by Christoph Lameter's avatar Christoph Lameter Committed by Linus Torvalds

SLUB: Force inlining for functions in slub_def.h

Some compilers (especially older gcc releases) may skip inlining
sometimes which will lead to link failures.  Force the inlining of
keyfunctions in slub_def.h to avoid these issues.
Signed-off-by: default avatarChristoph Lameter <clameter@sgi.com>
Acked-by: default avatarJan Dittmer <jdi@l4x.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent d66ac475
...@@ -78,7 +78,7 @@ extern struct kmem_cache kmalloc_caches[KMALLOC_SHIFT_HIGH + 1]; ...@@ -78,7 +78,7 @@ extern struct kmem_cache kmalloc_caches[KMALLOC_SHIFT_HIGH + 1];
* Sorry that the following has to be that ugly but some versions of GCC * Sorry that the following has to be that ugly but some versions of GCC
* have trouble with constant propagation and loops. * have trouble with constant propagation and loops.
*/ */
static inline int kmalloc_index(size_t size) static __always_inline int kmalloc_index(size_t size)
{ {
if (!size) if (!size)
return 0; return 0;
...@@ -133,7 +133,7 @@ static inline int kmalloc_index(size_t size) ...@@ -133,7 +133,7 @@ static inline int kmalloc_index(size_t size)
* This ought to end up with a global pointer to the right cache * This ought to end up with a global pointer to the right cache
* in kmalloc_caches. * in kmalloc_caches.
*/ */
static inline struct kmem_cache *kmalloc_slab(size_t size) static __always_inline struct kmem_cache *kmalloc_slab(size_t size)
{ {
int index = kmalloc_index(size); int index = kmalloc_index(size);
...@@ -166,7 +166,7 @@ static inline struct kmem_cache *kmalloc_slab(size_t size) ...@@ -166,7 +166,7 @@ static inline struct kmem_cache *kmalloc_slab(size_t size)
void *kmem_cache_alloc(struct kmem_cache *, gfp_t); void *kmem_cache_alloc(struct kmem_cache *, gfp_t);
void *__kmalloc(size_t size, gfp_t flags); void *__kmalloc(size_t size, gfp_t flags);
static inline void *kmalloc(size_t size, gfp_t flags) static __always_inline void *kmalloc(size_t size, gfp_t flags)
{ {
if (__builtin_constant_p(size) && !(flags & SLUB_DMA)) { if (__builtin_constant_p(size) && !(flags & SLUB_DMA)) {
struct kmem_cache *s = kmalloc_slab(size); struct kmem_cache *s = kmalloc_slab(size);
...@@ -183,7 +183,7 @@ static inline void *kmalloc(size_t size, gfp_t flags) ...@@ -183,7 +183,7 @@ static inline void *kmalloc(size_t size, gfp_t flags)
void *__kmalloc_node(size_t size, gfp_t flags, int node); void *__kmalloc_node(size_t size, gfp_t flags, int node);
void *kmem_cache_alloc_node(struct kmem_cache *, gfp_t flags, int node); void *kmem_cache_alloc_node(struct kmem_cache *, gfp_t flags, int node);
static inline void *kmalloc_node(size_t size, gfp_t flags, int node) static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node)
{ {
if (__builtin_constant_p(size) && !(flags & SLUB_DMA)) { if (__builtin_constant_p(size) && !(flags & SLUB_DMA)) {
struct kmem_cache *s = kmalloc_slab(size); struct kmem_cache *s = kmalloc_slab(size);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment