Subversion Repositories shark

Rev

Rev 422 | Blame | Compare with Previous | Last modification | View Log | RSS feed

/*
 * linux/mm/slab.h
 * Written by Mark Hemment, 1996.
 * (markhe@nextd.demon.co.uk)
 */


#if     !defined(_LINUX_SLAB_H)
#define _LINUX_SLAB_H

#if     defined(__KERNEL__)

typedef struct kmem_cache_s kmem_cache_t;

#include        <linux/config.h>        /* kmalloc_sizes.h needs CONFIG_ options */
#include        <linux/gfp.h>
#include        <linux/types.h>
#include        <asm/page.h>            /* kmalloc_sizes.h needs PAGE_SIZE */
#include        <asm/cache.h>           /* kmalloc_sizes.h needs L1_CACHE_BYTES */

/* flags for kmem_cache_alloc() */
#define SLAB_NOFS               GFP_NOFS
#define SLAB_NOIO               GFP_NOIO
#define SLAB_ATOMIC             GFP_ATOMIC
#define SLAB_USER               GFP_USER
#define SLAB_KERNEL             GFP_KERNEL
#define SLAB_DMA                GFP_DMA

#define SLAB_LEVEL_MASK         (__GFP_WAIT|__GFP_HIGH|__GFP_IO|__GFP_FS|\
                                __GFP_COLD|__GFP_NOWARN|__GFP_REPEAT|\
                                __GFP_NOFAIL|__GFP_NORETRY)


#define SLAB_NO_GROW            __GFP_NO_GROW   /* don't grow a cache */

/* flags to pass to kmem_cache_create().
 * The first 3 are only valid when the allocator as been build
 * SLAB_DEBUG_SUPPORT.
 */

#define SLAB_DEBUG_FREE         0x00000100UL    /* Peform (expensive) checks on free */
#define SLAB_DEBUG_INITIAL      0x00000200UL    /* Call constructor (as verifier) */
#define SLAB_RED_ZONE           0x00000400UL    /* Red zone objs in a cache */
#define SLAB_POISON             0x00000800UL    /* Poison objects */
#define SLAB_NO_REAP            0x00001000UL    /* never reap from the cache */
#define SLAB_HWCACHE_ALIGN      0x00002000UL    /* align objs on a h/w cache lines */
#define SLAB_CACHE_DMA          0x00004000UL    /* use GFP_DMA memory */
#define SLAB_MUST_HWCACHE_ALIGN 0x00008000UL    /* force alignment */
#define SLAB_STORE_USER         0x00010000UL    /* store the last owner for bug hunting */
#define SLAB_RECLAIM_ACCOUNT    0x00020000UL    /* track pages allocated to indicate
                                                   what is reclaimable later*/

/* flags passed to a constructor func */
#define SLAB_CTOR_CONSTRUCTOR   0x001UL         /* if not set, then deconstructor */
#define SLAB_CTOR_ATOMIC        0x002UL         /* tell constructor it can't sleep */
#define SLAB_CTOR_VERIFY        0x004UL         /* tell constructor it's a verify call */

/* prototypes */
extern void kmem_cache_init(void);

extern kmem_cache_t *kmem_find_general_cachep(size_t, int gfpflags);
extern kmem_cache_t *kmem_cache_create(const char *, size_t, size_t, unsigned long,
                                       void (*)(void *, kmem_cache_t *, unsigned long),
                                       void (*)(void *, kmem_cache_t *, unsigned long));
extern int kmem_cache_destroy(kmem_cache_t *);
extern int kmem_cache_shrink(kmem_cache_t *);
extern void *kmem_cache_alloc(kmem_cache_t *, int);
extern void kmem_cache_free(kmem_cache_t *, void *);
extern unsigned int kmem_cache_size(kmem_cache_t *);

/* Size description struct for general caches. */
struct cache_sizes {
        size_t           cs_size;
        kmem_cache_t    *cs_cachep;
        kmem_cache_t    *cs_dmacachep;
};
extern struct cache_sizes malloc_sizes[];
extern void *__kmalloc(size_t, int);

static inline void *kmalloc(size_t size, int flags)
{
        return __kmalloc(size, flags);
}

extern void kfree(const void *);
extern unsigned int ksize(const void *);

extern int FASTCALL(kmem_cache_reap(int));

/* System wide caches */
extern kmem_cache_t     *vm_area_cachep;
extern kmem_cache_t     *mm_cachep;
extern kmem_cache_t     *names_cachep;
extern kmem_cache_t     *files_cachep;
extern kmem_cache_t     *filp_cachep;
extern kmem_cache_t     *dquot_cachep;
extern kmem_cache_t     *fs_cachep;
extern kmem_cache_t     *signal_cachep;
extern kmem_cache_t     *sighand_cachep;
extern kmem_cache_t     *bio_cachep;

void ptrinfo(unsigned long addr);

extern atomic_t slab_reclaim_pages;

#endif  /* __KERNEL__ */

#endif  /* _LINUX_SLAB_H */