slab 1352 fs/bio.c *bvp = mempool_create_slab_pool(pool_entries, bp->slab);
slab 1414 fs/bio.c bvs->slab = kmem_cache_create(bvs->name, size, 0,
slab 930 fs/nfsd/nfs4state.c if (*slab == NULL)
slab 932 fs/nfsd/nfs4state.c kmem_cache_destroy(*slab);
slab 933 fs/nfsd/nfs4state.c *slab = NULL;
slab 391 include/linux/bio.h struct kmem_cache *slab;
slab 493 include/linux/i2o.h struct kmem_cache *slab;
slab 942 include/linux/i2o.h pool->slab =
slab 944 include/linux/i2o.h if (!pool->slab)
slab 947 include/linux/i2o.h pool->mempool = mempool_create_slab_pool(min_nr, pool->slab);
slab 954 include/linux/i2o.h kmem_cache_destroy(pool->slab);
slab 973 include/linux/i2o.h kmem_cache_destroy(pool->slab);
slab 73 include/linux/mm_types.h struct kmem_cache *slab; /* SLUB: Pointer to slab */
slab 33 include/net/request_sock.h struct kmem_cache *slab;
slab 63 include/net/request_sock.h struct request_sock *req = kmem_cache_alloc(ops->slab, GFP_ATOMIC);
slab 73 include/net/request_sock.h kmem_cache_free(req->rsk_ops->slab, req);
slab 590 include/net/sock.h struct kmem_cache *slab;
slab 364 mm/slab.c list_splice(&(cachep->nodelists[nodeid]->slab), listp); \
slab 596 mm/slab.c page->lru.prev = (struct list_head *)slab;
slab 599 mm/slab.c static inline struct slab *page_get_slab(struct page *page)
slab 602 mm/slab.c return (struct slab *)page->lru.prev;
slab 611 mm/slab.c static inline struct slab *virt_to_slab(const void *obj)
slab 620 mm/slab.c return slab->s_mem + cache->buffer_size * idx;
slab 632 mm/slab.c u32 offset = (obj - slab->s_mem);
slab 800 mm/slab.c return ALIGN(sizeof(struct slab)+nr_objs*sizeof(kmem_bufctl_t), align);
slab 844 mm/slab.c nr_objs = (slab_size - sizeof(struct slab)) /
slab 1122 mm/slab.c struct slab *slabp = virt_to_slab(objp);
slab 1508 mm/slab.c sizeof(struct slab), cache_line_size());
slab 1881 mm/slab.c struct slab *slabp = virt_to_slab(objp);
slab 2021 mm/slab.c offslab_limit = size - sizeof(struct slab);
slab 2329 mm/slab.c + sizeof(struct slab), align);
slab 2343 mm/slab.c cachep->num * sizeof(kmem_bufctl_t) + sizeof(struct slab);
slab 2474 mm/slab.c struct slab *slabp;
slab 2486 mm/slab.c slabp = list_entry(p, struct slab, list);
slab 2602 mm/slab.c static struct slab *alloc_slabmgmt(struct kmem_cache *cachep, void *objp,
slab 2606 mm/slab.c struct slab *slabp;
slab 2744 mm/slab.c page_set_slab(page, slab);
slab 2756 mm/slab.c struct slab *slabp;
slab 2874 mm/slab.c struct slab *slabp;
slab 2980 mm/slab.c struct slab *slabp;
slab 2990 mm/slab.c slabp = list_entry(entry, struct slab, list);
slab 3085 mm/slab.c struct slab *slabp;
slab 3309 mm/slab.c struct slab *slabp;
slab 3328 mm/slab.c slabp = list_entry(entry, struct slab, list);
slab 3486 mm/slab.c struct slab *slabp;
slab 3557 mm/slab.c struct slab *slabp;
slab 3559 mm/slab.c slabp = list_entry(p, struct slab, list);
slab 4154 mm/slab.c struct slab *slabp;
slab 4389 mm/slab.c struct slab *slabp;
slab 926 mm/slub.c if (unlikely(s != page->slab)) {
slab 930 mm/slub.c } else if (!page->slab) {
slab 1124 mm/slub.c page->slab = s;
slab 1180 mm/slub.c __free_slab(page->slab, page);
slab 2342 mm/slub.c if (!page || s != page->slab)
slab 2708 mm/slub.c s = page->slab;
slab 2746 mm/slub.c slab_free(page->slab, page, object, __builtin_return_address(0));
slab 888 net/core/sock.c struct kmem_cache *slab;
slab 890 net/core/sock.c slab = prot->slab;
slab 891 net/core/sock.c if (slab != NULL)
slab 892 net/core/sock.c sk = kmem_cache_alloc(slab, priority);
slab 909 net/core/sock.c if (slab != NULL)
slab 910 net/core/sock.c kmem_cache_free(slab, sk);
slab 918 net/core/sock.c struct kmem_cache *slab;
slab 922 net/core/sock.c slab = prot->slab;
slab 925 net/core/sock.c if (slab != NULL)
slab 926 net/core/sock.c kmem_cache_free(slab, sk);
slab 2044 net/core/sock.c prot->slab = kmem_cache_create(prot->name, prot->obj_size, 0,
slab 2047 net/core/sock.c if (prot->slab == NULL) {
slab 2061 net/core/sock.c prot->rsk_prot->slab = kmem_cache_create(request_sock_slab_name,
slab 2065 net/core/sock.c if (prot->rsk_prot->slab == NULL) {
slab 2100 net/core/sock.c if (prot->rsk_prot && prot->rsk_prot->slab) {
slab 2101 net/core/sock.c kmem_cache_destroy(prot->rsk_prot->slab);
slab 2102 net/core/sock.c prot->rsk_prot->slab = NULL;
slab 2107 net/core/sock.c kmem_cache_destroy(prot->slab);
slab 2108 net/core/sock.c prot->slab = NULL;
slab 2122 net/core/sock.c if (prot->slab != NULL) {
slab 2123 net/core/sock.c kmem_cache_destroy(prot->slab);
slab 2124 net/core/sock.c prot->slab = NULL;
slab 2127 net/core/sock.c if (prot->rsk_prot != NULL && prot->rsk_prot->slab != NULL) {
slab 2128 net/core/sock.c const char *name = kmem_cache_name(prot->rsk_prot->slab);
slab 2130 net/core/sock.c kmem_cache_destroy(prot->rsk_prot->slab);
slab 2132 net/core/sock.c prot->rsk_prot->slab = NULL;
slab 2180 net/core/sock.c proto->slab == NULL ? "no" : "yes",
slab 61 net/dccp/ccid.c struct kmem_cache *slab;
slab 72 net/dccp/ccid.c slab = kmem_cache_create(slab_name, sizeof(struct ccid) + obj_size, 0,
slab 74 net/dccp/ccid.c if (slab == NULL)
slab 76 net/dccp/ccid.c return slab;
slab 81 net/dccp/ccid.c if (slab != NULL) {
slab 82 net/dccp/ccid.c const char *name = kmem_cache_name(slab);
slab 84 net/dccp/ccid.c kmem_cache_destroy(slab);
slab 341 net/ipv4/af_inet.c WARN_ON(answer_prot->slab == NULL);
slab 154 net/ipv6/af_inet6.c WARN_ON(answer_prot->slab == NULL);