slab 1352 fs/bio.c *bvp = mempool_create_slab_pool(pool_entries, bp->slab); slab 1414 fs/bio.c bvs->slab = kmem_cache_create(bvs->name, size, 0, slab 930 fs/nfsd/nfs4state.c if (*slab == NULL) slab 932 fs/nfsd/nfs4state.c kmem_cache_destroy(*slab); slab 933 fs/nfsd/nfs4state.c *slab = NULL; slab 391 include/linux/bio.h struct kmem_cache *slab; slab 493 include/linux/i2o.h struct kmem_cache *slab; slab 942 include/linux/i2o.h pool->slab = slab 944 include/linux/i2o.h if (!pool->slab) slab 947 include/linux/i2o.h pool->mempool = mempool_create_slab_pool(min_nr, pool->slab); slab 954 include/linux/i2o.h kmem_cache_destroy(pool->slab); slab 973 include/linux/i2o.h kmem_cache_destroy(pool->slab); slab 73 include/linux/mm_types.h struct kmem_cache *slab; /* SLUB: Pointer to slab */ slab 33 include/net/request_sock.h struct kmem_cache *slab; slab 63 include/net/request_sock.h struct request_sock *req = kmem_cache_alloc(ops->slab, GFP_ATOMIC); slab 73 include/net/request_sock.h kmem_cache_free(req->rsk_ops->slab, req); slab 590 include/net/sock.h struct kmem_cache *slab; slab 364 mm/slab.c list_splice(&(cachep->nodelists[nodeid]->slab), listp); \ slab 596 mm/slab.c page->lru.prev = (struct list_head *)slab; slab 599 mm/slab.c static inline struct slab *page_get_slab(struct page *page) slab 602 mm/slab.c return (struct slab *)page->lru.prev; slab 611 mm/slab.c static inline struct slab *virt_to_slab(const void *obj) slab 620 mm/slab.c return slab->s_mem + cache->buffer_size * idx; slab 632 mm/slab.c u32 offset = (obj - slab->s_mem); slab 800 mm/slab.c return ALIGN(sizeof(struct slab)+nr_objs*sizeof(kmem_bufctl_t), align); slab 844 mm/slab.c nr_objs = (slab_size - sizeof(struct slab)) / slab 1122 mm/slab.c struct slab *slabp = virt_to_slab(objp); slab 1508 mm/slab.c sizeof(struct slab), cache_line_size()); slab 1881 mm/slab.c struct slab *slabp = virt_to_slab(objp); slab 2021 mm/slab.c offslab_limit = size - sizeof(struct slab); slab 2329 mm/slab.c + sizeof(struct slab), align); slab 2343 mm/slab.c cachep->num * sizeof(kmem_bufctl_t) + sizeof(struct slab); slab 2474 mm/slab.c struct slab *slabp; slab 2486 mm/slab.c slabp = list_entry(p, struct slab, list); slab 2602 mm/slab.c static struct slab *alloc_slabmgmt(struct kmem_cache *cachep, void *objp, slab 2606 mm/slab.c struct slab *slabp; slab 2744 mm/slab.c page_set_slab(page, slab); slab 2756 mm/slab.c struct slab *slabp; slab 2874 mm/slab.c struct slab *slabp; slab 2980 mm/slab.c struct slab *slabp; slab 2990 mm/slab.c slabp = list_entry(entry, struct slab, list); slab 3085 mm/slab.c struct slab *slabp; slab 3309 mm/slab.c struct slab *slabp; slab 3328 mm/slab.c slabp = list_entry(entry, struct slab, list); slab 3486 mm/slab.c struct slab *slabp; slab 3557 mm/slab.c struct slab *slabp; slab 3559 mm/slab.c slabp = list_entry(p, struct slab, list); slab 4154 mm/slab.c struct slab *slabp; slab 4389 mm/slab.c struct slab *slabp; slab 926 mm/slub.c if (unlikely(s != page->slab)) { slab 930 mm/slub.c } else if (!page->slab) { slab 1124 mm/slub.c page->slab = s; slab 1180 mm/slub.c __free_slab(page->slab, page); slab 2342 mm/slub.c if (!page || s != page->slab) slab 2708 mm/slub.c s = page->slab; slab 2746 mm/slub.c slab_free(page->slab, page, object, __builtin_return_address(0)); slab 888 net/core/sock.c struct kmem_cache *slab; slab 890 net/core/sock.c slab = prot->slab; slab 891 net/core/sock.c if (slab != NULL) slab 892 net/core/sock.c sk = kmem_cache_alloc(slab, priority); slab 909 net/core/sock.c if (slab != NULL) slab 910 net/core/sock.c kmem_cache_free(slab, sk); slab 918 net/core/sock.c struct kmem_cache *slab; slab 922 net/core/sock.c slab = prot->slab; slab 925 net/core/sock.c if (slab != NULL) slab 926 net/core/sock.c kmem_cache_free(slab, sk); slab 2044 net/core/sock.c prot->slab = kmem_cache_create(prot->name, prot->obj_size, 0, slab 2047 net/core/sock.c if (prot->slab == NULL) { slab 2061 net/core/sock.c prot->rsk_prot->slab = kmem_cache_create(request_sock_slab_name, slab 2065 net/core/sock.c if (prot->rsk_prot->slab == NULL) { slab 2100 net/core/sock.c if (prot->rsk_prot && prot->rsk_prot->slab) { slab 2101 net/core/sock.c kmem_cache_destroy(prot->rsk_prot->slab); slab 2102 net/core/sock.c prot->rsk_prot->slab = NULL; slab 2107 net/core/sock.c kmem_cache_destroy(prot->slab); slab 2108 net/core/sock.c prot->slab = NULL; slab 2122 net/core/sock.c if (prot->slab != NULL) { slab 2123 net/core/sock.c kmem_cache_destroy(prot->slab); slab 2124 net/core/sock.c prot->slab = NULL; slab 2127 net/core/sock.c if (prot->rsk_prot != NULL && prot->rsk_prot->slab != NULL) { slab 2128 net/core/sock.c const char *name = kmem_cache_name(prot->rsk_prot->slab); slab 2130 net/core/sock.c kmem_cache_destroy(prot->rsk_prot->slab); slab 2132 net/core/sock.c prot->rsk_prot->slab = NULL; slab 2180 net/core/sock.c proto->slab == NULL ? "no" : "yes", slab 61 net/dccp/ccid.c struct kmem_cache *slab; slab 72 net/dccp/ccid.c slab = kmem_cache_create(slab_name, sizeof(struct ccid) + obj_size, 0, slab 74 net/dccp/ccid.c if (slab == NULL) slab 76 net/dccp/ccid.c return slab; slab 81 net/dccp/ccid.c if (slab != NULL) { slab 82 net/dccp/ccid.c const char *name = kmem_cache_name(slab); slab 84 net/dccp/ccid.c kmem_cache_destroy(slab); slab 341 net/ipv4/af_inet.c WARN_ON(answer_prot->slab == NULL); slab 154 net/ipv6/af_inet6.c WARN_ON(answer_prot->slab == NULL);