- page = allocate_slab(s, flags & GFP_LEVEL_MASK, node);
+ page = allocate_slab(s,
+ flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node);
struct kmem_cache_node *n = s->node[node];
if (n && n != &s->local_node)
kmem_cache_free(kmalloc_caches, n);
struct kmem_cache_node *n = s->node[node];
if (n && n != &s->local_node)
kmem_cache_free(kmalloc_caches, n);
struct kmem_cache_node *n = get_node(s, node);
n->nr_partial -= free_list(s, n, &n->partial);
struct kmem_cache_node *n = get_node(s, node);
n->nr_partial -= free_list(s, n, &n->partial);
struct kmem_cache_node *n = get_node(s, node);
count += validate_slab_node(s, n, map);
struct kmem_cache_node *n = get_node(s, node);
count += validate_slab_node(s, n, map);
struct kmem_cache_node *n = get_node(s, node);
if (flags & SO_PARTIAL) {
struct kmem_cache_node *n = get_node(s, node);
if (flags & SO_PARTIAL) {