Merge tag 'memblock-v5.11-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/rppt...
[linux-2.6-microblaze.git] / kernel / bpf / stackmap.c
index 06065fa..aea96b6 100644 (file)
@@ -90,7 +90,6 @@ static struct bpf_map *stack_map_alloc(union bpf_attr *attr)
 {
        u32 value_size = attr->value_size;
        struct bpf_stack_map *smap;
-       struct bpf_map_memory mem;
        u64 cost, n_buckets;
        int err;
 
@@ -119,15 +118,9 @@ static struct bpf_map *stack_map_alloc(union bpf_attr *attr)
 
        cost = n_buckets * sizeof(struct stack_map_bucket *) + sizeof(*smap);
        cost += n_buckets * (value_size + sizeof(struct stack_map_bucket));
-       err = bpf_map_charge_init(&mem, cost);
-       if (err)
-               return ERR_PTR(err);
-
        smap = bpf_map_area_alloc(cost, bpf_map_attr_numa_node(attr));
-       if (!smap) {
-               bpf_map_charge_finish(&mem);
+       if (!smap)
                return ERR_PTR(-ENOMEM);
-       }
 
        bpf_map_init_from_attr(&smap->map, attr);
        smap->map.value_size = value_size;
@@ -135,20 +128,17 @@ static struct bpf_map *stack_map_alloc(union bpf_attr *attr)
 
        err = get_callchain_buffers(sysctl_perf_event_max_stack);
        if (err)
-               goto free_charge;
+               goto free_smap;
 
        err = prealloc_elems_and_freelist(smap);
        if (err)
                goto put_buffers;
 
-       bpf_map_charge_move(&smap->map.memory, &mem);
-
        return &smap->map;
 
 put_buffers:
        put_callchain_buffers();
-free_charge:
-       bpf_map_charge_finish(&mem);
+free_smap:
        bpf_map_area_free(smap);
        return ERR_PTR(err);
 }
@@ -298,7 +288,7 @@ static void stack_map_get_build_id_offset(struct bpf_stack_build_id *id_offs,
        if (irqs_disabled()) {
                if (!IS_ENABLED(CONFIG_PREEMPT_RT)) {
                        work = this_cpu_ptr(&up_read_work);
-                       if (atomic_read(&work->irq_work.flags) & IRQ_WORK_BUSY) {
+                       if (irq_work_is_busy(&work->irq_work)) {
                                /* cannot queue more up_read, fallback */
                                irq_work_busy = true;
                        }