bpf: add bpf_map_lookup_percpu_elem for percpu map
authorFeng Zhou <zhoufeng.zf@bytedance.com>
Wed, 11 May 2022 09:38:53 +0000 (17:38 +0800)
committerAlexei Starovoitov <ast@kernel.org>
Thu, 12 May 2022 01:16:54 +0000 (18:16 -0700)
Add new ebpf helpers bpf_map_lookup_percpu_elem.

The implementation method is relatively simple, refer to the implementation
method of map_lookup_elem of percpu map, increase the parameters of cpu, and
obtain it according to the specified cpu.

Signed-off-by: Feng Zhou <zhoufeng.zf@bytedance.com>
Link: https://lore.kernel.org/r/20220511093854.411-2-zhoufeng.zf@bytedance.com
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
include/linux/bpf.h
include/uapi/linux/bpf.h
kernel/bpf/arraymap.c
kernel/bpf/core.c
kernel/bpf/hashtab.c
kernel/bpf/helpers.c
kernel/bpf/verifier.c
kernel/trace/bpf_trace.c
tools/include/uapi/linux/bpf.h

index 3ded871..5061ccd 100644 (file)
@@ -89,6 +89,7 @@ struct bpf_map_ops {
        int (*map_push_elem)(struct bpf_map *map, void *value, u64 flags);
        int (*map_pop_elem)(struct bpf_map *map, void *value);
        int (*map_peek_elem)(struct bpf_map *map, void *value);
+       void *(*map_lookup_percpu_elem)(struct bpf_map *map, void *key, u32 cpu);
 
        /* funcs called by prog_array and perf_event_array map */
        void *(*map_fd_get_ptr)(struct bpf_map *map, struct file *map_file,
@@ -2184,6 +2185,7 @@ extern const struct bpf_func_proto bpf_map_delete_elem_proto;
 extern const struct bpf_func_proto bpf_map_push_elem_proto;
 extern const struct bpf_func_proto bpf_map_pop_elem_proto;
 extern const struct bpf_func_proto bpf_map_peek_elem_proto;
+extern const struct bpf_func_proto bpf_map_lookup_percpu_elem_proto;
 
 extern const struct bpf_func_proto bpf_get_prandom_u32_proto;
 extern const struct bpf_func_proto bpf_get_smp_processor_id_proto;
index bc7f899..0210f85 100644 (file)
@@ -5164,6 +5164,14 @@ union bpf_attr {
  *             if not NULL, is a reference which must be released using its
  *             corresponding release function, or moved into a BPF map before
  *             program exit.
+ *
+ * void *bpf_map_lookup_percpu_elem(struct bpf_map *map, const void *key, u32 cpu)
+ *     Description
+ *             Perform a lookup in *percpu map* for an entry associated to
+ *             *key* on *cpu*.
+ *     Return
+ *             Map value associated to *key* on *cpu*, or **NULL** if no entry
+ *             was found or *cpu* is invalid.
  */
 #define __BPF_FUNC_MAPPER(FN)          \
        FN(unspec),                     \
@@ -5361,6 +5369,7 @@ union bpf_attr {
        FN(skb_set_tstamp),             \
        FN(ima_file_hash),              \
        FN(kptr_xchg),                  \
+       FN(map_lookup_percpu_elem),     \
        /* */
 
 /* integer value in 'imm' field of BPF_CALL instruction selects which helper
index 724613d..fe40d3b 100644 (file)
@@ -243,6 +243,20 @@ static void *percpu_array_map_lookup_elem(struct bpf_map *map, void *key)
        return this_cpu_ptr(array->pptrs[index & array->index_mask]);
 }
 
+static void *percpu_array_map_lookup_percpu_elem(struct bpf_map *map, void *key, u32 cpu)
+{
+       struct bpf_array *array = container_of(map, struct bpf_array, map);
+       u32 index = *(u32 *)key;
+
+       if (cpu >= nr_cpu_ids)
+               return NULL;
+
+       if (unlikely(index >= array->map.max_entries))
+               return NULL;
+
+       return per_cpu_ptr(array->pptrs[index & array->index_mask], cpu);
+}
+
 int bpf_percpu_array_copy(struct bpf_map *map, void *key, void *value)
 {
        struct bpf_array *array = container_of(map, struct bpf_array, map);
@@ -725,6 +739,7 @@ const struct bpf_map_ops percpu_array_map_ops = {
        .map_lookup_elem = percpu_array_map_lookup_elem,
        .map_update_elem = array_map_update_elem,
        .map_delete_elem = array_map_delete_elem,
+       .map_lookup_percpu_elem = percpu_array_map_lookup_percpu_elem,
        .map_seq_show_elem = percpu_array_map_seq_show_elem,
        .map_check_btf = array_map_check_btf,
        .map_lookup_batch = generic_map_lookup_batch,
index 13e9dbe..76f68d0 100644 (file)
@@ -2619,6 +2619,7 @@ const struct bpf_func_proto bpf_map_delete_elem_proto __weak;
 const struct bpf_func_proto bpf_map_push_elem_proto __weak;
 const struct bpf_func_proto bpf_map_pop_elem_proto __weak;
 const struct bpf_func_proto bpf_map_peek_elem_proto __weak;
+const struct bpf_func_proto bpf_map_lookup_percpu_elem_proto __weak;
 const struct bpf_func_proto bpf_spin_lock_proto __weak;
 const struct bpf_func_proto bpf_spin_unlock_proto __weak;
 const struct bpf_func_proto bpf_jiffies64_proto __weak;
index 7058412..17fb69c 100644 (file)
@@ -2199,6 +2199,20 @@ static void *htab_percpu_map_lookup_elem(struct bpf_map *map, void *key)
                return NULL;
 }
 
+static void *htab_percpu_map_lookup_percpu_elem(struct bpf_map *map, void *key, u32 cpu)
+{
+       struct htab_elem *l;
+
+       if (cpu >= nr_cpu_ids)
+               return NULL;
+
+       l = __htab_map_lookup_elem(map, key);
+       if (l)
+               return per_cpu_ptr(htab_elem_get_ptr(l, map->key_size), cpu);
+       else
+               return NULL;
+}
+
 static void *htab_lru_percpu_map_lookup_elem(struct bpf_map *map, void *key)
 {
        struct htab_elem *l = __htab_map_lookup_elem(map, key);
@@ -2211,6 +2225,22 @@ static void *htab_lru_percpu_map_lookup_elem(struct bpf_map *map, void *key)
        return NULL;
 }
 
+static void *htab_lru_percpu_map_lookup_percpu_elem(struct bpf_map *map, void *key, u32 cpu)
+{
+       struct htab_elem *l;
+
+       if (cpu >= nr_cpu_ids)
+               return NULL;
+
+       l = __htab_map_lookup_elem(map, key);
+       if (l) {
+               bpf_lru_node_set_ref(&l->lru_node);
+               return per_cpu_ptr(htab_elem_get_ptr(l, map->key_size), cpu);
+       }
+
+       return NULL;
+}
+
 int bpf_percpu_hash_copy(struct bpf_map *map, void *key, void *value)
 {
        struct htab_elem *l;
@@ -2300,6 +2330,7 @@ const struct bpf_map_ops htab_percpu_map_ops = {
        .map_lookup_and_delete_elem = htab_percpu_map_lookup_and_delete_elem,
        .map_update_elem = htab_percpu_map_update_elem,
        .map_delete_elem = htab_map_delete_elem,
+       .map_lookup_percpu_elem = htab_percpu_map_lookup_percpu_elem,
        .map_seq_show_elem = htab_percpu_map_seq_show_elem,
        .map_set_for_each_callback_args = map_set_for_each_callback_args,
        .map_for_each_callback = bpf_for_each_hash_elem,
@@ -2318,6 +2349,7 @@ const struct bpf_map_ops htab_lru_percpu_map_ops = {
        .map_lookup_and_delete_elem = htab_lru_percpu_map_lookup_and_delete_elem,
        .map_update_elem = htab_lru_percpu_map_update_elem,
        .map_delete_elem = htab_lru_map_delete_elem,
+       .map_lookup_percpu_elem = htab_lru_percpu_map_lookup_percpu_elem,
        .map_seq_show_elem = htab_percpu_map_seq_show_elem,
        .map_set_for_each_callback_args = map_set_for_each_callback_args,
        .map_for_each_callback = bpf_for_each_hash_elem,
index 3e709fe..d5f104a 100644 (file)
@@ -119,6 +119,22 @@ const struct bpf_func_proto bpf_map_peek_elem_proto = {
        .arg2_type      = ARG_PTR_TO_UNINIT_MAP_VALUE,
 };
 
+BPF_CALL_3(bpf_map_lookup_percpu_elem, struct bpf_map *, map, void *, key, u32, cpu)
+{
+       WARN_ON_ONCE(!rcu_read_lock_held() && !rcu_read_lock_bh_held());
+       return (unsigned long) map->ops->map_lookup_percpu_elem(map, key, cpu);
+}
+
+const struct bpf_func_proto bpf_map_lookup_percpu_elem_proto = {
+       .func           = bpf_map_lookup_percpu_elem,
+       .gpl_only       = false,
+       .pkt_access     = true,
+       .ret_type       = RET_PTR_TO_MAP_VALUE_OR_NULL,
+       .arg1_type      = ARG_CONST_MAP_PTR,
+       .arg2_type      = ARG_PTR_TO_MAP_KEY,
+       .arg3_type      = ARG_ANYTHING,
+};
+
 const struct bpf_func_proto bpf_get_prandom_u32_proto = {
        .func           = bpf_user_rnd_u32,
        .gpl_only       = false,
@@ -1420,6 +1436,8 @@ bpf_base_func_proto(enum bpf_func_id func_id)
                return &bpf_map_pop_elem_proto;
        case BPF_FUNC_map_peek_elem:
                return &bpf_map_peek_elem_proto;
+       case BPF_FUNC_map_lookup_percpu_elem:
+               return &bpf_map_lookup_percpu_elem_proto;
        case BPF_FUNC_get_prandom_u32:
                return &bpf_get_prandom_u32_proto;
        case BPF_FUNC_get_smp_processor_id:
index c27fee7..05c1b66 100644 (file)
@@ -6137,6 +6137,12 @@ static int check_map_func_compatibility(struct bpf_verifier_env *env,
                    map->map_type != BPF_MAP_TYPE_BLOOM_FILTER)
                        goto error;
                break;
+       case BPF_FUNC_map_lookup_percpu_elem:
+               if (map->map_type != BPF_MAP_TYPE_PERCPU_ARRAY &&
+                   map->map_type != BPF_MAP_TYPE_PERCPU_HASH &&
+                   map->map_type != BPF_MAP_TYPE_LRU_PERCPU_HASH)
+                       goto error;
+               break;
        case BPF_FUNC_sk_storage_get:
        case BPF_FUNC_sk_storage_delete:
                if (map->map_type != BPF_MAP_TYPE_SK_STORAGE)
@@ -6750,7 +6756,8 @@ record_func_map(struct bpf_verifier_env *env, struct bpf_call_arg_meta *meta,
            func_id != BPF_FUNC_map_pop_elem &&
            func_id != BPF_FUNC_map_peek_elem &&
            func_id != BPF_FUNC_for_each_map_elem &&
-           func_id != BPF_FUNC_redirect_map)
+           func_id != BPF_FUNC_redirect_map &&
+           func_id != BPF_FUNC_map_lookup_percpu_elem)
                return 0;
 
        if (map == NULL) {
@@ -13810,7 +13817,8 @@ static int do_misc_fixups(struct bpf_verifier_env *env)
                     insn->imm == BPF_FUNC_map_pop_elem    ||
                     insn->imm == BPF_FUNC_map_peek_elem   ||
                     insn->imm == BPF_FUNC_redirect_map    ||
-                    insn->imm == BPF_FUNC_for_each_map_elem)) {
+                    insn->imm == BPF_FUNC_for_each_map_elem ||
+                    insn->imm == BPF_FUNC_map_lookup_percpu_elem)) {
                        aux = &env->insn_aux_data[i + delta];
                        if (bpf_map_ptr_poisoned(aux))
                                goto patch_call_imm;
@@ -13859,6 +13867,8 @@ static int do_misc_fixups(struct bpf_verifier_env *env)
                                              bpf_callback_t callback_fn,
                                              void *callback_ctx,
                                              u64 flags))NULL));
+                       BUILD_BUG_ON(!__same_type(ops->map_lookup_percpu_elem,
+                                    (void *(*)(struct bpf_map *map, void *key, u32 cpu))NULL));
 
 patch_map_ops_generic:
                        switch (insn->imm) {
@@ -13886,6 +13896,9 @@ patch_map_ops_generic:
                        case BPF_FUNC_for_each_map_elem:
                                insn->imm = BPF_CALL_IMM(ops->map_for_each_callback);
                                continue;
+                       case BPF_FUNC_map_lookup_percpu_elem:
+                               insn->imm = BPF_CALL_IMM(ops->map_lookup_percpu_elem);
+                               continue;
                        }
 
                        goto patch_call_imm;
index 2eaac09..7141ca8 100644 (file)
@@ -1197,6 +1197,8 @@ bpf_tracing_func_proto(enum bpf_func_id func_id, const struct bpf_prog *prog)
                return &bpf_map_pop_elem_proto;
        case BPF_FUNC_map_peek_elem:
                return &bpf_map_peek_elem_proto;
+       case BPF_FUNC_map_lookup_percpu_elem:
+               return &bpf_map_lookup_percpu_elem_proto;
        case BPF_FUNC_ktime_get_ns:
                return &bpf_ktime_get_ns_proto;
        case BPF_FUNC_ktime_get_boot_ns:
index bc7f899..0210f85 100644 (file)
@@ -5164,6 +5164,14 @@ union bpf_attr {
  *             if not NULL, is a reference which must be released using its
  *             corresponding release function, or moved into a BPF map before
  *             program exit.
+ *
+ * void *bpf_map_lookup_percpu_elem(struct bpf_map *map, const void *key, u32 cpu)
+ *     Description
+ *             Perform a lookup in *percpu map* for an entry associated to
+ *             *key* on *cpu*.
+ *     Return
+ *             Map value associated to *key* on *cpu*, or **NULL** if no entry
+ *             was found or *cpu* is invalid.
  */
 #define __BPF_FUNC_MAPPER(FN)          \
        FN(unspec),                     \
@@ -5361,6 +5369,7 @@ union bpf_attr {
        FN(skb_set_tstamp),             \
        FN(ima_file_hash),              \
        FN(kptr_xchg),                  \
+       FN(map_lookup_percpu_elem),     \
        /* */
 
 /* integer value in 'imm' field of BPF_CALL instruction selects which helper