Commit 496f4f1b authored by Alexei Starovoitov's avatar Alexei Starovoitov

Merge branch 'Don't invoke KPTR_REF destructor on NULL xchg'

David Vernet says:

====================

When a map value is being freed, we loop over all of the fields of the
corresponding BPF object and issue the appropriate cleanup calls
corresponding to the field's type. If the field is a referenced kptr, we
atomically xchg the value out of the map, and invoke the kptr's
destructor on whatever was there before.

Currently, we always invoke the destructor (or bpf_obj_drop() for a
local kptr) on any kptr, including if no value was xchg'd out of the
map. This means that any function serving as the kptr's KF_RELEASE
destructor must always treat the argument as possibly NULL, and we
invoke unnecessary (and seemingly unsafe) cleanup logic for the local
kptr path as well.

This is an odd requirement -- KF_RELEASE kfuncs that are invoked by BPF
programs do not have this restriction, and the verifier will fail to
load the program if the register containing the to-be-released type has
any untrusted modifiers (e.g. PTR_UNTRUSTED or PTR_MAYBE_NULL). So as to
simplify the expectations required for a KF_RELEASE kfunc, this patch
set updates the KPTR_REF destructor logic to only be invoked when a
non-NULL value is xchg'd out of the map.

Additionally, the patch removes now-unnecessary KF_RELEASE calls from
several kfuncs, and finally, updates the verifier to have KF_RELEASE
automatically imply KF_TRUSTED_ARGS. This restriction was already
implicitly happening because of the aforementioned logic in the verifier
to reject any regs with untrusted modifiers, and to enforce that
KF_RELEASE args are passed with a 0 offset. This change just updates the
behavior to match that of other trusted args. This patch is left to the
end of the series in case it happens to be controversial, as it arguably
is slightly orthogonal to the purpose of the rest of the series.
====================
Signed-off-by: default avatarAlexei Starovoitov <ast@kernel.org>
parents 55fbae05 6c831c46
...@@ -179,9 +179,10 @@ both are orthogonal to each other. ...@@ -179,9 +179,10 @@ both are orthogonal to each other.
--------------------- ---------------------
The KF_RELEASE flag is used to indicate that the kfunc releases the pointer The KF_RELEASE flag is used to indicate that the kfunc releases the pointer
passed in to it. There can be only one referenced pointer that can be passed in. passed in to it. There can be only one referenced pointer that can be passed
All copies of the pointer being released are invalidated as a result of invoking in. All copies of the pointer being released are invalidated as a result of
kfunc with this flag. invoking kfunc with this flag. KF_RELEASE kfuncs automatically receive the
protection afforded by the KF_TRUSTED_ARGS flag described below.
2.4.4 KF_KPTR_GET flag 2.4.4 KF_KPTR_GET flag
---------------------- ----------------------
......
...@@ -342,9 +342,6 @@ hid_bpf_release_context(struct hid_bpf_ctx *ctx) ...@@ -342,9 +342,6 @@ hid_bpf_release_context(struct hid_bpf_ctx *ctx)
{ {
struct hid_bpf_ctx_kern *ctx_kern; struct hid_bpf_ctx_kern *ctx_kern;
if (!ctx)
return;
ctx_kern = container_of(ctx, struct hid_bpf_ctx_kern, ctx); ctx_kern = container_of(ctx, struct hid_bpf_ctx_kern, ctx);
kfree(ctx_kern); kfree(ctx_kern);
......
...@@ -102,9 +102,6 @@ static void cpumask_free_cb(struct rcu_head *head) ...@@ -102,9 +102,6 @@ static void cpumask_free_cb(struct rcu_head *head)
*/ */
__bpf_kfunc void bpf_cpumask_release(struct bpf_cpumask *cpumask) __bpf_kfunc void bpf_cpumask_release(struct bpf_cpumask *cpumask)
{ {
if (!cpumask)
return;
if (refcount_dec_and_test(&cpumask->usage)) if (refcount_dec_and_test(&cpumask->usage))
call_rcu(&cpumask->rcu, cpumask_free_cb); call_rcu(&cpumask->rcu, cpumask_free_cb);
} }
...@@ -405,7 +402,7 @@ __diag_pop(); ...@@ -405,7 +402,7 @@ __diag_pop();
BTF_SET8_START(cpumask_kfunc_btf_ids) BTF_SET8_START(cpumask_kfunc_btf_ids)
BTF_ID_FLAGS(func, bpf_cpumask_create, KF_ACQUIRE | KF_RET_NULL) BTF_ID_FLAGS(func, bpf_cpumask_create, KF_ACQUIRE | KF_RET_NULL)
BTF_ID_FLAGS(func, bpf_cpumask_release, KF_RELEASE | KF_TRUSTED_ARGS) BTF_ID_FLAGS(func, bpf_cpumask_release, KF_RELEASE)
BTF_ID_FLAGS(func, bpf_cpumask_acquire, KF_ACQUIRE | KF_TRUSTED_ARGS) BTF_ID_FLAGS(func, bpf_cpumask_acquire, KF_ACQUIRE | KF_TRUSTED_ARGS)
BTF_ID_FLAGS(func, bpf_cpumask_first, KF_RCU) BTF_ID_FLAGS(func, bpf_cpumask_first, KF_RCU)
BTF_ID_FLAGS(func, bpf_cpumask_first_zero, KF_RCU) BTF_ID_FLAGS(func, bpf_cpumask_first_zero, KF_RCU)
......
...@@ -2089,9 +2089,6 @@ __bpf_kfunc struct task_struct *bpf_task_kptr_get(struct task_struct **pp) ...@@ -2089,9 +2089,6 @@ __bpf_kfunc struct task_struct *bpf_task_kptr_get(struct task_struct **pp)
*/ */
__bpf_kfunc void bpf_task_release(struct task_struct *p) __bpf_kfunc void bpf_task_release(struct task_struct *p)
{ {
if (!p)
return;
put_task_struct(p); put_task_struct(p);
} }
...@@ -2148,9 +2145,6 @@ __bpf_kfunc struct cgroup *bpf_cgroup_kptr_get(struct cgroup **cgrpp) ...@@ -2148,9 +2145,6 @@ __bpf_kfunc struct cgroup *bpf_cgroup_kptr_get(struct cgroup **cgrpp)
*/ */
__bpf_kfunc void bpf_cgroup_release(struct cgroup *cgrp) __bpf_kfunc void bpf_cgroup_release(struct cgroup *cgrp)
{ {
if (!cgrp)
return;
cgroup_put(cgrp); cgroup_put(cgrp);
} }
......
...@@ -677,6 +677,9 @@ void bpf_obj_free_fields(const struct btf_record *rec, void *obj) ...@@ -677,6 +677,9 @@ void bpf_obj_free_fields(const struct btf_record *rec, void *obj)
break; break;
case BPF_KPTR_REF: case BPF_KPTR_REF:
xchgd_field = (void *)xchg((unsigned long *)field_ptr, 0); xchgd_field = (void *)xchg((unsigned long *)field_ptr, 0);
if (!xchgd_field)
break;
if (!btf_is_kernel(field->kptr.btf)) { if (!btf_is_kernel(field->kptr.btf)) {
pointee_struct_meta = btf_find_struct_meta(field->kptr.btf, pointee_struct_meta = btf_find_struct_meta(field->kptr.btf,
field->kptr.btf_id); field->kptr.btf_id);
......
...@@ -9307,7 +9307,7 @@ static bool is_kfunc_release(struct bpf_kfunc_call_arg_meta *meta) ...@@ -9307,7 +9307,7 @@ static bool is_kfunc_release(struct bpf_kfunc_call_arg_meta *meta)
static bool is_kfunc_trusted_args(struct bpf_kfunc_call_arg_meta *meta) static bool is_kfunc_trusted_args(struct bpf_kfunc_call_arg_meta *meta)
{ {
return meta->kfunc_flags & KF_TRUSTED_ARGS; return (meta->kfunc_flags & KF_TRUSTED_ARGS) || is_kfunc_release(meta);
} }
static bool is_kfunc_sleepable(struct bpf_kfunc_call_arg_meta *meta) static bool is_kfunc_sleepable(struct bpf_kfunc_call_arg_meta *meta)
......
...@@ -606,6 +606,11 @@ bpf_kfunc_call_test_acquire(unsigned long *scalar_ptr) ...@@ -606,6 +606,11 @@ bpf_kfunc_call_test_acquire(unsigned long *scalar_ptr)
return &prog_test_struct; return &prog_test_struct;
} }
__bpf_kfunc void bpf_kfunc_call_test_offset(struct prog_test_ref_kfunc *p)
{
WARN_ON_ONCE(1);
}
__bpf_kfunc struct prog_test_member * __bpf_kfunc struct prog_test_member *
bpf_kfunc_call_memb_acquire(void) bpf_kfunc_call_memb_acquire(void)
{ {
...@@ -615,9 +620,6 @@ bpf_kfunc_call_memb_acquire(void) ...@@ -615,9 +620,6 @@ bpf_kfunc_call_memb_acquire(void)
__bpf_kfunc void bpf_kfunc_call_test_release(struct prog_test_ref_kfunc *p) __bpf_kfunc void bpf_kfunc_call_test_release(struct prog_test_ref_kfunc *p)
{ {
if (!p)
return;
refcount_dec(&p->cnt); refcount_dec(&p->cnt);
} }
...@@ -803,6 +805,7 @@ BTF_ID_FLAGS(func, bpf_kfunc_call_test_mem_len_fail2) ...@@ -803,6 +805,7 @@ BTF_ID_FLAGS(func, bpf_kfunc_call_test_mem_len_fail2)
BTF_ID_FLAGS(func, bpf_kfunc_call_test_ref, KF_TRUSTED_ARGS | KF_RCU) BTF_ID_FLAGS(func, bpf_kfunc_call_test_ref, KF_TRUSTED_ARGS | KF_RCU)
BTF_ID_FLAGS(func, bpf_kfunc_call_test_destructive, KF_DESTRUCTIVE) BTF_ID_FLAGS(func, bpf_kfunc_call_test_destructive, KF_DESTRUCTIVE)
BTF_ID_FLAGS(func, bpf_kfunc_call_test_static_unused_arg) BTF_ID_FLAGS(func, bpf_kfunc_call_test_static_unused_arg)
BTF_ID_FLAGS(func, bpf_kfunc_call_test_offset)
BTF_SET8_END(test_sk_check_kfunc_ids) BTF_SET8_END(test_sk_check_kfunc_ids)
static void *bpf_test_init(const union bpf_attr *kattr, u32 user_size, static void *bpf_test_init(const union bpf_attr *kattr, u32 user_size,
......
...@@ -401,8 +401,6 @@ __bpf_kfunc struct nf_conn *bpf_ct_insert_entry(struct nf_conn___init *nfct_i) ...@@ -401,8 +401,6 @@ __bpf_kfunc struct nf_conn *bpf_ct_insert_entry(struct nf_conn___init *nfct_i)
*/ */
__bpf_kfunc void bpf_ct_release(struct nf_conn *nfct) __bpf_kfunc void bpf_ct_release(struct nf_conn *nfct)
{ {
if (!nfct)
return;
nf_ct_put(nfct); nf_ct_put(nfct);
} }
......
...@@ -206,7 +206,7 @@ int BPF_PROG(cgrp_kfunc_get_unreleased, struct cgroup *cgrp, const char *path) ...@@ -206,7 +206,7 @@ int BPF_PROG(cgrp_kfunc_get_unreleased, struct cgroup *cgrp, const char *path)
} }
SEC("tp_btf/cgroup_mkdir") SEC("tp_btf/cgroup_mkdir")
__failure __msg("expects refcounted") __failure __msg("Possibly NULL pointer passed to trusted arg0")
int BPF_PROG(cgrp_kfunc_release_untrusted, struct cgroup *cgrp, const char *path) int BPF_PROG(cgrp_kfunc_release_untrusted, struct cgroup *cgrp, const char *path)
{ {
struct __cgrps_kfunc_map_value *v; struct __cgrps_kfunc_map_value *v;
...@@ -234,7 +234,7 @@ int BPF_PROG(cgrp_kfunc_release_fp, struct cgroup *cgrp, const char *path) ...@@ -234,7 +234,7 @@ int BPF_PROG(cgrp_kfunc_release_fp, struct cgroup *cgrp, const char *path)
} }
SEC("tp_btf/cgroup_mkdir") SEC("tp_btf/cgroup_mkdir")
__failure __msg("arg#0 is ptr_or_null_ expected ptr_ or socket") __failure __msg("Possibly NULL pointer passed to trusted arg0")
int BPF_PROG(cgrp_kfunc_release_null, struct cgroup *cgrp, const char *path) int BPF_PROG(cgrp_kfunc_release_null, struct cgroup *cgrp, const char *path)
{ {
struct __cgrps_kfunc_map_value local, *v; struct __cgrps_kfunc_map_value local, *v;
......
...@@ -206,7 +206,7 @@ int BPF_PROG(task_kfunc_get_unreleased, struct task_struct *task, u64 clone_flag ...@@ -206,7 +206,7 @@ int BPF_PROG(task_kfunc_get_unreleased, struct task_struct *task, u64 clone_flag
} }
SEC("tp_btf/task_newtask") SEC("tp_btf/task_newtask")
__failure __msg("arg#0 is untrusted_ptr_or_null_ expected ptr_ or socket") __failure __msg("Possibly NULL pointer passed to trusted arg0")
int BPF_PROG(task_kfunc_release_untrusted, struct task_struct *task, u64 clone_flags) int BPF_PROG(task_kfunc_release_untrusted, struct task_struct *task, u64 clone_flags)
{ {
struct __tasks_kfunc_map_value *v; struct __tasks_kfunc_map_value *v;
...@@ -234,7 +234,7 @@ int BPF_PROG(task_kfunc_release_fp, struct task_struct *task, u64 clone_flags) ...@@ -234,7 +234,7 @@ int BPF_PROG(task_kfunc_release_fp, struct task_struct *task, u64 clone_flags)
} }
SEC("tp_btf/task_newtask") SEC("tp_btf/task_newtask")
__failure __msg("arg#0 is ptr_or_null_ expected ptr_ or socket") __failure __msg("Possibly NULL pointer passed to trusted arg0")
int BPF_PROG(task_kfunc_release_null, struct task_struct *task, u64 clone_flags) int BPF_PROG(task_kfunc_release_null, struct task_struct *task, u64 clone_flags)
{ {
struct __tasks_kfunc_map_value local, *v; struct __tasks_kfunc_map_value local, *v;
...@@ -277,7 +277,7 @@ int BPF_PROG(task_kfunc_release_unacquired, struct task_struct *task, u64 clone_ ...@@ -277,7 +277,7 @@ int BPF_PROG(task_kfunc_release_unacquired, struct task_struct *task, u64 clone_
} }
SEC("tp_btf/task_newtask") SEC("tp_btf/task_newtask")
__failure __msg("arg#0 is ptr_or_null_ expected ptr_ or socket") __failure __msg("Possibly NULL pointer passed to trusted arg0")
int BPF_PROG(task_kfunc_from_pid_no_null_check, struct task_struct *task, u64 clone_flags) int BPF_PROG(task_kfunc_from_pid_no_null_check, struct task_struct *task, u64 clone_flags)
{ {
struct task_struct *acquired; struct task_struct *acquired;
......
...@@ -109,7 +109,7 @@ ...@@ -109,7 +109,7 @@
}, },
.prog_type = BPF_PROG_TYPE_SCHED_CLS, .prog_type = BPF_PROG_TYPE_SCHED_CLS,
.result = REJECT, .result = REJECT,
.errstr = "arg#0 is ptr_or_null_ expected ptr_ or socket", .errstr = "Possibly NULL pointer passed to trusted arg0",
.fixup_kfunc_btf_id = { .fixup_kfunc_btf_id = {
{ "bpf_kfunc_call_test_acquire", 3 }, { "bpf_kfunc_call_test_acquire", 3 },
{ "bpf_kfunc_call_test_release", 5 }, { "bpf_kfunc_call_test_release", 5 },
...@@ -165,19 +165,23 @@ ...@@ -165,19 +165,23 @@
BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8),
BPF_ST_MEM(BPF_DW, BPF_REG_1, 0, 0), BPF_ST_MEM(BPF_DW, BPF_REG_1, 0, 0),
BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, BPF_PSEUDO_KFUNC_CALL, 0, 0), BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, BPF_PSEUDO_KFUNC_CALL, 0, 0),
BPF_MOV64_REG(BPF_REG_2, BPF_REG_0),
BPF_JMP_IMM(BPF_JNE, BPF_REG_0, 0, 1), BPF_JMP_IMM(BPF_JNE, BPF_REG_0, 0, 1),
BPF_EXIT_INSN(), BPF_EXIT_INSN(),
BPF_MOV64_REG(BPF_REG_1, BPF_REG_0), BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
BPF_LDX_MEM(BPF_DW, BPF_REG_1, BPF_REG_1, 16),
BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -4), BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -4),
BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, BPF_PSEUDO_KFUNC_CALL, 0, 0), BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, BPF_PSEUDO_KFUNC_CALL, 0, 0),
BPF_MOV64_IMM(BPF_REG_0, 0), BPF_MOV64_IMM(BPF_REG_0, 0),
BPF_MOV64_REG(BPF_REG_1, BPF_REG_2),
BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, BPF_PSEUDO_KFUNC_CALL, 0, 0),
BPF_MOV64_IMM(BPF_REG_0, 0),
BPF_EXIT_INSN(), BPF_EXIT_INSN(),
}, },
.prog_type = BPF_PROG_TYPE_SCHED_CLS, .prog_type = BPF_PROG_TYPE_SCHED_CLS,
.fixup_kfunc_btf_id = { .fixup_kfunc_btf_id = {
{ "bpf_kfunc_call_test_acquire", 3 }, { "bpf_kfunc_call_test_acquire", 3 },
{ "bpf_kfunc_call_test_release", 9 }, { "bpf_kfunc_call_test_offset", 9 },
{ "bpf_kfunc_call_test_release", 12 },
}, },
.result_unpriv = REJECT, .result_unpriv = REJECT,
.result = REJECT, .result = REJECT,
......
...@@ -142,7 +142,7 @@ ...@@ -142,7 +142,7 @@
.kfunc = "bpf", .kfunc = "bpf",
.expected_attach_type = BPF_LSM_MAC, .expected_attach_type = BPF_LSM_MAC,
.flags = BPF_F_SLEEPABLE, .flags = BPF_F_SLEEPABLE,
.errstr = "arg#0 is ptr_or_null_ expected ptr_ or socket", .errstr = "Possibly NULL pointer passed to trusted arg0",
.fixup_kfunc_btf_id = { .fixup_kfunc_btf_id = {
{ "bpf_lookup_user_key", 2 }, { "bpf_lookup_user_key", 2 },
{ "bpf_key_put", 4 }, { "bpf_key_put", 4 },
...@@ -163,7 +163,7 @@ ...@@ -163,7 +163,7 @@
.kfunc = "bpf", .kfunc = "bpf",
.expected_attach_type = BPF_LSM_MAC, .expected_attach_type = BPF_LSM_MAC,
.flags = BPF_F_SLEEPABLE, .flags = BPF_F_SLEEPABLE,
.errstr = "arg#0 is ptr_or_null_ expected ptr_ or socket", .errstr = "Possibly NULL pointer passed to trusted arg0",
.fixup_kfunc_btf_id = { .fixup_kfunc_btf_id = {
{ "bpf_lookup_system_key", 1 }, { "bpf_lookup_system_key", 1 },
{ "bpf_key_put", 3 }, { "bpf_key_put", 3 },
...@@ -182,7 +182,7 @@ ...@@ -182,7 +182,7 @@
.kfunc = "bpf", .kfunc = "bpf",
.expected_attach_type = BPF_LSM_MAC, .expected_attach_type = BPF_LSM_MAC,
.flags = BPF_F_SLEEPABLE, .flags = BPF_F_SLEEPABLE,
.errstr = "arg#0 pointer type STRUCT bpf_key must point to scalar, or struct with scalar", .errstr = "Possibly NULL pointer passed to trusted arg0",
.fixup_kfunc_btf_id = { .fixup_kfunc_btf_id = {
{ "bpf_key_put", 1 }, { "bpf_key_put", 1 },
}, },
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment