Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

selftests/bpf: Add cgroup kfunc / kptr selftests

This patch adds a selftest suite to validate the cgroup kfuncs that were
added in the prior patch.

Signed-off-by: David Vernet <void@manifault.com>
Link: https://lore.kernel.org/r/20221122055458.173143-3-void@manifault.com
Signed-off-by: Alexei Starovoitov <ast@kernel.org>

authored by

David Vernet and committed by
Alexei Starovoitov
f583ddf1 fda01efc

+631
+1
tools/testing/selftests/bpf/DENYLIST.s390x
··· 10 10 bpf_tcp_ca # JIT does not support calling kernel function (kfunc) 11 11 cb_refs # expected error message unexpected error: -524 (trampoline) 12 12 cgroup_hierarchical_stats # JIT does not support calling kernel function (kfunc) 13 + cgrp_kfunc # JIT does not support calling kernel function 13 14 cgrp_local_storage # prog_attach unexpected error: -524 (trampoline) 14 15 core_read_macros # unknown func bpf_probe_read#4 (overlapping) 15 16 d_path # failed to auto-attach program 'prog_stat': -524 (trampoline)
+174
tools/testing/selftests/bpf/prog_tests/cgrp_kfunc.c
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + /* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */ 3 + 4 + #define _GNU_SOURCE 5 + #include <cgroup_helpers.h> 6 + #include <test_progs.h> 7 + 8 + #include "cgrp_kfunc_failure.skel.h" 9 + #include "cgrp_kfunc_success.skel.h" 10 + 11 + static size_t log_buf_sz = 1 << 20; /* 1 MB */ 12 + static char obj_log_buf[1048576]; 13 + 14 + static struct cgrp_kfunc_success *open_load_cgrp_kfunc_skel(void) 15 + { 16 + struct cgrp_kfunc_success *skel; 17 + int err; 18 + 19 + skel = cgrp_kfunc_success__open(); 20 + if (!ASSERT_OK_PTR(skel, "skel_open")) 21 + return NULL; 22 + 23 + skel->bss->pid = getpid(); 24 + 25 + err = cgrp_kfunc_success__load(skel); 26 + if (!ASSERT_OK(err, "skel_load")) 27 + goto cleanup; 28 + 29 + return skel; 30 + 31 + cleanup: 32 + cgrp_kfunc_success__destroy(skel); 33 + return NULL; 34 + } 35 + 36 + static int mkdir_rm_test_dir(void) 37 + { 38 + int fd; 39 + const char *cgrp_path = "cgrp_kfunc"; 40 + 41 + fd = create_and_get_cgroup(cgrp_path); 42 + if (!ASSERT_GT(fd, 0, "mkdir_cgrp_fd")) 43 + return -1; 44 + 45 + close(fd); 46 + remove_cgroup(cgrp_path); 47 + 48 + return 0; 49 + } 50 + 51 + static void run_success_test(const char *prog_name) 52 + { 53 + struct cgrp_kfunc_success *skel; 54 + struct bpf_program *prog; 55 + struct bpf_link *link = NULL; 56 + 57 + skel = open_load_cgrp_kfunc_skel(); 58 + if (!ASSERT_OK_PTR(skel, "open_load_skel")) 59 + return; 60 + 61 + if (!ASSERT_OK(skel->bss->err, "pre_mkdir_err")) 62 + goto cleanup; 63 + 64 + prog = bpf_object__find_program_by_name(skel->obj, prog_name); 65 + if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name")) 66 + goto cleanup; 67 + 68 + link = bpf_program__attach(prog); 69 + if (!ASSERT_OK_PTR(link, "attached_link")) 70 + goto cleanup; 71 + 72 + ASSERT_EQ(skel->bss->invocations, 0, "pre_rmdir_count"); 73 + if (!ASSERT_OK(mkdir_rm_test_dir(), "cgrp_mkdir")) 74 + goto cleanup; 75 + 76 + ASSERT_EQ(skel->bss->invocations, 1, "post_rmdir_count"); 77 + ASSERT_OK(skel->bss->err, "post_rmdir_err"); 78 + 79 + cleanup: 80 + bpf_link__destroy(link); 81 + cgrp_kfunc_success__destroy(skel); 82 + } 83 + 84 + static const char * const success_tests[] = { 85 + "test_cgrp_acquire_release_argument", 86 + "test_cgrp_acquire_leave_in_map", 87 + "test_cgrp_xchg_release", 88 + "test_cgrp_get_release", 89 + }; 90 + 91 + static struct { 92 + const char *prog_name; 93 + const char *expected_err_msg; 94 + } failure_tests[] = { 95 + {"cgrp_kfunc_acquire_untrusted", "R1 must be referenced or trusted"}, 96 + {"cgrp_kfunc_acquire_fp", "arg#0 pointer type STRUCT cgroup must point"}, 97 + {"cgrp_kfunc_acquire_unsafe_kretprobe", "reg type unsupported for arg#0 function"}, 98 + {"cgrp_kfunc_acquire_trusted_walked", "R1 must be referenced or trusted"}, 99 + {"cgrp_kfunc_acquire_null", "arg#0 pointer type STRUCT cgroup must point"}, 100 + {"cgrp_kfunc_acquire_unreleased", "Unreleased reference"}, 101 + {"cgrp_kfunc_get_non_kptr_param", "arg#0 expected pointer to map value"}, 102 + {"cgrp_kfunc_get_non_kptr_acquired", "arg#0 expected pointer to map value"}, 103 + {"cgrp_kfunc_get_null", "arg#0 expected pointer to map value"}, 104 + {"cgrp_kfunc_xchg_unreleased", "Unreleased reference"}, 105 + {"cgrp_kfunc_get_unreleased", "Unreleased reference"}, 106 + {"cgrp_kfunc_release_untrusted", "arg#0 is untrusted_ptr_or_null_ expected ptr_ or socket"}, 107 + {"cgrp_kfunc_release_fp", "arg#0 pointer type STRUCT cgroup must point"}, 108 + {"cgrp_kfunc_release_null", "arg#0 is ptr_or_null_ expected ptr_ or socket"}, 109 + {"cgrp_kfunc_release_unacquired", "release kernel function bpf_cgroup_release expects"}, 110 + }; 111 + 112 + static void verify_fail(const char *prog_name, const char *expected_err_msg) 113 + { 114 + LIBBPF_OPTS(bpf_object_open_opts, opts); 115 + struct cgrp_kfunc_failure *skel; 116 + int err, i; 117 + 118 + opts.kernel_log_buf = obj_log_buf; 119 + opts.kernel_log_size = log_buf_sz; 120 + opts.kernel_log_level = 1; 121 + 122 + skel = cgrp_kfunc_failure__open_opts(&opts); 123 + if (!ASSERT_OK_PTR(skel, "cgrp_kfunc_failure__open_opts")) 124 + goto cleanup; 125 + 126 + for (i = 0; i < ARRAY_SIZE(failure_tests); i++) { 127 + struct bpf_program *prog; 128 + const char *curr_name = failure_tests[i].prog_name; 129 + 130 + prog = bpf_object__find_program_by_name(skel->obj, curr_name); 131 + if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name")) 132 + goto cleanup; 133 + 134 + bpf_program__set_autoload(prog, !strcmp(curr_name, prog_name)); 135 + } 136 + 137 + err = cgrp_kfunc_failure__load(skel); 138 + if (!ASSERT_ERR(err, "unexpected load success")) 139 + goto cleanup; 140 + 141 + if (!ASSERT_OK_PTR(strstr(obj_log_buf, expected_err_msg), "expected_err_msg")) { 142 + fprintf(stderr, "Expected err_msg: %s\n", expected_err_msg); 143 + fprintf(stderr, "Verifier output: %s\n", obj_log_buf); 144 + } 145 + 146 + cleanup: 147 + cgrp_kfunc_failure__destroy(skel); 148 + } 149 + 150 + void test_cgrp_kfunc(void) 151 + { 152 + int i, err; 153 + 154 + err = setup_cgroup_environment(); 155 + if (!ASSERT_OK(err, "cgrp_env_setup")) 156 + goto cleanup; 157 + 158 + for (i = 0; i < ARRAY_SIZE(success_tests); i++) { 159 + if (!test__start_subtest(success_tests[i])) 160 + continue; 161 + 162 + run_success_test(success_tests[i]); 163 + } 164 + 165 + for (i = 0; i < ARRAY_SIZE(failure_tests); i++) { 166 + if (!test__start_subtest(failure_tests[i].prog_name)) 167 + continue; 168 + 169 + verify_fail(failure_tests[i].prog_name, failure_tests[i].expected_err_msg); 170 + } 171 + 172 + cleanup: 173 + cleanup_cgroup_environment(); 174 + }
+71
tools/testing/selftests/bpf/progs/cgrp_kfunc_common.h
··· 1 + /* SPDX-License-Identifier: GPL-2.0 */ 2 + /* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */ 3 + 4 + #ifndef _CGRP_KFUNC_COMMON_H 5 + #define _CGRP_KFUNC_COMMON_H 6 + 7 + #include <errno.h> 8 + #include <vmlinux.h> 9 + #include <bpf/bpf_helpers.h> 10 + #include <bpf/bpf_tracing.h> 11 + 12 + struct __cgrps_kfunc_map_value { 13 + struct cgroup __kptr_ref * cgrp; 14 + }; 15 + 16 + struct hash_map { 17 + __uint(type, BPF_MAP_TYPE_HASH); 18 + __type(key, int); 19 + __type(value, struct __cgrps_kfunc_map_value); 20 + __uint(max_entries, 1); 21 + } __cgrps_kfunc_map SEC(".maps"); 22 + 23 + struct cgroup *bpf_cgroup_acquire(struct cgroup *p) __ksym; 24 + struct cgroup *bpf_cgroup_kptr_get(struct cgroup **pp) __ksym; 25 + void bpf_cgroup_release(struct cgroup *p) __ksym; 26 + 27 + static inline struct __cgrps_kfunc_map_value *cgrps_kfunc_map_value_lookup(struct cgroup *cgrp) 28 + { 29 + s32 id; 30 + long status; 31 + 32 + status = bpf_probe_read_kernel(&id, sizeof(id), &cgrp->self.id); 33 + if (status) 34 + return NULL; 35 + 36 + return bpf_map_lookup_elem(&__cgrps_kfunc_map, &id); 37 + } 38 + 39 + static inline int cgrps_kfunc_map_insert(struct cgroup *cgrp) 40 + { 41 + struct __cgrps_kfunc_map_value local, *v; 42 + long status; 43 + struct cgroup *acquired, *old; 44 + s32 id; 45 + 46 + status = bpf_probe_read_kernel(&id, sizeof(id), &cgrp->self.id); 47 + if (status) 48 + return status; 49 + 50 + local.cgrp = NULL; 51 + status = bpf_map_update_elem(&__cgrps_kfunc_map, &id, &local, BPF_NOEXIST); 52 + if (status) 53 + return status; 54 + 55 + v = bpf_map_lookup_elem(&__cgrps_kfunc_map, &id); 56 + if (!v) { 57 + bpf_map_delete_elem(&__cgrps_kfunc_map, &id); 58 + return -ENOENT; 59 + } 60 + 61 + acquired = bpf_cgroup_acquire(cgrp); 62 + old = bpf_kptr_xchg(&v->cgrp, acquired); 63 + if (old) { 64 + bpf_cgroup_release(old); 65 + return -EEXIST; 66 + } 67 + 68 + return 0; 69 + } 70 + 71 + #endif /* _CGRP_KFUNC_COMMON_H */
+260
tools/testing/selftests/bpf/progs/cgrp_kfunc_failure.c
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + /* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */ 3 + 4 + #include <vmlinux.h> 5 + #include <bpf/bpf_tracing.h> 6 + #include <bpf/bpf_helpers.h> 7 + 8 + #include "cgrp_kfunc_common.h" 9 + 10 + char _license[] SEC("license") = "GPL"; 11 + 12 + /* Prototype for all of the program trace events below: 13 + * 14 + * TRACE_EVENT(cgroup_mkdir, 15 + * TP_PROTO(struct cgroup *cgrp, const char *path), 16 + * TP_ARGS(cgrp, path) 17 + */ 18 + 19 + static struct __cgrps_kfunc_map_value *insert_lookup_cgrp(struct cgroup *cgrp) 20 + { 21 + int status; 22 + 23 + status = cgrps_kfunc_map_insert(cgrp); 24 + if (status) 25 + return NULL; 26 + 27 + return cgrps_kfunc_map_value_lookup(cgrp); 28 + } 29 + 30 + SEC("tp_btf/cgroup_mkdir") 31 + int BPF_PROG(cgrp_kfunc_acquire_untrusted, struct cgroup *cgrp, const char *path) 32 + { 33 + struct cgroup *acquired; 34 + struct __cgrps_kfunc_map_value *v; 35 + 36 + v = insert_lookup_cgrp(cgrp); 37 + if (!v) 38 + return 0; 39 + 40 + /* Can't invoke bpf_cgroup_acquire() on an untrusted pointer. */ 41 + acquired = bpf_cgroup_acquire(v->cgrp); 42 + bpf_cgroup_release(acquired); 43 + 44 + return 0; 45 + } 46 + 47 + SEC("tp_btf/cgroup_mkdir") 48 + int BPF_PROG(cgrp_kfunc_acquire_fp, struct cgroup *cgrp, const char *path) 49 + { 50 + struct cgroup *acquired, *stack_cgrp = (struct cgroup *)&path; 51 + 52 + /* Can't invoke bpf_cgroup_acquire() on a random frame pointer. */ 53 + acquired = bpf_cgroup_acquire((struct cgroup *)&stack_cgrp); 54 + bpf_cgroup_release(acquired); 55 + 56 + return 0; 57 + } 58 + 59 + SEC("kretprobe/cgroup_destroy_locked") 60 + int BPF_PROG(cgrp_kfunc_acquire_unsafe_kretprobe, struct cgroup *cgrp) 61 + { 62 + struct cgroup *acquired; 63 + 64 + /* Can't acquire an untrusted struct cgroup * pointer. */ 65 + acquired = bpf_cgroup_acquire(cgrp); 66 + bpf_cgroup_release(acquired); 67 + 68 + return 0; 69 + } 70 + 71 + SEC("tp_btf/cgroup_mkdir") 72 + int BPF_PROG(cgrp_kfunc_acquire_trusted_walked, struct cgroup *cgrp, const char *path) 73 + { 74 + struct cgroup *acquired; 75 + 76 + /* Can't invoke bpf_cgroup_acquire() on a pointer obtained from walking a trusted cgroup. */ 77 + acquired = bpf_cgroup_acquire(cgrp->old_dom_cgrp); 78 + bpf_cgroup_release(acquired); 79 + 80 + return 0; 81 + } 82 + 83 + 84 + SEC("tp_btf/cgroup_mkdir") 85 + int BPF_PROG(cgrp_kfunc_acquire_null, struct cgroup *cgrp, const char *path) 86 + { 87 + struct cgroup *acquired; 88 + 89 + /* Can't invoke bpf_cgroup_acquire() on a NULL pointer. */ 90 + acquired = bpf_cgroup_acquire(NULL); 91 + if (!acquired) 92 + return 0; 93 + bpf_cgroup_release(acquired); 94 + 95 + return 0; 96 + } 97 + 98 + SEC("tp_btf/cgroup_mkdir") 99 + int BPF_PROG(cgrp_kfunc_acquire_unreleased, struct cgroup *cgrp, const char *path) 100 + { 101 + struct cgroup *acquired; 102 + 103 + acquired = bpf_cgroup_acquire(cgrp); 104 + 105 + /* Acquired cgroup is never released. */ 106 + 107 + return 0; 108 + } 109 + 110 + SEC("tp_btf/cgroup_mkdir") 111 + int BPF_PROG(cgrp_kfunc_get_non_kptr_param, struct cgroup *cgrp, const char *path) 112 + { 113 + struct cgroup *kptr; 114 + 115 + /* Cannot use bpf_cgroup_kptr_get() on a non-kptr, even on a valid cgroup. */ 116 + kptr = bpf_cgroup_kptr_get(&cgrp); 117 + if (!kptr) 118 + return 0; 119 + 120 + bpf_cgroup_release(kptr); 121 + 122 + return 0; 123 + } 124 + 125 + SEC("tp_btf/cgroup_mkdir") 126 + int BPF_PROG(cgrp_kfunc_get_non_kptr_acquired, struct cgroup *cgrp, const char *path) 127 + { 128 + struct cgroup *kptr, *acquired; 129 + 130 + acquired = bpf_cgroup_acquire(cgrp); 131 + 132 + /* Cannot use bpf_cgroup_kptr_get() on a non-map-value, even if the kptr was acquired. */ 133 + kptr = bpf_cgroup_kptr_get(&acquired); 134 + bpf_cgroup_release(acquired); 135 + if (!kptr) 136 + return 0; 137 + 138 + bpf_cgroup_release(kptr); 139 + 140 + return 0; 141 + } 142 + 143 + SEC("tp_btf/cgroup_mkdir") 144 + int BPF_PROG(cgrp_kfunc_get_null, struct cgroup *cgrp, const char *path) 145 + { 146 + struct cgroup *kptr; 147 + 148 + /* Cannot use bpf_cgroup_kptr_get() on a NULL pointer. */ 149 + kptr = bpf_cgroup_kptr_get(NULL); 150 + if (!kptr) 151 + return 0; 152 + 153 + bpf_cgroup_release(kptr); 154 + 155 + return 0; 156 + } 157 + 158 + SEC("tp_btf/cgroup_mkdir") 159 + int BPF_PROG(cgrp_kfunc_xchg_unreleased, struct cgroup *cgrp, const char *path) 160 + { 161 + struct cgroup *kptr; 162 + struct __cgrps_kfunc_map_value *v; 163 + 164 + v = insert_lookup_cgrp(cgrp); 165 + if (!v) 166 + return 0; 167 + 168 + kptr = bpf_kptr_xchg(&v->cgrp, NULL); 169 + if (!kptr) 170 + return 0; 171 + 172 + /* Kptr retrieved from map is never released. */ 173 + 174 + return 0; 175 + } 176 + 177 + SEC("tp_btf/cgroup_mkdir") 178 + int BPF_PROG(cgrp_kfunc_get_unreleased, struct cgroup *cgrp, const char *path) 179 + { 180 + struct cgroup *kptr; 181 + struct __cgrps_kfunc_map_value *v; 182 + 183 + v = insert_lookup_cgrp(cgrp); 184 + if (!v) 185 + return 0; 186 + 187 + kptr = bpf_cgroup_kptr_get(&v->cgrp); 188 + if (!kptr) 189 + return 0; 190 + 191 + /* Kptr acquired above is never released. */ 192 + 193 + return 0; 194 + } 195 + 196 + SEC("tp_btf/cgroup_mkdir") 197 + int BPF_PROG(cgrp_kfunc_release_untrusted, struct cgroup *cgrp, const char *path) 198 + { 199 + struct __cgrps_kfunc_map_value *v; 200 + 201 + v = insert_lookup_cgrp(cgrp); 202 + if (!v) 203 + return 0; 204 + 205 + /* Can't invoke bpf_cgroup_release() on an untrusted pointer. */ 206 + bpf_cgroup_release(v->cgrp); 207 + 208 + return 0; 209 + } 210 + 211 + SEC("tp_btf/cgroup_mkdir") 212 + int BPF_PROG(cgrp_kfunc_release_fp, struct cgroup *cgrp, const char *path) 213 + { 214 + struct cgroup *acquired = (struct cgroup *)&path; 215 + 216 + /* Cannot release random frame pointer. */ 217 + bpf_cgroup_release(acquired); 218 + 219 + return 0; 220 + } 221 + 222 + SEC("tp_btf/cgroup_mkdir") 223 + int BPF_PROG(cgrp_kfunc_release_null, struct cgroup *cgrp, const char *path) 224 + { 225 + struct __cgrps_kfunc_map_value local, *v; 226 + long status; 227 + struct cgroup *acquired, *old; 228 + s32 id; 229 + 230 + status = bpf_probe_read_kernel(&id, sizeof(id), &cgrp->self.id); 231 + if (status) 232 + return 0; 233 + 234 + local.cgrp = NULL; 235 + status = bpf_map_update_elem(&__cgrps_kfunc_map, &id, &local, BPF_NOEXIST); 236 + if (status) 237 + return status; 238 + 239 + v = bpf_map_lookup_elem(&__cgrps_kfunc_map, &id); 240 + if (!v) 241 + return -ENOENT; 242 + 243 + acquired = bpf_cgroup_acquire(cgrp); 244 + 245 + old = bpf_kptr_xchg(&v->cgrp, acquired); 246 + 247 + /* old cannot be passed to bpf_cgroup_release() without a NULL check. */ 248 + bpf_cgroup_release(old); 249 + 250 + return 0; 251 + } 252 + 253 + SEC("tp_btf/cgroup_mkdir") 254 + int BPF_PROG(cgrp_kfunc_release_unacquired, struct cgroup *cgrp, const char *path) 255 + { 256 + /* Cannot release trusted cgroup pointer which was not acquired. */ 257 + bpf_cgroup_release(cgrp); 258 + 259 + return 0; 260 + }
+125
tools/testing/selftests/bpf/progs/cgrp_kfunc_success.c
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + /* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */ 3 + 4 + #include <vmlinux.h> 5 + #include <bpf/bpf_tracing.h> 6 + #include <bpf/bpf_helpers.h> 7 + 8 + #include "cgrp_kfunc_common.h" 9 + 10 + char _license[] SEC("license") = "GPL"; 11 + 12 + int err, pid, invocations; 13 + 14 + /* Prototype for all of the program trace events below: 15 + * 16 + * TRACE_EVENT(cgroup_mkdir, 17 + * TP_PROTO(struct cgroup *cgrp, const char *path), 18 + * TP_ARGS(cgrp, path) 19 + */ 20 + 21 + static bool is_test_kfunc_task(void) 22 + { 23 + int cur_pid = bpf_get_current_pid_tgid() >> 32; 24 + bool same = pid == cur_pid; 25 + 26 + if (same) 27 + __sync_fetch_and_add(&invocations, 1); 28 + 29 + return same; 30 + } 31 + 32 + SEC("tp_btf/cgroup_mkdir") 33 + int BPF_PROG(test_cgrp_acquire_release_argument, struct cgroup *cgrp, const char *path) 34 + { 35 + struct cgroup *acquired; 36 + 37 + if (!is_test_kfunc_task()) 38 + return 0; 39 + 40 + acquired = bpf_cgroup_acquire(cgrp); 41 + bpf_cgroup_release(acquired); 42 + 43 + return 0; 44 + } 45 + 46 + SEC("tp_btf/cgroup_mkdir") 47 + int BPF_PROG(test_cgrp_acquire_leave_in_map, struct cgroup *cgrp, const char *path) 48 + { 49 + long status; 50 + 51 + if (!is_test_kfunc_task()) 52 + return 0; 53 + 54 + status = cgrps_kfunc_map_insert(cgrp); 55 + if (status) 56 + err = 1; 57 + 58 + return 0; 59 + } 60 + 61 + SEC("tp_btf/cgroup_mkdir") 62 + int BPF_PROG(test_cgrp_xchg_release, struct cgroup *cgrp, const char *path) 63 + { 64 + struct cgroup *kptr; 65 + struct __cgrps_kfunc_map_value *v; 66 + long status; 67 + 68 + if (!is_test_kfunc_task()) 69 + return 0; 70 + 71 + status = cgrps_kfunc_map_insert(cgrp); 72 + if (status) { 73 + err = 1; 74 + return 0; 75 + } 76 + 77 + v = cgrps_kfunc_map_value_lookup(cgrp); 78 + if (!v) { 79 + err = 2; 80 + return 0; 81 + } 82 + 83 + kptr = bpf_kptr_xchg(&v->cgrp, NULL); 84 + if (!kptr) { 85 + err = 3; 86 + return 0; 87 + } 88 + 89 + bpf_cgroup_release(kptr); 90 + 91 + return 0; 92 + } 93 + 94 + SEC("tp_btf/cgroup_mkdir") 95 + int BPF_PROG(test_cgrp_get_release, struct cgroup *cgrp, const char *path) 96 + { 97 + struct cgroup *kptr; 98 + struct __cgrps_kfunc_map_value *v; 99 + long status; 100 + 101 + if (!is_test_kfunc_task()) 102 + return 0; 103 + 104 + status = cgrps_kfunc_map_insert(cgrp); 105 + if (status) { 106 + err = 1; 107 + return 0; 108 + } 109 + 110 + v = cgrps_kfunc_map_value_lookup(cgrp); 111 + if (!v) { 112 + err = 2; 113 + return 0; 114 + } 115 + 116 + kptr = bpf_cgroup_kptr_get(&v->cgrp); 117 + if (!kptr) { 118 + err = 3; 119 + return 0; 120 + } 121 + 122 + bpf_cgroup_release(kptr); 123 + 124 + return 0; 125 + }