// SPDX-License-Identifier: GPL-2.0 /* Copyright (c) 2019 Facebook */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include static struct bpf_create_map_attr xattr = { .name = "sk_storage_map", .map_type = BPF_MAP_TYPE_SK_STORAGE, .map_flags = BPF_F_NO_PREALLOC, .max_entries = 0, .key_size = 4, .value_size = 8, .btf_key_type_id = 1, .btf_value_type_id = 3, .btf_fd = -1, }; static unsigned int nr_sk_threads_done; static unsigned int nr_sk_threads_err; static unsigned int nr_sk_per_thread = 4096; static unsigned int nr_sk_threads = 4; static int sk_storage_map = -1; static unsigned int stop; static int runtime_s = 5; static bool is_stopped(void) { return READ_ONCE(stop); } static unsigned int threads_err(void) { return READ_ONCE(nr_sk_threads_err); } static void notify_thread_err(void) { __sync_add_and_fetch(&nr_sk_threads_err, 1); } static bool wait_for_threads_err(void) { while (!is_stopped() && !threads_err()) usleep(500); return !is_stopped(); } static unsigned int threads_done(void) { return READ_ONCE(nr_sk_threads_done); } static void notify_thread_done(void) { __sync_add_and_fetch(&nr_sk_threads_done, 1); } static void notify_thread_redo(void) { __sync_sub_and_fetch(&nr_sk_threads_done, 1); } static bool wait_for_threads_done(void) { while (threads_done() != nr_sk_threads && !is_stopped() && !threads_err()) usleep(50); return !is_stopped() && !threads_err(); } static bool wait_for_threads_redo(void) { while (threads_done() && !is_stopped() && !threads_err()) usleep(50); return !is_stopped() && !threads_err(); } static bool wait_for_map(void) { while (READ_ONCE(sk_storage_map) == -1 && !is_stopped()) usleep(50); return !is_stopped(); } static bool wait_for_map_close(void) { while (READ_ONCE(sk_storage_map) != -1 && !is_stopped()) ; return !is_stopped(); } static int load_btf(void) { const char btf_str_sec[] = "\0bpf_spin_lock\0val\0cnt\0l"; __u32 btf_raw_types[] = { /* int */ BTF_TYPE_INT_ENC(0, BTF_INT_SIGNED, 0, 32, 4), /* [1] */ /* struct bpf_spin_lock */ /* [2] */ BTF_TYPE_ENC(1, BTF_INFO_ENC(BTF_KIND_STRUCT, 0, 1), 4), BTF_MEMBER_ENC(15, 1, 0), /* int val; */ /* struct val */ /* [3] */ BTF_TYPE_ENC(15, BTF_INFO_ENC(BTF_KIND_STRUCT, 0, 2), 8), BTF_MEMBER_ENC(19, 1, 0), /* int cnt; */ BTF_MEMBER_ENC(23, 2, 32),/* struct bpf_spin_lock l; */ }; struct btf_header btf_hdr = { .magic = BTF_MAGIC, .version = BTF_VERSION, .hdr_len = sizeof(struct btf_header), .type_len = sizeof(btf_raw_types), .str_off = sizeof(btf_raw_types), .str_len = sizeof(btf_str_sec), }; __u8 raw_btf[sizeof(struct btf_header) + sizeof(btf_raw_types) + sizeof(btf_str_sec)]; memcpy(raw_btf, &btf_hdr, sizeof(btf_hdr)); memcpy(raw_btf + sizeof(btf_hdr), btf_raw_types, sizeof(btf_raw_types)); memcpy(raw_btf + sizeof(btf_hdr) + sizeof(btf_raw_types), btf_str_sec, sizeof(btf_str_sec)); return bpf_load_btf(raw_btf, sizeof(raw_btf), 0, 0, 0); } static int create_sk_storage_map(void) { int btf_fd, map_fd; btf_fd = load_btf(); CHECK(btf_fd == -1, "bpf_load_btf", "btf_fd:%d errno:%d\n", btf_fd, errno); xattr.btf_fd = btf_fd; map_fd = bpf_create_map_xattr(&xattr); xattr.btf_fd = -1; close(btf_fd); CHECK(map_fd == -1, "bpf_create_map_xattr()", "errno:%d\n", errno); return map_fd; } static void *insert_close_thread(void *arg) { struct { int cnt; int lock; } value = { .cnt = 0xeB9F, .lock = 0, }; int i, map_fd, err, *sk_fds; sk_fds = malloc(sizeof(*sk_fds) * nr_sk_per_thread); if (!sk_fds) { notify_thread_err(); return ERR_PTR(-ENOMEM); } for (i = 0; i < nr_sk_per_thread; i++) sk_fds[i] = -1; while (!is_stopped()) { if (!wait_for_map()) goto close_all; map_fd = READ_ONCE(sk_storage_map); for (i = 0; i < nr_sk_per_thread && !is_stopped(); i++) { sk_fds[i] = socket(AF_INET6, SOCK_STREAM, 0); if (sk_fds[i] == -1) { err = -errno; fprintf(stderr, "socket(): errno:%d\n", errno); goto errout; } err = bpf_map_update_elem(map_fd, &sk_fds[i], &value, BPF_NOEXIST); if (err) { err = -errno; fprintf(stderr, "bpf_map_update_elem(): errno:%d\n", errno); goto errout; } } notify_thread_done(); wait_for_map_close(); close_all: for (i = 0; i < nr_sk_per_thread; i++) { close(sk_fds[i]); sk_fds[i] = -1; } notify_thread_redo(); } free(sk_fds); return NULL; errout: for (i = 0; i < nr_sk_per_thread && sk_fds[i] != -1; i++) close(sk_fds[i]); free(sk_fds); notify_thread_err(); return ERR_PTR(err); } static int do_sk_storage_map_stress_free(void) { int i, map_fd = -1, err = 0, nr_threads_created = 0; pthread_t *sk_thread_ids; void *thread_ret; sk_thread_ids = malloc(sizeof(pthread_t) * nr_sk_threads); if (!sk_thread_ids) { fprintf(stderr, "malloc(sk_threads): NULL\n"); return -ENOMEM; } for (i = 0; i < nr_sk_threads; i++) { err = pthread_create(&sk_thread_ids[i], NULL, insert_close_thread, NULL); if (err) { err = -errno; goto done; } nr_threads_created++; } while (!is_stopped()) { map_fd = create_sk_storage_map(); WRITE_ONCE(sk_storage_map, map_fd); if (!wait_for_threads_done()) break; WRITE_ONCE(sk_storage_map, -1); close(map_fd); map_fd = -1; if (!wait_for_threads_redo()) break; } done: WRITE_ONCE(stop, 1); for (i = 0; i < nr_threads_created; i++) { pthread_join(sk_thread_ids[i], &thread_ret); if (IS_ERR(thread_ret) && !err) { err = PTR_ERR(thread_ret); fprintf(stderr, "threads#%u: err:%d\n", i, err); } } free(sk_thread_ids); if (map_fd != -1) close(map_fd); return err; } static void *update_thread(void *arg) { struct { int cnt; int lock; } value = { .cnt = 0xeB9F, .lock = 0, }; int map_fd = READ_ONCE(sk_storage_map); int sk_fd = *(int *)arg; int err = 0; /* Suppress compiler false alarm */ while (!is_stopped()) { err = bpf_map_update_elem(map_fd, &sk_fd, &value, 0); if (err && errno != EAGAIN) { err = -errno; fprintf(stderr, "bpf_map_update_elem: %d %d\n", err, errno); break; } } if (!is_stopped()) { notify_thread_err(); return ERR_PTR(err); } return NULL; } static void *delete_thread(void *arg) { int map_fd = READ_ONCE(sk_storage_map); int sk_fd = *(int *)arg; int err = 0; /* Suppress compiler false alarm */ while (!is_stopped()) { err = bpf_map_delete_elem(map_fd, &sk_fd); if (err && errno != ENOENT) { err = -errno; fprintf(stderr, "bpf_map_delete_elem: %d %d\n", err, errno); break; } } if (!is_stopped()) { notify_thread_err(); return ERR_PTR(err); } return NULL; } static int do_sk_storage_map_stress_change(void) { int i, sk_fd, map_fd = -1, err = 0, nr_threads_created = 0; pthread_t *sk_thread_ids; void *thread_ret; sk_thread_ids = malloc(sizeof(pthread_t) * nr_sk_threads); if (!sk_thread_ids) { fprintf(stderr, "malloc(sk_threads): NULL\n"); return -ENOMEM; } sk_fd = socket(AF_INET6, SOCK_STREAM, 0); if (sk_fd == -1) { err = -errno; goto done; } map_fd = create_sk_storage_map(); WRITE_ONCE(sk_storage_map, map_fd); for (i = 0; i < nr_sk_threads; i++) { if (i & 0x1) err = pthread_create(&sk_thread_ids[i], NULL, update_thread, &sk_fd); else err = pthread_create(&sk_thread_ids[i], NULL, delete_thread, &sk_fd); if (err) { err = -errno; goto done; } nr_threads_created++; } wait_for_threads_err(); done: WRITE_ONCE(stop, 1); for (i = 0; i < nr_threads_created; i++) { pthread_join(sk_thread_ids[i], &thread_ret); if (IS_ERR(thread_ret) && !err) { err = PTR_ERR(thread_ret); fprintf(stderr, "threads#%u: err:%d\n", i, err); } } free(sk_thread_ids); if (sk_fd != -1) close(sk_fd); close(map_fd); return err; } static void stop_handler(int signum) { if (signum != SIGALRM) printf("stopping...\n"); WRITE_ONCE(stop, 1); } #define BPF_SK_STORAGE_MAP_TEST_NR_THREADS "BPF_SK_STORAGE_MAP_TEST_NR_THREADS" #define BPF_SK_STORAGE_MAP_TEST_SK_PER_THREAD "BPF_SK_STORAGE_MAP_TEST_SK_PER_THREAD" #define BPF_SK_STORAGE_MAP_TEST_RUNTIME_S "BPF_SK_STORAGE_MAP_TEST_RUNTIME_S" #define BPF_SK_STORAGE_MAP_TEST_NAME "BPF_SK_STORAGE_MAP_TEST_NAME" static void test_sk_storage_map_stress_free(void) { struct rlimit rlim_old, rlim_new = {}; int err; getrlimit(RLIMIT_NOFILE, &rlim_old); signal(SIGTERM, stop_handler); signal(SIGINT, stop_handler); if (runtime_s > 0) { signal(SIGALRM, stop_handler); alarm(runtime_s); } if (rlim_old.rlim_cur < nr_sk_threads * nr_sk_per_thread) { rlim_new.rlim_cur = nr_sk_threads * nr_sk_per_thread + 128; rlim_new.rlim_max = rlim_new.rlim_cur + 128; err = setrlimit(RLIMIT_NOFILE, &rlim_new); CHECK(err, "setrlimit(RLIMIT_NOFILE)", "rlim_new:%lu errno:%d", rlim_new.rlim_cur, errno); } err = do_sk_storage_map_stress_free(); signal(SIGTERM, SIG_DFL); signal(SIGINT, SIG_DFL); if (runtime_s > 0) { signal(SIGALRM, SIG_DFL); alarm(0); } if (rlim_new.rlim_cur) setrlimit(RLIMIT_NOFILE, &rlim_old); CHECK(err, "test_sk_storage_map_stress_free", "err:%d\n", err); } static void test_sk_storage_map_stress_change(void) { int err; signal(SIGTERM, stop_handler); signal(SIGINT, stop_handler); if (runtime_s > 0) { signal(SIGALRM, stop_handler); alarm(runtime_s); } err = do_sk_storage_map_stress_change(); signal(SIGTERM, SIG_DFL); signal(SIGINT, SIG_DFL); if (runtime_s > 0) { signal(SIGALRM, SIG_DFL); alarm(0); } CHECK(err, "test_sk_storage_map_stress_change", "err:%d\n", err); } static void test_sk_storage_map_basic(void) { struct { int cnt; int lock; } value = { .cnt = 0xeB9f, .lock = 0, }, lookup_value; struct bpf_create_map_attr bad_xattr; int btf_fd, map_fd, sk_fd, err; btf_fd = load_btf(); CHECK(btf_fd == -1, "bpf_load_btf", "btf_fd:%d errno:%d\n", btf_fd, errno); xattr.btf_fd = btf_fd; sk_fd = socket(AF_INET6, SOCK_STREAM, 0); CHECK(sk_fd == -1, "socket()", "sk_fd:%d errno:%d\n", sk_fd, errno); map_fd = bpf_create_map_xattr(&xattr); CHECK(map_fd == -1, "bpf_create_map_xattr(good_xattr)", "map_fd:%d errno:%d\n", map_fd, errno); /* Add new elem */ memcpy(&lookup_value, &value, sizeof(value)); err = bpf_map_update_elem(map_fd, &sk_fd, &value, BPF_NOEXIST | BPF_F_LOCK); CHECK(err, "bpf_map_update_elem(BPF_NOEXIST|BPF_F_LOCK)", "err:%d errno:%d\n", err, errno); err = bpf_map_lookup_elem_flags(map_fd, &sk_fd, &lookup_value, BPF_F_LOCK); CHECK(err || lookup_value.cnt != value.cnt, "bpf_map_lookup_elem_flags(BPF_F_LOCK)", "err:%d errno:%d cnt:%x(%x)\n", err, errno, lookup_value.cnt, value.cnt); /* Bump the cnt and update with BPF_EXIST | BPF_F_LOCK */ value.cnt += 1; err = bpf_map_update_elem(map_fd, &sk_fd, &value, BPF_EXIST | BPF_F_LOCK); CHECK(err, "bpf_map_update_elem(BPF_EXIST|BPF_F_LOCK)", "err:%d errno:%d\n", err, errno); err = bpf_map_lookup_elem_flags(map_fd, &sk_fd, &lookup_value, BPF_F_LOCK); CHECK(err || lookup_value.cnt != value.cnt, "bpf_map_lookup_elem_flags(BPF_F_LOCK)", "err:%d errno:%d cnt:%x(%x)\n", err, errno, lookup_value.cnt, value.cnt); /* Bump the cnt and update with BPF_EXIST */ value.cnt += 1; err = bpf_map_update_elem(map_fd, &sk_fd, &value, BPF_EXIST); CHECK(err, "bpf_map_update_elem(BPF_EXIST)", "err:%d errno:%d\n", err, errno); err = bpf_map_lookup_elem_flags(map_fd, &sk_fd, &lookup_value, BPF_F_LOCK); CHECK(err || lookup_value.cnt != value.cnt, "bpf_map_lookup_elem_flags(BPF_F_LOCK)", "err:%d errno:%d cnt:%x(%x)\n", err, errno, lookup_value.cnt, value.cnt); /* Update with BPF_NOEXIST */ value.cnt += 1; err = bpf_map_update_elem(map_fd, &sk_fd, &value, BPF_NOEXIST | BPF_F_LOCK); CHECK(!err || errno != EEXIST, "bpf_map_update_elem(BPF_NOEXIST|BPF_F_LOCK)", "err:%d errno:%d\n", err, errno); err = bpf_map_update_elem(map_fd, &sk_fd, &value, BPF_NOEXIST); CHECK(!err || errno != EEXIST, "bpf_map_update_elem(BPF_NOEXIST)", "err:%d errno:%d\n", err, errno); value.cnt -= 1; err = bpf_map_lookup_elem_flags(map_fd, &sk_fd, &lookup_value, BPF_F_LOCK); CHECK(err || lookup_value.cnt != value.cnt, "bpf_map_lookup_elem_flags(BPF_F_LOCK)", "err:%d errno:%d cnt:%x(%x)\n", err, errno, lookup_value.cnt, value.cnt); /* Bump the cnt again and update with map_flags == 0 */ value.cnt += 1; err = bpf_map_update_elem(map_fd, &sk_fd, &value, 0); CHECK(err, "bpf_map_update_elem()", "err:%d errno:%d\n", err, errno); err = bpf_map_lookup_elem_flags(map_fd, &sk_fd, &lookup_value, BPF_F_LOCK); CHECK(err || lookup_value.cnt != value.cnt, "bpf_map_lookup_elem_flags(BPF_F_LOCK)", "err:%d errno:%d cnt:%x(%x)\n", err, errno, lookup_value.cnt, value.cnt); /* Test delete elem */ err = bpf_map_delete_elem(map_fd, &sk_fd); CHECK(err, "bpf_map_delete_elem()", "err:%d errno:%d\n", err, errno); err = bpf_map_lookup_elem_flags(map_fd, &sk_fd, &lookup_value, BPF_F_LOCK); CHECK(!err || errno != ENOENT, "bpf_map_lookup_elem_flags(BPF_F_LOCK)", "err:%d errno:%d\n", err, errno); err = bpf_map_delete_elem(map_fd, &sk_fd); CHECK(!err || errno != ENOENT, "bpf_map_delete_elem()", "err:%d errno:%d\n", err, errno); memcpy(&bad_xattr, &xattr, sizeof(xattr)); bad_xattr.btf_key_type_id = 0; err = bpf_create_map_xattr(&bad_xattr); CHECK(!err || errno != EINVAL, "bap_create_map_xattr(bad_xattr)", "err:%d errno:%d\n", err, errno); memcpy(&bad_xattr, &xattr, sizeof(xattr)); bad_xattr.btf_key_type_id = 3; err = bpf_create_map_xattr(&bad_xattr); CHECK(!err || errno != EINVAL, "bap_create_map_xattr(bad_xattr)", "err:%d errno:%d\n", err, errno); memcpy(&bad_xattr, &xattr, sizeof(xattr)); bad_xattr.max_entries = 1; err = bpf_create_map_xattr(&bad_xattr); CHECK(!err || errno != EINVAL, "bap_create_map_xattr(bad_xattr)", "err:%d errno:%d\n", err, errno); memcpy(&bad_xattr, &xattr, sizeof(xattr)); bad_xattr.map_flags = 0; err = bpf_create_map_xattr(&bad_xattr); CHECK(!err || errno != EINVAL, "bap_create_map_xattr(bad_xattr)", "err:%d errno:%d\n", err, errno); xattr.btf_fd = -1; close(btf_fd); close(map_fd); close(sk_fd); } void test_sk_storage_map(void) { const char *test_name, *env_opt; bool test_ran = false; test_name = getenv(BPF_SK_STORAGE_MAP_TEST_NAME); env_opt = getenv(BPF_SK_STORAGE_MAP_TEST_NR_THREADS); if (env_opt) nr_sk_threads = atoi(env_opt); env_opt = getenv(BPF_SK_STORAGE_MAP_TEST_SK_PER_THREAD); if (env_opt) nr_sk_per_thread = atoi(env_opt); env_opt = getenv(BPF_SK_STORAGE_MAP_TEST_RUNTIME_S); if (env_opt) runtime_s = atoi(env_opt); if (!test_name || !strcmp(test_name, "basic")) { test_sk_storage_map_basic(); test_ran = true; } if (!test_name || !strcmp(test_name, "stress_free")) { test_sk_storage_map_stress_free(); test_ran = true; } if (!test_name || !strcmp(test_name, "stress_change")) { test_sk_storage_map_stress_change(); test_ran = true; } if (test_ran) printf("%s:PASS\n", __func__); else CHECK(1, "Invalid test_name", "%s\n", test_name); }