// SPDX-License-Identifier: GPL-2.0 /* Copyright (c) 2023 Meta Platforms, Inc. and affiliates. */ #include #include #include #include #include "bpf_misc.h" #include "bpf_experimental.h" extern void bpf_rcu_read_lock(void) __ksym; extern void bpf_rcu_read_unlock(void) __ksym; struct node_data { long key; long list_data; struct bpf_rb_node r; struct bpf_list_node l; struct bpf_refcount ref; }; struct map_value { struct node_data __kptr *node; }; struct { __uint(type, BPF_MAP_TYPE_ARRAY); __type(key, int); __type(value, struct map_value); __uint(max_entries, 2); } stashed_nodes SEC(".maps"); struct node_acquire { long key; long data; struct bpf_rb_node node; struct bpf_refcount refcount; }; #define private(name) SEC(".bss." #name) __hidden __attribute__((aligned(8))) private(A) struct bpf_spin_lock lock; private(A) struct bpf_rb_root root __contains(node_data, r); private(A) struct bpf_list_head head __contains(node_data, l); private(B) struct bpf_spin_lock alock; private(B) struct bpf_rb_root aroot __contains(node_acquire, node); private(C) struct bpf_spin_lock block; private(C) struct bpf_rb_root broot __contains(node_data, r); static bool less(struct bpf_rb_node *node_a, const struct bpf_rb_node *node_b) { struct node_data *a; struct node_data *b; a = container_of(node_a, struct node_data, r); b = container_of(node_b, struct node_data, r); return a->key < b->key; } static bool less_a(struct bpf_rb_node *a, const struct bpf_rb_node *b) { struct node_acquire *node_a; struct node_acquire *node_b; node_a = container_of(a, struct node_acquire, node); node_b = container_of(b, struct node_acquire, node); return node_a->key < node_b->key; } static long __insert_in_tree_and_list(struct bpf_list_head *head, struct bpf_rb_root *root, struct bpf_spin_lock *lock) { struct node_data *n, *m; n = bpf_obj_new(typeof(*n)); if (!n) return -1; m = bpf_refcount_acquire(n); m->key = 123; m->list_data = 456; bpf_spin_lock(lock); if (bpf_rbtree_add(root, &n->r, less)) { /* Failure to insert - unexpected */ bpf_spin_unlock(lock); bpf_obj_drop(m); return -2; } bpf_spin_unlock(lock); bpf_spin_lock(lock); if (bpf_list_push_front(head, &m->l)) { /* Failure to insert - unexpected */ bpf_spin_unlock(lock); return -3; } bpf_spin_unlock(lock); return 0; } static long __stash_map_insert_tree(int idx, int val, struct bpf_rb_root *root, struct bpf_spin_lock *lock) { struct map_value *mapval; struct node_data *n, *m; mapval = bpf_map_lookup_elem(&stashed_nodes, &idx); if (!mapval) return -1; n = bpf_obj_new(typeof(*n)); if (!n) return -2; n->key = val; m = bpf_refcount_acquire(n); n = bpf_kptr_xchg(&mapval->node, n); if (n) { bpf_obj_drop(n); bpf_obj_drop(m); return -3; } bpf_spin_lock(lock); if (bpf_rbtree_add(root, &m->r, less)) { /* Failure to insert - unexpected */ bpf_spin_unlock(lock); return -4; } bpf_spin_unlock(lock); return 0; } static long __read_from_tree(struct bpf_rb_root *root, struct bpf_spin_lock *lock, bool remove_from_tree) { struct bpf_rb_node *rb; struct node_data *n; long res = -99; bpf_spin_lock(lock); rb = bpf_rbtree_first(root); if (!rb) { bpf_spin_unlock(lock); return -1; } n = container_of(rb, struct node_data, r); res = n->key; if (!remove_from_tree) { bpf_spin_unlock(lock); return res; } rb = bpf_rbtree_remove(root, rb); bpf_spin_unlock(lock); if (!rb) return -2; n = container_of(rb, struct node_data, r); bpf_obj_drop(n); return res; } static long __read_from_list(struct bpf_list_head *head, struct bpf_spin_lock *lock, bool remove_from_list) { struct bpf_list_node *l; struct node_data *n; long res = -99; bpf_spin_lock(lock); l = bpf_list_pop_front(head); if (!l) { bpf_spin_unlock(lock); return -1; } n = container_of(l, struct node_data, l); res = n->list_data; if (!remove_from_list) { if (bpf_list_push_back(head, &n->l)) { bpf_spin_unlock(lock); return -2; } } bpf_spin_unlock(lock); if (remove_from_list) bpf_obj_drop(n); return res; } static long __read_from_unstash(int idx) { struct node_data *n = NULL; struct map_value *mapval; long val = -99; mapval = bpf_map_lookup_elem(&stashed_nodes, &idx); if (!mapval) return -1; n = bpf_kptr_xchg(&mapval->node, n); if (!n) return -2; val = n->key; bpf_obj_drop(n); return val; } #define INSERT_READ_BOTH(rem_tree, rem_list, desc) \ SEC("tc") \ __description(desc) \ __success __retval(579) \ long insert_and_remove_tree_##rem_tree##_list_##rem_list(void *ctx) \ { \ long err, tree_data, list_data; \ \ err = __insert_in_tree_and_list(&head, &root, &lock); \ if (err) \ return err; \ \ err = __read_from_tree(&root, &lock, rem_tree); \ if (err < 0) \ return err; \ else \ tree_data = err; \ \ err = __read_from_list(&head, &lock, rem_list); \ if (err < 0) \ return err; \ else \ list_data = err; \ \ return tree_data + list_data; \ } /* After successful insert of struct node_data into both collections: * - it should have refcount = 2 * - removing / not removing the node_data from a collection after * reading should have no effect on ability to read / remove from * the other collection */ INSERT_READ_BOTH(true, true, "insert_read_both: remove from tree + list"); INSERT_READ_BOTH(false, false, "insert_read_both: remove from neither"); INSERT_READ_BOTH(true, false, "insert_read_both: remove from tree"); INSERT_READ_BOTH(false, true, "insert_read_both: remove from list"); #undef INSERT_READ_BOTH #define INSERT_READ_BOTH(rem_tree, rem_list, desc) \ SEC("tc") \ __description(desc) \ __success __retval(579) \ long insert_and_remove_lf_tree_##rem_tree##_list_##rem_list(void *ctx) \ { \ long err, tree_data, list_data; \ \ err = __insert_in_tree_and_list(&head, &root, &lock); \ if (err) \ return err; \ \ err = __read_from_list(&head, &lock, rem_list); \ if (err < 0) \ return err; \ else \ list_data = err; \ \ err = __read_from_tree(&root, &lock, rem_tree); \ if (err < 0) \ return err; \ else \ tree_data = err; \ \ return tree_data + list_data; \ } /* Similar to insert_read_both, but list data is read and possibly removed * first * * Results should be no different than reading and possibly removing rbtree * node first */ INSERT_READ_BOTH(true, true, "insert_read_both_list_first: remove from tree + list"); INSERT_READ_BOTH(false, false, "insert_read_both_list_first: remove from neither"); INSERT_READ_BOTH(true, false, "insert_read_both_list_first: remove from tree"); INSERT_READ_BOTH(false, true, "insert_read_both_list_first: remove from list"); #define INSERT_DOUBLE_READ_AND_DEL(read_fn, read_root, desc) \ SEC("tc") \ __description(desc) \ __success __retval(-1) \ long insert_double_##read_fn##_and_del_##read_root(void *ctx) \ { \ long err, list_data; \ \ err = __insert_in_tree_and_list(&head, &root, &lock); \ if (err) \ return err; \ \ err = read_fn(&read_root, &lock, true); \ if (err < 0) \ return err; \ else \ list_data = err; \ \ err = read_fn(&read_root, &lock, true); \ if (err < 0) \ return err; \ \ return err + list_data; \ } /* Insert into both tree and list, then try reading-and-removing from either twice * * The second read-and-remove should fail on read step since the node has * already been removed */ INSERT_DOUBLE_READ_AND_DEL(__read_from_tree, root, "insert_double_del: 2x read-and-del from tree"); INSERT_DOUBLE_READ_AND_DEL(__read_from_list, head, "insert_double_del: 2x read-and-del from list"); #define INSERT_STASH_READ(rem_tree, desc) \ SEC("tc") \ __description(desc) \ __success __retval(84) \ long insert_rbtree_and_stash__del_tree_##rem_tree(void *ctx) \ { \ long err, tree_data, map_data; \ \ err = __stash_map_insert_tree(0, 42, &root, &lock); \ if (err) \ return err; \ \ err = __read_from_tree(&root, &lock, rem_tree); \ if (err < 0) \ return err; \ else \ tree_data = err; \ \ err = __read_from_unstash(0); \ if (err < 0) \ return err; \ else \ map_data = err; \ \ return tree_data + map_data; \ } /* Stash a refcounted node in map_val, insert same node into tree, then try * reading data from tree then unstashed map_val, possibly removing from tree * * Removing from tree should have no effect on map_val kptr validity */ INSERT_STASH_READ(true, "insert_stash_read: remove from tree"); INSERT_STASH_READ(false, "insert_stash_read: don't remove from tree"); SEC("tc") __success long rbtree_refcounted_node_ref_escapes(void *ctx) { struct node_acquire *n, *m; n = bpf_obj_new(typeof(*n)); if (!n) return 1; bpf_spin_lock(&alock); bpf_rbtree_add(&aroot, &n->node, less_a); m = bpf_refcount_acquire(n); bpf_spin_unlock(&alock); if (!m) return 2; m->key = 2; bpf_obj_drop(m); return 0; } SEC("tc") __success long rbtree_refcounted_node_ref_escapes_owning_input(void *ctx) { struct node_acquire *n, *m; n = bpf_obj_new(typeof(*n)); if (!n) return 1; m = bpf_refcount_acquire(n); m->key = 2; bpf_spin_lock(&alock); bpf_rbtree_add(&aroot, &n->node, less_a); bpf_spin_unlock(&alock); bpf_obj_drop(m); return 0; } static long __stash_map_empty_xchg(struct node_data *n, int idx) { struct map_value *mapval = bpf_map_lookup_elem(&stashed_nodes, &idx); if (!mapval) { bpf_obj_drop(n); return 1; } n = bpf_kptr_xchg(&mapval->node, n); if (n) { bpf_obj_drop(n); return 2; } return 0; } SEC("tc") long rbtree_wrong_owner_remove_fail_a1(void *ctx) { struct node_data *n, *m; n = bpf_obj_new(typeof(*n)); if (!n) return 1; m = bpf_refcount_acquire(n); if (__stash_map_empty_xchg(n, 0)) { bpf_obj_drop(m); return 2; } if (__stash_map_empty_xchg(m, 1)) return 3; return 0; } SEC("tc") long rbtree_wrong_owner_remove_fail_b(void *ctx) { struct map_value *mapval; struct node_data *n; int idx = 0; mapval = bpf_map_lookup_elem(&stashed_nodes, &idx); if (!mapval) return 1; n = bpf_kptr_xchg(&mapval->node, NULL); if (!n) return 2; bpf_spin_lock(&block); bpf_rbtree_add(&broot, &n->r, less); bpf_spin_unlock(&block); return 0; } SEC("tc") long rbtree_wrong_owner_remove_fail_a2(void *ctx) { struct map_value *mapval; struct bpf_rb_node *res; struct node_data *m; int idx = 1; mapval = bpf_map_lookup_elem(&stashed_nodes, &idx); if (!mapval) return 1; m = bpf_kptr_xchg(&mapval->node, NULL); if (!m) return 2; bpf_spin_lock(&lock); /* make m non-owning ref */ bpf_list_push_back(&head, &m->l); res = bpf_rbtree_remove(&root, &m->r); bpf_spin_unlock(&lock); if (res) { bpf_obj_drop(container_of(res, struct node_data, r)); return 3; } return 0; } SEC("?fentry.s/bpf_testmod_test_read") __success int BPF_PROG(rbtree_sleepable_rcu, struct file *file, struct kobject *kobj, struct bin_attribute *bin_attr, char *buf, loff_t off, size_t len) { struct bpf_rb_node *rb; struct node_data *n, *m = NULL; n = bpf_obj_new(typeof(*n)); if (!n) return 0; bpf_rcu_read_lock(); bpf_spin_lock(&lock); bpf_rbtree_add(&root, &n->r, less); rb = bpf_rbtree_first(&root); if (!rb) goto err_out; rb = bpf_rbtree_remove(&root, rb); if (!rb) goto err_out; m = container_of(rb, struct node_data, r); err_out: bpf_spin_unlock(&lock); bpf_rcu_read_unlock(); if (m) bpf_obj_drop(m); return 0; } SEC("?fentry.s/bpf_testmod_test_read") __success int BPF_PROG(rbtree_sleepable_rcu_no_explicit_rcu_lock, struct file *file, struct kobject *kobj, struct bin_attribute *bin_attr, char *buf, loff_t off, size_t len) { struct bpf_rb_node *rb; struct node_data *n, *m = NULL; n = bpf_obj_new(typeof(*n)); if (!n) return 0; /* No explicit bpf_rcu_read_lock */ bpf_spin_lock(&lock); bpf_rbtree_add(&root, &n->r, less); rb = bpf_rbtree_first(&root); if (!rb) goto err_out; rb = bpf_rbtree_remove(&root, rb); if (!rb) goto err_out; m = container_of(rb, struct node_data, r); err_out: bpf_spin_unlock(&lock); /* No explicit bpf_rcu_read_unlock */ if (m) bpf_obj_drop(m); return 0; } char _license[] SEC("license") = "GPL";