Commit f4d05259 authored by Martin KaFai Lau's avatar Martin KaFai Lau Committed by Daniel Borkmann

bpf: Add map_meta_equal map ops

Some properties of the inner map is used in the verification time.
When an inner map is inserted to an outer map at runtime,
bpf_map_meta_equal() is currently used to ensure those properties
of the inserting inner map stays the same as the verification
time.

In particular, the current bpf_map_meta_equal() checks max_entries which
turns out to be too restrictive for most of the maps which do not use
max_entries during the verification time.  It limits the use case that
wants to replace a smaller inner map with a larger inner map.  There are
some maps do use max_entries during verification though.  For example,
the map_gen_lookup in array_map_ops uses the max_entries to generate
the inline lookup code.

To accommodate differences between maps, the map_meta_equal is added
to bpf_map_ops.  Each map-type can decide what to check when its
map is used as an inner map during runtime.

Also, some map types cannot be used as an inner map and they are
currently black listed in bpf_map_meta_alloc() in map_in_map.c.
It is not unusual that the new map types may not aware that such
blacklist exists.  This patch enforces an explicit opt-in
and only allows a map to be used as an inner map if it has
implemented the map_meta_equal ops.  It is based on the
discussion in [1].

All maps that support inner map has its map_meta_equal points
to bpf_map_meta_equal in this patch.  A later patch will
relax the max_entries check for most maps.  bpf_types.h
counts 28 map types.  This patch adds 23 ".map_meta_equal"
by using coccinelle.  -5 for
	BPF_MAP_TYPE_PROG_ARRAY
	BPF_MAP_TYPE_(PERCPU)_CGROUP_STORAGE
	BPF_MAP_TYPE_STRUCT_OPS
	BPF_MAP_TYPE_ARRAY_OF_MAPS
	BPF_MAP_TYPE_HASH_OF_MAPS

The "if (inner_map->inner_map_meta)" check in bpf_map_meta_alloc()
is moved such that the same error is returned.

[1]: https://lore.kernel.org/bpf/20200522022342.899756-1-kafai@fb.com/Signed-off-by: default avatarMartin KaFai Lau <kafai@fb.com>
Signed-off-by: default avatarDaniel Borkmann <daniel@iogearbox.net>
Link: https://lore.kernel.org/bpf/20200828011806.1970400-1-kafai@fb.com
parent b0c9eb37
...@@ -112,6 +112,19 @@ struct bpf_map_ops { ...@@ -112,6 +112,19 @@ struct bpf_map_ops {
void (*map_local_storage_uncharge)(struct bpf_local_storage_map *smap, void (*map_local_storage_uncharge)(struct bpf_local_storage_map *smap,
void *owner, u32 size); void *owner, u32 size);
struct bpf_local_storage __rcu ** (*map_owner_storage_ptr)(void *owner); struct bpf_local_storage __rcu ** (*map_owner_storage_ptr)(void *owner);
/* map_meta_equal must be implemented for maps that can be
* used as an inner map. It is a runtime check to ensure
* an inner map can be inserted to an outer map.
*
* Some properties of the inner map has been used during the
* verification time. When inserting an inner map at the runtime,
* map_meta_equal has to ensure the inserting map has the same
* properties that the verifier has used earlier.
*/
bool (*map_meta_equal)(const struct bpf_map *meta0,
const struct bpf_map *meta1);
/* BTF name and id of struct allocated by map_alloc */ /* BTF name and id of struct allocated by map_alloc */
const char * const map_btf_name; const char * const map_btf_name;
int *map_btf_id; int *map_btf_id;
...@@ -235,6 +248,9 @@ int map_check_no_btf(const struct bpf_map *map, ...@@ -235,6 +248,9 @@ int map_check_no_btf(const struct bpf_map *map,
const struct btf_type *key_type, const struct btf_type *key_type,
const struct btf_type *value_type); const struct btf_type *value_type);
bool bpf_map_meta_equal(const struct bpf_map *meta0,
const struct bpf_map *meta1);
extern const struct bpf_map_ops bpf_map_offload_ops; extern const struct bpf_map_ops bpf_map_offload_ops;
/* function argument constraints */ /* function argument constraints */
......
...@@ -625,6 +625,7 @@ static const struct bpf_iter_seq_info iter_seq_info = { ...@@ -625,6 +625,7 @@ static const struct bpf_iter_seq_info iter_seq_info = {
static int array_map_btf_id; static int array_map_btf_id;
const struct bpf_map_ops array_map_ops = { const struct bpf_map_ops array_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = array_map_alloc_check, .map_alloc_check = array_map_alloc_check,
.map_alloc = array_map_alloc, .map_alloc = array_map_alloc,
.map_free = array_map_free, .map_free = array_map_free,
...@@ -647,6 +648,7 @@ const struct bpf_map_ops array_map_ops = { ...@@ -647,6 +648,7 @@ const struct bpf_map_ops array_map_ops = {
static int percpu_array_map_btf_id; static int percpu_array_map_btf_id;
const struct bpf_map_ops percpu_array_map_ops = { const struct bpf_map_ops percpu_array_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = array_map_alloc_check, .map_alloc_check = array_map_alloc_check,
.map_alloc = array_map_alloc, .map_alloc = array_map_alloc,
.map_free = array_map_free, .map_free = array_map_free,
...@@ -1003,6 +1005,11 @@ static void prog_array_map_free(struct bpf_map *map) ...@@ -1003,6 +1005,11 @@ static void prog_array_map_free(struct bpf_map *map)
fd_array_map_free(map); fd_array_map_free(map);
} }
/* prog_array->aux->{type,jited} is a runtime binding.
* Doing static check alone in the verifier is not enough.
* Thus, prog_array_map cannot be used as an inner_map
* and map_meta_equal is not implemented.
*/
static int prog_array_map_btf_id; static int prog_array_map_btf_id;
const struct bpf_map_ops prog_array_map_ops = { const struct bpf_map_ops prog_array_map_ops = {
.map_alloc_check = fd_array_map_alloc_check, .map_alloc_check = fd_array_map_alloc_check,
...@@ -1101,6 +1108,7 @@ static void perf_event_fd_array_release(struct bpf_map *map, ...@@ -1101,6 +1108,7 @@ static void perf_event_fd_array_release(struct bpf_map *map,
static int perf_event_array_map_btf_id; static int perf_event_array_map_btf_id;
const struct bpf_map_ops perf_event_array_map_ops = { const struct bpf_map_ops perf_event_array_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = fd_array_map_alloc_check, .map_alloc_check = fd_array_map_alloc_check,
.map_alloc = array_map_alloc, .map_alloc = array_map_alloc,
.map_free = fd_array_map_free, .map_free = fd_array_map_free,
...@@ -1137,6 +1145,7 @@ static void cgroup_fd_array_free(struct bpf_map *map) ...@@ -1137,6 +1145,7 @@ static void cgroup_fd_array_free(struct bpf_map *map)
static int cgroup_array_map_btf_id; static int cgroup_array_map_btf_id;
const struct bpf_map_ops cgroup_array_map_ops = { const struct bpf_map_ops cgroup_array_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = fd_array_map_alloc_check, .map_alloc_check = fd_array_map_alloc_check,
.map_alloc = array_map_alloc, .map_alloc = array_map_alloc,
.map_free = cgroup_fd_array_free, .map_free = cgroup_fd_array_free,
......
...@@ -235,6 +235,7 @@ static void inode_storage_map_free(struct bpf_map *map) ...@@ -235,6 +235,7 @@ static void inode_storage_map_free(struct bpf_map *map)
static int inode_storage_map_btf_id; static int inode_storage_map_btf_id;
const struct bpf_map_ops inode_storage_map_ops = { const struct bpf_map_ops inode_storage_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = bpf_local_storage_map_alloc_check, .map_alloc_check = bpf_local_storage_map_alloc_check,
.map_alloc = inode_storage_map_alloc, .map_alloc = inode_storage_map_alloc,
.map_free = inode_storage_map_free, .map_free = inode_storage_map_free,
......
...@@ -658,6 +658,7 @@ static int cpu_map_get_next_key(struct bpf_map *map, void *key, void *next_key) ...@@ -658,6 +658,7 @@ static int cpu_map_get_next_key(struct bpf_map *map, void *key, void *next_key)
static int cpu_map_btf_id; static int cpu_map_btf_id;
const struct bpf_map_ops cpu_map_ops = { const struct bpf_map_ops cpu_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = cpu_map_alloc, .map_alloc = cpu_map_alloc,
.map_free = cpu_map_free, .map_free = cpu_map_free,
.map_delete_elem = cpu_map_delete_elem, .map_delete_elem = cpu_map_delete_elem,
......
...@@ -751,6 +751,7 @@ static int dev_map_hash_update_elem(struct bpf_map *map, void *key, void *value, ...@@ -751,6 +751,7 @@ static int dev_map_hash_update_elem(struct bpf_map *map, void *key, void *value,
static int dev_map_btf_id; static int dev_map_btf_id;
const struct bpf_map_ops dev_map_ops = { const struct bpf_map_ops dev_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = dev_map_alloc, .map_alloc = dev_map_alloc,
.map_free = dev_map_free, .map_free = dev_map_free,
.map_get_next_key = dev_map_get_next_key, .map_get_next_key = dev_map_get_next_key,
...@@ -764,6 +765,7 @@ const struct bpf_map_ops dev_map_ops = { ...@@ -764,6 +765,7 @@ const struct bpf_map_ops dev_map_ops = {
static int dev_map_hash_map_btf_id; static int dev_map_hash_map_btf_id;
const struct bpf_map_ops dev_map_hash_ops = { const struct bpf_map_ops dev_map_hash_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = dev_map_alloc, .map_alloc = dev_map_alloc,
.map_free = dev_map_free, .map_free = dev_map_free,
.map_get_next_key = dev_map_hash_get_next_key, .map_get_next_key = dev_map_hash_get_next_key,
......
...@@ -1810,6 +1810,7 @@ static const struct bpf_iter_seq_info iter_seq_info = { ...@@ -1810,6 +1810,7 @@ static const struct bpf_iter_seq_info iter_seq_info = {
static int htab_map_btf_id; static int htab_map_btf_id;
const struct bpf_map_ops htab_map_ops = { const struct bpf_map_ops htab_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = htab_map_alloc_check, .map_alloc_check = htab_map_alloc_check,
.map_alloc = htab_map_alloc, .map_alloc = htab_map_alloc,
.map_free = htab_map_free, .map_free = htab_map_free,
...@@ -1827,6 +1828,7 @@ const struct bpf_map_ops htab_map_ops = { ...@@ -1827,6 +1828,7 @@ const struct bpf_map_ops htab_map_ops = {
static int htab_lru_map_btf_id; static int htab_lru_map_btf_id;
const struct bpf_map_ops htab_lru_map_ops = { const struct bpf_map_ops htab_lru_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = htab_map_alloc_check, .map_alloc_check = htab_map_alloc_check,
.map_alloc = htab_map_alloc, .map_alloc = htab_map_alloc,
.map_free = htab_map_free, .map_free = htab_map_free,
...@@ -1947,6 +1949,7 @@ static void htab_percpu_map_seq_show_elem(struct bpf_map *map, void *key, ...@@ -1947,6 +1949,7 @@ static void htab_percpu_map_seq_show_elem(struct bpf_map *map, void *key,
static int htab_percpu_map_btf_id; static int htab_percpu_map_btf_id;
const struct bpf_map_ops htab_percpu_map_ops = { const struct bpf_map_ops htab_percpu_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = htab_map_alloc_check, .map_alloc_check = htab_map_alloc_check,
.map_alloc = htab_map_alloc, .map_alloc = htab_map_alloc,
.map_free = htab_map_free, .map_free = htab_map_free,
...@@ -1963,6 +1966,7 @@ const struct bpf_map_ops htab_percpu_map_ops = { ...@@ -1963,6 +1966,7 @@ const struct bpf_map_ops htab_percpu_map_ops = {
static int htab_lru_percpu_map_btf_id; static int htab_lru_percpu_map_btf_id;
const struct bpf_map_ops htab_lru_percpu_map_ops = { const struct bpf_map_ops htab_lru_percpu_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = htab_map_alloc_check, .map_alloc_check = htab_map_alloc_check,
.map_alloc = htab_map_alloc, .map_alloc = htab_map_alloc,
.map_free = htab_map_free, .map_free = htab_map_free,
......
...@@ -732,6 +732,7 @@ static int trie_check_btf(const struct bpf_map *map, ...@@ -732,6 +732,7 @@ static int trie_check_btf(const struct bpf_map *map,
static int trie_map_btf_id; static int trie_map_btf_id;
const struct bpf_map_ops trie_map_ops = { const struct bpf_map_ops trie_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = trie_alloc, .map_alloc = trie_alloc,
.map_free = trie_free, .map_free = trie_free,
.map_get_next_key = trie_get_next_key, .map_get_next_key = trie_get_next_key,
......
...@@ -17,23 +17,17 @@ struct bpf_map *bpf_map_meta_alloc(int inner_map_ufd) ...@@ -17,23 +17,17 @@ struct bpf_map *bpf_map_meta_alloc(int inner_map_ufd)
if (IS_ERR(inner_map)) if (IS_ERR(inner_map))
return inner_map; return inner_map;
/* prog_array->aux->{type,jited} is a runtime binding.
* Doing static check alone in the verifier is not enough.
*/
if (inner_map->map_type == BPF_MAP_TYPE_PROG_ARRAY ||
inner_map->map_type == BPF_MAP_TYPE_CGROUP_STORAGE ||
inner_map->map_type == BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE ||
inner_map->map_type == BPF_MAP_TYPE_STRUCT_OPS) {
fdput(f);
return ERR_PTR(-ENOTSUPP);
}
/* Does not support >1 level map-in-map */ /* Does not support >1 level map-in-map */
if (inner_map->inner_map_meta) { if (inner_map->inner_map_meta) {
fdput(f); fdput(f);
return ERR_PTR(-EINVAL); return ERR_PTR(-EINVAL);
} }
if (!inner_map->ops->map_meta_equal) {
fdput(f);
return ERR_PTR(-ENOTSUPP);
}
if (map_value_has_spin_lock(inner_map)) { if (map_value_has_spin_lock(inner_map)) {
fdput(f); fdput(f);
return ERR_PTR(-ENOTSUPP); return ERR_PTR(-ENOTSUPP);
...@@ -89,7 +83,7 @@ void *bpf_map_fd_get_ptr(struct bpf_map *map, ...@@ -89,7 +83,7 @@ void *bpf_map_fd_get_ptr(struct bpf_map *map,
struct file *map_file /* not used */, struct file *map_file /* not used */,
int ufd) int ufd)
{ {
struct bpf_map *inner_map; struct bpf_map *inner_map, *inner_map_meta;
struct fd f; struct fd f;
f = fdget(ufd); f = fdget(ufd);
...@@ -97,7 +91,8 @@ void *bpf_map_fd_get_ptr(struct bpf_map *map, ...@@ -97,7 +91,8 @@ void *bpf_map_fd_get_ptr(struct bpf_map *map,
if (IS_ERR(inner_map)) if (IS_ERR(inner_map))
return inner_map; return inner_map;
if (bpf_map_meta_equal(map->inner_map_meta, inner_map)) inner_map_meta = map->inner_map_meta;
if (inner_map_meta->ops->map_meta_equal(inner_map_meta, inner_map))
bpf_map_inc(inner_map); bpf_map_inc(inner_map);
else else
inner_map = ERR_PTR(-EINVAL); inner_map = ERR_PTR(-EINVAL);
......
...@@ -11,8 +11,6 @@ struct bpf_map; ...@@ -11,8 +11,6 @@ struct bpf_map;
struct bpf_map *bpf_map_meta_alloc(int inner_map_ufd); struct bpf_map *bpf_map_meta_alloc(int inner_map_ufd);
void bpf_map_meta_free(struct bpf_map *map_meta); void bpf_map_meta_free(struct bpf_map *map_meta);
bool bpf_map_meta_equal(const struct bpf_map *meta0,
const struct bpf_map *meta1);
void *bpf_map_fd_get_ptr(struct bpf_map *map, struct file *map_file, void *bpf_map_fd_get_ptr(struct bpf_map *map, struct file *map_file,
int ufd); int ufd);
void bpf_map_fd_put_ptr(void *ptr); void bpf_map_fd_put_ptr(void *ptr);
......
...@@ -257,6 +257,7 @@ static int queue_stack_map_get_next_key(struct bpf_map *map, void *key, ...@@ -257,6 +257,7 @@ static int queue_stack_map_get_next_key(struct bpf_map *map, void *key,
static int queue_map_btf_id; static int queue_map_btf_id;
const struct bpf_map_ops queue_map_ops = { const struct bpf_map_ops queue_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = queue_stack_map_alloc_check, .map_alloc_check = queue_stack_map_alloc_check,
.map_alloc = queue_stack_map_alloc, .map_alloc = queue_stack_map_alloc,
.map_free = queue_stack_map_free, .map_free = queue_stack_map_free,
...@@ -273,6 +274,7 @@ const struct bpf_map_ops queue_map_ops = { ...@@ -273,6 +274,7 @@ const struct bpf_map_ops queue_map_ops = {
static int stack_map_btf_id; static int stack_map_btf_id;
const struct bpf_map_ops stack_map_ops = { const struct bpf_map_ops stack_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = queue_stack_map_alloc_check, .map_alloc_check = queue_stack_map_alloc_check,
.map_alloc = queue_stack_map_alloc, .map_alloc = queue_stack_map_alloc,
.map_free = queue_stack_map_free, .map_free = queue_stack_map_free,
......
...@@ -351,6 +351,7 @@ static int reuseport_array_get_next_key(struct bpf_map *map, void *key, ...@@ -351,6 +351,7 @@ static int reuseport_array_get_next_key(struct bpf_map *map, void *key,
static int reuseport_array_map_btf_id; static int reuseport_array_map_btf_id;
const struct bpf_map_ops reuseport_array_ops = { const struct bpf_map_ops reuseport_array_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = reuseport_array_alloc_check, .map_alloc_check = reuseport_array_alloc_check,
.map_alloc = reuseport_array_alloc, .map_alloc = reuseport_array_alloc,
.map_free = reuseport_array_free, .map_free = reuseport_array_free,
......
...@@ -287,6 +287,7 @@ static __poll_t ringbuf_map_poll(struct bpf_map *map, struct file *filp, ...@@ -287,6 +287,7 @@ static __poll_t ringbuf_map_poll(struct bpf_map *map, struct file *filp,
static int ringbuf_map_btf_id; static int ringbuf_map_btf_id;
const struct bpf_map_ops ringbuf_map_ops = { const struct bpf_map_ops ringbuf_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = ringbuf_map_alloc, .map_alloc = ringbuf_map_alloc,
.map_free = ringbuf_map_free, .map_free = ringbuf_map_free,
.map_mmap = ringbuf_map_mmap, .map_mmap = ringbuf_map_mmap,
......
...@@ -839,6 +839,7 @@ static void stack_map_free(struct bpf_map *map) ...@@ -839,6 +839,7 @@ static void stack_map_free(struct bpf_map *map)
static int stack_trace_map_btf_id; static int stack_trace_map_btf_id;
const struct bpf_map_ops stack_trace_map_ops = { const struct bpf_map_ops stack_trace_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = stack_map_alloc, .map_alloc = stack_map_alloc,
.map_free = stack_map_free, .map_free = stack_map_free,
.map_get_next_key = stack_map_get_next_key, .map_get_next_key = stack_map_get_next_key,
......
...@@ -90,6 +90,7 @@ int bpf_check_uarg_tail_zero(void __user *uaddr, ...@@ -90,6 +90,7 @@ int bpf_check_uarg_tail_zero(void __user *uaddr,
} }
const struct bpf_map_ops bpf_map_offload_ops = { const struct bpf_map_ops bpf_map_offload_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = bpf_map_offload_map_alloc, .map_alloc = bpf_map_offload_map_alloc,
.map_free = bpf_map_offload_map_free, .map_free = bpf_map_offload_map_free,
.map_check_btf = map_check_no_btf, .map_check_btf = map_check_no_btf,
......
...@@ -335,6 +335,7 @@ sk_storage_ptr(void *owner) ...@@ -335,6 +335,7 @@ sk_storage_ptr(void *owner)
static int sk_storage_map_btf_id; static int sk_storage_map_btf_id;
const struct bpf_map_ops sk_storage_map_ops = { const struct bpf_map_ops sk_storage_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc_check = bpf_local_storage_map_alloc_check, .map_alloc_check = bpf_local_storage_map_alloc_check,
.map_alloc = sk_storage_map_alloc, .map_alloc = sk_storage_map_alloc,
.map_free = sk_storage_map_free, .map_free = sk_storage_map_free,
......
...@@ -705,6 +705,7 @@ const struct bpf_func_proto bpf_msg_redirect_map_proto = { ...@@ -705,6 +705,7 @@ const struct bpf_func_proto bpf_msg_redirect_map_proto = {
static int sock_map_btf_id; static int sock_map_btf_id;
const struct bpf_map_ops sock_map_ops = { const struct bpf_map_ops sock_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = sock_map_alloc, .map_alloc = sock_map_alloc,
.map_free = sock_map_free, .map_free = sock_map_free,
.map_get_next_key = sock_map_get_next_key, .map_get_next_key = sock_map_get_next_key,
...@@ -1200,6 +1201,7 @@ const struct bpf_func_proto bpf_msg_redirect_hash_proto = { ...@@ -1200,6 +1201,7 @@ const struct bpf_func_proto bpf_msg_redirect_hash_proto = {
static int sock_hash_map_btf_id; static int sock_hash_map_btf_id;
const struct bpf_map_ops sock_hash_ops = { const struct bpf_map_ops sock_hash_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = sock_hash_alloc, .map_alloc = sock_hash_alloc,
.map_free = sock_hash_free, .map_free = sock_hash_free,
.map_get_next_key = sock_hash_get_next_key, .map_get_next_key = sock_hash_get_next_key,
......
...@@ -256,6 +256,7 @@ void xsk_map_try_sock_delete(struct xsk_map *map, struct xdp_sock *xs, ...@@ -256,6 +256,7 @@ void xsk_map_try_sock_delete(struct xsk_map *map, struct xdp_sock *xs,
static int xsk_map_btf_id; static int xsk_map_btf_id;
const struct bpf_map_ops xsk_map_ops = { const struct bpf_map_ops xsk_map_ops = {
.map_meta_equal = bpf_map_meta_equal,
.map_alloc = xsk_map_alloc, .map_alloc = xsk_map_alloc,
.map_free = xsk_map_free, .map_free = xsk_map_free,
.map_get_next_key = xsk_map_get_next_key, .map_get_next_key = xsk_map_get_next_key,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment