Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/tools/lib/bpf/skel_internal.h
29278 views
1
/* SPDX-License-Identifier: (LGPL-2.1 OR BSD-2-Clause) */
2
/* Copyright (c) 2021 Facebook */
3
#ifndef __SKEL_INTERNAL_H
4
#define __SKEL_INTERNAL_H
5
6
#ifdef __KERNEL__
7
#include <linux/fdtable.h>
8
#include <linux/mm.h>
9
#include <linux/mman.h>
10
#include <linux/slab.h>
11
#include <linux/bpf.h>
12
#else
13
#include <unistd.h>
14
#include <sys/syscall.h>
15
#include <sys/mman.h>
16
#include <linux/keyctl.h>
17
#include <stdlib.h>
18
#include "bpf.h"
19
#endif
20
21
#ifndef SHA256_DIGEST_LENGTH
22
#define SHA256_DIGEST_LENGTH 32
23
#endif
24
25
#ifndef __NR_bpf
26
# if defined(__mips__) && defined(_ABIO32)
27
# define __NR_bpf 4355
28
# elif defined(__mips__) && defined(_ABIN32)
29
# define __NR_bpf 6319
30
# elif defined(__mips__) && defined(_ABI64)
31
# define __NR_bpf 5315
32
# endif
33
#endif
34
35
/* This file is a base header for auto-generated *.lskel.h files.
36
* Its contents will change and may become part of auto-generation in the future.
37
*
38
* The layout of bpf_[map|prog]_desc and bpf_loader_ctx is feature dependent
39
* and will change from one version of libbpf to another and features
40
* requested during loader program generation.
41
*/
42
struct bpf_map_desc {
43
/* output of the loader prog */
44
int map_fd;
45
/* input for the loader prog */
46
__u32 max_entries;
47
__aligned_u64 initial_value;
48
};
49
struct bpf_prog_desc {
50
int prog_fd;
51
};
52
53
enum {
54
BPF_SKEL_KERNEL = (1ULL << 0),
55
};
56
57
struct bpf_loader_ctx {
58
__u32 sz;
59
__u32 flags;
60
__u32 log_level;
61
__u32 log_size;
62
__u64 log_buf;
63
};
64
65
struct bpf_load_and_run_opts {
66
struct bpf_loader_ctx *ctx;
67
const void *data;
68
const void *insns;
69
__u32 data_sz;
70
__u32 insns_sz;
71
const char *errstr;
72
void *signature;
73
__u32 signature_sz;
74
__s32 keyring_id;
75
void *excl_prog_hash;
76
__u32 excl_prog_hash_sz;
77
};
78
79
long kern_sys_bpf(__u32 cmd, void *attr, __u32 attr_size);
80
81
static inline int skel_sys_bpf(enum bpf_cmd cmd, union bpf_attr *attr,
82
unsigned int size)
83
{
84
#ifdef __KERNEL__
85
return kern_sys_bpf(cmd, attr, size);
86
#else
87
return syscall(__NR_bpf, cmd, attr, size);
88
#endif
89
}
90
91
#ifdef __KERNEL__
92
static inline int close(int fd)
93
{
94
return close_fd(fd);
95
}
96
97
static inline void *skel_alloc(size_t size)
98
{
99
struct bpf_loader_ctx *ctx = kzalloc(size, GFP_KERNEL);
100
101
if (!ctx)
102
return NULL;
103
ctx->flags |= BPF_SKEL_KERNEL;
104
return ctx;
105
}
106
107
static inline void skel_free(const void *p)
108
{
109
kfree(p);
110
}
111
112
/* skel->bss/rodata maps are populated the following way:
113
*
114
* For kernel use:
115
* skel_prep_map_data() allocates kernel memory that kernel module can directly access.
116
* Generated lskel stores the pointer in skel->rodata and in skel->maps.rodata.initial_value.
117
* The loader program will perform probe_read_kernel() from maps.rodata.initial_value.
118
* skel_finalize_map_data() sets skel->rodata to point to actual value in a bpf map and
119
* does maps.rodata.initial_value = ~0ULL to signal skel_free_map_data() that kvfree
120
* is not necessary.
121
*
122
* For user space:
123
* skel_prep_map_data() mmaps anon memory into skel->rodata that can be accessed directly.
124
* Generated lskel stores the pointer in skel->rodata and in skel->maps.rodata.initial_value.
125
* The loader program will perform copy_from_user() from maps.rodata.initial_value.
126
* skel_finalize_map_data() remaps bpf array map value from the kernel memory into
127
* skel->rodata address.
128
*
129
* The "bpftool gen skeleton -L" command generates lskel.h that is suitable for
130
* both kernel and user space. The generated loader program does
131
* either bpf_probe_read_kernel() or bpf_copy_from_user() from initial_value
132
* depending on bpf_loader_ctx->flags.
133
*/
134
static inline void skel_free_map_data(void *p, __u64 addr, size_t sz)
135
{
136
if (addr != ~0ULL)
137
kvfree(p);
138
/* When addr == ~0ULL the 'p' points to
139
* ((struct bpf_array *)map)->value. See skel_finalize_map_data.
140
*/
141
}
142
143
static inline void *skel_prep_map_data(const void *val, size_t mmap_sz, size_t val_sz)
144
{
145
void *addr;
146
147
addr = kvmalloc(val_sz, GFP_KERNEL);
148
if (!addr)
149
return NULL;
150
memcpy(addr, val, val_sz);
151
return addr;
152
}
153
154
static inline void *skel_finalize_map_data(__u64 *init_val, size_t mmap_sz, int flags, int fd)
155
{
156
struct bpf_map *map;
157
void *addr = NULL;
158
159
kvfree((void *) (long) *init_val);
160
*init_val = ~0ULL;
161
162
/* At this point bpf_load_and_run() finished without error and
163
* 'fd' is a valid bpf map FD. All sanity checks below should succeed.
164
*/
165
map = bpf_map_get(fd);
166
if (IS_ERR(map))
167
return NULL;
168
if (map->map_type != BPF_MAP_TYPE_ARRAY)
169
goto out;
170
addr = ((struct bpf_array *)map)->value;
171
/* the addr stays valid, since FD is not closed */
172
out:
173
bpf_map_put(map);
174
return addr;
175
}
176
177
#else
178
179
static inline void *skel_alloc(size_t size)
180
{
181
return calloc(1, size);
182
}
183
184
static inline void skel_free(void *p)
185
{
186
free(p);
187
}
188
189
static inline void skel_free_map_data(void *p, __u64 addr, size_t sz)
190
{
191
munmap(p, sz);
192
}
193
194
static inline void *skel_prep_map_data(const void *val, size_t mmap_sz, size_t val_sz)
195
{
196
void *addr;
197
198
addr = mmap(NULL, mmap_sz, PROT_READ | PROT_WRITE,
199
MAP_SHARED | MAP_ANONYMOUS, -1, 0);
200
if (addr == (void *) -1)
201
return NULL;
202
memcpy(addr, val, val_sz);
203
return addr;
204
}
205
206
static inline void *skel_finalize_map_data(__u64 *init_val, size_t mmap_sz, int flags, int fd)
207
{
208
void *addr;
209
210
addr = mmap((void *) (long) *init_val, mmap_sz, flags, MAP_SHARED | MAP_FIXED, fd, 0);
211
if (addr == (void *) -1)
212
return NULL;
213
return addr;
214
}
215
#endif
216
217
static inline int skel_closenz(int fd)
218
{
219
if (fd > 0)
220
return close(fd);
221
return -EINVAL;
222
}
223
224
#ifndef offsetofend
225
#define offsetofend(TYPE, MEMBER) \
226
(offsetof(TYPE, MEMBER) + sizeof((((TYPE *)0)->MEMBER)))
227
#endif
228
229
static inline int skel_map_create(enum bpf_map_type map_type,
230
const char *map_name,
231
__u32 key_size,
232
__u32 value_size,
233
__u32 max_entries,
234
const void *excl_prog_hash,
235
__u32 excl_prog_hash_sz)
236
{
237
const size_t attr_sz = offsetofend(union bpf_attr, excl_prog_hash_size);
238
union bpf_attr attr;
239
240
memset(&attr, 0, attr_sz);
241
242
attr.map_type = map_type;
243
attr.excl_prog_hash = (unsigned long) excl_prog_hash;
244
attr.excl_prog_hash_size = excl_prog_hash_sz;
245
246
strncpy(attr.map_name, map_name, sizeof(attr.map_name));
247
attr.key_size = key_size;
248
attr.value_size = value_size;
249
attr.max_entries = max_entries;
250
251
return skel_sys_bpf(BPF_MAP_CREATE, &attr, attr_sz);
252
}
253
254
static inline int skel_map_update_elem(int fd, const void *key,
255
const void *value, __u64 flags)
256
{
257
const size_t attr_sz = offsetofend(union bpf_attr, flags);
258
union bpf_attr attr;
259
260
memset(&attr, 0, attr_sz);
261
attr.map_fd = fd;
262
attr.key = (long) key;
263
attr.value = (long) value;
264
attr.flags = flags;
265
266
return skel_sys_bpf(BPF_MAP_UPDATE_ELEM, &attr, attr_sz);
267
}
268
269
static inline int skel_map_delete_elem(int fd, const void *key)
270
{
271
const size_t attr_sz = offsetofend(union bpf_attr, flags);
272
union bpf_attr attr;
273
274
memset(&attr, 0, attr_sz);
275
attr.map_fd = fd;
276
attr.key = (long)key;
277
278
return skel_sys_bpf(BPF_MAP_DELETE_ELEM, &attr, attr_sz);
279
}
280
281
static inline int skel_map_get_fd_by_id(__u32 id)
282
{
283
const size_t attr_sz = offsetofend(union bpf_attr, flags);
284
union bpf_attr attr;
285
286
memset(&attr, 0, attr_sz);
287
attr.map_id = id;
288
289
return skel_sys_bpf(BPF_MAP_GET_FD_BY_ID, &attr, attr_sz);
290
}
291
292
static inline int skel_raw_tracepoint_open(const char *name, int prog_fd)
293
{
294
const size_t attr_sz = offsetofend(union bpf_attr, raw_tracepoint.prog_fd);
295
union bpf_attr attr;
296
297
memset(&attr, 0, attr_sz);
298
attr.raw_tracepoint.name = (long) name;
299
attr.raw_tracepoint.prog_fd = prog_fd;
300
301
return skel_sys_bpf(BPF_RAW_TRACEPOINT_OPEN, &attr, attr_sz);
302
}
303
304
static inline int skel_link_create(int prog_fd, int target_fd,
305
enum bpf_attach_type attach_type)
306
{
307
const size_t attr_sz = offsetofend(union bpf_attr, link_create.iter_info_len);
308
union bpf_attr attr;
309
310
memset(&attr, 0, attr_sz);
311
attr.link_create.prog_fd = prog_fd;
312
attr.link_create.target_fd = target_fd;
313
attr.link_create.attach_type = attach_type;
314
315
return skel_sys_bpf(BPF_LINK_CREATE, &attr, attr_sz);
316
}
317
318
static inline int skel_obj_get_info_by_fd(int fd)
319
{
320
const size_t attr_sz = offsetofend(union bpf_attr, info);
321
__u8 sha[SHA256_DIGEST_LENGTH];
322
struct bpf_map_info info;
323
__u32 info_len = sizeof(info);
324
union bpf_attr attr;
325
326
memset(&info, 0, sizeof(info));
327
info.hash = (long) &sha;
328
info.hash_size = SHA256_DIGEST_LENGTH;
329
330
memset(&attr, 0, attr_sz);
331
attr.info.bpf_fd = fd;
332
attr.info.info = (long) &info;
333
attr.info.info_len = info_len;
334
return skel_sys_bpf(BPF_OBJ_GET_INFO_BY_FD, &attr, attr_sz);
335
}
336
337
static inline int skel_map_freeze(int fd)
338
{
339
const size_t attr_sz = offsetofend(union bpf_attr, map_fd);
340
union bpf_attr attr;
341
342
memset(&attr, 0, attr_sz);
343
attr.map_fd = fd;
344
345
return skel_sys_bpf(BPF_MAP_FREEZE, &attr, attr_sz);
346
}
347
#ifdef __KERNEL__
348
#define set_err
349
#else
350
#define set_err err = -errno
351
#endif
352
353
static inline int bpf_load_and_run(struct bpf_load_and_run_opts *opts)
354
{
355
const size_t prog_load_attr_sz = offsetofend(union bpf_attr, keyring_id);
356
const size_t test_run_attr_sz = offsetofend(union bpf_attr, test);
357
int map_fd = -1, prog_fd = -1, key = 0, err;
358
union bpf_attr attr;
359
360
err = map_fd = skel_map_create(BPF_MAP_TYPE_ARRAY, "__loader.map", 4, opts->data_sz, 1,
361
opts->excl_prog_hash, opts->excl_prog_hash_sz);
362
if (map_fd < 0) {
363
opts->errstr = "failed to create loader map";
364
set_err;
365
goto out;
366
}
367
368
err = skel_map_update_elem(map_fd, &key, opts->data, 0);
369
if (err < 0) {
370
opts->errstr = "failed to update loader map";
371
set_err;
372
goto out;
373
}
374
375
#ifndef __KERNEL__
376
err = skel_map_freeze(map_fd);
377
if (err < 0) {
378
opts->errstr = "failed to freeze map";
379
set_err;
380
goto out;
381
}
382
err = skel_obj_get_info_by_fd(map_fd);
383
if (err < 0) {
384
opts->errstr = "failed to fetch obj info";
385
set_err;
386
goto out;
387
}
388
#endif
389
390
memset(&attr, 0, prog_load_attr_sz);
391
attr.prog_type = BPF_PROG_TYPE_SYSCALL;
392
attr.insns = (long) opts->insns;
393
attr.insn_cnt = opts->insns_sz / sizeof(struct bpf_insn);
394
attr.license = (long) "Dual BSD/GPL";
395
#ifndef __KERNEL__
396
attr.signature = (long) opts->signature;
397
attr.signature_size = opts->signature_sz;
398
#else
399
if (opts->signature || opts->signature_sz)
400
pr_warn("signatures are not supported from bpf_preload\n");
401
#endif
402
attr.keyring_id = opts->keyring_id;
403
memcpy(attr.prog_name, "__loader.prog", sizeof("__loader.prog"));
404
attr.fd_array = (long) &map_fd;
405
attr.log_level = opts->ctx->log_level;
406
attr.log_size = opts->ctx->log_size;
407
attr.log_buf = opts->ctx->log_buf;
408
attr.prog_flags = BPF_F_SLEEPABLE;
409
err = prog_fd = skel_sys_bpf(BPF_PROG_LOAD, &attr, prog_load_attr_sz);
410
if (prog_fd < 0) {
411
opts->errstr = "failed to load loader prog";
412
set_err;
413
goto out;
414
}
415
416
memset(&attr, 0, test_run_attr_sz);
417
attr.test.prog_fd = prog_fd;
418
attr.test.ctx_in = (long) opts->ctx;
419
attr.test.ctx_size_in = opts->ctx->sz;
420
err = skel_sys_bpf(BPF_PROG_RUN, &attr, test_run_attr_sz);
421
if (err < 0 || (int)attr.test.retval < 0) {
422
if (err < 0) {
423
opts->errstr = "failed to execute loader prog";
424
set_err;
425
} else {
426
opts->errstr = "error returned by loader prog";
427
err = (int)attr.test.retval;
428
#ifndef __KERNEL__
429
errno = -err;
430
#endif
431
}
432
goto out;
433
}
434
err = 0;
435
out:
436
if (map_fd >= 0)
437
close(map_fd);
438
if (prog_fd >= 0)
439
close(prog_fd);
440
return err;
441
}
442
443
#endif
444
445