mirror of
https://github.com/Qortal/Brooklyn.git
synced 2025-02-07 14:54:17 +00:00
a94b3d14aa
Changes included (and more): 1. Dynamic RAM merge 2. Real-time page scan and allocation 3. Cache compression 4. Real-time IRQ checks 5. Dynamic I/O allocation for Java heap 6. Java page migration 7. Contiguous memory allocation 8. Idle pages tracking 9. Per CPU RAM usage tracking 10. ARM NEON scalar multiplication library 11. NEON/ARMv8 crypto extensions 12. NEON SHA, Blake, RIPEMD crypto extensions 13. Parallel NEON crypto engine for multi-algo based CPU stress reduction
119 lines
3.0 KiB
C
119 lines
3.0 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <vmlinux.h>
|
|
#include <bpf/bpf_helpers.h>
|
|
|
|
#define EAFNOSUPPORT 97
|
|
#define EPROTO 71
|
|
#define ENONET 64
|
|
#define EINVAL 22
|
|
#define ENOENT 2
|
|
|
|
int test_einval_bpf_tuple = 0;
|
|
int test_einval_reserved = 0;
|
|
int test_einval_netns_id = 0;
|
|
int test_einval_len_opts = 0;
|
|
int test_eproto_l4proto = 0;
|
|
int test_enonet_netns_id = 0;
|
|
int test_enoent_lookup = 0;
|
|
int test_eafnosupport = 0;
|
|
|
|
struct nf_conn;
|
|
|
|
struct bpf_ct_opts___local {
|
|
s32 netns_id;
|
|
s32 error;
|
|
u8 l4proto;
|
|
u8 reserved[3];
|
|
} __attribute__((preserve_access_index));
|
|
|
|
struct nf_conn *bpf_xdp_ct_lookup(struct xdp_md *, struct bpf_sock_tuple *, u32,
|
|
struct bpf_ct_opts___local *, u32) __ksym;
|
|
struct nf_conn *bpf_skb_ct_lookup(struct __sk_buff *, struct bpf_sock_tuple *, u32,
|
|
struct bpf_ct_opts___local *, u32) __ksym;
|
|
void bpf_ct_release(struct nf_conn *) __ksym;
|
|
|
|
static __always_inline void
|
|
nf_ct_test(struct nf_conn *(*func)(void *, struct bpf_sock_tuple *, u32,
|
|
struct bpf_ct_opts___local *, u32),
|
|
void *ctx)
|
|
{
|
|
struct bpf_ct_opts___local opts_def = { .l4proto = IPPROTO_TCP, .netns_id = -1 };
|
|
struct bpf_sock_tuple bpf_tuple;
|
|
struct nf_conn *ct;
|
|
|
|
__builtin_memset(&bpf_tuple, 0, sizeof(bpf_tuple.ipv4));
|
|
|
|
ct = func(ctx, NULL, 0, &opts_def, sizeof(opts_def));
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_einval_bpf_tuple = opts_def.error;
|
|
|
|
opts_def.reserved[0] = 1;
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4), &opts_def, sizeof(opts_def));
|
|
opts_def.reserved[0] = 0;
|
|
opts_def.l4proto = IPPROTO_TCP;
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_einval_reserved = opts_def.error;
|
|
|
|
opts_def.netns_id = -2;
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4), &opts_def, sizeof(opts_def));
|
|
opts_def.netns_id = -1;
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_einval_netns_id = opts_def.error;
|
|
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4), &opts_def, sizeof(opts_def) - 1);
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_einval_len_opts = opts_def.error;
|
|
|
|
opts_def.l4proto = IPPROTO_ICMP;
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4), &opts_def, sizeof(opts_def));
|
|
opts_def.l4proto = IPPROTO_TCP;
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_eproto_l4proto = opts_def.error;
|
|
|
|
opts_def.netns_id = 0xf00f;
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4), &opts_def, sizeof(opts_def));
|
|
opts_def.netns_id = -1;
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_enonet_netns_id = opts_def.error;
|
|
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4), &opts_def, sizeof(opts_def));
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_enoent_lookup = opts_def.error;
|
|
|
|
ct = func(ctx, &bpf_tuple, sizeof(bpf_tuple.ipv4) - 1, &opts_def, sizeof(opts_def));
|
|
if (ct)
|
|
bpf_ct_release(ct);
|
|
else
|
|
test_eafnosupport = opts_def.error;
|
|
}
|
|
|
|
SEC("xdp")
|
|
int nf_xdp_ct_test(struct xdp_md *ctx)
|
|
{
|
|
nf_ct_test((void *)bpf_xdp_ct_lookup, ctx);
|
|
return 0;
|
|
}
|
|
|
|
SEC("tc")
|
|
int nf_skb_ct_test(struct __sk_buff *ctx)
|
|
{
|
|
nf_ct_test((void *)bpf_skb_ct_lookup, ctx);
|
|
return 0;
|
|
}
|
|
|
|
char _license[] SEC("license") = "GPL";
|