|
@@ -0,0 +1,86 @@
|
|
|
+//! eBPF bindings generated by rust-bindgen
|
|
|
+
|
|
|
+#![allow(
|
|
|
+ dead_code,
|
|
|
+ non_camel_case_types,
|
|
|
+ non_snake_case,
|
|
|
+ clippy::all,
|
|
|
+ missing_docs
|
|
|
+)]
|
|
|
+
|
|
|
+mod btf_internal_bindings;
|
|
|
+#[cfg(target_arch = "aarch64")]
|
|
|
+mod linux_bindings_aarch64;
|
|
|
+#[cfg(target_arch = "arm")]
|
|
|
+mod linux_bindings_armv7;
|
|
|
+#[cfg(target_arch = "riscv64")]
|
|
|
+mod linux_bindings_riscv64;
|
|
|
+#[cfg(target_arch = "x86_64")]
|
|
|
+mod linux_bindings_x86_64;
|
|
|
+
|
|
|
+pub use btf_internal_bindings::*;
|
|
|
+
|
|
|
+#[cfg(target_arch = "x86_64")]
|
|
|
+pub use linux_bindings_x86_64::*;
|
|
|
+
|
|
|
+#[cfg(target_arch = "arm")]
|
|
|
+pub use linux_bindings_armv7::*;
|
|
|
+
|
|
|
+#[cfg(target_arch = "aarch64")]
|
|
|
+pub use linux_bindings_aarch64::*;
|
|
|
+
|
|
|
+#[cfg(target_arch = "riscv64")]
|
|
|
+pub use linux_bindings_riscv64::*;
|
|
|
+
|
|
|
+pub mod maps {
|
|
|
+ /// Invalid map type encontered
|
|
|
+ pub struct InvalidMapTypeError {
|
|
|
+ /// The map type
|
|
|
+ pub map_type: u32,
|
|
|
+ }
|
|
|
+
|
|
|
+ impl TryFrom<u32> for super::bpf_map_type {
|
|
|
+ type Error = InvalidMapTypeError;
|
|
|
+
|
|
|
+ fn try_from(map_type: u32) -> Result<Self, Self::Error> {
|
|
|
+ use super::bpf_map_type::*;
|
|
|
+ Ok(match map_type {
|
|
|
+ x if x == BPF_MAP_TYPE_UNSPEC as u32 => BPF_MAP_TYPE_UNSPEC,
|
|
|
+ x if x == BPF_MAP_TYPE_HASH as u32 => BPF_MAP_TYPE_HASH,
|
|
|
+ x if x == BPF_MAP_TYPE_ARRAY as u32 => BPF_MAP_TYPE_ARRAY,
|
|
|
+ x if x == BPF_MAP_TYPE_PROG_ARRAY as u32 => BPF_MAP_TYPE_PROG_ARRAY,
|
|
|
+ x if x == BPF_MAP_TYPE_PERF_EVENT_ARRAY as u32 => BPF_MAP_TYPE_PERF_EVENT_ARRAY,
|
|
|
+ x if x == BPF_MAP_TYPE_PERCPU_HASH as u32 => BPF_MAP_TYPE_PERCPU_HASH,
|
|
|
+ x if x == BPF_MAP_TYPE_PERCPU_ARRAY as u32 => BPF_MAP_TYPE_PERCPU_ARRAY,
|
|
|
+ x if x == BPF_MAP_TYPE_STACK_TRACE as u32 => BPF_MAP_TYPE_STACK_TRACE,
|
|
|
+ x if x == BPF_MAP_TYPE_CGROUP_ARRAY as u32 => BPF_MAP_TYPE_CGROUP_ARRAY,
|
|
|
+ x if x == BPF_MAP_TYPE_LRU_HASH as u32 => BPF_MAP_TYPE_LRU_HASH,
|
|
|
+ x if x == BPF_MAP_TYPE_LRU_PERCPU_HASH as u32 => BPF_MAP_TYPE_LRU_PERCPU_HASH,
|
|
|
+ x if x == BPF_MAP_TYPE_LPM_TRIE as u32 => BPF_MAP_TYPE_LPM_TRIE,
|
|
|
+ x if x == BPF_MAP_TYPE_BLOOM_FILTER as u32 => BPF_MAP_TYPE_BLOOM_FILTER,
|
|
|
+ x if x == BPF_MAP_TYPE_ARRAY_OF_MAPS as u32 => BPF_MAP_TYPE_ARRAY_OF_MAPS,
|
|
|
+ x if x == BPF_MAP_TYPE_HASH_OF_MAPS as u32 => BPF_MAP_TYPE_HASH_OF_MAPS,
|
|
|
+ x if x == BPF_MAP_TYPE_DEVMAP as u32 => BPF_MAP_TYPE_DEVMAP,
|
|
|
+ x if x == BPF_MAP_TYPE_SOCKMAP as u32 => BPF_MAP_TYPE_SOCKMAP,
|
|
|
+ x if x == BPF_MAP_TYPE_CPUMAP as u32 => BPF_MAP_TYPE_CPUMAP,
|
|
|
+ x if x == BPF_MAP_TYPE_XSKMAP as u32 => BPF_MAP_TYPE_XSKMAP,
|
|
|
+ x if x == BPF_MAP_TYPE_SOCKHASH as u32 => BPF_MAP_TYPE_SOCKHASH,
|
|
|
+ x if x == BPF_MAP_TYPE_CGROUP_STORAGE as u32 => BPF_MAP_TYPE_CGROUP_STORAGE,
|
|
|
+ x if x == BPF_MAP_TYPE_REUSEPORT_SOCKARRAY as u32 => BPF_MAP_TYPE_REUSEPORT_SOCKARRAY,
|
|
|
+ x if x == BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE as u32 => {
|
|
|
+ BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE
|
|
|
+ }
|
|
|
+ x if x == BPF_MAP_TYPE_QUEUE as u32 => BPF_MAP_TYPE_QUEUE,
|
|
|
+ x if x == BPF_MAP_TYPE_STACK as u32 => BPF_MAP_TYPE_STACK,
|
|
|
+ x if x == BPF_MAP_TYPE_SK_STORAGE as u32 => BPF_MAP_TYPE_SK_STORAGE,
|
|
|
+ x if x == BPF_MAP_TYPE_DEVMAP_HASH as u32 => BPF_MAP_TYPE_DEVMAP_HASH,
|
|
|
+ x if x == BPF_MAP_TYPE_STRUCT_OPS as u32 => BPF_MAP_TYPE_STRUCT_OPS,
|
|
|
+ x if x == BPF_MAP_TYPE_RINGBUF as u32 => BPF_MAP_TYPE_RINGBUF,
|
|
|
+ x if x == BPF_MAP_TYPE_INODE_STORAGE as u32 => BPF_MAP_TYPE_INODE_STORAGE,
|
|
|
+ x if x == BPF_MAP_TYPE_TASK_STORAGE as u32 => BPF_MAP_TYPE_TASK_STORAGE,
|
|
|
+ _ => return Err(InvalidMapTypeError { map_type }),
|
|
|
+ })
|
|
|
+ }
|
|
|
+ }
|
|
|
+}
|
|
|
+
|