Home
last modified time | relevance | path

Searched refs:skb (Results 1 – 13 of 13) sorted by relevance

/packages/modules/Connectivity/bpf/progs/
Dnetd.c179 static __always_inline inline void update_##the_stats_map(const struct __sk_buff* const skb, \ in DEFINE_BPF_MAP_RO_NETD()
192 uint64_t bytes = skb->len; \ in DEFINE_BPF_MAP_RO_NETD()
194 bool is_ipv6 = (skb->protocol == htons(ETH_P_IPV6)); \ in DEFINE_BPF_MAP_RO_NETD()
218 static __always_inline inline int bpf_skb_load_bytes_net(const struct __sk_buff* const skb,
240 ? bpf_skb_load_bytes_relative(skb, L3_off, to, len, BPF_HDR_START_NET)
241 : bpf_skb_load_bytes(skb, L3_off, to, len);
259 static __always_inline inline bool should_block_local_network_packets(struct __sk_buff *skb, in should_block_local_network_packets() argument
271 if (skb->protocol == htons(ETH_P_IP)) { in should_block_local_network_packets()
276 (void)bpf_skb_load_bytes_net(skb, remote_ip_ofs, &remote_ip6.s6_addr32[3], 4, kver); in should_block_local_network_packets()
277 (void)bpf_skb_load_bytes_net(skb, IP4_OFFSET(protocol), &ip_proto, sizeof(ip_proto), kver); in should_block_local_network_packets()
[all …]
Doffload.c64 static inline __always_inline int do_forward6(struct __sk_buff* skb,
71 if (skb->protocol != htons(ETH_P_IPV6)) return TC_ACT_PIPE;
74 if (is_ethernet && (skb->pkt_type != PACKET_HOST)) return TC_ACT_PIPE;
81 if (bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN)) return TC_ACT_PIPE;
83 void* data = (void*)(long)skb->data;
84 const void* data_end = (void*)(long)skb->data_end;
105 bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN + TCP_HLEN);
107 data = (void*)(long)skb->data;
108 data_end = (void*)(long)skb->data_end;
139 .iif = skb->ifindex,
[all …]
Doffload@mainline.c64 static inline __always_inline int do_forward6(struct __sk_buff* skb,
71 if (skb->protocol != htons(ETH_P_IPV6)) return TC_ACT_PIPE;
74 if (is_ethernet && (skb->pkt_type != PACKET_HOST)) return TC_ACT_PIPE;
81 if (bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN)) return TC_ACT_PIPE;
83 void* data = (void*)(long)skb->data;
84 const void* data_end = (void*)(long)skb->data_end;
105 bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN + TCP_HLEN);
107 data = (void*)(long)skb->data;
108 data_end = (void*)(long)skb->data_end;
139 .iif = skb->ifindex,
[all …]
Dclatd.c26 static inline __always_inline int nat64(struct __sk_buff* skb, in nat64() argument
32 if (is_ethernet && (skb->pkt_type != PACKET_HOST)) return TC_ACT_PIPE; in nat64()
35 if (skb->protocol != htons(ETH_P_IPV6)) return TC_ACT_PIPE; in nat64()
42 try_make_writable(skb, l2_header_size + sizeof(struct ipv6hdr)); in nat64()
44 void* data = (void*)(long)skb->data; in nat64()
45 const void* data_end = (void*)(long)skb->data_end; in nat64()
68 .iif = skb->ifindex, in nat64()
123 skb->mark = CLAT_MARK; in nat64()
169 if (bpf_skb_change_proto(skb, htons(ETH_P_IP), 0)) { in nat64()
173 skb->mark = CLAT_MARK; in nat64()
[all …]
Dbpf_net_helpers.h90 static uint64_t (*bpf_get_socket_cookie)(struct __sk_buff* skb) = (void*)BPF_FUNC_get_socket_cookie;
93 static uint32_t (*bpf_get_socket_uid)(struct __sk_buff* skb) = (void*)BPF_FUNC_get_socket_uid;
95 static int (*bpf_skb_pull_data)(struct __sk_buff* skb, __u32 len) = (void*)BPF_FUNC_skb_pull_data;
97 static int (*bpf_skb_load_bytes)(const struct __sk_buff* skb, int off, void* to,
100 static int (*bpf_skb_load_bytes_relative)(const struct __sk_buff* skb, int off, void* to, int len,
103 static int (*bpf_skb_store_bytes)(struct __sk_buff* skb, __u32 offset, const void* from, __u32 len,
109 static int64_t (*bpf_csum_update)(struct __sk_buff* skb, __wsum csum) = (void*)BPF_FUNC_csum_update;
111 static int (*bpf_skb_change_proto)(struct __sk_buff* skb, __be16 proto,
113 static int (*bpf_l3_csum_replace)(struct __sk_buff* skb, __u32 offset, __u64 from, __u64 to,
115 static int (*bpf_l4_csum_replace)(struct __sk_buff* skb, __u32 offset, __u64 from, __u64 to,
[all …]
DdscpPolicy.c39 static inline __always_inline void match_policy(struct __sk_buff* skb, const bool ipv4) { in match_policy() argument
40 void* data = (void*)(long)skb->data; in match_policy()
41 const void* data_end = (void*)(long)skb->data_end; in match_policy()
51 uint64_t cookie = bpf_get_socket_cookie(skb); in match_policy()
126 nomatch |= (skb->ifindex ^ existing_rule->ifindex); in match_policy()
147 bpf_l3_csum_replace(skb, l2_header_size + IP4_OFFSET(check), htons(tos), htons(newTos), in match_policy()
149 bpf_skb_store_bytes(skb, l2_header_size + IP4_OFFSET(tos), &newTos, sizeof(newTos), 0); in match_policy()
153 bpf_skb_store_bytes(skb, l2_header_size, &new_first_be32, sizeof(__be32), in match_policy()
192 nomatch |= (policy->ifindex ^ skb->ifindex); in match_policy()
243 .ifindex = skb->ifindex, in match_policy()
[all …]
Dtest.c41 (__unused struct __sk_buff* skb) {
Dtest@mainline.c41 (__unused struct __sk_buff* skb) {
/packages/modules/UprobeStats/src/bpf/headers/include/
Dbpf_helpers.h391 unsigned long long load_byte(void* skb, unsigned long long off) asm("llvm.bpf.load.byte");
392 unsigned long long load_half(void* skb, unsigned long long off) asm("llvm.bpf.load.half");
393 unsigned long long load_word(void* skb, unsigned long long off) asm("llvm.bpf.load.word");
/packages/modules/Connectivity/bpf/headers/include/
Dbpf_helpers.h439 unsigned long long load_byte(void* skb, unsigned long long off) asm("llvm.bpf.load.byte");
440 unsigned long long load_half(void* skb, unsigned long long off) asm("llvm.bpf.load.half");
441 unsigned long long load_word(void* skb, unsigned long long off) asm("llvm.bpf.load.word");
/packages/modules/AdServices/adservices/tests/unittest/service-core/topics/assets/classifier/
Dprecomputed_test_app_list.csv9096 com.skb.btvmobile 10055,10283
/packages/modules/AdServices/adservices/apk/assets/classifier/
Dprecomputed_app_list.csv9096 com.skb.btvmobile 10055,10283
/packages/modules/AdServices/adservices/tests/unittest/service-core/assets/mobiledatadownload/
Dbcde93568f2064cfc803d77873c5b4028435234e9096 com.skb.btvmobile 10055,10283