/packages/modules/Connectivity/bpf/progs/ |
D | netd.c | 179 static __always_inline inline void update_##the_stats_map(const struct __sk_buff* const skb, \ in DEFINE_BPF_MAP_RO_NETD() 192 uint64_t bytes = skb->len; \ in DEFINE_BPF_MAP_RO_NETD() 194 bool is_ipv6 = (skb->protocol == htons(ETH_P_IPV6)); \ in DEFINE_BPF_MAP_RO_NETD() 218 static __always_inline inline int bpf_skb_load_bytes_net(const struct __sk_buff* const skb, 240 ? bpf_skb_load_bytes_relative(skb, L3_off, to, len, BPF_HDR_START_NET) 241 : bpf_skb_load_bytes(skb, L3_off, to, len); 259 static __always_inline inline bool should_block_local_network_packets(struct __sk_buff *skb, in should_block_local_network_packets() argument 271 if (skb->protocol == htons(ETH_P_IP)) { in should_block_local_network_packets() 276 (void)bpf_skb_load_bytes_net(skb, remote_ip_ofs, &remote_ip6.s6_addr32[3], 4, kver); in should_block_local_network_packets() 277 (void)bpf_skb_load_bytes_net(skb, IP4_OFFSET(protocol), &ip_proto, sizeof(ip_proto), kver); in should_block_local_network_packets() [all …]
|
D | offload.c | 64 static inline __always_inline int do_forward6(struct __sk_buff* skb, 71 if (skb->protocol != htons(ETH_P_IPV6)) return TC_ACT_PIPE; 74 if (is_ethernet && (skb->pkt_type != PACKET_HOST)) return TC_ACT_PIPE; 81 if (bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN)) return TC_ACT_PIPE; 83 void* data = (void*)(long)skb->data; 84 const void* data_end = (void*)(long)skb->data_end; 105 bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN + TCP_HLEN); 107 data = (void*)(long)skb->data; 108 data_end = (void*)(long)skb->data_end; 139 .iif = skb->ifindex, [all …]
|
D | offload@mainline.c | 64 static inline __always_inline int do_forward6(struct __sk_buff* skb, 71 if (skb->protocol != htons(ETH_P_IPV6)) return TC_ACT_PIPE; 74 if (is_ethernet && (skb->pkt_type != PACKET_HOST)) return TC_ACT_PIPE; 81 if (bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN)) return TC_ACT_PIPE; 83 void* data = (void*)(long)skb->data; 84 const void* data_end = (void*)(long)skb->data_end; 105 bpf_skb_pull_data(skb, l2_header_size + IP6_HLEN + TCP_HLEN); 107 data = (void*)(long)skb->data; 108 data_end = (void*)(long)skb->data_end; 139 .iif = skb->ifindex, [all …]
|
D | clatd.c | 26 static inline __always_inline int nat64(struct __sk_buff* skb, in nat64() argument 32 if (is_ethernet && (skb->pkt_type != PACKET_HOST)) return TC_ACT_PIPE; in nat64() 35 if (skb->protocol != htons(ETH_P_IPV6)) return TC_ACT_PIPE; in nat64() 42 try_make_writable(skb, l2_header_size + sizeof(struct ipv6hdr)); in nat64() 44 void* data = (void*)(long)skb->data; in nat64() 45 const void* data_end = (void*)(long)skb->data_end; in nat64() 68 .iif = skb->ifindex, in nat64() 123 skb->mark = CLAT_MARK; in nat64() 169 if (bpf_skb_change_proto(skb, htons(ETH_P_IP), 0)) { in nat64() 173 skb->mark = CLAT_MARK; in nat64() [all …]
|
D | bpf_net_helpers.h | 90 static uint64_t (*bpf_get_socket_cookie)(struct __sk_buff* skb) = (void*)BPF_FUNC_get_socket_cookie; 93 static uint32_t (*bpf_get_socket_uid)(struct __sk_buff* skb) = (void*)BPF_FUNC_get_socket_uid; 95 static int (*bpf_skb_pull_data)(struct __sk_buff* skb, __u32 len) = (void*)BPF_FUNC_skb_pull_data; 97 static int (*bpf_skb_load_bytes)(const struct __sk_buff* skb, int off, void* to, 100 static int (*bpf_skb_load_bytes_relative)(const struct __sk_buff* skb, int off, void* to, int len, 103 static int (*bpf_skb_store_bytes)(struct __sk_buff* skb, __u32 offset, const void* from, __u32 len, 109 static int64_t (*bpf_csum_update)(struct __sk_buff* skb, __wsum csum) = (void*)BPF_FUNC_csum_update; 111 static int (*bpf_skb_change_proto)(struct __sk_buff* skb, __be16 proto, 113 static int (*bpf_l3_csum_replace)(struct __sk_buff* skb, __u32 offset, __u64 from, __u64 to, 115 static int (*bpf_l4_csum_replace)(struct __sk_buff* skb, __u32 offset, __u64 from, __u64 to, [all …]
|
D | dscpPolicy.c | 39 static inline __always_inline void match_policy(struct __sk_buff* skb, const bool ipv4) { in match_policy() argument 40 void* data = (void*)(long)skb->data; in match_policy() 41 const void* data_end = (void*)(long)skb->data_end; in match_policy() 51 uint64_t cookie = bpf_get_socket_cookie(skb); in match_policy() 126 nomatch |= (skb->ifindex ^ existing_rule->ifindex); in match_policy() 147 bpf_l3_csum_replace(skb, l2_header_size + IP4_OFFSET(check), htons(tos), htons(newTos), in match_policy() 149 bpf_skb_store_bytes(skb, l2_header_size + IP4_OFFSET(tos), &newTos, sizeof(newTos), 0); in match_policy() 153 bpf_skb_store_bytes(skb, l2_header_size, &new_first_be32, sizeof(__be32), in match_policy() 192 nomatch |= (policy->ifindex ^ skb->ifindex); in match_policy() 243 .ifindex = skb->ifindex, in match_policy() [all …]
|
D | test.c | 41 (__unused struct __sk_buff* skb) {
|
D | test@mainline.c | 41 (__unused struct __sk_buff* skb) {
|
/packages/modules/UprobeStats/src/bpf/headers/include/ |
D | bpf_helpers.h | 391 unsigned long long load_byte(void* skb, unsigned long long off) asm("llvm.bpf.load.byte"); 392 unsigned long long load_half(void* skb, unsigned long long off) asm("llvm.bpf.load.half"); 393 unsigned long long load_word(void* skb, unsigned long long off) asm("llvm.bpf.load.word");
|
/packages/modules/Connectivity/bpf/headers/include/ |
D | bpf_helpers.h | 439 unsigned long long load_byte(void* skb, unsigned long long off) asm("llvm.bpf.load.byte"); 440 unsigned long long load_half(void* skb, unsigned long long off) asm("llvm.bpf.load.half"); 441 unsigned long long load_word(void* skb, unsigned long long off) asm("llvm.bpf.load.word");
|
/packages/modules/AdServices/adservices/tests/unittest/service-core/topics/assets/classifier/ |
D | precomputed_test_app_list.csv | 9096 com.skb.btvmobile 10055,10283
|
/packages/modules/AdServices/adservices/apk/assets/classifier/ |
D | precomputed_app_list.csv | 9096 com.skb.btvmobile 10055,10283
|
/packages/modules/AdServices/adservices/tests/unittest/service-core/assets/mobiledatadownload/ |
D | bcde93568f2064cfc803d77873c5b4028435234e | 9096 com.skb.btvmobile 10055,10283
|