Home
last modified time | relevance | path

Searched refs:skb (Results 1 – 25 of 883) sorted by relevance

12345678910>>...36

/net/bridge/
Dbr_netfilter.c45 #define skb_origaddr(skb) (((struct bridge_skb_cb *) \ argument
46 (skb->nf_bridge->data))->daddr.ipv4)
47 #define store_orig_dstaddr(skb) (skb_origaddr(skb) = ip_hdr(skb)->daddr) argument
48 #define dnat_took_place(skb) (skb_origaddr(skb) != ip_hdr(skb)->daddr) argument
67 #define IS_IP(skb) \ argument
68 (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_IP))
70 #define IS_IPV6(skb) \ argument
71 (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_IPV6))
73 #define IS_ARP(skb) \ argument
74 (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_ARP))
[all …]
Dbr_forward.c25 struct sk_buff *skb,
27 struct sk_buff *skb));
31 const struct sk_buff *skb) in should_deliver() argument
33 return ((p->flags & BR_HAIRPIN_MODE) || skb->dev != p->dev) && in should_deliver()
34 br_allowed_egress(p->br, nbp_get_vlan_info(p), skb) && in should_deliver()
38 int br_dev_queue_push_xmit(struct sk_buff *skb) in br_dev_queue_push_xmit() argument
41 if (nf_bridge_maybe_copy_header(skb) || in br_dev_queue_push_xmit()
42 !is_skb_forwardable(skb->dev, skb)) { in br_dev_queue_push_xmit()
43 kfree_skb(skb); in br_dev_queue_push_xmit()
45 skb_push(skb, ETH_HLEN); in br_dev_queue_push_xmit()
[all …]
Dbr_input.c27 static int br_pass_frame_up(struct sk_buff *skb) in br_pass_frame_up() argument
29 struct net_device *indev, *brdev = BR_INPUT_SKB_CB(skb)->brdev; in br_pass_frame_up()
36 brstats->rx_bytes += skb->len; in br_pass_frame_up()
45 !br_allowed_egress(br, pv, skb)) { in br_pass_frame_up()
46 kfree_skb(skb); in br_pass_frame_up()
50 indev = skb->dev; in br_pass_frame_up()
51 skb->dev = brdev; in br_pass_frame_up()
52 skb = br_handle_vlan(br, pv, skb); in br_pass_frame_up()
53 if (!skb) in br_pass_frame_up()
56 return NF_HOOK(NFPROTO_BRIDGE, NF_BR_LOCAL_IN, skb, indev, NULL, in br_pass_frame_up()
[all …]
/net/openvswitch/
Dactions.c41 static int do_execute_actions(struct datapath *dp, struct sk_buff *skb,
46 struct sk_buff *skb; member
92 static struct deferred_action *add_deferred_actions(struct sk_buff *skb, in add_deferred_actions() argument
102 da->skb = skb; in add_deferred_actions()
110 static int make_writable(struct sk_buff *skb, int write_len) in make_writable() argument
112 if (!pskb_may_pull(skb, write_len)) in make_writable()
115 if (!skb_cloned(skb) || skb_clone_writable(skb, write_len)) in make_writable()
118 return pskb_expand_head(skb, 0, 0, GFP_ATOMIC); in make_writable()
122 static int __pop_vlan_tci(struct sk_buff *skb, __be16 *current_tci) in __pop_vlan_tci() argument
127 err = make_writable(skb, VLAN_ETH_HLEN); in __pop_vlan_tci()
[all …]
/net/core/
Dskbuff.c95 static void skb_panic(struct sk_buff *skb, unsigned int sz, void *addr, in skb_panic() argument
99 msg, addr, skb->len, sz, skb->head, skb->data, in skb_panic()
100 (unsigned long)skb->tail, (unsigned long)skb->end, in skb_panic()
101 skb->dev ? skb->dev->name : "<NULL>"); in skb_panic()
105 static void skb_over_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_over_panic() argument
107 skb_panic(skb, sz, addr, __func__); in skb_over_panic()
110 static void skb_under_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_under_panic() argument
112 skb_panic(skb, sz, addr, __func__); in skb_under_panic()
160 struct sk_buff *skb; in __alloc_skb_head() local
163 skb = kmem_cache_alloc_node(skbuff_head_cache, in __alloc_skb_head()
[all …]
/net/ipv6/
Dxfrm6_output.c23 int xfrm6_find_1stfragopt(struct xfrm_state *x, struct sk_buff *skb, in xfrm6_find_1stfragopt() argument
26 return ip6_find_1stfragopt(skb, prevhdr); in xfrm6_find_1stfragopt()
30 static int xfrm6_local_dontfrag(struct sk_buff *skb) in xfrm6_local_dontfrag() argument
33 struct sock *sk = skb->sk; in xfrm6_local_dontfrag()
47 static void xfrm6_local_rxpmtu(struct sk_buff *skb, u32 mtu) in xfrm6_local_rxpmtu() argument
50 struct sock *sk = skb->sk; in xfrm6_local_rxpmtu()
53 fl6.daddr = ipv6_hdr(skb)->daddr; in xfrm6_local_rxpmtu()
58 void xfrm6_local_error(struct sk_buff *skb, u32 mtu) in xfrm6_local_error() argument
62 struct sock *sk = skb->sk; in xfrm6_local_error()
64 hdr = skb->encapsulation ? inner_ipv6_hdr(skb) : ipv6_hdr(skb); in xfrm6_local_error()
[all …]
Dexthdrs.c62 bool (*func)(struct sk_buff *skb, int offset);
71 static bool ip6_tlvopt_unknown(struct sk_buff *skb, int optoff) in ip6_tlvopt_unknown() argument
73 switch ((skb_network_header(skb)[optoff] & 0xC0) >> 6) { in ip6_tlvopt_unknown()
84 if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) in ip6_tlvopt_unknown()
87 icmpv6_param_prob(skb, ICMPV6_UNK_OPTION, optoff); in ip6_tlvopt_unknown()
91 kfree_skb(skb); in ip6_tlvopt_unknown()
97 static bool ip6_parse_tlv(const struct tlvtype_proc *procs, struct sk_buff *skb) in ip6_parse_tlv() argument
100 const unsigned char *nh = skb_network_header(skb); in ip6_parse_tlv()
101 int off = skb_network_header_len(skb); in ip6_parse_tlv()
102 int len = (skb_transport_header(skb)[1] + 1) << 3; in ip6_parse_tlv()
[all …]
Dip6_input.c50 int ip6_rcv_finish(struct sk_buff *skb) in ip6_rcv_finish() argument
52 if (sysctl_ip_early_demux && !skb_dst(skb) && skb->sk == NULL) { in ip6_rcv_finish()
55 ipprot = rcu_dereference(inet6_protos[ipv6_hdr(skb)->nexthdr]); in ip6_rcv_finish()
57 ipprot->early_demux(skb); in ip6_rcv_finish()
59 if (!skb_dst(skb)) in ip6_rcv_finish()
60 ip6_route_input(skb); in ip6_rcv_finish()
62 return dst_input(skb); in ip6_rcv_finish()
65 int ipv6_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device… in ipv6_rcv() argument
70 struct net *net = dev_net(skb->dev); in ipv6_rcv()
72 if (skb->pkt_type == PACKET_OTHERHOST) { in ipv6_rcv()
[all …]
Dip6_output.c59 static int ip6_finish_output2(struct sk_buff *skb) in ip6_finish_output2() argument
61 struct dst_entry *dst = skb_dst(skb); in ip6_finish_output2()
67 skb->protocol = htons(ETH_P_IPV6); in ip6_finish_output2()
68 skb->dev = dev; in ip6_finish_output2()
70 if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) { in ip6_finish_output2()
71 struct inet6_dev *idev = ip6_dst_idev(skb_dst(skb)); in ip6_finish_output2()
73 if (!(dev->flags & IFF_LOOPBACK) && sk_mc_loop(skb->sk) && in ip6_finish_output2()
74 ((mroute6_socket(dev_net(dev), skb) && in ip6_finish_output2()
75 !(IP6CB(skb)->flags & IP6SKB_FORWARDED)) || in ip6_finish_output2()
76 ipv6_chk_mcast_addr(dev, &ipv6_hdr(skb)->daddr, in ip6_finish_output2()
[all …]
Dip6_offload.c22 static int ipv6_gso_pull_exthdrs(struct sk_buff *skb, int proto) in ipv6_gso_pull_exthdrs() argument
40 if (unlikely(!pskb_may_pull(skb, 8))) in ipv6_gso_pull_exthdrs()
43 opth = (void *)skb->data; in ipv6_gso_pull_exthdrs()
46 if (unlikely(!pskb_may_pull(skb, len))) in ipv6_gso_pull_exthdrs()
49 opth = (void *)skb->data; in ipv6_gso_pull_exthdrs()
51 __skb_pull(skb, len); in ipv6_gso_pull_exthdrs()
57 static struct sk_buff *ipv6_gso_segment(struct sk_buff *skb, in ipv6_gso_segment() argument
70 if (unlikely(skb_shinfo(skb)->gso_type & in ipv6_gso_segment()
86 skb_reset_network_header(skb); in ipv6_gso_segment()
87 nhoff = skb_network_header(skb) - skb_mac_header(skb); in ipv6_gso_segment()
[all …]
/net/lapb/
Dlapb_subr.c53 struct sk_buff *skb; in lapb_frames_acked() local
63 skb = skb_dequeue(&lapb->ack_queue); in lapb_frames_acked()
64 kfree_skb(skb); in lapb_frames_acked()
71 struct sk_buff *skb, *skb_prev = NULL; in lapb_requeue_frames() local
78 while ((skb = skb_dequeue(&lapb->ack_queue)) != NULL) { in lapb_requeue_frames()
80 skb_queue_head(&lapb->write_queue, skb); in lapb_requeue_frames()
82 skb_append(skb_prev, skb, &lapb->write_queue); in lapb_requeue_frames()
83 skb_prev = skb; in lapb_requeue_frames()
111 int lapb_decode(struct lapb_cb *lapb, struct sk_buff *skb, in lapb_decode() argument
118 skb->data[0], skb->data[1], skb->data[2]); in lapb_decode()
[all …]
/net/ipv4/
Dtcp_offload.c17 static void tcp_gso_tstamp(struct sk_buff *skb, unsigned int ts_seq, in tcp_gso_tstamp() argument
20 while (skb) { in tcp_gso_tstamp()
22 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp()
23 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp()
27 skb = skb->next; in tcp_gso_tstamp()
32 struct sk_buff *tcp4_gso_segment(struct sk_buff *skb, in tcp4_gso_segment() argument
35 if (!pskb_may_pull(skb, sizeof(struct tcphdr))) in tcp4_gso_segment()
38 if (unlikely(skb->ip_summed != CHECKSUM_PARTIAL)) { in tcp4_gso_segment()
39 const struct iphdr *iph = ip_hdr(skb); in tcp4_gso_segment()
40 struct tcphdr *th = tcp_hdr(skb); in tcp4_gso_segment()
[all …]
Dudp_offload.c28 static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb, in __skb_udp_tunnel_segment() argument
30 struct sk_buff *(*gso_inner_segment)(struct sk_buff *skb, in __skb_udp_tunnel_segment() argument
35 u16 mac_offset = skb->mac_header; in __skb_udp_tunnel_segment()
36 int mac_len = skb->mac_len; in __skb_udp_tunnel_segment()
37 int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb); in __skb_udp_tunnel_segment()
38 __be16 protocol = skb->protocol; in __skb_udp_tunnel_segment()
44 oldlen = (u16)~skb->len; in __skb_udp_tunnel_segment()
46 if (unlikely(!pskb_may_pull(skb, tnl_hlen))) in __skb_udp_tunnel_segment()
49 skb->encapsulation = 0; in __skb_udp_tunnel_segment()
50 __skb_pull(skb, tnl_hlen); in __skb_udp_tunnel_segment()
[all …]
Dip_output.c94 int __ip_local_out(struct sk_buff *skb) in __ip_local_out() argument
96 struct iphdr *iph = ip_hdr(skb); in __ip_local_out()
98 iph->tot_len = htons(skb->len); in __ip_local_out()
100 return nf_hook(NFPROTO_IPV4, NF_INET_LOCAL_OUT, skb, NULL, in __ip_local_out()
101 skb_dst(skb)->dev, dst_output); in __ip_local_out()
104 int ip_local_out_sk(struct sock *sk, struct sk_buff *skb) in ip_local_out_sk() argument
108 err = __ip_local_out(skb); in ip_local_out_sk()
110 err = dst_output_sk(sk, skb); in ip_local_out_sk()
129 int ip_build_and_send_pkt(struct sk_buff *skb, struct sock *sk, in ip_build_and_send_pkt() argument
133 struct rtable *rt = skb_rtable(skb); in ip_build_and_send_pkt()
[all …]
Dip_input.c153 bool ip_call_ra_chain(struct sk_buff *skb) in ip_call_ra_chain() argument
156 u8 protocol = ip_hdr(skb)->protocol; in ip_call_ra_chain()
158 struct net_device *dev = skb->dev; in ip_call_ra_chain()
170 if (ip_is_fragment(ip_hdr(skb))) { in ip_call_ra_chain()
171 if (ip_defrag(skb, IP_DEFRAG_CALL_RA_CHAIN)) in ip_call_ra_chain()
175 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_call_ra_chain()
184 raw_rcv(last, skb); in ip_call_ra_chain()
190 static int ip_local_deliver_finish(struct sk_buff *skb) in ip_local_deliver_finish() argument
192 struct net *net = dev_net(skb->dev); in ip_local_deliver_finish()
194 __skb_pull(skb, skb_network_header_len(skb)); in ip_local_deliver_finish()
[all …]
Dgre_offload.c18 static struct sk_buff *gre_gso_segment(struct sk_buff *skb, in gre_gso_segment() argument
25 u16 mac_offset = skb->mac_header; in gre_gso_segment()
26 int mac_len = skb->mac_len; in gre_gso_segment()
27 __be16 protocol = skb->protocol; in gre_gso_segment()
31 if (unlikely(skb_shinfo(skb)->gso_type & in gre_gso_segment()
43 if (!skb->encapsulation) in gre_gso_segment()
46 if (unlikely(!pskb_may_pull(skb, sizeof(*greh)))) in gre_gso_segment()
49 greh = (struct gre_base_hdr *)skb_transport_header(skb); in gre_gso_segment()
51 ghl = skb_inner_mac_header(skb) - skb_transport_header(skb); in gre_gso_segment()
57 skb->encap_hdr_csum = 1; in gre_gso_segment()
[all …]
Dip_forward.c42 static bool ip_may_fragment(const struct sk_buff *skb) in ip_may_fragment() argument
44 return unlikely((ip_hdr(skb)->frag_off & htons(IP_DF)) == 0) || in ip_may_fragment()
45 skb->ignore_df; in ip_may_fragment()
48 static bool ip_exceeds_mtu(const struct sk_buff *skb, unsigned int mtu) in ip_exceeds_mtu() argument
50 if (skb->len <= mtu) in ip_exceeds_mtu()
53 if (skb_is_gso(skb) && skb_gso_network_seglen(skb) <= mtu) in ip_exceeds_mtu()
60 static int ip_forward_finish(struct sk_buff *skb) in ip_forward_finish() argument
62 struct ip_options *opt = &(IPCB(skb)->opt); in ip_forward_finish()
64 IP_INC_STATS_BH(dev_net(skb_dst(skb)->dev), IPSTATS_MIB_OUTFORWDATAGRAMS); in ip_forward_finish()
65 IP_ADD_STATS_BH(dev_net(skb_dst(skb)->dev), IPSTATS_MIB_OUTOCTETS, skb->len); in ip_forward_finish()
[all …]
Dxfrm4_output.c21 static int xfrm4_tunnel_check_size(struct sk_buff *skb) in xfrm4_tunnel_check_size() argument
25 if (IPCB(skb)->flags & IPSKB_XFRM_TUNNEL_SIZE) in xfrm4_tunnel_check_size()
28 if (!(ip_hdr(skb)->frag_off & htons(IP_DF)) || skb->ignore_df) in xfrm4_tunnel_check_size()
31 mtu = dst_mtu(skb_dst(skb)); in xfrm4_tunnel_check_size()
32 if (skb->len > mtu) { in xfrm4_tunnel_check_size()
33 if (skb->sk) in xfrm4_tunnel_check_size()
34 xfrm_local_error(skb, mtu); in xfrm4_tunnel_check_size()
36 icmp_send(skb, ICMP_DEST_UNREACH, in xfrm4_tunnel_check_size()
44 int xfrm4_extract_output(struct xfrm_state *x, struct sk_buff *skb) in xfrm4_extract_output() argument
48 err = xfrm4_tunnel_check_size(skb); in xfrm4_extract_output()
[all …]
Dnetfilter.c20 int ip_route_me_harder(struct sk_buff *skb, unsigned int addr_type) in ip_route_me_harder() argument
22 struct net *net = dev_net(skb_dst(skb)->dev); in ip_route_me_harder()
23 const struct iphdr *iph = ip_hdr(skb); in ip_route_me_harder()
27 __u8 flags = skb->sk ? inet_sk_flowi_flags(skb->sk) : 0; in ip_route_me_harder()
43 fl4.flowi4_oif = skb->sk ? skb->sk->sk_bound_dev_if : 0; in ip_route_me_harder()
44 fl4.flowi4_mark = skb->mark; in ip_route_me_harder()
51 skb_dst_drop(skb); in ip_route_me_harder()
52 skb_dst_set(skb, &rt->dst); in ip_route_me_harder()
54 if (skb_dst(skb)->error) in ip_route_me_harder()
55 return skb_dst(skb)->error; in ip_route_me_harder()
[all …]
/net/ax25/
Dax25_in.c38 static int ax25_rx_fragment(ax25_cb *ax25, struct sk_buff *skb) in ax25_rx_fragment() argument
43 if (!(*skb->data & AX25_SEG_FIRST)) { in ax25_rx_fragment()
44 if ((ax25->fragno - 1) == (*skb->data & AX25_SEG_REM)) { in ax25_rx_fragment()
46 ax25->fragno = *skb->data & AX25_SEG_REM; in ax25_rx_fragment()
47 skb_pull(skb, 1); /* skip fragno */ in ax25_rx_fragment()
48 ax25->fraglen += skb->len; in ax25_rx_fragment()
49 skb_queue_tail(&ax25->frag_queue, skb); in ax25_rx_fragment()
86 if (*skb->data & AX25_SEG_FIRST) { in ax25_rx_fragment()
88 ax25->fragno = *skb->data & AX25_SEG_REM; in ax25_rx_fragment()
89 skb_pull(skb, 1); /* skip fragno */ in ax25_rx_fragment()
[all …]
/net/netfilter/ipvs/
Dip_vs_xmit.c108 __mtu_check_toobig_v6(const struct sk_buff *skb, u32 mtu) in __mtu_check_toobig_v6() argument
110 if (IP6CB(skb)->frag_max_size) { in __mtu_check_toobig_v6()
114 if (IP6CB(skb)->frag_max_size > mtu) in __mtu_check_toobig_v6()
117 else if (skb->len > mtu && !skb_is_gso(skb)) { in __mtu_check_toobig_v6()
166 static inline bool crosses_local_route_boundary(int skb_af, struct sk_buff *skb, in crosses_local_route_boundary() argument
178 int addr_type = ipv6_addr_type(&ipv6_hdr(skb)->saddr); in crosses_local_route_boundary()
181 (!skb->dev || skb->dev->flags & IFF_LOOPBACK) && in crosses_local_route_boundary()
184 (struct rt6_info *)skb_dst(skb)); in crosses_local_route_boundary()
188 source_is_loopback = ipv4_is_loopback(ip_hdr(skb)->saddr); in crosses_local_route_boundary()
189 old_rt_is_local = skb_rtable(skb)->rt_flags & RTCF_LOCAL; in crosses_local_route_boundary()
[all …]
/net/llc/
Dllc_c_ev.c78 struct sk_buff *skb; in llc_util_nr_inside_tx_window() local
88 skb = skb_peek(&llc->pdu_unack_q); in llc_util_nr_inside_tx_window()
89 pdu = llc_pdu_sn_hdr(skb); in llc_util_nr_inside_tx_window()
91 skb = skb_peek_tail(&llc->pdu_unack_q); in llc_util_nr_inside_tx_window()
92 pdu = llc_pdu_sn_hdr(skb); in llc_util_nr_inside_tx_window()
99 int llc_conn_ev_conn_req(struct sock *sk, struct sk_buff *skb) in llc_conn_ev_conn_req() argument
101 const struct llc_conn_state_ev *ev = llc_conn_ev(skb); in llc_conn_ev_conn_req()
107 int llc_conn_ev_data_req(struct sock *sk, struct sk_buff *skb) in llc_conn_ev_data_req() argument
109 const struct llc_conn_state_ev *ev = llc_conn_ev(skb); in llc_conn_ev_data_req()
115 int llc_conn_ev_disc_req(struct sock *sk, struct sk_buff *skb) in llc_conn_ev_disc_req() argument
[all …]
/net/decnet/
Ddn_nsp_in.c81 static void dn_log_martian(struct sk_buff *skb, const char *msg) in dn_log_martian() argument
84 char *devname = skb->dev ? skb->dev->name : "???"; in dn_log_martian()
85 struct dn_skb_cb *cb = DN_SKB_CB(skb); in dn_log_martian()
100 static void dn_ack(struct sock *sk, struct sk_buff *skb, unsigned short ack) in dn_ack() argument
110 wakeup |= dn_nsp_check_xmit_queue(sk, skb, in dn_ack()
120 wakeup |= dn_nsp_check_xmit_queue(sk, skb, in dn_ack()
136 static int dn_process_ack(struct sock *sk, struct sk_buff *skb, int oth) in dn_process_ack() argument
138 __le16 *ptr = (__le16 *)skb->data; in dn_process_ack()
142 if (skb->len < 2) in dn_process_ack()
146 skb_pull(skb, 2); in dn_process_ack()
[all …]
/net/x25/
Dx25_dev.c31 static int x25_receive_data(struct sk_buff *skb, struct x25_neigh *nb) in x25_receive_data() argument
37 if (!pskb_may_pull(skb, X25_STD_MIN_LEN)) in x25_receive_data()
40 frametype = skb->data[2]; in x25_receive_data()
41 lci = ((skb->data[0] << 8) & 0xF00) + ((skb->data[1] << 0) & 0x0FF); in x25_receive_data()
48 x25_link_control(skb, nb, frametype); in x25_receive_data()
58 skb_reset_transport_header(skb); in x25_receive_data()
61 queued = x25_process_rx_frame(sk, skb); in x25_receive_data()
63 queued = !sk_add_backlog(sk, skb, sk->sk_rcvbuf); in x25_receive_data()
74 return x25_rx_call_request(skb, nb, lci); in x25_receive_data()
81 if (x25_forward_data(lci, nb, skb)) { in x25_receive_data()
[all …]
/net/dsa/
Dtag_edsa.c19 static netdev_tx_t edsa_xmit(struct sk_buff *skb, struct net_device *dev) in edsa_xmit() argument
25 dev->stats.tx_bytes += skb->len; in edsa_xmit()
33 if (skb->protocol == htons(ETH_P_8021Q)) { in edsa_xmit()
34 if (skb_cow_head(skb, DSA_HLEN) < 0) in edsa_xmit()
36 skb_push(skb, DSA_HLEN); in edsa_xmit()
38 memmove(skb->data, skb->data + DSA_HLEN, 2 * ETH_ALEN); in edsa_xmit()
43 edsa_header = skb->data + 2 * ETH_ALEN; in edsa_xmit()
59 if (skb_cow_head(skb, EDSA_HLEN) < 0) in edsa_xmit()
61 skb_push(skb, EDSA_HLEN); in edsa_xmit()
63 memmove(skb->data, skb->data + EDSA_HLEN, 2 * ETH_ALEN); in edsa_xmit()
[all …]

12345678910>>...36