/net/rds/ |
D | iw_send.c | 502 struct rds_iw_send_work *first; in rds_iw_xmit() local 623 first = send; in rds_iw_xmit() 746 failed_wr = &first->s_wr; in rds_iw_xmit() 747 ret = ib_post_send(ic->i_cm_id->qp, &first->s_wr, &failed_wr); in rds_iw_xmit() 749 first, &first->s_wr, ret, failed_wr); in rds_iw_xmit() 750 BUG_ON(failed_wr != &first->s_wr); in rds_iw_xmit() 792 struct rds_iw_send_work *first; in rds_iw_xmit_rdma() local 851 first = prev = &ic->i_sends[fr_pos]; in rds_iw_xmit_rdma() 853 first = send; in rds_iw_xmit_rdma() 932 first->s_wr.send_flags = IB_SEND_SIGNALED; in rds_iw_xmit_rdma() [all …]
|
D | ib_send.c | 536 struct rds_ib_send_work *first; in rds_ib_xmit() local 664 first = send; in rds_ib_xmit() 761 failed_wr = &first->s_wr; in rds_ib_xmit() 762 ret = ib_post_send(ic->i_cm_id->qp, &first->s_wr, &failed_wr); in rds_ib_xmit() 764 first, &first->s_wr, ret, failed_wr); in rds_ib_xmit() 765 BUG_ON(failed_wr != &first->s_wr); in rds_ib_xmit() 884 struct rds_ib_send_work *first; in rds_ib_xmit_rdma() local 930 first = send; in rds_ib_xmit_rdma() 995 failed_wr = &first->s_wr; in rds_ib_xmit_rdma() 996 ret = ib_post_send(ic->i_cm_id->qp, &first->s_wr, &failed_wr); in rds_ib_xmit_rdma() [all …]
|
D | ib_recv.c | 111 head->first = NULL; in rds_ib_recv_alloc_cache() 142 if (head->first) { in rds_ib_cache_splice_all_lists() 143 list_splice_entire_tail(head->first, caller_list); in rds_ib_cache_splice_all_lists() 144 head->first = NULL; in rds_ib_cache_splice_all_lists() 428 chpfirst = __this_cpu_read(cache->percpu->first); in rds_ib_recv_cache_put() 434 __this_cpu_write(cache->percpu->first, new_item); in rds_ib_recv_cache_put() 454 __this_cpu_write(cache->percpu->first, NULL); in rds_ib_recv_cache_put()
|
D | ib.h | 48 struct list_head *first; member
|
/net/sched/ |
D | sch_fq.c | 77 struct fq_flow *first; member 156 if (head->first) in fq_flow_add_tail() 159 head->first = flow; in fq_flow_add_tail() 429 if (!head->first) { in fq_dequeue() 431 if (!head->first) { in fq_dequeue() 439 f = head->first; in fq_dequeue() 443 head->first = f->next; in fq_dequeue() 449 head->first = f->next; in fq_dequeue() 456 head->first = f->next; in fq_dequeue() 458 if ((head == &q->new_flows) && q->old_flows.first) { in fq_dequeue() [all …]
|
D | cls_api.c | 111 u32 first = TC_H_MAKE(0xC0000000U, 0U); in tcf_auto_prio() local 114 first = tp->prio - 1; in tcf_auto_prio() 116 return first; in tcf_auto_prio()
|
D | Kconfig | 10 device, it has to decide which ones to send first, which ones to 16 is a FIFO (first come, first served). If you say Y here, you will be 334 command, only packets up to the first plug are released for delivery.
|
/net/netrom/ |
D | nr_route.c | 600 struct net_device *dev, *first = NULL; in nr_dev_first() local 605 if (first == NULL || strncmp(dev->name, first->name, 3) < 0) in nr_dev_first() 606 first = dev; in nr_dev_first() 608 if (first) in nr_dev_first() 609 dev_hold(first); in nr_dev_first() 612 return first; in nr_dev_first()
|
/net/ax25/ |
D | ax25_out.c | 124 int frontlen, len, fragno, ka9qfrag, first = 1; in ax25_output() local 171 if (first) { in ax25_output() 173 first = 0; in ax25_output()
|
/net/netfilter/ipset/ |
D | ip_set_bitmap_gen.h | 205 u32 id, first = cb->args[IPSET_CB_ARG0]; in mtype_list() local 223 if (id == first) { in mtype_list() 245 if (unlikely(id == first)) { in mtype_list()
|
D | ip_set_list_set.c | 505 u32 i, first = cb->args[IPSET_CB_ARG0]; in list_set_list() local 522 if (i == first) { in list_set_list() 543 if (unlikely(i == first)) { in list_set_list()
|
D | ip_set_hash_gen.h | 960 u32 first = cb->args[IPSET_CB_ARG0]; in mtype_list() local 984 if (cb->args[IPSET_CB_ARG0] == first) { in mtype_list() 1005 if (unlikely(first == cb->args[IPSET_CB_ARG0])) { in mtype_list()
|
/net/rose/ |
D | rose_route.c | 610 struct net_device *dev, *first = NULL; in rose_dev_first() local 615 if (first == NULL || strncmp(dev->name, first->name, 3) < 0) in rose_dev_first() 616 first = dev; in rose_dev_first() 620 return first; in rose_dev_first()
|
/net/tipc/ |
D | link.c | 859 u32 first = buf_seqno(buf); in tipc_link_push_packet() local 861 while (buf && less(first, r_q_head)) { in tipc_link_push_packet() 862 first = mod(first + 1); in tipc_link_push_packet() 865 l_ptr->retransm_queue_head = r_q_head = first; in tipc_link_push_packet() 866 l_ptr->retransm_queue_size = r_q_size = mod(last - first); in tipc_link_push_packet() 897 u32 first = buf_seqno(l_ptr->first_out); in tipc_link_push_packet() local 899 if (mod(next - first) < l_ptr->queue_limit[0]) { in tipc_link_push_packet()
|
/net/caif/ |
D | cfctrl.c | 143 struct cfctrl_request_info *p, *tmp, *first; in cfctrl_remove_req() local 145 first = list_first_entry(&ctrl->list, struct cfctrl_request_info, list); in cfctrl_remove_req() 149 if (p != first) in cfctrl_remove_req()
|
/net/decnet/ |
D | Kconfig | 36 filtering". The first two are required to allow configuration via
|
/net/sunrpc/auth_gss/ |
D | auth_gss.c | 1730 int first, last, i; in alloc_enc_pages() local 1737 first = snd_buf->page_base >> PAGE_CACHE_SHIFT; in alloc_enc_pages() 1739 rqstp->rq_enc_pages_num = last - first + 1 + 1; in alloc_enc_pages() 1770 int first; in gss_wrap_req_priv() local 1784 first = snd_buf->page_base >> PAGE_CACHE_SHIFT; in gss_wrap_req_priv() 1785 inpages = snd_buf->pages + first; in gss_wrap_req_priv() 1787 snd_buf->page_base -= first << PAGE_CACHE_SHIFT; in gss_wrap_req_priv()
|
/net/ipv4/ |
D | udp.c | 222 unsigned short first, last; in udp_lib_get_port() local 229 first = reciprocal_scale(rand, remaining) + low; in udp_lib_get_port() 234 last = first + udptable->mask + 1; in udp_lib_get_port() 236 hslot = udp_hashslot(udptable, net, first); in udp_lib_get_port() 242 snum = first; in udp_lib_get_port() 254 } while (snum != first); in udp_lib_get_port() 256 } while (++first != last); in udp_lib_get_port()
|
D | igmp.c | 454 int scount, stotal, first, isquery, truncate; in add_grec() local 487 first = 1; in add_grec() 504 first*sizeof(struct igmpv3_grec)) { in add_grec() 505 if (truncate && !first) in add_grec() 512 first = 1; in add_grec() 515 if (first) { in add_grec() 517 first = 0; in add_grec()
|
/net/ipv6/ |
D | mcast.c | 1704 int scount, stotal, first, isquery, truncate; in add_grec() local 1737 first = 1; in add_grec() 1754 first*sizeof(struct mld2_grec)) { in add_grec() 1755 if (truncate && !first) in add_grec() 1762 first = 1; in add_grec() 1765 if (first) { in add_grec() 1767 first = 0; in add_grec()
|
/net/rxrpc/ |
D | rxkad.c | 164 __be32 first; /* first four bytes of data and padding */ in rxkad_secure_packet_auth() member 176 memcpy(&tmpbuf.first, sechdr + 4, sizeof(tmpbuf.first)); in rxkad_secure_packet_auth()
|
/net/batman-adv/ |
D | fragmentation.c | 248 entry = hlist_entry(chain->first, struct batadv_frag_list_entry, list); in batadv_frag_merge_packets()
|
/net/netfilter/ |
D | nfnetlink_queue_core.c | 1210 return q->instance_table[st->bucket].first; in get_first() 1228 h = q->instance_table[st->bucket].first; in get_next()
|
/net/sunrpc/ |
D | auth.c | 412 cred = hlist_entry(head->first, struct rpc_cred, cr_hash); in rpcauth_clear_credcache()
|
/net/ipv6/netfilter/ |
D | Kconfig | 253 first in the netfilter framework and hooks in at the PREROUTING
|