| /kernel/linux/linux-5.10/net/rds/ |
| D | ib_send.c | 151 struct rds_ib_send_work *send, in rds_ib_send_unmap_op() argument 157 switch (send->s_wr.opcode) { in rds_ib_send_unmap_op() 159 if (send->s_op) { in rds_ib_send_unmap_op() 160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 161 rds_ib_send_unmap_data(ic, send->s_op, wc_status); in rds_ib_send_unmap_op() 166 if (send->s_op) { in rds_ib_send_unmap_op() 167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 168 rds_ib_send_unmap_rdma(ic, send->s_op, wc_status); in rds_ib_send_unmap_op() 173 if (send->s_op) { in rds_ib_send_unmap_op() 174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() [all …]
|
| /kernel/linux/linux-4.19/net/rds/ |
| D | ib_send.c | 155 struct rds_ib_send_work *send, in rds_ib_send_unmap_op() argument 161 switch (send->s_wr.opcode) { in rds_ib_send_unmap_op() 163 if (send->s_op) { in rds_ib_send_unmap_op() 164 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 165 rds_ib_send_unmap_data(ic, send->s_op, wc_status); in rds_ib_send_unmap_op() 170 if (send->s_op) { in rds_ib_send_unmap_op() 171 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 172 rds_ib_send_unmap_rdma(ic, send->s_op, wc_status); in rds_ib_send_unmap_op() 177 if (send->s_op) { in rds_ib_send_unmap_op() 178 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() [all …]
|
| /kernel/linux/linux-5.10/net/llc/ |
| D | llc_s_ac.c | 46 * @skb: the event to send 69 * llc_sap_action_send_xid_c - send XID PDU as response to XID REQ 71 * @skb: the event to send 73 * Send a XID command PDU to MAC layer in response to a XID REQUEST 94 * llc_sap_action_send_xid_r - send XID PDU resp to MAC for received XID 96 * @skb: the event to send 98 * Send XID response PDU to MAC in response to an earlier received XID 125 * llc_sap_action_send_test_c - send TEST PDU to MAC in resp to TEST REQ 127 * @skb: the event to send 129 * Send a TEST command PDU to the MAC layer in response to a TEST REQUEST [all …]
|
| /kernel/linux/linux-4.19/net/llc/ |
| D | llc_s_ac.c | 46 * @skb: the event to send 69 * llc_sap_action_send_xid_c - send XID PDU as response to XID REQ 71 * @skb: the event to send 73 * Send a XID command PDU to MAC layer in response to a XID REQUEST 94 * llc_sap_action_send_xid_r - send XID PDU resp to MAC for received XID 96 * @skb: the event to send 98 * Send XID response PDU to MAC in response to an earlier received XID 125 * llc_sap_action_send_test_c - send TEST PDU to MAC in resp to TEST REQ 127 * @skb: the event to send 129 * Send a TEST command PDU to the MAC layer in response to a TEST REQUEST [all …]
|
| /kernel/linux/linux-4.19/tools/testing/selftests/net/ |
| D | udpgso.c | 66 int tlen; /* send() buffer size, may exceed mss */ 67 bool tfail; /* send() call is expected to fail */ 78 /* no GSO: send a single byte */ 83 /* no GSO: send a single MSS */ 88 /* no GSO: send a single MSS + 1B: fail */ 93 /* send a single MSS: will fall back to no GSO */ 99 /* send a single MSS + 1B */ 106 /* send exactly 2 MSS */ 112 /* send 2 MSS + 1B */ 119 /* send MAX segs */ [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/net/ |
| D | udpgso.c | 65 int tlen; /* send() buffer size, may exceed mss */ 66 bool tfail; /* send() call is expected to fail */ 77 /* no GSO: send a single byte */ 82 /* no GSO: send a single MSS */ 87 /* no GSO: send a single MSS + 1B: fail */ 92 /* send a single MSS: will fall back to no GSO */ 98 /* send a single MSS + 1B */ 105 /* send exactly 2 MSS */ 111 /* send 2 MSS + 1B */ 118 /* send MAX segs */ [all …]
|
| /kernel/linux/linux-4.19/drivers/isdn/hardware/mISDN/ |
| D | netjet.c | 89 struct tiger_dma send; member 194 bc->bch.nr, fill, cnt, idx, card->send.idx); in fill_mem() 201 val = card->send.start[idx]; in fill_mem() 204 card->send.start[idx++] = val; in fill_mem() 205 if (idx >= card->send.size) in fill_mem() 221 fill_mem(bc, 0, card->send.size, 0xff); in mode_tiger() 240 bc->free = card->send.size / 2; in mode_tiger() 254 bc->free = card->send.size / 2; in mode_tiger() 271 card->send.dmacur = inl(card->base + NJ_DMA_READ_ADR); in mode_tiger() 273 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; in mode_tiger() [all …]
|
| /kernel/linux/linux-5.10/drivers/isdn/hardware/mISDN/ |
| D | netjet.c | 76 struct tiger_dma send; member 181 bc->bch.nr, fill, cnt, idx, card->send.idx); in fill_mem() 188 val = card->send.start[idx]; in fill_mem() 191 card->send.start[idx++] = val; in fill_mem() 192 if (idx >= card->send.size) in fill_mem() 208 fill_mem(bc, 0, card->send.size, 0xff); in mode_tiger() 227 bc->free = card->send.size / 2; in mode_tiger() 241 bc->free = card->send.size / 2; in mode_tiger() 258 card->send.dmacur = inl(card->base + NJ_DMA_READ_ADR); in mode_tiger() 260 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; in mode_tiger() [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/bpf/ |
| D | test_flow_dissector.sh | 111 # Send 10 IPv4/UDP packets from port 8. Filter should not drop any. 113 # Send 10 IPv4/UDP packets from port 9. Filter should drop all. 115 # Send 10 IPv4/UDP packets from port 10. Filter should not drop any. 119 # Send 10 IPv4/IPv4/UDP packets from port 8. Filter should not drop any. 122 # Send 10 IPv4/IPv4/UDP packets from port 9. Filter should drop all. 125 # Send 10 IPv4/IPv4/UDP packets from port 10. Filter should not drop any. 130 # Send 10 IPv4/GRE/IPv4/UDP packets from port 8. Filter should not drop any. 133 # Send 10 IPv4/GRE/IPv4/UDP packets from port 9. Filter should drop all. 136 # Send 10 IPv4/GRE/IPv4/UDP packets from port 10. Filter should not drop any. 147 # Send 10 IPv4/UDP packets from port 7. Filter should not drop any. [all …]
|
| /kernel/linux/linux-4.19/include/rdma/ |
| D | rdmavt_qp.h | 73 * RVT_S_SIGNAL_REQ_WR - set if QP send WRs contain completion signaled 74 * RVT_S_BUSY - send tasklet is processing the QP 83 * RVT_S_WAIT_DMA - waiting for send DMA queue to drain before generating 84 * next send completion entry not via send DMA 85 * RVT_S_WAIT_PIO - waiting for a send buffer to be available 90 * RVT_S_WAIT_PSN - waiting for a packet to exit the send DMA queue 92 * RVT_S_SEND_ONE - send one packet, request ACK, then wait for ACK 93 * RVT_S_ECN - a BECN was queued to the send engine 130 * Wait flags that would prevent send work requests from making progress. 155 * Internal send flags [all …]
|
| D | ib_mad.h | 465 * @seg_count: The number of RMPP segments allocated for this send. 546 * @mad_send_wc: Send work completion information on the sent MAD. 554 * @send_buf: send MAD data buffer. 556 * only for snooping that occurs on a send completion. 568 * @send_buf: Send buffer if found, else NULL 571 * MADs received in response to a send request operation will be handed to 572 * the user before the send operation completes. All data buffers given 618 * ib_mad_send_wc - MAD send completion information. 619 * @send_buf: Send MAD data buffer associated with the send MAD request. 656 * for the corresponding send request. [all …]
|
| /kernel/linux/linux-5.10/include/rdma/ |
| D | rdmavt_qp.h | 59 * RVT_S_SIGNAL_REQ_WR - set if QP send WRs contain completion signaled 60 * RVT_S_BUSY - send tasklet is processing the QP 69 * RVT_S_WAIT_DMA - waiting for send DMA queue to drain before generating 70 * next send completion entry not via send DMA 71 * RVT_S_WAIT_PIO - waiting for a send buffer to be available 75 * RVT_S_WAIT_PSN - waiting for a packet to exit the send DMA queue 77 * RVT_S_SEND_ONE - send one packet, request ACK, then wait for ACK 78 * RVT_S_ECN - a BECN was queued to the send engine 115 * Wait flags that would prevent send work requests from making progress. 140 * Internal send flags [all …]
|
| /kernel/linux/linux-5.10/drivers/misc/vmw_vmci/ |
| D | vmci_route.c | 37 * them again when we do the actual send to ensure that we do in vmci_route() 50 * cannot send it to the hypervisor. It must come in vmci_route() 57 * We must be acting as a guest in order to send to in vmci_route() 63 /* And we cannot send if the source is the host context. */ in vmci_route() 79 /* Send from local client down to the hypervisor. */ in vmci_route() 88 * guest, then we need to send it down to the host. in vmci_route() 98 * may send vmci event datagrams to the host in vmci_route() 99 * itself, but it will never send datagrams to in vmci_route() 116 /* Send it from local client down to the host. */ in vmci_route() 186 * send it down, since we have no guest in vmci_route() [all …]
|
| /kernel/linux/linux-5.10/Documentation/crypto/ |
| D | userspace-if.rst | 62 send()/write() system call family. The result of the cipher operation is 79 particular cipher instance. When invoking send/write or recv/read 80 system calls to send data to the kernel or obtain data from the 88 the input buffer used for the send/write system call and the output 120 Using the send() system call, the application provides the data that 121 should be processed with the message digest. The send system call allows 124 - MSG_MORE: If this flag is set, the send system call acts like a 126 calculated. If the flag is not set, the send system call calculates 154 Before data can be sent to the kernel using the write/send system call 165 send/recv system call family. That cmsghdr data structure holds the [all …]
|
| /kernel/linux/linux-4.19/drivers/misc/vmw_vmci/ |
| D | vmci_route.c | 45 * them again when we do the actual send to ensure that we do in vmci_route() 58 * cannot send it to the hypervisor. It must come in vmci_route() 65 * We must be acting as a guest in order to send to in vmci_route() 71 /* And we cannot send if the source is the host context. */ in vmci_route() 87 /* Send from local client down to the hypervisor. */ in vmci_route() 96 * guest, then we need to send it down to the host. in vmci_route() 106 * may send vmci event datagrams to the host in vmci_route() 107 * itself, but it will never send datagrams to in vmci_route() 124 /* Send it from local client down to the host. */ in vmci_route() 194 * send it down, since we have no guest in vmci_route() [all …]
|
| /kernel/linux/linux-4.19/Documentation/crypto/ |
| D | userspace-if.rst | 62 send()/write() system call family. The result of the cipher operation is 79 particular cipher instance. When invoking send/write or recv/read 80 system calls to send data to the kernel or obtain data from the 88 the input buffer used for the send/write system call and the output 120 Using the send() system call, the application provides the data that 121 should be processed with the message digest. The send system call allows 124 - MSG_MORE: If this flag is set, the send system call acts like a 126 calculated. If the flag is not set, the send system call calculates 154 Before data can be sent to the kernel using the write/send system call 165 send/recv system call family. That cmsghdr data structure holds the [all …]
|
| /kernel/linux/linux-4.19/Documentation/networking/ |
| D | rds.txt | 48 Sockets must be bound before you can send or receive data. 71 setsockopt(SOL_SOCKET): send and receive buffer size 72 RDS honors the send and receive buffer size socket options. 82 mark the port as "congested" and send a congestion update to 96 An attempt to send a message that exceeds SO_SNDSIZE will 99 An attempt to send a message that would take the total number 103 An attempt to send a message to a destination that is marked 127 POLLOUT is a little harder. Since you can essentially send 129 there's room on the send queue (ie the number of bytes queued 144 it detects a timeout. For instance, if it tried to send a message, [all …]
|
| D | tls.txt | 59 For example, we can send an encrypted hello world record as follows: 62 send(sock, msg, strlen(msg)); 64 send() data is directly encrypted from the userspace buffer provided 65 to the encrypted kernel send buffer if possible. 67 The sendfile system call will send the file's data over TLS records of maximum 74 TLS records are created and sent after each send() call, unless 79 This buffer is allocated at the time send() is called, such that 80 either the entire send() call will return -ENOMEM (or block waiting 81 for memory), or the encryption will always succeed. If send() returns 107 Send TLS control messages [all …]
|
| /kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
| D | svc_rdma_sendto.c | 53 * a Send WR conveying the transport header and the RPC message itself to 68 * The logic here depends on Send Queue and completion ordering. Since 69 * the Send WR is always posted last, it will always complete last. Thus 78 * When the Send WR is constructed, it also gets its own svc_rdma_send_ctxt. 80 * ctxt, the Send WR is posted, and sendto returns. 82 * The svc_rdma_send_ctxt is presented when the Send WR completes. The 83 * Send completion handler finally releases the Reply's pages. 85 * This mechanism also assumes that completions on the transport's Send 87 * and Send completion running at the same time could release pages that 92 * - If the Send WR is posted successfully, it will either complete [all …]
|
| /kernel/linux/linux-5.10/Documentation/networking/ |
| D | rds.rst | 55 Sockets must be bound before you can send or receive data. 79 setsockopt(SOL_SOCKET): send and receive buffer size 80 RDS honors the send and receive buffer size socket options. 90 mark the port as "congested" and send a congestion update to 104 An attempt to send a message that exceeds SO_SNDSIZE will 107 An attempt to send a message that would take the total number 111 An attempt to send a message to a destination that is marked 135 POLLOUT is a little harder. Since you can essentially send 137 there's room on the send queue (ie the number of bytes queued 152 it detects a timeout. For instance, if it tried to send a message, [all …]
|
| D | tls.rst | 69 For example, we can send an encrypted hello world record as follows: 74 send(sock, msg, strlen(msg)); 76 send() data is directly encrypted from the userspace buffer provided 77 to the encrypted kernel send buffer if possible. 79 The sendfile system call will send the file's data over TLS records of maximum 88 TLS records are created and sent after each send() call, unless 93 This buffer is allocated at the time send() is called, such that 94 either the entire send() call will return -ENOMEM (or block waiting 95 for memory), or the encryption will always succeed. If send() returns 123 Send TLS control messages [all …]
|
| /kernel/linux/linux-4.19/Documentation/infiniband/ |
| D | tag_matching.txt | 4 source send operations to destination receives. The following parameters must 10 The ordering rules require that when more than one pair of send and receive 11 message envelopes may match, the pair that includes the earliest posted-send 27 1. The Eager protocol- the complete message is sent when the send is 28 processed by the sender. A completion send is received in the send_cq 41 to the MPI receive routines in the posted receive list and posts send messages 42 using the MPI send routines. The head of the posted receive list may be 45 When send is initiated and arrives at the receive side, if there is no
|
| /kernel/linux/linux-4.19/drivers/infiniband/hw/hfi1/ |
| D | pio.c | 55 * Send Context functions 82 /* global control of PIO send */ 139 /* number of send context memory pools */ 142 /* Send Context Size (SCS) wildcards */ 146 /* Send Context Count (SCC) wildcards */ 151 /* Send Context Size (SCS) constants */ 159 /* default send context sizes */ 172 /* send context memory pool configuration */ 227 * Read the send context memory pool configuration and send context 229 * counts and sizes for the send context types. [all …]
|
| /kernel/linux/linux-5.10/Documentation/infiniband/ |
| D | tag_matching.rst | 6 source send operations to destination receives. The following parameters must 14 The ordering rules require that when more than one pair of send and receive 15 message envelopes may match, the pair that includes the earliest posted-send 31 1. The Eager protocol- the complete message is sent when the send is 32 processed by the sender. A completion send is received in the send_cq 46 to the MPI receive routines in the posted receive list and posts send messages 47 using the MPI send routines. The head of the posted receive list may be 50 When send is initiated and arrives at the receive side, if there is no
|
| /kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
| D | pio.c | 55 * Send Context functions 74 /* global control of PIO send */ 131 /* number of send context memory pools */ 134 /* Send Context Size (SCS) wildcards */ 138 /* Send Context Count (SCC) wildcards */ 143 /* Send Context Size (SCS) constants */ 151 /* default send context sizes */ 164 /* send context memory pool configuration */ 219 * Read the send context memory pool configuration and send context 221 * counts and sizes for the send context types. [all …]
|