/kernel/linux/linux-5.10/Documentation/scsi/ |
D | hptiop.rst | 110 All queued requests are handled via inbound/outbound queue port. 115 - Get a free request packet by reading the inbound queue port or 118 The value returned from the inbound queue port is an offset 125 - Post the packet to IOP by writing it to inbound queue. For requests 126 allocated in IOP memory, write the offset to inbound queue port. For 128 to the inbound queue port. 147 Non-queued requests (reset/flush etc) can be sent via inbound message 149 of an inbound message. 155 All queued requests are handled via inbound/outbound list. 165 Fill a free inbound list unit with the physical address and the size of [all …]
|
D | arcmsr_spec.rst | 39 offset 0xa00 for inbound message code message_rwbuffer 60 (inbound queue port) 151 ->offset 0xa00 :for inbound message code message_rwbuffer 166 ->offset 0xa00 for inbound message code message_rwbuffer
|
/kernel/linux/linux-5.10/Documentation/devicetree/bindings/soc/qcom/ |
D | qcom,smp2p.txt | 38 Definition: two identifiers of the inbound and outbound smem items used 52 Each SMP2P pair contain a set of inbound and outbound entries, these are 59 Definition: specifies the name of this entry, for inbound entries this 67 Definition: marks the entry as inbound; the node should be specified 85 inbound and one outbound entry:
|
/kernel/linux/linux-5.10/Documentation/arm/ |
D | cluster-pm-race-avoidance.rst | 90 state value into two parts (the "cluster" state and "inbound" state) and 222 as seen by a CPU tearing the cluster down. The "inbound side" is the 237 "inbound" state: The state of the cluster on the inbound side. 266 Transitions ===##> can only be made by the inbound CPU, and only 267 involve changes to the "inbound" state, except where there is no 280 The parallelism of the inbound and outbound CPUs is observed by 296 <cluster state>/<inbound state> (<transitioner>) 299 can occur; either the inbound or the outbound side. 304 CLUSTER_DOWN/INBOUND_COMING_UP (inbound) 317 In this state, an inbound CPU sets up the cluster, including [all …]
|
/kernel/linux/linux-5.10/drivers/soc/qcom/ |
D | smp2p.c | 147 struct list_head inbound; member 202 list_for_each_entry(entry, &smp2p->inbound, node) { in qcom_smp2p_intr() 213 list_for_each_entry(entry, &smp2p->inbound, node) { in qcom_smp2p_intr() 457 INIT_LIST_HEAD(&smp2p->inbound); in qcom_smp2p_probe() 519 list_add(&entry->node, &smp2p->inbound); in qcom_smp2p_probe() 545 list_for_each_entry(entry, &smp2p->inbound, node) in qcom_smp2p_probe() 568 list_for_each_entry(entry, &smp2p->inbound, node) in qcom_smp2p_remove()
|
/kernel/linux/linux-5.10/drivers/hv/ |
D | vmbus_drv.c | 424 struct hv_ring_buffer_debug_info inbound; in in_intr_mask_show() local 430 ret = hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound); in in_intr_mask_show() 434 return sprintf(buf, "%d\n", inbound.current_interrupt_mask); in in_intr_mask_show() 442 struct hv_ring_buffer_debug_info inbound; in in_read_index_show() local 448 ret = hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound); in in_read_index_show() 452 return sprintf(buf, "%d\n", inbound.current_read_index); in in_read_index_show() 460 struct hv_ring_buffer_debug_info inbound; in in_write_index_show() local 466 ret = hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound); in in_write_index_show() 470 return sprintf(buf, "%d\n", inbound.current_write_index); in in_write_index_show() 479 struct hv_ring_buffer_debug_info inbound; in in_read_bytes_avail_show() local [all …]
|
D | ring_buffer.c | 187 mutex_init(&channel->inbound.ring_buffer_mutex); in hv_ringbuffer_pre_init() 418 struct hv_ring_buffer_info *rbi = &channel->inbound; in hv_pkt_iter_first() 443 struct hv_ring_buffer_info *rbi = &channel->inbound; in __hv_pkt_iter_next() 491 struct hv_ring_buffer_info *rbi = &channel->inbound; in hv_pkt_iter_close()
|
D | connection.c | 351 if (likely(hv_end_read(&channel->inbound) == 0)) in vmbus_on_event() 354 hv_begin_read(&channel->inbound); in vmbus_on_event()
|
D | channel.c | 152 hv_ringbuffer_cleanup(&channel->inbound); in vmbus_free_ring() 534 err = hv_ringbuffer_init(&newchannel->inbound, in __vmbus_open() 628 hv_ringbuffer_cleanup(&newchannel->inbound); in __vmbus_open()
|
/kernel/linux/linux-5.10/Documentation/devicetree/bindings/pci/ |
D | mobiveil-pcie.txt | 4 has up to 8 outbound and inbound windows for the address translation. 25 - ppio-wins : number of requested ppio inbound windows 26 default 1 inbound memory window is configured.
|
D | brcm,iproc-pcie.txt | 30 - dma-ranges: Some PAXB-based root complexes do not have inbound mapping done 32 the mapping, based on inbound memory regions specified by this property.
|
D | xgene-pci.txt | 15 - dma-ranges: ranges for the inbound memory regions.
|
D | pci-rcar-gen2.txt | 41 - dma-ranges: a single range for the inbound memory region. If not supplied,
|
/kernel/linux/linux-5.10/drivers/uio/ |
D | uio_hv_generic.c | 84 dev->channel->inbound.ring_buffer->interrupt_mask = !irq_state; in hv_uio_irqcontrol() 99 chan->inbound.ring_buffer->interrupt_mask = 1; in hv_uio_channel_cb() 168 new_sc->inbound.ring_buffer->interrupt_mask = 1; in hv_uio_new_channel() 213 dev->channel->inbound.ring_buffer->interrupt_mask = 1; in hv_uio_open()
|
/kernel/linux/linux-5.10/drivers/net/ppp/ |
D | ppp_async.c | 107 int len, int inbound); 952 int len, int inbound) in async_lcp_peek() argument 968 if (code == (inbound? CONFACK: CONFREQ)) { in async_lcp_peek() 977 if (!inbound) { in async_lcp_peek() 988 } else if (inbound) in async_lcp_peek() 999 if (inbound) in async_lcp_peek() 1006 if (inbound) in async_lcp_peek()
|
/kernel/linux/linux-5.10/Documentation/ABI/testing/ |
D | sysfs-bus-rpmsg | 23 inbound messages arrive to this address, the rpmsg core 41 inbound messages arrive to this address, the rpmsg core 49 and it is listening for inbound messages that may be sent
|
/kernel/linux/linux-5.10/arch/arm/common/ |
D | mcpm_entry.c | 92 sync_cache_r(&c->inbound); in __mcpm_outbound_enter_critical() 93 if (c->inbound == INBOUND_COMING_UP) in __mcpm_outbound_enter_critical() 437 mcpm_sync.clusters[i].inbound = INBOUND_NOT_COMING_UP; in mcpm_sync_init()
|
/kernel/linux/linux-5.10/Documentation/driver-api/rapidio/ |
D | tsi721.rst | 9 It supports maintenance read and write operations, inbound and outbound RapidIO 10 doorbells, inbound maintenance port-writes and RapidIO messaging. 17 For inbound messages this driver uses destination ID matching to forward messages
|
/kernel/linux/linux-5.10/arch/arm/include/asm/ |
D | mcpm.h | 294 s8 inbound __aligned(__CACHE_WRITEBACK_GRANULE);
|
/kernel/linux/linux-5.10/drivers/pci/controller/mobiveil/ |
D | Kconfig | 22 Soft IP. It has up to 8 outbound and inbound windows
|
/kernel/linux/linux-5.10/Documentation/networking/ |
D | secid.rst | 19 If this is an inbound flow, the label is derived from the IPSec security
|
D | xfrm_proc.rst | 33 i.e. Either inbound SPI, address, or IPsec protocol at SA is wrong
|
/kernel/linux/linux-5.10/Documentation/ABI/stable/ |
D | firewire-cdev | 42 link layers, reception of inbound requests to such 44 to inbound requests
|
/kernel/linux/linux-5.10/include/net/sctp/ |
D | ulpevent.h | 79 __u16 inbound,
|
/kernel/linux/linux-5.10/Documentation/staging/ |
D | rpmsg.rst | 60 a unique rpmsg local address (a 32-bit integer). This way when inbound messages 63 with the payload of the inbound message). 200 inbound messages arrive, they are dispatched by the rpmsg bus using the 214 relevant inbound messages arrive (i.e. messages which their dst address
|