/linux/Documentation/scsi/ |
H A D | hptiop.rst | 110 All queued requests are handled via inbound/outbound queue port. 115 - Get a free request packet by reading the inbound queue port or 118 The value returned from the inbound queue port is an offset 125 - Post the packet to IOP by writing it to inbound queue. For requests 126 allocated in IOP memory, write the offset to inbound queue port. For 128 to the inbound queue port. 147 Non-queued requests (reset/flush etc) can be sent via inbound message 149 of an inbound message. 155 All queued requests are handled via inbound/outbound list. 165 Fill a free inbound lis [all...] |
H A D | arcmsr_spec.rst | 40 offset 0xa00 for inbound message code message_rwbuffer 61 (inbound queue port) 152 ->offset 0xa00 :for inbound message code message_rwbuffer 167 ->offset 0xa00 for inbound message code message_rwbuffer
|
/linux/Documentation/arch/arm/ |
H A D | cluster-pm-race-avoidance.rst | 90 state value into two parts (the "cluster" state and "inbound" state) and 222 as seen by a CPU tearing the cluster down. The "inbound side" is the 237 "inbound" state: The state of the cluster on the inbound side. 266 Transitions ===##> can only be made by the inbound CPU, and only 267 involve changes to the "inbound" state, except where there is no 280 The parallelism of the inbound and outbound CPUs is observed by 296 <cluster state>/<inbound state> (<transitioner>) 299 can occur; either the inbound or the outbound side. 304 CLUSTER_DOWN/INBOUND_COMING_UP (inbound) [all...] |
/linux/drivers/soc/qcom/ |
H A D | smp2p.c | 38 * inbound entry. 88 * @domain: irq_domain for inbound entries 119 * @in: pointer to the inbound smem item 122 * @valid_entries: already scanned inbound entries 126 * @local_pid: processor id of the inbound edge 133 * @inbound: list of inbound entries 160 struct list_head inbound; member 239 list_for_each_entry(entry, &smp2p->inbound, node) { in qcom_smp2p_notify_in() 250 list_for_each_entry(entry, &smp2p->inbound, nod in qcom_smp2p_notify_in() [all...] |
/linux/drivers/hv/ |
H A D | vmbus_drv.c | 390 struct hv_ring_buffer_debug_info inbound; in in_intr_mask_show() local 396 ret = hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound); in in_intr_mask_show() 400 return sysfs_emit(buf, "%d\n", inbound.current_interrupt_mask); in in_intr_mask_show() 408 struct hv_ring_buffer_debug_info inbound; in in_read_index_show() local 414 ret = hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound); in in_read_index_show() 418 return sysfs_emit(buf, "%d\n", inbound.current_read_index); in in_read_index_show() 426 struct hv_ring_buffer_debug_info inbound; in in_write_index_show() local 432 ret = hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, in in_write_index_show() 445 struct hv_ring_buffer_debug_info inbound; in_read_bytes_avail_show() local 464 struct hv_ring_buffer_debug_info inbound; in_write_bytes_avail_show() local [all...] |
H A D | ring_buffer.c | 181 mutex_init(&channel->inbound.ring_buffer_mutex); in hv_ringbuffer_pre_init() 466 struct hv_ring_buffer_info *rbi = &channel->inbound; in hv_pkt_iter_first() 526 struct hv_ring_buffer_info *rbi = &channel->inbound; in __hv_pkt_iter_next() 574 struct hv_ring_buffer_info *rbi = &channel->inbound; in hv_pkt_iter_close()
|
H A D | connection.c | 426 if (likely(hv_end_read(&channel->inbound) == 0)) in vmbus_on_event() 429 hv_begin_read(&channel->inbound); in vmbus_on_event()
|
H A D | channel.c | 154 hv_ringbuffer_cleanup(&channel->inbound); in vmbus_free_ring() 684 err = hv_ringbuffer_init(&newchannel->inbound, &page[send_pages], in __vmbus_open() 767 hv_ringbuffer_cleanup(&newchannel->inbound); in __vmbus_open() 884 * the former is accessing channel->inbound.ring_buffer, the latter in vmbus_reset_channel_cb()
|
/linux/Documentation/admin-guide/perf/ |
H A D | mrvl-pem-pmu.rst | 9 The counters track inbound and outbound transactions which 11 Also, inbound and outbound memory read requests along with their 17 tlps in inbound and outbound transactions. ATS events are measured by
|
/linux/Documentation/trace/ |
H A D | hisi-ptt.rst | 99 type. Rx means the inbound while Tx means outbound. The packets will 132 direction of inbound and traced data format of 8DW. 177 You can specify multiple types when tracing inbound TLP headers, but can only 185 This is optional and the default parameter is inbound. The parameter value 189 - 4'b0000: inbound TLPs (P, NP, CPL) 191 - 4'b0010: outbound TLPs (P, NP, CPL) and inbound TLPs (P, NP, CPL B) 192 - 4'b0011: outbound TLPs (P, NP, CPL) and inbound TLPs (CPL A) 199 - 4'b0010: inbound TLPs (P, NP, CPL B) 200 - 4'b0011: inbound TLPs (CPL A)
|
/linux/drivers/net/ppp/ |
H A D | ppp_async.c | 108 int len, int inbound); 945 int len, int inbound) in async_lcp_peek() argument 961 if (code == (inbound? CONFACK: CONFREQ)) { in async_lcp_peek() 970 if (!inbound) { in async_lcp_peek() 981 } else if (inbound) in async_lcp_peek() 992 if (inbound) in async_lcp_peek() 999 if (inbound) in async_lcp_peek()
|
/linux/arch/arm/common/ |
H A D | mcpm_entry.c | 77 * false: the critical section was not entered because an inbound CPU was 87 /* Warn inbound CPUs that the cluster is being torn down: */ in __mcpm_outbound_enter_critical() 91 /* Back out if the inbound cluster is already in the critical region: */ in __mcpm_outbound_enter_critical() 92 sync_cache_r(&c->inbound); in __mcpm_outbound_enter_critical() 93 if (c->inbound == INBOUND_COMING_UP) in __mcpm_outbound_enter_critical() 179 * for an outbound CPU to call power_down() after its inbound counterpart 437 mcpm_sync.clusters[i].inbound = INBOUND_NOT_COMING_UP; in mcpm_sync_init()
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-bus-rpmsg | 23 inbound messages arrive to this address, the rpmsg core 41 inbound messages arrive to this address, the rpmsg core 49 and it is listening for inbound messages that may be sent
|
/linux/Documentation/driver-api/rapidio/ |
H A D | tsi721.rst | 9 It supports maintenance read and write operations, inbound and outbound RapidIO 10 doorbells, inbound maintenance port-writes and RapidIO messaging. 17 For inbound messages this driver uses destination ID matching to forward messages
|
/linux/drivers/pci/controller/mobiveil/ |
H A D | Kconfig | 32 Soft IP. It has up to 8 outbound and inbound windows
|
/linux/arch/arm/include/asm/ |
H A D | mcpm.h | 293 /* inbound-side state */ 294 s8 inbound __aligned(__CACHE_WRITEBACK_GRANULE);
|
/linux/Documentation/networking/ |
H A D | secid.rst | 19 If this is an inbound flow, the label is derived from the IPSec security
|
H A D | xfrm_proc.rst | 33 i.e. Either inbound SPI, address, or IPsec protocol at SA is wrong
|
/linux/Documentation/ABI/stable/ |
H A D | firewire-cdev | 42 link layers, reception of inbound requests to such 44 to inbound requests
|
/linux/include/net/sctp/ |
H A D | ulpevent.h | 79 __u16 inbound,
|
/linux/Documentation/staging/ |
H A D | rpmsg.rst | 60 a unique rpmsg local address (a 32-bit integer). This way when inbound messages 63 with the payload of the inbound message). 158 inbound messages arrive, they are dispatched by the rpmsg bus using the 172 relevant inbound messages arrive (i.e. messages which their dst address
|
/linux/include/linux/netfilter/ |
H A D | nf_conntrack_pptp.h | 309 int (*inbound)(struct sk_buff *skb, member
|
/linux/net/vmw_vsock/ |
H A D | hyperv_transport.c | 187 u32 readable = hv_get_bytes_to_read(&chan->inbound); in hvs_channel_readable() 195 u32 readable = hv_get_bytes_to_read(&chan->inbound); in hvs_channel_readable_payload()
|
/linux/Documentation/devicetree/bindings/pci/ |
H A D | ti-pci.txt | 52 - num-ib-windows : number of inbound address translation windows
|
/linux/net/ipv4/netfilter/ |
H A D | nf_nat_pptp.c | 229 /* inbound packets == from PAC to PNS */ 270 pr_debug("unknown inbound packet %s\n", pptp_msg_name(msg)); in pptp_inbound_pkt() 300 .inbound = pptp_inbound_pkt,
|