| /linux/drivers/i2c/busses/ |
| H A D | i2c-nct6694.c | 22 /* Command 00h - I2C Deliver */ 60 struct nct6694_i2c_deliver deliver; member 68 struct nct6694_i2c_deliver *deliver = &data->deliver; in nct6694_i2c_xfer() local 73 .len = cpu_to_le16(sizeof(*deliver)) in nct6694_i2c_xfer() 80 memset(deliver, 0, sizeof(*deliver)); in nct6694_i2c_xfer() 82 deliver->port = data->port; in nct6694_i2c_xfer() 83 deliver->br = data->br; in nct6694_i2c_xfer() 84 deliver->addr = i2c_8bit_addr_from_msg(msg_temp); in nct6694_i2c_xfer() 86 deliver->r_cnt = msg_temp->len; in nct6694_i2c_xfer() 87 ret = nct6694_write_msg(data->nct6694, &cmd_hd, deliver); in nct6694_i2c_xfer() [all …]
|
| /linux/fs/afs/ |
| H A D | cmservice.c | 39 .deliver = afs_deliver_cb_callback, 49 .deliver = afs_deliver_cb_init_call_back_state, 59 .deliver = afs_deliver_cb_init_call_back_state3, 69 .deliver = afs_deliver_cb_probe, 79 .deliver = afs_deliver_cb_probe_uuid, 89 .deliver = afs_deliver_cb_tell_me_about_yourself, 99 .deliver = afs_deliver_yfs_cb_callback, 188 * deliver request data to a CB.CallBack call 301 * deliver request data to a CB.InitCallBackState call 312 * deliver request data to a CB.InitCallBackState3 call [all …]
|
| H A D | fsclient.c | 236 * deliver reply data to an FS.FetchStatus 265 .deliver = afs_deliver_fs_fetch_status, 299 * deliver reply data to an FS.FetchData 412 .deliver = afs_deliver_fs_fetch_data, 421 .deliver = afs_deliver_fs_fetch_data, 495 * deliver reply data to an FS.CreateFile or an FS.MakeDir 527 .deliver = afs_deliver_fs_create_vnode, 581 .deliver = afs_deliver_fs_create_vnode, 633 * Deliver reply data to any operation that returns status and volume sync. 661 .deliver = afs_deliver_fs_file_status_and_vol, [all …]
|
| H A D | yfsclient.c | 328 * Deliver reply data to operations that just return a file status and a volume 350 * Deliver reply data to an YFS.FetchData64. 460 .deliver = yfs_deliver_fs_fetch_data64, 508 * Deliver reply data for YFS.CreateFile or YFS.MakeDir. 542 .deliver = yfs_deliver_fs_create_vnode, 593 .deliver = yfs_deliver_fs_create_vnode, 640 * Deliver reply data to a YFS.RemoveFile2 operation. 683 .deliver = yfs_deliver_fs_remove_file2, 726 * Deliver reply data to a YFS.RemoveFile or YFS.RemoveDir operation. 753 .deliver = yfs_deliver_fs_remove, [all …]
|
| H A D | vlclient.c | 15 * Deliver reply data to a VL.GetEntryByNameU call. 102 .deliver = afs_deliver_vl_get_entry_by_name_u, 165 * Deliver reply data to a VL.GetAddrsU call. 250 .deliver = afs_deliver_vl_get_addrs_u, 315 * Deliver reply data to an VL.GetCapabilities operation. 372 .deliver = afs_deliver_vl_get_capabilities, 423 * Deliver reply data to a YFSVL.GetEndpoints call. 626 .deliver = afs_deliver_yfsvl_get_endpoints, 678 * Deliver reply data to a YFSVL.GetCellName operation. 745 .deliver = afs_deliver_yfsvl_get_cell_name,
|
| /linux/drivers/iommu/iommufd/ |
| H A D | eventq.c | 33 list_for_each_entry_safe(group, next, &fault->common.deliver, node) { in iommufd_auto_response_faults() 70 list_for_each_entry_safe(group, next, &fault->common.deliver, node) { in iommufd_fault_destroy() 99 /* Fetch the first node out of the fault->deliver list */ 103 struct list_head *list = &fault->common.deliver; in iommufd_fault_deliver_fetch() 115 /* Restore a node back to the head of the fault->deliver list */ 120 list_add(&group->node, &fault->common.deliver); in iommufd_fault_deliver_restore() 231 list_for_each_entry_safe(cur, next, &eventq->deliver, node) { in iommufd_veventq_abort() 255 struct list_head *list = &eventq->deliver; in iommufd_veventq_deliver_fetch() 284 struct list_head *list = &eventq->deliver; in iommufd_veventq_deliver_restore() 365 if (!list_empty(&eventq->deliver)) in iommufd_eventq_fops_poll() [all …]
|
| H A D | iommufd_private.h | 555 spinlock_t lock; /* protects the deliver list */ 556 struct list_head deliver; member 570 * An iommufd_fault object represents an interface to deliver I/O page faults 603 struct list_head node; /* for iommufd_eventq::deliver */ 612 * An iommufd_veventq object represents an interface to deliver vIOMMU events to 661 if (list_is_last(&veventq->lost_events_header.node, &eventq->deliver)) in iommufd_vevent_handler() 663 list_add_tail(&vevent->node, &eventq->deliver); in iommufd_vevent_handler()
|
| /linux/arch/x86/kvm/ |
| H A D | xen.c | 692 /* For the per-vCPU lapic vector, deliver it as MSI. */ in kvm_xen_inject_pending_events() 1857 * and prod the vCPU to deliver it for itself. in kvm_xen_set_evtchn_fast() 1879 /* For the per-vCPU lapic vector, deliver it as MSI. */ in kvm_xen_set_evtchn_fast() 2057 } deliver; member 2087 if (!evtchnfd->deliver.port.port || in kvm_xen_eventfd_update() 2088 evtchnfd->deliver.port.port != data->u.evtchn.deliver.port.port) in kvm_xen_eventfd_update() 2092 if (data->u.evtchn.deliver.port.priority != KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL) in kvm_xen_eventfd_update() 2095 evtchnfd->deliver.port.priority = data->u.evtchn.deliver.port.priority; in kvm_xen_eventfd_update() 2096 if (evtchnfd->deliver.port.vcpu_id != data->u.evtchn.deliver.port.vcpu) { in kvm_xen_eventfd_update() 2097 evtchnfd->deliver.port.vcpu_id = data->u.evtchn.deliver.port.vcpu; in kvm_xen_eventfd_update() [all …]
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | xen_shinfo_test.c | 230 /* Attempt to deliver a *masked* interrupt */ in guest_code() 238 /* Now deliver an *unmasked* interrupt */ in guest_code() 243 /* Change memslots and deliver an interrupt */ in guest_code() 248 /* Deliver event channel with KVM_XEN_HVM_EVTCHN_SEND */ in guest_code() 255 /* Our turn. Deliver event channel (to ourselves) with in guest_code() 274 /* Deliver "outbound" event channel to an eventfd which in guest_code() 590 .u.evtchn.deliver.port.port = EVTCHN_TEST1, in main() 591 .u.evtchn.deliver.port.vcpu = vcpu->id + 1, in main() 592 .u.evtchn.deliver.port.priority = KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL, in main() 598 inj.u.evtchn.deliver.port.vcpu = vcpu->id; in main() [all …]
|
| /linux/net/ipv6/ |
| H A D | ip6_input.c | 361 * Deliver the packet to the host 514 bool deliver; in ip6_mc_input() local 530 deliver = ipv6_chk_mcast_addr(dev, &hdr->daddr, NULL); in ip6_mc_input() 558 deliver = false; in ip6_mc_input() 570 deliver = true; in ip6_mc_input() 577 if (deliver) { in ip6_mc_input() 589 if (likely(deliver)) { in ip6_mc_input()
|
| /linux/drivers/irqchip/ |
| H A D | irq-csky-mpintc.c | 143 * The csky,mpintc could support auto irq deliver, but it only in csky_irq_set_affinity() 144 * could deliver external irq to one cpu or all cpus. So it in csky_irq_set_affinity() 145 * doesn't support deliver external irq to a group of cpus in csky_irq_set_affinity() 147 * SO we only use auto deliver mode when affinity mask_val is in csky_irq_set_affinity()
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s.c | 193 /* might as well deliver this straight away */ in kvmppc_core_queue_machine_check() 206 /* might as well deliver this straight away */ in kvmppc_core_queue_program() 213 /* might as well deliver this straight away */ in kvmppc_core_queue_fpunavail() 219 /* might as well deliver this straight away */ in kvmppc_core_queue_vec_unavail() 225 /* might as well deliver this straight away */ in kvmppc_core_queue_vsx_unavail() 299 int deliver = 1; in kvmppc_book3s_irqprio_deliver() local 305 deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit; in kvmppc_book3s_irqprio_deliver() 309 deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit; in kvmppc_book3s_irqprio_deliver() 358 deliver = 0; in kvmppc_book3s_irqprio_deliver() 363 if (deliver) in kvmppc_book3s_irqprio_deliver() [all …]
|
| H A D | book3s_xics.c | 76 XICS_DBG("ics deliver %#x (level: %d)\n", irq, level); in ics_deliver_irq() 144 bool deliver; in write_xive() local 153 deliver = false; in write_xive() 157 deliver = true; in write_xive() 163 return deliver; in write_xive() 342 XICS_DBG("try deliver %#x(P:%#x) to server %#lx\n", irq, priority, in icp_try_to_deliver() 350 /* See if we can deliver */ in icp_try_to_deliver() 365 * If we failed to deliver we set need_resend in icp_try_to_deliver() 481 * We failed to deliver the interrupt we need to set the in icp_deliver_irq()
|
| /linux/net/mctp/test/ |
| H A D | route-test.c | 219 bool deliver; member 242 if (params->deliver) { in mctp_test_route_input_sk() 266 { .hdr = RX_HDR(1, 10, 8, FL_S | FL_E | FL_TO), .type = 0, .deliver = true }, 267 { .hdr = RX_HDR(1, 10, 8, FL_S | FL_E | FL_TO), .type = 1, .deliver = false }, 268 { .hdr = RX_HDR(1, 10, 8, FL_S | FL_E), .type = 0, .deliver = false }, 269 { .hdr = RX_HDR(1, 10, 8, FL_E | FL_TO), .type = 0, .deliver = false }, 270 { .hdr = RX_HDR(1, 10, 8, FL_TO), .type = 0, .deliver = false }, 271 { .hdr = RX_HDR(1, 10, 8, 0), .type = 0, .deliver = false }, 421 bool deliver; member 477 if (params->deliver) in mctp_test_route_input_sk_keys() [all …]
|
| /linux/tools/perf/util/ |
| H A D | ordered-events.c | 245 ret = oe->deliver(oe, iter); in do_flush() 361 void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t deliver, in ordered_events__init() argument 369 oe->deliver = deliver; in ordered_events__init() 412 ordered_events__deliver_t old_deliver = oe->deliver; in ordered_events__reinit()
|
| H A D | ordered-events.h | 47 ordered_events__deliver_t deliver; member 61 void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t deliver,
|
| /linux/net/rxrpc/ |
| H A D | rxperf.c | 66 int (*deliver)(struct rxperf_call *call); member 160 call->deliver = rxperf_deliver_param_block; in rxperf_charge_preallocation() 278 * deliver messages to a call 301 ret = call->deliver(call); in rxperf_deliver_to_call() 405 call->deliver = rxperf_deliver_request; in rxperf_deliver_param_block() 435 return call->deliver(call); in rxperf_deliver_param_block() 439 * Deliver the request data.
|
| /linux/drivers/s390/net/ |
| H A D | Kconfig | 30 prompt "Deliver IUCV special messages as uevents (VM only)" 33 Select this option to deliver CP special messages (SMSGs) as
|
| /linux/Documentation/networking/ |
| H A D | x25-iface.rst | 80 call "netif_rx" to deliver the received packets. Instead, it should 81 call "netif_receive_skb_core" from softirq context to deliver them.
|
| /linux/net/can/ |
| H A D | af_can.c | 573 static inline void deliver(struct sk_buff *skb, struct receiver *rcv) in deliver() function 593 deliver(skb, rcv); in can_rcv_filter() 602 deliver(skb, rcv); in can_rcv_filter() 609 deliver(skb, rcv); in can_rcv_filter() 617 deliver(skb, rcv); in can_rcv_filter() 629 deliver(skb, rcv); in can_rcv_filter() 636 deliver(skb, rcv); in can_rcv_filter() 669 /* deliver the packet to sockets listening on all devices */ in can_receive()
|
| /linux/arch/um/kernel/ |
| H A D | signal.c | 102 /* Whee! Actually deliver the signal. */ in do_signal() 124 * if there's no signal to deliver, we just put the saved sigmask in do_signal()
|
| /linux/include/linux/ |
| H A D | mii_timestamper.h | 20 * the MII time stamping device promises to deliver it using 26 * device promises to deliver it using skb_complete_tx_timestamp()
|
| /linux/Documentation/devicetree/bindings/timer/ |
| H A D | arm,arch_timer.yaml | 17 The per-core architected timer is attached to a GIC to deliver its 19 to deliver its interrupts via SPIs.
|
| /linux/include/xen/ |
| H A D | events.h | 94 irq will be disabled so it won't deliver an interrupt. */ 98 * the irq will be disabled so it won't deliver an interrupt. */
|
| /linux/tools/testing/selftests/x86/ |
| H A D | mov_ss_trap.c | 7 * up a #DB trap but will not actually deliver that trap. The trap 13 * deliver #DB. 14 * - ICEBP: enter the kernel but do not deliver the watchpoint trap
|