Lines Matching +full:default +full:- +full:state
10 * your option) any later version. See the COPYING file in the top-level
18 #include "qemu/error-report.h"
22 #include "hw/s390x/s390-pci-inst.h"
23 #include "hw/s390x/s390-pci-bus.h"
24 #include "hw/s390x/s390-pci-kvm.h"
25 #include "hw/s390x/s390-pci-vfio.h"
32 if (iommu->dma_limit) { in inc_dma_avail()
33 iommu->dma_limit->avail++; in inc_dma_avail()
39 if (iommu->dma_limit) { in dec_dma_avail()
40 iommu->dma_limit->avail--; in dec_dma_avail()
47 env->regs[r] &= ~0xff000000ULL; in s390_set_status_code()
48 env->regs[r] |= (status_code & 0xff) << 24; in s390_set_status_code()
60 if (lduw_be_p(&rrb->request.hdr.len) != 32) { in list_pci()
62 rc = -EINVAL; in list_pci()
66 if ((ldl_be_p(&rrb->request.fmt) & CLP_MASK_FMT) != 0) { in list_pci()
68 rc = -EINVAL; in list_pci()
72 if ((ldl_be_p(&rrb->request.fmt) & ~CLP_MASK_FMT) != 0 || in list_pci()
73 ldq_be_p(&rrb->request.reserved1) != 0) { in list_pci()
75 rc = -EINVAL; in list_pci()
79 resume_token = ldq_be_p(&rrb->request.resume_token); in list_pci()
85 rc = -EINVAL; in list_pci()
92 if (lduw_be_p(&rrb->response.hdr.len) < 48) { in list_pci()
94 rc = -EINVAL; in list_pci()
98 initial_l2 = lduw_be_p(&rrb->response.hdr.len); in list_pci()
99 if ((initial_l2 - LIST_PCI_HDR_LEN) % sizeof(ClpFhListEntry) in list_pci()
102 rc = -EINVAL; in list_pci()
107 stl_be_p(&rrb->response.fmt, 0); in list_pci()
108 stq_be_p(&rrb->response.reserved1, 0); in list_pci()
109 stl_be_p(&rrb->response.mdd, FH_MASK_SHM); in list_pci()
110 stw_be_p(&rrb->response.max_fn, PCI_MAX_FUNCTIONS); in list_pci()
111 rrb->response.flags = UID_CHECKING_ENABLED; in list_pci()
112 rrb->response.entry_size = sizeof(ClpFhListEntry); in list_pci()
117 stw_be_p(&rrb->response.fh_list[i].device_id, in list_pci()
118 pci_get_word(pbdev->pdev->config + PCI_DEVICE_ID)); in list_pci()
119 stw_be_p(&rrb->response.fh_list[i].vendor_id, in list_pci()
120 pci_get_word(pbdev->pdev->config + PCI_VENDOR_ID)); in list_pci()
122 stl_be_p(&rrb->response.fh_list[i].config, in list_pci()
123 pbdev->state == ZPCI_FS_STANDBY ? 0 : 1 << 31); in list_pci()
124 stl_be_p(&rrb->response.fh_list[i].fid, pbdev->fid); in list_pci()
125 stl_be_p(&rrb->response.fh_list[i].fh, pbdev->fh); in list_pci()
130 lduw_be_p(&rrb->response.fh_list[i].vendor_id), in list_pci()
131 lduw_be_p(&rrb->response.fh_list[i].device_id), in list_pci()
132 ldl_be_p(&rrb->response.fh_list[i].fid), in list_pci()
133 ldl_be_p(&rrb->response.fh_list[i].fh)); in list_pci()
141 resume_token = pbdev->fh & FH_MASK_INDEX; in list_pci()
143 stq_be_p(&rrb->response.resume_token, resume_token); in list_pci()
144 stw_be_p(&rrb->response.hdr.len, g_l2); in list_pci()
145 stw_be_p(&rrb->response.hdr.rsp, CLP_RC_OK); in list_pci()
149 stw_be_p(&rrb->response.hdr.rsp, res_code); in list_pci()
163 CPUS390XState *env = &cpu->env; in clp_service_call()
167 if (env->psw.mask & PSW_MASK_PSTATE) { in clp_service_call()
172 if (s390_cpu_virt_mem_read(cpu, env->regs[r2], r2, buffer, sizeof(*reqh))) { in clp_service_call()
177 req_len = lduw_be_p(&reqh->len); in clp_service_call()
183 if (s390_cpu_virt_mem_read(cpu, env->regs[r2], r2, buffer, in clp_service_call()
189 res_len = lduw_be_p(&resh->len); in clp_service_call()
199 if (s390_cpu_virt_mem_read(cpu, env->regs[r2], r2, buffer, in clp_service_call()
206 stw_be_p(&resh->rsp, CLP_RC_LEN); in clp_service_call()
210 switch (lduw_be_p(&reqh->cmd)) { in clp_service_call()
220 pbdev = s390_pci_find_dev_by_fh(s, ldl_be_p(&reqsetpci->fh)); in clp_service_call()
222 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_FH); in clp_service_call()
226 switch (reqsetpci->oc) { in clp_service_call()
228 switch (reqsetpci->ndas) { in clp_service_call()
230 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_DMAAS); in clp_service_call()
234 default: in clp_service_call()
235 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_RES); in clp_service_call()
239 if (pbdev->fh & FH_MASK_ENABLE) { in clp_service_call()
240 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_FHOP); in clp_service_call()
250 if (pbdev->interp) { in clp_service_call()
252 if (!s390_pci_get_host_fh(pbdev, &pbdev->fh) || in clp_service_call()
253 !(pbdev->fh & FH_MASK_ENABLE)) { in clp_service_call()
254 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_FH); in clp_service_call()
258 pbdev->fh |= FH_MASK_ENABLE; in clp_service_call()
259 pbdev->state = ZPCI_FS_ENABLED; in clp_service_call()
260 stl_be_p(&ressetpci->fh, pbdev->fh); in clp_service_call()
261 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_OK); in clp_service_call()
264 if (!(pbdev->fh & FH_MASK_ENABLE)) { in clp_service_call()
265 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_FHOP); in clp_service_call()
269 pbdev->fh &= ~FH_MASK_ENABLE; in clp_service_call()
270 pbdev->state = ZPCI_FS_DISABLED; in clp_service_call()
271 stl_be_p(&ressetpci->fh, pbdev->fh); in clp_service_call()
272 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_OK); in clp_service_call()
274 default: in clp_service_call()
275 trace_s390_pci_unknown("set-pci", reqsetpci->oc); in clp_service_call()
276 stw_be_p(&ressetpci->hdr.rsp, CLP_RC_SETPCIFN_FHOP); in clp_service_call()
285 pbdev = s390_pci_find_dev_by_fh(s, ldl_be_p(&reqquery->fh)); in clp_service_call()
287 trace_s390_pci_nodev("query", ldl_be_p(&reqquery->fh)); in clp_service_call()
288 stw_be_p(&resquery->hdr.rsp, CLP_RC_SETPCIFN_FH); in clp_service_call()
292 stq_be_p(&resquery->sdma, pbdev->zpci_fn.sdma); in clp_service_call()
293 stq_be_p(&resquery->edma, pbdev->zpci_fn.edma); in clp_service_call()
294 stw_be_p(&resquery->pchid, pbdev->zpci_fn.pchid); in clp_service_call()
295 stw_be_p(&resquery->vfn, pbdev->zpci_fn.vfn); in clp_service_call()
296 resquery->flags = pbdev->zpci_fn.flags; in clp_service_call()
297 resquery->pfgid = pbdev->zpci_fn.pfgid; in clp_service_call()
298 resquery->pft = pbdev->zpci_fn.pft; in clp_service_call()
299 resquery->fmbl = pbdev->zpci_fn.fmbl; in clp_service_call()
300 stl_be_p(&resquery->fid, pbdev->zpci_fn.fid); in clp_service_call()
301 stl_be_p(&resquery->uid, pbdev->zpci_fn.uid); in clp_service_call()
302 memcpy(resquery->pfip, pbdev->zpci_fn.pfip, CLP_PFIP_NR_SEGMENTS); in clp_service_call()
303 memcpy(resquery->util_str, pbdev->zpci_fn.util_str, CLP_UTIL_STR_LEN); in clp_service_call()
306 uint32_t data = pci_get_long(pbdev->pdev->config + in clp_service_call()
309 stl_be_p(&resquery->bar[i], data); in clp_service_call()
310 resquery->bar_size[i] = pbdev->pdev->io_regions[i].size ? in clp_service_call()
311 ctz64(pbdev->pdev->io_regions[i].size) : 0; in clp_service_call()
313 ldl_be_p(&resquery->bar[i]), in clp_service_call()
314 pbdev->pdev->io_regions[i].size, in clp_service_call()
315 resquery->bar_size[i]); in clp_service_call()
318 stw_be_p(&resquery->hdr.rsp, CLP_RC_OK); in clp_service_call()
327 group = s390_group_find(reqgrp->g); in clp_service_call()
331 stw_be_p(&resgrp->hdr.rsp, CLP_RC_QUERYPCIFG_PFGID); in clp_service_call()
334 resgrp->fr = group->zpci_group.fr; in clp_service_call()
335 stq_be_p(&resgrp->dasm, group->zpci_group.dasm); in clp_service_call()
336 stq_be_p(&resgrp->msia, group->zpci_group.msia); in clp_service_call()
337 stw_be_p(&resgrp->mui, group->zpci_group.mui); in clp_service_call()
338 stw_be_p(&resgrp->i, group->zpci_group.i); in clp_service_call()
339 stw_be_p(&resgrp->maxstbl, group->zpci_group.maxstbl); in clp_service_call()
340 resgrp->version = group->zpci_group.version; in clp_service_call()
341 resgrp->dtsm = group->zpci_group.dtsm; in clp_service_call()
342 stw_be_p(&resgrp->hdr.rsp, CLP_RC_OK); in clp_service_call()
345 default: in clp_service_call()
346 trace_s390_pci_unknown("clp", lduw_be_p(&reqh->cmd)); in clp_service_call()
347 stw_be_p(&resh->rsp, CLP_RC_CMD); in clp_service_call()
352 if (s390_cpu_virt_mem_write(cpu, env->regs[r2], r2, buffer, in clp_service_call()
384 default: in zpci_endian_swap()
385 return -EINVAL; in zpci_endian_swap()
397 QTAILQ_FOREACH(subregion, &mr->subregions, subregions_link) { in s390_get_subregion()
398 subregion_size = int128_get64(subregion->size); in s390_get_subregion()
399 if ((offset >= subregion->addr) && in s390_get_subregion()
400 (offset + len) <= (subregion->addr + subregion_size)) { in s390_get_subregion()
413 mr = pbdev->pdev->io_regions[pcias].memory; in zpci_read_bar()
415 offset -= mr->addr; in zpci_read_bar()
423 CPUS390XState *env = &cpu->env; in pcilg_service_call()
432 if (env->psw.mask & PSW_MASK_PSTATE) { in pcilg_service_call()
442 fh = env->regs[r2] >> 32; in pcilg_service_call()
443 pcias = (env->regs[r2] >> 16) & 0xf; in pcilg_service_call()
444 len = env->regs[r2] & 0xf; in pcilg_service_call()
445 offset = env->regs[r2 + 1]; in pcilg_service_call()
459 switch (pbdev->state) { in pcilg_service_call()
465 default: in pcilg_service_call()
471 if (!len || (len > (8 - (offset & 0x7)))) { in pcilg_service_call()
482 if (!len || (len > (4 - (offset & 0x3))) || len == 3) { in pcilg_service_call()
487 pbdev->pdev, offset, pci_config_size(pbdev->pdev), len); in pcilg_service_call()
494 default: in pcilg_service_call()
501 pbdev->fmb.counter[ZPCI_FMB_CNT_LD]++; in pcilg_service_call()
503 env->regs[r1] = data; in pcilg_service_call()
513 mr = pbdev->pdev->io_regions[pcias].memory; in zpci_write_bar()
515 offset -= mr->addr; in zpci_write_bar()
523 CPUS390XState *env = &cpu->env; in pcistg_service_call()
531 if (env->psw.mask & PSW_MASK_PSTATE) { in pcistg_service_call()
541 fh = env->regs[r2] >> 32; in pcistg_service_call()
542 pcias = (env->regs[r2] >> 16) & 0xf; in pcistg_service_call()
543 len = env->regs[r2] & 0xf; in pcistg_service_call()
544 offset = env->regs[r2 + 1]; in pcistg_service_call()
545 data = env->regs[r1]; in pcistg_service_call()
559 switch (pbdev->state) { in pcistg_service_call()
568 default: in pcistg_service_call()
578 if (!len || (len > (8 - (offset & 0x7)))) { in pcistg_service_call()
592 if (!len || (len > (4 - (offset & 0x3))) || len == 3) { in pcistg_service_call()
598 pci_host_config_write_common(pbdev->pdev, offset, in pcistg_service_call()
599 pci_config_size(pbdev->pdev), in pcistg_service_call()
602 default: in pcistg_service_call()
609 pbdev->fmb.counter[ZPCI_FMB_CNT_ST]++; in pcistg_service_call()
618 S390IOTLBEntry *cache = g_hash_table_lookup(iommu->iotlb, &entry->iova); in s390_pci_update_iotlb()
620 .type = entry->perm ? IOMMU_NOTIFIER_MAP : IOMMU_NOTIFIER_UNMAP, in s390_pci_update_iotlb()
623 .iova = entry->iova, in s390_pci_update_iotlb()
624 .translated_addr = entry->translated_addr, in s390_pci_update_iotlb()
625 .perm = entry->perm, in s390_pci_update_iotlb()
634 g_hash_table_remove(iommu->iotlb, &entry->iova); in s390_pci_update_iotlb()
640 if (cache->perm == entry->perm && in s390_pci_update_iotlb()
641 cache->translated_addr == entry->translated_addr) { in s390_pci_update_iotlb()
647 memory_region_notify_iommu(&iommu->iommu_mr, 0, event); in s390_pci_update_iotlb()
649 event.entry.perm = entry->perm; in s390_pci_update_iotlb()
653 cache->iova = entry->iova; in s390_pci_update_iotlb()
654 cache->translated_addr = entry->translated_addr; in s390_pci_update_iotlb()
655 cache->len = TARGET_PAGE_SIZE; in s390_pci_update_iotlb()
656 cache->perm = entry->perm; in s390_pci_update_iotlb()
657 g_hash_table_replace(iommu->iotlb, &cache->iova, cache); in s390_pci_update_iotlb()
665 memory_region_notify_iommu(&iommu->iommu_mr, 0, event); in s390_pci_update_iotlb()
668 return iommu->dma_limit ? iommu->dma_limit->avail : 1; in s390_pci_update_iotlb()
674 uint64_t remain = len, start = iova, end = start + len - 1, mask, size; in s390_pci_batch_unmap()
689 memory_region_notify_iommu(&iommu->iommu_mr, 0, event); in s390_pci_batch_unmap()
691 remain -= size; in s390_pci_batch_unmap()
697 CPUS390XState *env = &cpu->env; in rpcit_service_call()
708 if (env->psw.mask & PSW_MASK_PSTATE) { in rpcit_service_call()
718 fh = env->regs[r1] >> 32; in rpcit_service_call()
719 sstart = start = env->regs[r2]; in rpcit_service_call()
720 end = start + env->regs[r2 + 1]; in rpcit_service_call()
729 switch (pbdev->state) { in rpcit_service_call()
740 default: in rpcit_service_call()
744 iommu = pbdev->iommu; in rpcit_service_call()
745 if (iommu->dma_limit) { in rpcit_service_call()
746 dma_avail = iommu->dma_limit->avail; in rpcit_service_call()
750 if (!iommu->g_iota) { in rpcit_service_call()
755 if (end < iommu->pba || start > iommu->pal) { in rpcit_service_call()
764 error = s390_guest_io_table_walk(iommu->g_iota, start, &entry); in rpcit_service_call()
810 pbdev->state = ZPCI_FS_ERROR; in rpcit_service_call()
813 s390_pci_generate_error_event(error, pbdev->fh, pbdev->fid, start, 0); in rpcit_service_call()
815 pbdev->fmb.counter[ZPCI_FMB_CNT_RPCIT]++; in rpcit_service_call()
830 CPUS390XState *env = &cpu->env; in pcistb_service_call()
841 if (env->psw.mask & PSW_MASK_PSTATE) { in pcistb_service_call()
846 fh = env->regs[r1] >> 32; in pcistb_service_call()
847 pcias = (env->regs[r1] >> 16) & 0xf; in pcistb_service_call()
848 len = env->regs[r1] & 0x1fff; in pcistb_service_call()
849 offset = env->regs[r3]; in pcistb_service_call()
863 switch (pbdev->state) { in pcistb_service_call()
869 default: in pcistb_service_call()
888 (len > pbdev->pci_group->zpci_group.maxstbl)) { in pcistb_service_call()
891 /* Do not cross a 4K-byte boundary */ in pcistb_service_call()
900 mr = pbdev->pdev->io_regions[pcias].memory; in pcistb_service_call()
902 offset -= mr->addr; in pcistb_service_call()
927 pbdev->fmb.counter[ZPCI_FMB_CNT_STB]++; in pcistb_service_call()
942 pbdev->routes.adapter.adapter_id = css_get_adapter_id( in reg_irqs()
944 pbdev->summary_ind = get_indicator(ldq_be_p(&fib.aisb), sizeof(uint64_t)); in reg_irqs()
946 pbdev->indicator = get_indicator(ldq_be_p(&fib.aibv), len); in reg_irqs()
948 ret = map_indicator(&pbdev->routes.adapter, pbdev->summary_ind); in reg_irqs()
953 ret = map_indicator(&pbdev->routes.adapter, pbdev->indicator); in reg_irqs()
958 pbdev->routes.adapter.summary_addr = ldq_be_p(&fib.aisb); in reg_irqs()
959 pbdev->routes.adapter.summary_offset = FIB_DATA_AISBO(ldl_be_p(&fib.data)); in reg_irqs()
960 pbdev->routes.adapter.ind_addr = ldq_be_p(&fib.aibv); in reg_irqs()
961 pbdev->routes.adapter.ind_offset = FIB_DATA_AIBVO(ldl_be_p(&fib.data)); in reg_irqs()
962 pbdev->isc = isc; in reg_irqs()
963 pbdev->noi = FIB_DATA_NOI(ldl_be_p(&fib.data)); in reg_irqs()
964 pbdev->sum = FIB_DATA_SUM(ldl_be_p(&fib.data)); in reg_irqs()
966 trace_s390_pci_irqs("register", pbdev->routes.adapter.adapter_id); in reg_irqs()
969 release_indicator(&pbdev->routes.adapter, pbdev->summary_ind); in reg_irqs()
970 release_indicator(&pbdev->routes.adapter, pbdev->indicator); in reg_irqs()
971 pbdev->summary_ind = NULL; in reg_irqs()
972 pbdev->indicator = NULL; in reg_irqs()
978 release_indicator(&pbdev->routes.adapter, pbdev->summary_ind); in pci_dereg_irqs()
979 release_indicator(&pbdev->routes.adapter, pbdev->indicator); in pci_dereg_irqs()
981 pbdev->summary_ind = NULL; in pci_dereg_irqs()
982 pbdev->indicator = NULL; in pci_dereg_irqs()
983 pbdev->routes.adapter.summary_addr = 0; in pci_dereg_irqs()
984 pbdev->routes.adapter.summary_offset = 0; in pci_dereg_irqs()
985 pbdev->routes.adapter.ind_addr = 0; in pci_dereg_irqs()
986 pbdev->routes.adapter.ind_offset = 0; in pci_dereg_irqs()
987 pbdev->isc = 0; in pci_dereg_irqs()
988 pbdev->noi = 0; in pci_dereg_irqs()
989 pbdev->sum = 0; in pci_dereg_irqs()
991 trace_s390_pci_irqs("unregister", pbdev->routes.adapter.adapter_id); in pci_dereg_irqs()
998 S390PCIIOMMU *iommu = pbdev->iommu; in reg_ioat()
1007 if (pba > pal || pba < pbdev->zpci_fn.sdma || pal > pbdev->zpci_fn.edma) { in reg_ioat()
1009 return -EINVAL; in reg_ioat()
1016 return -EINVAL; in reg_ioat()
1017 } else if (!t && !pbdev->rtr_avail) { in reg_ioat()
1020 return -EINVAL; in reg_ioat()
1023 iommu->pba = pba; in reg_ioat()
1024 iommu->pal = pal; in reg_ioat()
1025 iommu->g_iota = g_iota; in reg_ioat()
1039 iommu->pba = 0; in pci_dereg_ioat()
1040 iommu->pal = 0; in pci_dereg_ioat()
1041 iommu->g_iota = 0; in pci_dereg_ioat()
1046 if (pbdev->fmb_timer) { in fmb_timer_free()
1047 timer_free(pbdev->fmb_timer); in fmb_timer_free()
1048 pbdev->fmb_timer = NULL; in fmb_timer_free()
1050 pbdev->fmb_addr = 0; in fmb_timer_free()
1051 memset(&pbdev->fmb, 0, sizeof(ZpciFmb)); in fmb_timer_free()
1058 uint64_t dst = pbdev->fmb_addr + offset; in fmb_do_update()
1081 default: in fmb_do_update()
1086 s390_pci_generate_error_event(ERR_EVENT_FMBA, pbdev->fh, pbdev->fid, in fmb_do_update()
1087 pbdev->fmb_addr, 0); in fmb_do_update()
1101 pbdev->fmb.last_update *= 2; in fmb_update()
1102 pbdev->fmb.last_update |= UPDATE_U_BIT; in fmb_update()
1104 pbdev->fmb.last_update, in fmb_update()
1105 sizeof(pbdev->fmb.last_update))) { in fmb_update()
1111 pbdev->fmb.sample++, in fmb_update()
1112 sizeof(pbdev->fmb.sample))) { in fmb_update()
1119 pbdev->fmb.counter[i], in fmb_update()
1120 sizeof(pbdev->fmb.counter[0]))) { in fmb_update()
1126 pbdev->fmb.last_update = time2tod(qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL)); in fmb_update()
1127 pbdev->fmb.last_update *= 2; in fmb_update()
1129 pbdev->fmb.last_update, in fmb_update()
1130 sizeof(pbdev->fmb.last_update))) { in fmb_update()
1133 timer_mod(pbdev->fmb_timer, t + pbdev->pci_group->zpci_group.mui); in fmb_update()
1140 rc = s390_pci_kvm_aif_enable(pbdev, fib, pbdev->forwarding_assist); in mpcifc_reg_int_interp()
1165 CPUS390XState *env = &cpu->env; in mpcifc_service_call()
1172 if (env->psw.mask & PSW_MASK_PSTATE) { in mpcifc_service_call()
1177 oc = env->regs[r1] & 0xff; in mpcifc_service_call()
1178 dmaas = (env->regs[r1] >> 16) & 0xff; in mpcifc_service_call()
1179 fh = env->regs[r1] >> 32; in mpcifc_service_call()
1193 switch (pbdev->state) { in mpcifc_service_call()
1200 default: in mpcifc_service_call()
1216 if (pbdev->interp) { in mpcifc_service_call()
1221 } else if (pbdev->summary_ind) { in mpcifc_service_call()
1230 if (pbdev->interp) { in mpcifc_service_call()
1235 } else if (!pbdev->summary_ind) { in mpcifc_service_call()
1246 } else if (pbdev->iommu->enabled) { in mpcifc_service_call()
1258 } else if (!pbdev->iommu->enabled) { in mpcifc_service_call()
1262 pci_dereg_ioat(pbdev->iommu); in mpcifc_service_call()
1269 } else if (!pbdev->iommu->enabled) { in mpcifc_service_call()
1273 pci_dereg_ioat(pbdev->iommu); in mpcifc_service_call()
1281 switch (pbdev->state) { in mpcifc_service_call()
1284 pbdev->state = ZPCI_FS_ENABLED; in mpcifc_service_call()
1286 default: in mpcifc_service_call()
1292 switch (pbdev->state) { in mpcifc_service_call()
1294 pbdev->state = ZPCI_FS_BLOCKED; in mpcifc_service_call()
1296 default: in mpcifc_service_call()
1306 s390_pci_generate_error_event(ERR_EVENT_FMBPRO, pbdev->fh, in mpcifc_service_call()
1307 pbdev->fid, fmb_addr, 0); in mpcifc_service_call()
1318 if (!pbdev->fmb_timer) { in mpcifc_service_call()
1319 pbdev->fmb_timer = timer_new_ms(QEMU_CLOCK_VIRTUAL, in mpcifc_service_call()
1321 } else if (timer_pending(pbdev->fmb_timer)) { in mpcifc_service_call()
1323 timer_del(pbdev->fmb_timer); in mpcifc_service_call()
1325 pbdev->fmb_addr = fmb_addr; in mpcifc_service_call()
1326 timer_mod(pbdev->fmb_timer, in mpcifc_service_call()
1328 pbdev->pci_group->zpci_group.mui); in mpcifc_service_call()
1331 default: in mpcifc_service_call()
1332 s390_program_interrupt(&cpu->env, PGM_OPERAND, ra); in mpcifc_service_call()
1343 CPUS390XState *env = &cpu->env; in stpcifc_service_call()
1351 if (env->psw.mask & PSW_MASK_PSTATE) { in stpcifc_service_call()
1356 fh = env->regs[r1] >> 32; in stpcifc_service_call()
1357 dmaas = (env->regs[r1] >> 16) & 0xff; in stpcifc_service_call()
1378 switch (pbdev->state) { in stpcifc_service_call()
1399 if (pbdev->iommu->enabled) { in stpcifc_service_call()
1403 env->regs[r1] |= 1ULL << 63; in stpcifc_service_call()
1412 stq_be_p(&fib.pba, pbdev->iommu->pba); in stpcifc_service_call()
1413 stq_be_p(&fib.pal, pbdev->iommu->pal); in stpcifc_service_call()
1414 stq_be_p(&fib.iota, pbdev->iommu->g_iota); in stpcifc_service_call()
1415 stq_be_p(&fib.aibv, pbdev->routes.adapter.ind_addr); in stpcifc_service_call()
1416 stq_be_p(&fib.aisb, pbdev->routes.adapter.summary_addr); in stpcifc_service_call()
1417 stq_be_p(&fib.fmb_addr, pbdev->fmb_addr); in stpcifc_service_call()
1419 data = ((uint32_t)pbdev->isc << 28) | ((uint32_t)pbdev->noi << 16) | in stpcifc_service_call()
1420 ((uint32_t)pbdev->routes.adapter.ind_offset << 8) | in stpcifc_service_call()
1421 ((uint32_t)pbdev->sum << 7) | pbdev->routes.adapter.summary_offset; in stpcifc_service_call()