Lines Matching refs:mv
163 u32 inbound_head = readl(&hba->u.mv.mu->inbound_head);
169 memcpy_toio(&hba->u.mv.mu->inbound_q[inbound_head], &p, 8);
170 writel(head, &hba->u.mv.mu->inbound_head);
172 &hba->u.mv.regs->inbound_doorbell);
208 status = readl(&hba->u.mv.regs->outbound_doorbell);
209 writel(~status, &hba->u.mv.regs->outbound_doorbell);
213 msg = readl(&hba->u.mv.mu->outbound_msg);
222 while ((tag = mv_outbound_read(hba->u.mv.mu)))
323 struct hpt_iop_request_header *reqhdr = hba->u.mv.internal_req;
328 mv_inbound_write(hba->u.mv.internal_req_phy |
368 writel(msg, &hba->u.mv.mu->inbound_msg);
369 writel(MVIOP_MU_INBOUND_INT_MSG, &hba->u.mv.regs->inbound_doorbell);
370 readl(&hba->u.mv.regs->inbound_doorbell);
431 struct hpt_iop_request_get_config *req = hba->u.mv.internal_req;
507 struct hpt_iop_request_set_config *req = hba->u.mv.internal_req;
558 &hba->u.mv.regs->outbound_intmask);
637 hba->u.mv.regs = hptiop_map_pci_bar(hba, 0);
638 if (hba->u.mv.regs == NULL)
641 hba->u.mv.mu = hptiop_map_pci_bar(hba, 2);
642 if (hba->u.mv.mu == NULL) {
643 iounmap(hba->u.mv.regs);
667 iounmap(hba->u.mv.regs);
668 iounmap(hba->u.mv.mu);
1184 hba->u.mv.internal_req = dma_alloc_coherent(&hba->pcidev->dev,
1185 0x800, &hba->u.mv.internal_req_phy, GFP_KERNEL);
1186 if (hba->u.mv.internal_req)
1251 if (hba->u.mv.internal_req) {
1253 hba->u.mv.internal_req, hba->u.mv.internal_req_phy);
1531 writel(0, &hba->u.mv.regs->outbound_intmask);
1532 readl(&hba->u.mv.regs->outbound_intmask);