/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_gen6_ras.c | 10 static void enable_errsou_reporting(void __iomem *csr) in enable_errsou_reporting() argument 13 ADF_CSR_WR(csr, ADF_GEN6_ERRMSK0, 0); in enable_errsou_reporting() 16 ADF_CSR_WR(csr, ADF_GEN6_ERRMSK1, 0); in enable_errsou_reporting() 22 ADF_CSR_WR(csr, ADF_GEN6_ERRMSK2, ADF_GEN6_ERRSOU2_PM_INT_BIT); in enable_errsou_reporting() 25 ADF_CSR_WR(csr, ADF_GEN6_ERRMSK3, 0); in enable_errsou_reporting() 28 static void enable_ae_error_reporting(struct adf_accel_dev *accel_dev, void __iomem *csr) in enable_ae_error_reporting() argument 33 ADF_CSR_WR(csr, ADF_GEN6_HIAECORERRLOGENABLE_CPP0, ae_mask); in enable_ae_error_reporting() 36 ADF_CSR_WR(csr, ADF_GEN6_HIAEUNCERRLOGENABLE_CPP0, ae_mask); in enable_ae_error_reporting() 39 static void enable_cpp_error_reporting(struct adf_accel_dev *accel_dev, void __iomem *csr) in enable_cpp_error_reporting() argument 42 ADF_CSR_WR(csr, ADF_GEN6_HICPPAGENTCMDPARERRLOGENABLE, in enable_cpp_error_reporting() [all …]
|
H A D | adf_gen4_ras.c | 10 static void enable_errsou_reporting(void __iomem *csr) in enable_errsou_reporting() argument 13 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK0, 0); in enable_errsou_reporting() 16 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK1, 0); in enable_errsou_reporting() 22 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK2, in enable_errsou_reporting() 30 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK3, in enable_errsou_reporting() 35 static void disable_errsou_reporting(void __iomem *csr) in disable_errsou_reporting() argument 40 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK0, ADF_GEN4_ERRSOU0_BIT); in disable_errsou_reporting() 43 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK1, ADF_GEN4_ERRSOU1_BITMASK); in disable_errsou_reporting() 46 val = ADF_CSR_RD(csr, ADF_GEN4_ERRMSK2); in disable_errsou_reporting() 48 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK2, val); in disable_errsou_reporting() [all …]
|
H A D | icp_qat_hw_20_comp.h | 23 ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER(struct icp_qat_hw_comp_20_config_csr_lower csr) in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() argument 27 QAT_FIELD_SET(val32, csr.algo, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 30 QAT_FIELD_SET(val32, csr.sd, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 33 QAT_FIELD_SET(val32, csr.edmm, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 36 QAT_FIELD_SET(val32, csr.hbs, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 39 QAT_FIELD_SET(val32, csr.lllbd, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 42 QAT_FIELD_SET(val32, csr.mmctrl, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 45 QAT_FIELD_SET(val32, csr.hash_col, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 48 QAT_FIELD_SET(val32, csr.hash_update, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 51 QAT_FIELD_SET(val32, csr.skip_ctrl, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() [all …]
|
H A D | icp_qat_hw_51_comp.h | 20 ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER(struct icp_qat_hw_comp_51_config_csr_lower csr) in ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER() argument 24 QAT_FIELD_SET(val32, csr.abd, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER() 27 QAT_FIELD_SET(val32, csr.lllbd, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER() 30 QAT_FIELD_SET(val32, csr.sd, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER() 33 QAT_FIELD_SET(val32, csr.mmctrl, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER() 36 QAT_FIELD_SET(val32, csr.lbc, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_LOWER() 50 ICP_QAT_FW_COMP_51_BUILD_CONFIG_UPPER(struct icp_qat_hw_comp_51_config_csr_upper csr) in ICP_QAT_FW_COMP_51_BUILD_CONFIG_UPPER() argument 54 QAT_FIELD_SET(val32, csr.edmm, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_UPPER() 57 QAT_FIELD_SET(val32, csr.bms, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_UPPER() 60 QAT_FIELD_SET(val32, csr.scb_mode_reset, in ICP_QAT_FW_COMP_51_BUILD_CONFIG_UPPER() [all …]
|
/linux/arch/loongarch/kvm/ |
H A D | vcpu.c | 74 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_save_guest_pmu() local 76 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR0); in kvm_save_guest_pmu() 77 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR1); in kvm_save_guest_pmu() 78 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR2); in kvm_save_guest_pmu() 79 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR3); in kvm_save_guest_pmu() 80 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL0); in kvm_save_guest_pmu() 81 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL1); in kvm_save_guest_pmu() 82 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL2); in kvm_save_guest_pmu() 83 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL3); in kvm_save_guest_pmu() 88 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_restore_guest_pmu() local [all …]
|
H A D | timer.c | 46 kvm_write_sw_gcsr(vcpu->arch.csr, LOONGARCH_CSR_TVAL, 0); in kvm_init_timer() 57 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_restore_timer() local 60 * Set guest stable timer cfg csr in kvm_restore_timer() 61 * Disable timer before restore estat CSR register, avoid to in kvm_restore_timer() 64 cfg = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_TCFG); in kvm_restore_timer() 67 kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_ESTAT); in kvm_restore_timer() 68 kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TCFG); in kvm_restore_timer() 71 kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TVAL); in kvm_restore_timer() 83 * If oneshot timer is fired, CSR TVAL will be -1, there are two in kvm_restore_timer() 90 ticks = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_TVAL); in kvm_restore_timer() [all …]
|
/linux/arch/sparc/kernel/ |
H A D | ebus.c | 74 u32 csr = 0; in ebus_dma_irq() local 77 csr = readl(p->regs + EBDMA_CSR); in ebus_dma_irq() 78 writel(csr, p->regs + EBDMA_CSR); in ebus_dma_irq() 81 if (csr & EBDMA_CSR_ERR_PEND) { in ebus_dma_irq() 85 } else if (csr & EBDMA_CSR_INT_PEND) { in ebus_dma_irq() 87 (csr & EBDMA_CSR_TC) ? in ebus_dma_irq() 99 u32 csr; in ebus_dma_register() local 113 csr = EBDMA_CSR_BURST_SZ_16 | EBDMA_CSR_EN_CNT; in ebus_dma_register() 116 csr |= EBDMA_CSR_TCI_DIS; in ebus_dma_register() 118 writel(csr, p->regs + EBDMA_CSR); in ebus_dma_register() [all …]
|
/linux/arch/alpha/kernel/ |
H A D | core_tsunami.c | 181 volatile unsigned long *csr; in tsunami_pci_tbi() local 186 csr = &pchip->tlbia.csr; in tsunami_pci_tbi() 188 csr = &pchip->tlbiv.csr; in tsunami_pci_tbi() 194 *csr = value; in tsunami_pci_tbi() 196 *csr; in tsunami_pci_tbi() 227 TSUNAMI_cchip->misc.csr |= (1L << 28); /* clear NXM... */ in tsunami_probe_write() 231 if (TSUNAMI_cchip->misc.csr & (1L << 28)) { in tsunami_probe_write() 232 int source = (TSUNAMI_cchip->misc.csr >> 29) & 7; in tsunami_probe_write() 233 TSUNAMI_cchip->misc.csr |= (1L << 28); /* ...and unlock NXS. */ in tsunami_probe_write() 251 if (tsunami_probe_read(&pchip->pctl.csr) == 0) in tsunami_init_one_pchip() [all …]
|
H A D | core_wildfire.c | 121 pci->pci_window[0].wbase.csr = hose->sg_isa->dma_base | 3; in wildfire_init_hose() 122 pci->pci_window[0].wmask.csr = (hose->sg_isa->size - 1) & 0xfff00000; in wildfire_init_hose() 123 pci->pci_window[0].tbase.csr = virt_to_phys(hose->sg_isa->ptes); in wildfire_init_hose() 125 pci->pci_window[1].wbase.csr = 0x40000000 | 1; in wildfire_init_hose() 126 pci->pci_window[1].wmask.csr = (0x40000000 -1) & 0xfff00000; in wildfire_init_hose() 127 pci->pci_window[1].tbase.csr = 0; in wildfire_init_hose() 129 pci->pci_window[2].wbase.csr = 0x80000000 | 1; in wildfire_init_hose() 130 pci->pci_window[2].wmask.csr = (0x40000000 -1) & 0xfff00000; in wildfire_init_hose() 131 pci->pci_window[2].tbase.csr = 0x40000000; in wildfire_init_hose() 133 pci->pci_window[3].wbase.csr = hose->sg_pci->dma_base | 3; in wildfire_init_hose() [all …]
|
H A D | core_titan.c | 207 volatile unsigned long *csr; in titan_pci_tbi() local 220 csr = &port->port_specific.g.gtlbia.csr; in titan_pci_tbi() 222 csr = &port->port_specific.g.gtlbiv.csr; in titan_pci_tbi() 229 *csr = value; in titan_pci_tbi() 231 *csr; in titan_pci_tbi() 240 pctl.pctl_q_whole = port->pctl.csr; in titan_query_agp() 293 saved_config[index].wsba[0] = port->wsba[0].csr; in titan_init_one_pachip_port() 294 saved_config[index].wsm[0] = port->wsm[0].csr; in titan_init_one_pachip_port() 295 saved_config[index].tba[0] = port->tba[0].csr; in titan_init_one_pachip_port() 297 saved_config[index].wsba[1] = port->wsba[1].csr; in titan_init_one_pachip_port() [all …]
|
H A D | sys_marvel.c | 96 ctl = &io7->csrs->PO7_LSI_CTL[irq & 0xff].csr; /* assume LSI */ in io7_get_irq_ctl() 98 ctl = &io7->csrs->PO7_MSI_CTL[((irq - 0x80) >> 5) & 0x0f].csr; in io7_get_irq_ctl() 174 volatile unsigned long *csr, in io7_redirect_irq() argument 179 val = *csr; in io7_redirect_irq() 183 *csr = val; in io7_redirect_irq() 185 *csr; in io7_redirect_irq() 196 val = io7->csrs->PO7_LSI_CTL[which].csr; in io7_redirect_one_lsi() 200 io7->csrs->PO7_LSI_CTL[which].csr = val; in io7_redirect_one_lsi() 202 io7->csrs->PO7_LSI_CTL[which].csr; in io7_redirect_one_lsi() 213 val = io7->csrs->PO7_MSI_CTL[which].csr; in io7_redirect_one_msi() [all …]
|
H A D | err_ev7.c | 194 "DM CSR PH", "DM CSR PH", "DM CSR PH", 195 "DM CSR PH", "reserved", 198 "DM CSR PH", "DM CSR PH", "DM CSR PH", 199 "DM CSR PH", "reserved", 202 "DM CSR PH", "DM CSR PH", "DM CSR PH", 203 "DM CSR PH", "reserved", 206 "DM CSR PH", "DM CSR PH", "DM CSR PH", 207 "DM CSR PH", "reserved",
|
/linux/drivers/crypto/starfive/ |
H A D | jh7110-rsa.c | 82 rctx->csr.pka.v = 0; in starfive_rsa_montgomery_form() 84 writel(rctx->csr.pka.v, cryp->base + STARFIVE_PKA_CACR_OFFSET); in starfive_rsa_montgomery_form() 90 rctx->csr.pka.v = 0; in starfive_rsa_montgomery_form() 91 rctx->csr.pka.cln_done = 1; in starfive_rsa_montgomery_form() 92 rctx->csr.pka.opsize = opsize; in starfive_rsa_montgomery_form() 93 rctx->csr.pka.exposize = opsize; in starfive_rsa_montgomery_form() 94 rctx->csr.pka.cmd = CRYPTO_CMD_PRE; in starfive_rsa_montgomery_form() 95 rctx->csr.pka.start = 1; in starfive_rsa_montgomery_form() 96 rctx->csr.pka.not_r2 = 1; in starfive_rsa_montgomery_form() 97 rctx->csr.pka.ie = 1; in starfive_rsa_montgomery_form() [all …]
|
/linux/arch/riscv/kvm/ |
H A D | aia.c | 50 struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr; in kvm_riscv_vcpu_aia_flush_interrupts() local 60 csr->hviph &= ~mask; in kvm_riscv_vcpu_aia_flush_interrupts() 61 csr->hviph |= val; in kvm_riscv_vcpu_aia_flush_interrupts() 67 struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr; in kvm_riscv_vcpu_aia_sync_interrupts() local 70 csr->vsieh = ncsr_read(CSR_VSIEH); in kvm_riscv_vcpu_aia_sync_interrupts() 99 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_aia_update_hvip() local 107 ncsr_write(CSR_HVICTL, aia_hvictl_value(!!(csr->hvip & BIT(IRQ_VS_EXT)))); in kvm_riscv_vcpu_aia_update_hvip() 112 struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr; in kvm_riscv_vcpu_aia_load() local 120 nacl_csr_write(nsh, CSR_VSISELECT, csr->vsiselect); in kvm_riscv_vcpu_aia_load() 121 nacl_csr_write(nsh, CSR_HVIPRIO1, csr->hviprio1); in kvm_riscv_vcpu_aia_load() [all …]
|
H A D | vcpu.c | 58 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_context_reset() local 63 memset(csr, 0, sizeof(*csr)); in kvm_riscv_vcpu_context_reset() 348 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_flush_interrupts() local 355 csr->hvip &= ~mask; in kvm_riscv_vcpu_flush_interrupts() 356 csr->hvip |= val; in kvm_riscv_vcpu_flush_interrupts() 367 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_sync_interrupts() local 370 csr->vsie = ncsr_read(CSR_VSIE); in kvm_riscv_vcpu_sync_interrupts() 374 if ((csr->hvip ^ hvip) & (1UL << IRQ_VS_SOFT)) { in kvm_riscv_vcpu_sync_interrupts() 387 if ((csr->hvip ^ hvip) & (1UL << IRQ_PMU_OVF)) { in kvm_riscv_vcpu_sync_interrupts() 581 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_arch_vcpu_load() local [all …]
|
/linux/drivers/scsi/ |
H A D | sun3_scsi.c | 73 unsigned short csr; /* control/status reg */ member 117 /* bits in csr reg */ 190 // safe bits for the CSR 196 unsigned short csr = dregs->csr; in scsi_sun3_intr() local 200 dregs->csr &= ~CSR_DMA_ENABLE; in scsi_sun3_intr() 203 if(csr & ~CSR_GOOD) { in scsi_sun3_intr() 204 if (csr & CSR_DMA_BUSERR) in scsi_sun3_intr() 206 if (csr & CSR_DMA_CONFLICT) in scsi_sun3_intr() 211 if(csr & (CSR_SDB_INT | CSR_DMA_INT)) { in scsi_sun3_intr() 242 dregs->csr &= ~CSR_FIFO; in sun3scsi_dma_setup() [all …]
|
/linux/arch/loongarch/include/asm/ |
H A D | kvm_csr.h | 14 #define gcsr_read(csr) \ argument 20 : [reg] "i" (csr) \ 25 #define gcsr_write(v, csr) \ argument 31 : [reg] "i" (csr) \ 36 #define gcsr_xchg(v, m, csr) \ argument 42 : [mask] "r" (m), [reg] "i" (csr) \ 182 #define kvm_save_hw_gcsr(csr, gid) (csr->csrs[gid] = gcsr_read(gid)) argument 183 #define kvm_restore_hw_gcsr(csr, gid) (gcsr_write(csr->csrs[gid], gid)) argument 185 #define kvm_read_clear_hw_gcsr(csr, gid) (csr->csrs[gid] = gcsr_write(0, gid)) argument 189 static __always_inline unsigned long kvm_read_sw_gcsr(struct loongarch_csrs *csr, int gid) in kvm_read_sw_gcsr() argument [all …]
|
/linux/drivers/usb/musb/ |
H A D | musb_gadget_ep0.c | 243 u16 csr; in service_zero_data_request() local 266 csr = musb_readw(regs, MUSB_TXCSR); in service_zero_data_request() 267 csr |= MUSB_TXCSR_CLRDATATOG | in service_zero_data_request() 269 csr &= ~(MUSB_TXCSR_P_SENDSTALL | in service_zero_data_request() 272 musb_writew(regs, MUSB_TXCSR, csr); in service_zero_data_request() 274 csr = musb_readw(regs, MUSB_RXCSR); in service_zero_data_request() 275 csr |= MUSB_RXCSR_CLRDATATOG | in service_zero_data_request() 277 csr &= ~(MUSB_RXCSR_P_SENDSTALL | in service_zero_data_request() 279 musb_writew(regs, MUSB_RXCSR, csr); in service_zero_data_request() 403 u16 csr; in service_zero_data_request() local [all …]
|
H A D | musb_cppi41.c | 56 u16 csr; in save_rx_toggle() local 64 csr = musb_readw(cppi41_channel->hw_ep->regs, MUSB_RXCSR); in save_rx_toggle() 65 toggle = csr & MUSB_RXCSR_H_DATATOGGLE ? 1 : 0; in save_rx_toggle() 74 u16 csr; in update_rx_toggle() local 83 csr = musb_readw(hw_ep->regs, MUSB_RXCSR); in update_rx_toggle() 84 toggle = csr & MUSB_RXCSR_H_DATATOGGLE ? 1 : 0; in update_rx_toggle() 92 csr |= MUSB_RXCSR_H_DATATOGGLE | MUSB_RXCSR_H_WR_DATATOGGLE; in update_rx_toggle() 93 musb_writew(cppi41_channel->hw_ep->regs, MUSB_RXCSR, csr); in update_rx_toggle() 105 u16 csr; in musb_is_tx_fifo_empty() local 108 csr = musb_readw(epio, MUSB_TXCSR); in musb_is_tx_fifo_empty() [all …]
|
/linux/drivers/usb/gadget/udc/ |
H A D | at91_udc.c | 112 u32 csr; in proc_ep_show() local 119 csr = __raw_readl(ep->creg); in proc_ep_show() 132 seq_printf(s, "csr %08x rxbytes=%d %s %s %s" EIGHTBITS "\n", in proc_ep_show() 133 csr, in proc_ep_show() 134 (csr & 0x07ff0000) >> 16, in proc_ep_show() 135 str_enabled_disabled(csr & (1 << 15)), in proc_ep_show() 136 (csr & (1 << 11)) ? "DATA1" : "DATA0", in proc_ep_show() 137 types[(csr & 0x700) >> 8], in proc_ep_show() 140 (!(csr & 0x700)) in proc_ep_show() 141 ? ((csr & (1 << 7)) ? " IN" : " OUT") in proc_ep_show() [all …]
|
/linux/arch/sh/kernel/cpu/ |
H A D | adc.c | 16 unsigned char csr; in adc_single() local 22 csr = __raw_readb(ADCSR); in adc_single() 23 csr = channel | ADCSR_ADST | ADCSR_CKS; in adc_single() 24 __raw_writeb(csr, ADCSR); in adc_single() 27 csr = __raw_readb(ADCSR); in adc_single() 28 } while ((csr & ADCSR_ADF) == 0); in adc_single() 30 csr &= ~(ADCSR_ADF | ADCSR_ADST); in adc_single() 31 __raw_writeb(csr, ADCSR); in adc_single()
|
/linux/arch/alpha/include/asm/ |
H A D | core_t2.h | 206 unsigned long elcmc_bcc; /* CSR 0 */ 207 unsigned long elcmc_bcce; /* CSR 1 */ 208 unsigned long elcmc_bccea; /* CSR 2 */ 209 unsigned long elcmc_bcue; /* CSR 3 */ 210 unsigned long elcmc_bcuea; /* CSR 4 */ 211 unsigned long elcmc_dter; /* CSR 5 */ 212 unsigned long elcmc_cbctl; /* CSR 6 */ 213 unsigned long elcmc_cbe; /* CSR 7 */ 214 unsigned long elcmc_cbeal; /* CSR 8 */ 215 unsigned long elcmc_cbeah; /* CSR 9 */ [all …]
|
/linux/drivers/watchdog/ |
H A D | shwdt.c | 85 u8 csr; in sh_wdt_start() local 95 csr = sh_wdt_read_csr(); in sh_wdt_start() 96 csr |= WTCSR_WT | clock_division_ratio; in sh_wdt_start() 97 sh_wdt_write_csr(csr); in sh_wdt_start() 109 csr = sh_wdt_read_csr(); in sh_wdt_start() 110 csr |= WTCSR_TME; in sh_wdt_start() 111 csr &= ~WTCSR_RSTS; in sh_wdt_start() 112 sh_wdt_write_csr(csr); in sh_wdt_start() 115 csr = sh_wdt_read_rstcsr(); in sh_wdt_start() 116 csr &= ~RSTCSR_RSTS; in sh_wdt_start() [all …]
|
/linux/include/linux/ |
H A D | rio_regs.h | 103 #define RIO_MBOX_CSR 0x40 /* [II, <= 1.2] Mailbox CSR */ 129 #define RIO_WRITE_PORT_CSR 0x44 /* [I, <= 1.2] Write Port CSR */ 130 #define RIO_DOORBELL_CSR 0x44 /* [II, <= 1.2] Doorbell CSR */ 146 #define RIO_PELL_CTRL_CSR 0x4c /* [I] PE Logical Layer Control CSR */ 156 #define RIO_DID_CSR 0x60 /* [III] Base Device ID CSR */ 160 #define RIO_HOST_DID_LOCK_CSR 0x68 /* [III] Host Base Device ID Lock CSR */ 161 #define RIO_COMPONENT_TAG_CSR 0x6c /* [III] Component Tag CSR */ 301 #define RIO_EM_LTL_ERR_DETECT 0x008 /* Logical/Transport Layer Error Detect CSR */ 302 #define RIO_EM_LTL_ERR_EN 0x00c /* Logical/Transport Layer Error Enable CSR */ 307 #define RIO_EM_LTL_HIADDR_CAP 0x010 /* Logical/Transport Layer High Address Capture CSR */ [all …]
|
/linux/drivers/net/pcs/ |
H A D | pcs-xpcs-plat.c | 41 static u16 xpcs_mmio_addr_page(ptrdiff_t csr) in xpcs_mmio_addr_page() argument 43 return FIELD_GET(0x1fff00, csr); in xpcs_mmio_addr_page() 46 static ptrdiff_t xpcs_mmio_addr_offset(ptrdiff_t csr) in xpcs_mmio_addr_offset() argument 48 return FIELD_GET(0xff, csr); in xpcs_mmio_addr_offset() 54 ptrdiff_t csr, ofs; in xpcs_mmio_read_reg_indirect() local 58 csr = xpcs_mmio_addr_format(dev, reg); in xpcs_mmio_read_reg_indirect() 59 page = xpcs_mmio_addr_page(csr); in xpcs_mmio_read_reg_indirect() 60 ofs = xpcs_mmio_addr_offset(csr); in xpcs_mmio_read_reg_indirect() 85 ptrdiff_t csr, ofs; in xpcs_mmio_write_reg_indirect() local 89 csr = xpcs_mmio_addr_format(dev, reg); in xpcs_mmio_write_reg_indirect() [all …]
|