| /linux/drivers/crypto/intel/qat/qat_c62x/ |
| H A D | adf_c62x_hw_data.c | 116 void adf_init_hw_data_c62x(struct adf_hw_device_data *hw_data) in adf_init_hw_data_c62x() argument 118 hw_data->dev_class = &c62x_class; in adf_init_hw_data_c62x() 119 hw_data->instance_id = c62x_class.instances++; in adf_init_hw_data_c62x() 120 hw_data->num_banks = ADF_C62X_ETR_MAX_BANKS; in adf_init_hw_data_c62x() 121 hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK; in adf_init_hw_data_c62x() 122 hw_data->num_accel = ADF_C62X_MAX_ACCELERATORS; in adf_init_hw_data_c62x() 123 hw_data->num_logical_accel = 1; in adf_init_hw_data_c62x() 124 hw_data->num_engines = ADF_C62X_MAX_ACCELENGINES; in adf_init_hw_data_c62x() 125 hw_data->tx_rx_gap = ADF_GEN2_RX_RINGS_OFFSET; in adf_init_hw_data_c62x() 126 hw_data->tx_rings_mask = ADF_GEN2_TX_RINGS_MASK; in adf_init_hw_data_c62x() [all …]
|
| H A D | adf_drv.c | 60 struct adf_hw_device_data *hw_data; in adf_probe() local 100 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe() 102 if (!hw_data) { in adf_probe() 107 accel_dev->hw_device = hw_data; in adf_probe() 111 &hw_data->fuses[ADF_FUSECTL0]); in adf_probe() 113 &hw_data->straps); in adf_probe() 116 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 117 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 118 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe() 120 if (!hw_data->accel_mask || !hw_data->ae_mask || in adf_probe() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_c3xxx/ |
| H A D | adf_c3xxx_hw_data.c | 114 void adf_init_hw_data_c3xxx(struct adf_hw_device_data *hw_data) in adf_init_hw_data_c3xxx() argument 116 hw_data->dev_class = &c3xxx_class; in adf_init_hw_data_c3xxx() 117 hw_data->instance_id = c3xxx_class.instances++; in adf_init_hw_data_c3xxx() 118 hw_data->num_banks = ADF_C3XXX_ETR_MAX_BANKS; in adf_init_hw_data_c3xxx() 119 hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK; in adf_init_hw_data_c3xxx() 120 hw_data->num_accel = ADF_C3XXX_MAX_ACCELERATORS; in adf_init_hw_data_c3xxx() 121 hw_data->num_logical_accel = 1; in adf_init_hw_data_c3xxx() 122 hw_data->num_engines = ADF_C3XXX_MAX_ACCELENGINES; in adf_init_hw_data_c3xxx() 123 hw_data->tx_rx_gap = ADF_GEN2_RX_RINGS_OFFSET; in adf_init_hw_data_c3xxx() 124 hw_data->tx_rings_mask = ADF_GEN2_TX_RINGS_MASK; in adf_init_hw_data_c3xxx() [all …]
|
| H A D | adf_drv.c | 60 struct adf_hw_device_data *hw_data; in adf_probe() local 100 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe() 102 if (!hw_data) { in adf_probe() 107 accel_dev->hw_device = hw_data; in adf_probe() 111 &hw_data->fuses[ADF_FUSECTL0]); in adf_probe() 113 &hw_data->straps); in adf_probe() 116 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 117 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 118 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe() 120 if (!hw_data->accel_mask || !hw_data->ae_mask || in adf_probe() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_dh895xccvf/ |
| H A D | adf_dh895xccvf_hw_data.c | 61 void adf_init_hw_data_dh895xcciov(struct adf_hw_device_data *hw_data) in adf_init_hw_data_dh895xcciov() argument 63 hw_data->dev_class = &dh895xcciov_class; in adf_init_hw_data_dh895xcciov() 64 hw_data->num_banks = ADF_DH895XCCIOV_ETR_MAX_BANKS; in adf_init_hw_data_dh895xcciov() 65 hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK; in adf_init_hw_data_dh895xcciov() 66 hw_data->num_accel = ADF_DH895XCCIOV_MAX_ACCELERATORS; in adf_init_hw_data_dh895xcciov() 67 hw_data->num_logical_accel = 1; in adf_init_hw_data_dh895xcciov() 68 hw_data->num_engines = ADF_DH895XCCIOV_MAX_ACCELENGINES; in adf_init_hw_data_dh895xcciov() 69 hw_data->tx_rx_gap = ADF_DH895XCCIOV_RX_RINGS_OFFSET; in adf_init_hw_data_dh895xcciov() 70 hw_data->tx_rings_mask = ADF_DH895XCCIOV_TX_RINGS_MASK; in adf_init_hw_data_dh895xcciov() 71 hw_data->ring_to_svc_map = ADF_GEN2_DEFAULT_RING_TO_SRV_MAP; in adf_init_hw_data_dh895xcciov() [all …]
|
| H A D | adf_drv.c | 79 struct adf_hw_device_data *hw_data; in adf_probe() local 112 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe() 114 if (!hw_data) { in adf_probe() 118 accel_dev->hw_device = hw_data; in adf_probe() 122 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 123 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 124 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_c3xxxvf/ |
| H A D | adf_c3xxxvf_hw_data.c | 61 void adf_init_hw_data_c3xxxiov(struct adf_hw_device_data *hw_data) in adf_init_hw_data_c3xxxiov() argument 63 hw_data->dev_class = &c3xxxiov_class; in adf_init_hw_data_c3xxxiov() 64 hw_data->num_banks = ADF_C3XXXIOV_ETR_MAX_BANKS; in adf_init_hw_data_c3xxxiov() 65 hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK; in adf_init_hw_data_c3xxxiov() 66 hw_data->num_accel = ADF_C3XXXIOV_MAX_ACCELERATORS; in adf_init_hw_data_c3xxxiov() 67 hw_data->num_logical_accel = 1; in adf_init_hw_data_c3xxxiov() 68 hw_data->num_engines = ADF_C3XXXIOV_MAX_ACCELENGINES; in adf_init_hw_data_c3xxxiov() 69 hw_data->tx_rx_gap = ADF_C3XXXIOV_RX_RINGS_OFFSET; in adf_init_hw_data_c3xxxiov() 70 hw_data->tx_rings_mask = ADF_C3XXXIOV_TX_RINGS_MASK; in adf_init_hw_data_c3xxxiov() 71 hw_data->ring_to_svc_map = ADF_GEN2_DEFAULT_RING_TO_SRV_MAP; in adf_init_hw_data_c3xxxiov() [all …]
|
| H A D | adf_drv.c | 79 struct adf_hw_device_data *hw_data; in adf_probe() local 112 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe() 114 if (!hw_data) { in adf_probe() 118 accel_dev->hw_device = hw_data; in adf_probe() 122 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 123 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 124 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_c62xvf/ |
| H A D | adf_c62xvf_hw_data.c | 61 void adf_init_hw_data_c62xiov(struct adf_hw_device_data *hw_data) in adf_init_hw_data_c62xiov() argument 63 hw_data->dev_class = &c62xiov_class; in adf_init_hw_data_c62xiov() 64 hw_data->num_banks = ADF_C62XIOV_ETR_MAX_BANKS; in adf_init_hw_data_c62xiov() 65 hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK; in adf_init_hw_data_c62xiov() 66 hw_data->num_accel = ADF_C62XIOV_MAX_ACCELERATORS; in adf_init_hw_data_c62xiov() 67 hw_data->num_logical_accel = 1; in adf_init_hw_data_c62xiov() 68 hw_data->num_engines = ADF_C62XIOV_MAX_ACCELENGINES; in adf_init_hw_data_c62xiov() 69 hw_data->tx_rx_gap = ADF_C62XIOV_RX_RINGS_OFFSET; in adf_init_hw_data_c62xiov() 70 hw_data->tx_rings_mask = ADF_C62XIOV_TX_RINGS_MASK; in adf_init_hw_data_c62xiov() 71 hw_data->ring_to_svc_map = ADF_GEN2_DEFAULT_RING_TO_SRV_MAP; in adf_init_hw_data_c62xiov() [all …]
|
| H A D | adf_drv.c | 79 struct adf_hw_device_data *hw_data; in adf_probe() local 112 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe() 114 if (!hw_data) { in adf_probe() 118 accel_dev->hw_device = hw_data; in adf_probe() 122 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 123 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 124 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_4xxx/ |
| H A D | adf_4xxx_hw_data.c | 387 void adf_init_hw_data_4xxx(struct adf_hw_device_data *hw_data, u32 dev_id) in adf_init_hw_data_4xxx() argument 389 hw_data->dev_class = &adf_4xxx_class; in adf_init_hw_data_4xxx() 390 hw_data->instance_id = adf_4xxx_class.instances++; in adf_init_hw_data_4xxx() 391 hw_data->num_banks = ADF_GEN4_ETR_MAX_BANKS; in adf_init_hw_data_4xxx() 392 hw_data->num_banks_per_vf = ADF_GEN4_NUM_BANKS_PER_VF; in adf_init_hw_data_4xxx() 393 hw_data->num_rings_per_bank = ADF_GEN4_NUM_RINGS_PER_BANK; in adf_init_hw_data_4xxx() 394 hw_data->num_accel = ADF_GEN4_MAX_ACCELERATORS; in adf_init_hw_data_4xxx() 395 hw_data->num_engines = ADF_4XXX_MAX_ACCELENGINES; in adf_init_hw_data_4xxx() 396 hw_data->num_logical_accel = 1; in adf_init_hw_data_4xxx() 397 hw_data->tx_rx_gap = ADF_GEN4_RX_RINGS_OFFSET; in adf_init_hw_data_4xxx() [all …]
|
| H A D | adf_drv.c | 39 struct adf_hw_device_data *hw_data; in adf_probe() local 74 hw_data = devm_kzalloc(&pdev->dev, sizeof(*hw_data), GFP_KERNEL); in adf_probe() 75 if (!hw_data) { in adf_probe() 80 accel_dev->hw_device = hw_data; in adf_probe() 84 pci_read_config_dword(pdev, ADF_GEN4_FUSECTL4_OFFSET, &hw_data->fuses[ADF_FUSECTL4]); in adf_probe() 87 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 88 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 89 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe() 91 if (!hw_data->accel_mask || !hw_data->ae_mask || in adf_probe() 92 (~hw_data->ae_mask & 0x01)) { in adf_probe() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_420xx/ |
| H A D | adf_420xx_hw_data.c | 151 struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev); in update_ae_mask() local 157 ae_mask = get_ae_mask(hw_data); in update_ae_mask() 167 hw_data->ae_mask = ae_mask & config_ae_mask; in update_ae_mask() 412 void adf_init_hw_data_420xx(struct adf_hw_device_data *hw_data, u32 dev_id) in adf_init_hw_data_420xx() argument 414 hw_data->dev_class = &adf_420xx_class; in adf_init_hw_data_420xx() 415 hw_data->instance_id = adf_420xx_class.instances++; in adf_init_hw_data_420xx() 416 hw_data->num_banks = ADF_GEN4_ETR_MAX_BANKS; in adf_init_hw_data_420xx() 417 hw_data->num_banks_per_vf = ADF_GEN4_NUM_BANKS_PER_VF; in adf_init_hw_data_420xx() 418 hw_data->num_rings_per_bank = ADF_GEN4_NUM_RINGS_PER_BANK; in adf_init_hw_data_420xx() 419 hw_data->num_accel = ADF_GEN4_MAX_ACCELERATORS; in adf_init_hw_data_420xx() [all …]
|
| H A D | adf_drv.c | 37 struct adf_hw_device_data *hw_data; in adf_probe() local 72 hw_data = devm_kzalloc(&pdev->dev, sizeof(*hw_data), GFP_KERNEL); in adf_probe() 73 if (!hw_data) { in adf_probe() 78 accel_dev->hw_device = hw_data; in adf_probe() 82 pci_read_config_dword(pdev, ADF_GEN4_FUSECTL4_OFFSET, &hw_data->fuses[ADF_FUSECTL4]); in adf_probe() 85 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 86 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 87 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe() 89 if (!hw_data->accel_mask || !hw_data->ae_mask || in adf_probe() 90 (~hw_data->ae_mask & 0x01)) { in adf_probe() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_dh895xcc/ |
| H A D | adf_dh895xcc_hw_data.c | 214 void adf_init_hw_data_dh895xcc(struct adf_hw_device_data *hw_data) in adf_init_hw_data_dh895xcc() argument 216 hw_data->dev_class = &dh895xcc_class; in adf_init_hw_data_dh895xcc() 217 hw_data->instance_id = dh895xcc_class.instances++; in adf_init_hw_data_dh895xcc() 218 hw_data->num_banks = ADF_DH895XCC_ETR_MAX_BANKS; in adf_init_hw_data_dh895xcc() 219 hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK; in adf_init_hw_data_dh895xcc() 220 hw_data->num_accel = ADF_DH895XCC_MAX_ACCELERATORS; in adf_init_hw_data_dh895xcc() 221 hw_data->num_logical_accel = 1; in adf_init_hw_data_dh895xcc() 222 hw_data->num_engines = ADF_DH895XCC_MAX_ACCELENGINES; in adf_init_hw_data_dh895xcc() 223 hw_data->tx_rx_gap = ADF_GEN2_RX_RINGS_OFFSET; in adf_init_hw_data_dh895xcc() 224 hw_data->tx_rings_mask = ADF_GEN2_TX_RINGS_MASK; in adf_init_hw_data_dh895xcc() [all …]
|
| H A D | adf_drv.c | 60 struct adf_hw_device_data *hw_data; in adf_probe() local 100 hw_data = kzalloc_node(sizeof(*hw_data), GFP_KERNEL, in adf_probe() 102 if (!hw_data) { in adf_probe() 107 accel_dev->hw_device = hw_data; in adf_probe() 111 &hw_data->fuses[ADF_FUSECTL0]); in adf_probe() 114 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 115 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() 116 accel_pci_dev->sku = hw_data->get_sku(hw_data); in adf_probe() 118 if (!hw_data->accel_mask || !hw_data->ae_mask || in adf_probe() 119 ((~hw_data->ae_mask) & 0x01)) { in adf_probe() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_6xxx/ |
| H A D | adf_6xxx_hw_data.c | 444 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in ring_pair_reset() local 448 if (bank_number >= hw_data->num_banks) in ring_pair_reset() 515 struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev); in adf_gen6_init_thd2arb_map() local 516 u32 *thd2arb_map = hw_data->thd_to_arb_map; in adf_gen6_init_thd2arb_map() 519 for (i = 0; i < hw_data->num_engines; i++) { in adf_gen6_init_thd2arb_map() 622 struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev); in adf_gen6_set_vc() local 626 for (i = 0; i < hw_data->num_banks; i++) { in adf_gen6_set_vc() 876 void adf_init_hw_data_6xxx(struct adf_hw_device_data *hw_data) in adf_init_hw_data_6xxx() argument 878 hw_data->dev_class = &adf_6xxx_class; in adf_init_hw_data_6xxx() 879 hw_data->instance_id = adf_6xxx_class.instances++; in adf_init_hw_data_6xxx() [all …]
|
| H A D | adf_drv.c | 59 struct adf_hw_device_data *hw_data; in adf_probe() local 85 hw_data = devm_kzalloc(dev, sizeof(*hw_data), GFP_KERNEL); in adf_probe() 86 if (!hw_data) in adf_probe() 90 pci_read_config_dword(pdev, ADF_GEN6_FUSECTL4_OFFSET, &hw_data->fuses[ADF_FUSECTL4]); in adf_probe() 91 pci_read_config_dword(pdev, ADF_GEN6_FUSECTL0_OFFSET, &hw_data->fuses[ADF_FUSECTL0]); in adf_probe() 92 pci_read_config_dword(pdev, ADF_GEN6_FUSECTL1_OFFSET, &hw_data->fuses[ADF_FUSECTL1]); in adf_probe() 94 if (!(hw_data->fuses[ADF_FUSECTL1] & ICP_ACCEL_GEN6_MASK_WCP_WAT_SLICE)) in adf_probe() 110 accel_dev->hw_device = hw_data; in adf_probe() 118 hw_data->accel_mask = hw_data->get_accel_mask(hw_data); in adf_probe() 119 hw_data->ae_mask = hw_data->get_ae_mask(hw_data); in adf_probe() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | adf_init.c | 67 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_dev_init() local 70 if (!hw_data) { in adf_dev_init() 87 if (hw_data->init_device && hw_data->init_device(accel_dev)) { in adf_dev_init() 92 if (hw_data->init_admin_comms && hw_data->init_admin_comms(accel_dev)) { in adf_dev_init() 97 if (hw_data->init_arb && hw_data->init_arb(accel_dev)) { in adf_dev_init() 102 if (hw_data->get_ring_to_svc_map) in adf_dev_init() 103 hw_data->ring_to_svc_map = hw_data->get_ring_to_svc_map(accel_dev); in adf_dev_init() 119 if (hw_data->alloc_irq(accel_dev)) { in adf_dev_init() 125 if (hw_data->ras_ops.enable_ras_errors) in adf_dev_init() 126 hw_data->ras_ops.enable_ras_errors(accel_dev); in adf_dev_init() [all …]
|
| H A D | adf_hw_arbiter.c | 20 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_init_arb() local 22 unsigned long ae_mask = hw_data->ae_mask; in adf_init_arb() 28 hw_data->get_arb_info(&info); in adf_init_arb() 39 thd_2_arb_cfg = hw_data->get_arb_mapping(accel_dev); in adf_init_arb() 41 for_each_set_bit(i, &ae_mask, hw_data->num_engines) in adf_init_arb() 51 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_update_ring_arb() local 53 u32 tx_ring_mask = hw_data->tx_rings_mask; in adf_update_ring_arb() 54 u32 shift = hw_data->tx_rx_gap; in adf_update_ring_arb() 75 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_exit_arb() local 82 hw_data->get_arb_info(&info); in adf_exit_arb() [all …]
|
| H A D | adf_gen4_hw_data.c | 215 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_gen4_ring_pair_reset() local 219 if (bank_number >= hw_data->num_banks) in adf_gen4_ring_pair_reset() 284 struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev); in adf_gen4_init_thd2arb_map() local 285 u32 *thd2arb_map = hw_data->thd_to_arb_map; in adf_gen4_init_thd2arb_map() 292 if (!hw_data->get_rp_group || !hw_data->get_ena_thd_mask || in adf_gen4_init_thd2arb_map() 293 !hw_data->get_num_aes || !hw_data->uof_get_num_objs || in adf_gen4_init_thd2arb_map() 294 !hw_data->uof_get_ae_mask) in adf_gen4_init_thd2arb_map() 301 ae_cnt = hw_data->get_num_aes(hw_data); in adf_gen4_init_thd2arb_map() 302 worker_obj_cnt = hw_data->uof_get_num_objs(accel_dev) - in adf_gen4_init_thd2arb_map() 315 ae_mask = hw_data->uof_get_ae_mask(accel_dev, i); in adf_gen4_init_thd2arb_map() [all …]
|
| H A D | adf_isr.c | 24 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_enable_msix() local 25 u32 msix_num_entries = hw_data->num_banks + 1; in adf_enable_msix() 28 if (hw_data->set_msix_rttable) in adf_enable_msix() 29 hw_data->set_msix_rttable(accel_dev); in adf_enable_msix() 126 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_handle_pm_int() local 128 if (hw_data->handle_pm_interrupt && in adf_handle_pm_int() 129 hw_data->handle_pm_interrupt(accel_dev)) in adf_handle_pm_int() 180 struct adf_hw_device_data *hw_data = accel_dev->hw_device; in adf_free_irqs() local 183 int clust_irq = hw_data->num_banks; in adf_free_irqs() 187 for (i = 0; i < hw_data->num_banks; i++) { in adf_free_irqs() [all …]
|
| /linux/drivers/video/backlight/ |
| H A D | apple_bl.c | 31 struct hw_data { struct 40 static const struct hw_data *hw_data; argument 81 static const struct hw_data intel_chipset_data = { 126 static const struct hw_data nvidia_chipset_data = { 151 hw_data = &intel_chipset_data; in apple_bl_add() 153 hw_data = &nvidia_chipset_data; in apple_bl_add() 157 if (!hw_data) { in apple_bl_add() 164 intensity = hw_data->backlight_ops.get_brightness(NULL); in apple_bl_add() 167 hw_data->set_brightness(1); in apple_bl_add() 168 if (!hw_data->backlight_ops.get_brightness(NULL)) in apple_bl_add() [all …]
|
| /linux/drivers/clk/ |
| H A D | clk-lan966x.c | 226 struct clk_hw_onecell_data *hw_data, in lan966x_gate_clk_register() argument 235 hw_data->hws[i] = in lan966x_gate_clk_register() 241 if (IS_ERR(hw_data->hws[i])) in lan966x_gate_clk_register() 242 return dev_err_probe(dev, PTR_ERR(hw_data->hws[i]), in lan966x_gate_clk_register() 253 struct clk_hw_onecell_data *hw_data; in lan966x_clk_probe() local 263 hw_data = devm_kzalloc(dev, struct_size(hw_data, hws, data->num_total_clks), in lan966x_clk_probe() 265 if (!hw_data) in lan966x_clk_probe() 274 hw_data->num = data->num_generic_clks; in lan966x_clk_probe() 278 hw_data->hws[i] = lan966x_gck_clk_register(dev, i); in lan966x_clk_probe() 279 if (IS_ERR(hw_data->hws[i])) { in lan966x_clk_probe() [all …]
|
| /linux/net/core/ |
| H A D | drop_monitor.c | 304 net_dm_hw_reset_per_cpu_data(struct per_cpu_dm_data *hw_data) in net_dm_hw_reset_per_cpu_data() argument 315 mod_timer(&hw_data->send_timer, jiffies + HZ / 10); in net_dm_hw_reset_per_cpu_data() 318 raw_spin_lock_irqsave(&hw_data->lock, flags); in net_dm_hw_reset_per_cpu_data() 319 swap(hw_data->hw_entries, hw_entries); in net_dm_hw_reset_per_cpu_data() 320 raw_spin_unlock_irqrestore(&hw_data->lock, flags); in net_dm_hw_reset_per_cpu_data() 411 struct per_cpu_dm_data *hw_data; in net_dm_hw_summary_work() local 415 hw_data = container_of(work, struct per_cpu_dm_data, dm_alert_work); in net_dm_hw_summary_work() 417 hw_entries = net_dm_hw_reset_per_cpu_data(hw_data); in net_dm_hw_summary_work() 444 struct per_cpu_dm_data *hw_data; in net_dm_hw_trap_summary_probe() local 451 hw_data = this_cpu_ptr(&dm_hw_cpu_data); in net_dm_hw_trap_summary_probe() [all …]
|