| /linux/drivers/dma/idxd/ |
| H A D | init.c | 187 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 191 idxd->wq_enable_map = bitmap_zalloc_node(idxd->max_wqs, GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 198 wq = kzalloc_node(sizeof(*wq), GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 227 wq->wqcfg = kzalloc_node(idxd->wqcfg_size, GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 294 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_engines() 299 engine = kzalloc_node(sizeof(*engine), GFP_KERNEL, dev_to_node(dev)); in idxd_setup_engines() 358 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_groups() 363 group = kzalloc_node(sizeof(*group), GFP_KERNEL, dev_to_node(dev)); in idxd_setup_groups() 430 evl = kzalloc_node(sizeof(*evl), GFP_KERNEL, dev_to_node(dev)); in idxd_init_evl() 622 idxd = kzalloc_node(sizeof(*idxd), GFP_KERNEL, dev_to_node(dev)); in idxd_alloc() [all …]
|
| H A D | dma.c | 220 idxd_dma = kzalloc_node(sizeof(*idxd_dma), GFP_KERNEL, dev_to_node(dev)); in idxd_register_dma_device() 275 idxd_chan = kzalloc_node(sizeof(*idxd_chan), GFP_KERNEL, dev_to_node(dev)); in idxd_register_dma_channel()
|
| /linux/lib/ |
| H A D | devres.c | 36 dev_to_node(dev)); in __devm_ioremap() 286 dev_to_node(dev)); in devm_ioport_map() 337 dev_to_node(dev)); in devm_arch_phys_wc_add() 383 dev_to_node(dev)); in devm_arch_io_reserve_memtype_wc()
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | qat_compression.c | 62 tmp_dev_node = dev_to_node(&GET_DEV(tmp_dev)); in qat_compression_get_instance_node() 137 dev_to_node(&GET_DEV(accel_dev))); in qat_compression_create_instances() 199 dc_data = kzalloc_node(sizeof(*dc_data), GFP_KERNEL, dev_to_node(dev)); in qat_compression_alloc_dc_data() 203 obuff = kzalloc_node(ovf_buff_sz, GFP_KERNEL, dev_to_node(dev)); in qat_compression_alloc_dc_data()
|
| H A D | qat_crypto.c | 60 if ((node == dev_to_node(&GET_DEV(tmp_dev)) || in qat_crypto_get_instance_node() 61 dev_to_node(&GET_DEV(tmp_dev)) < 0) && in qat_crypto_get_instance_node() 153 dev_to_node(&GET_DEV(accel_dev))); in qat_crypto_create_instances()
|
| H A D | adf_transport.c | 407 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank() 430 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank() 483 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data() 490 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data()
|
| /linux/include/linux/ |
| H A D | async.h | 90 return async_schedule_node(func, dev, dev_to_node(dev)); in async_schedule_dev() 114 return async_schedule_node_domain(func, dev, dev_to_node(dev), domain); in async_schedule_dev_domain()
|
| /linux/drivers/crypto/intel/qat/qat_c62x/ |
| H A D | adf_drv.c | 73 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe() 82 dev_to_node(&pdev->dev)); in adf_probe() 101 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_dh895xcc/ |
| H A D | adf_drv.c | 73 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe() 82 dev_to_node(&pdev->dev)); in adf_probe() 101 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_c3xxx/ |
| H A D | adf_drv.c | 73 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe() 82 dev_to_node(&pdev->dev)); in adf_probe() 101 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/usb/host/ |
| H A D | xhci-mem.c | 38 seg = kzalloc_node(sizeof(*seg), flags, dev_to_node(dev)); in xhci_segment_alloc() 50 dev_to_node(dev)); in xhci_segment_alloc() 377 ring = kzalloc_node(sizeof(*ring), flags, dev_to_node(dev)); in xhci_ring_alloc() 461 ctx = kzalloc_node(sizeof(*ctx), flags, dev_to_node(dev)); in xhci_alloc_container_ctx() 493 ctx = kzalloc_node(sizeof(*ctx), flags, dev_to_node(dev)); in xhci_alloc_port_bw_ctx() 631 dev_to_node(dev)); in xhci_alloc_stream_info() 641 dev_to_node(dev)); in xhci_alloc_stream_info() 842 dev_to_node(dev)); in xhci_alloc_tt_info() 1656 dev_to_node(dev)); in scratchpad_alloc() 1667 flags, dev_to_node(dev)); in scratchpad_alloc() [all …]
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_mgr.c | 158 if ((node == dev_to_node(&tmp_dev->vdev->dev) || in virtcrypto_get_dev_node() 159 dev_to_node(&tmp_dev->vdev->dev) < 0) && in virtcrypto_get_dev_node()
|
| H A D | virtio_crypto_skcipher_algs.c | 355 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req() 360 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req() 417 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req()
|
| /linux/net/core/ |
| H A D | devmem.c | 208 dev_to_node(&dev->dev)); in net_devmem_bind_dmabuf() 257 dev_to_node(&dev->dev)); in net_devmem_bind_dmabuf() 271 dev_to_node(&dev->dev)); in net_devmem_bind_dmabuf() 283 dma_addr, len, dev_to_node(&dev->dev), in net_devmem_bind_dmabuf()
|
| /linux/drivers/crypto/intel/qat/qat_c3xxxvf/ |
| H A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_c62xvf/ |
| H A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_dh895xccvf/ |
| H A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/arch/riscv/include/asm/ |
| H A D | pci.h | 21 return dev_to_node(&bus->dev); in pcibus_to_node()
|
| /linux/drivers/base/test/ |
| H A D | test_async_driver_probe.c | 48 dev_to_node(dev) != numa_node_id()) { in test_probe() 50 dev_to_node(dev), numa_node_id()); in test_probe()
|
| /linux/arch/arm64/kernel/ |
| H A D | pci.c | 38 return dev_to_node(&bus->dev); in pcibus_to_node()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_managed.c | 147 dev_to_node(dev->dev)); in __drmm_add_action() 233 dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev)); in drmm_kmalloc()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | ptp.c | 355 node = dev_to_node(mlx5_core_dma_dev(mdev)); in mlx5e_ptp_alloc_txqsq() 490 err = mlx5e_ptp_alloc_traffic_db(ptpsq, dev_to_node(mlx5_core_dma_dev(c->mdev))); in mlx5e_ptp_open_txqsq() 572 ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev)); in mlx5e_ptp_open_tx_cqs() 622 ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev)); in mlx5e_ptp_open_rx_cq() 727 int node = dev_to_node(c->mdev->device); in mlx5e_ptp_open_rq() 885 c = kvzalloc_node(sizeof(*c), GFP_KERNEL, dev_to_node(mlx5_core_dma_dev(mdev))); in mlx5e_ptp_open()
|
| /linux/drivers/perf/ |
| H A D | fujitsu_uncore_pmu.c | 417 node = dev_to_node(uncorepmu->dev); in fujitsu_uncore_pmu_online_cpu() 434 node = dev_to_node(uncorepmu->dev); in fujitsu_uncore_pmu_offline_cpu() 466 uncorepmu->cpu = cpumask_local_spread(0, dev_to_node(dev)); in fujitsu_uncore_pmu_probe()
|
| /linux/kernel/ |
| H A D | iomem.c | 143 dev_to_node(dev)); in devm_memremap()
|
| /linux/drivers/nvdimm/ |
| H A D | of_pmem.c | 63 ndr_desc.numa_node = dev_to_node(&pdev->dev); in of_pmem_region_probe()
|