| /linux/drivers/vdpa/vdpa_user/ |
| H A D | iova_domain.c | 20 static int vduse_iotlb_add_range(struct vduse_iova_domain *domain, in vduse_iotlb_add_range() argument 35 ret = vhost_iotlb_add_range_ctx(domain->iotlb, start, last, in vduse_iotlb_add_range() 45 static void vduse_iotlb_del_range(struct vduse_iova_domain *domain, in vduse_iotlb_del_range() argument 51 while ((map = vhost_iotlb_itree_first(domain->iotlb, start, last))) { in vduse_iotlb_del_range() 55 vhost_iotlb_map_free(domain->iotlb, map); in vduse_iotlb_del_range() 59 int vduse_domain_set_map(struct vduse_iova_domain *domain, in vduse_domain_set_map() argument 67 spin_lock(&domain->iotlb_lock); in vduse_domain_set_map() 68 vduse_iotlb_del_range(domain, start, last); in vduse_domain_set_map() 73 ret = vduse_iotlb_add_range(domain, map->start, map->last, in vduse_domain_set_map() 80 spin_unlock(&domain in vduse_domain_set_map() 89 vduse_domain_clear_map(struct vduse_iova_domain * domain,struct vhost_iotlb * iotlb) vduse_domain_clear_map() argument 103 vduse_domain_map_bounce_page(struct vduse_iova_domain * domain,u64 iova,u64 size,u64 paddr) vduse_domain_map_bounce_page() argument 123 vduse_domain_unmap_bounce_page(struct vduse_iova_domain * domain,u64 iova,u64 size) vduse_domain_unmap_bounce_page() argument 160 vduse_domain_bounce(struct vduse_iova_domain * domain,dma_addr_t iova,size_t size,enum dma_data_direction dir) vduse_domain_bounce() argument 194 vduse_domain_get_coherent_page(struct vduse_iova_domain * domain,u64 iova) vduse_domain_get_coherent_page() argument 215 vduse_domain_get_bounce_page(struct vduse_iova_domain * domain,u64 iova) vduse_domain_get_bounce_page() argument 234 vduse_domain_free_kernel_bounce_pages(struct vduse_iova_domain * domain) vduse_domain_free_kernel_bounce_pages() argument 254 vduse_domain_add_user_bounce_pages(struct vduse_iova_domain * domain,struct page ** pages,int count) vduse_domain_add_user_bounce_pages() argument 289 vduse_domain_remove_user_bounce_pages(struct vduse_iova_domain * domain) vduse_domain_remove_user_bounce_pages() argument 320 vduse_domain_reset_bounce_map(struct vduse_iova_domain * domain) vduse_domain_reset_bounce_map() argument 335 vduse_domain_init_bounce_map(struct vduse_iova_domain * domain) vduse_domain_init_bounce_map() argument 379 vduse_domain_sync_single_for_device(struct vduse_iova_domain * domain,dma_addr_t dma_addr,size_t size,enum dma_data_direction dir) vduse_domain_sync_single_for_device() argument 389 vduse_domain_sync_single_for_cpu(struct vduse_iova_domain * domain,dma_addr_t dma_addr,size_t size,enum dma_data_direction dir) vduse_domain_sync_single_for_cpu() argument 399 vduse_domain_map_page(struct vduse_iova_domain * domain,struct page * page,unsigned long offset,size_t size,enum dma_data_direction dir,unsigned long attrs) vduse_domain_map_page() argument 433 vduse_domain_unmap_page(struct vduse_iova_domain * domain,dma_addr_t dma_addr,size_t size,enum dma_data_direction dir,unsigned long attrs) vduse_domain_unmap_page() argument 448 vduse_domain_alloc_coherent(struct vduse_iova_domain * domain,size_t size,dma_addr_t * dma_addr,gfp_t flag,unsigned long attrs) vduse_domain_alloc_coherent() argument 482 vduse_domain_free_coherent(struct vduse_iova_domain * domain,size_t size,void * vaddr,dma_addr_t dma_addr,unsigned long attrs) vduse_domain_free_coherent() argument 511 struct vduse_iova_domain *domain = vmf->vma->vm_private_data; vduse_domain_mmap_fault() local 537 struct vduse_iova_domain *domain = file->private_data; vduse_domain_mmap() local 548 struct vduse_iova_domain *domain = file->private_data; vduse_domain_release() local 570 vduse_domain_destroy(struct vduse_iova_domain * domain) vduse_domain_destroy() argument 578 struct vduse_iova_domain *domain; vduse_domain_create() local [all...] |
| /linux/drivers/scsi/elx/libefc/ |
| H A D | efc_domain.c | 17 struct efc_domain *domain = NULL; in efc_domain_cb() local 22 domain = data; in efc_domain_cb() 24 /* Accept domain callback events from the user driver */ in efc_domain_cb() 36 /* lookup domain, or allocate a new one */ in efc_domain_cb() 37 domain = efc->domain; in efc_domain_cb() 38 if (!domain) { in efc_domain_cb() 39 domain = efc_domain_alloc(efc, fcf_wwn); in efc_domain_cb() 40 if (!domain) { in efc_domain_cb() 45 efc_sm_transition(&domain in efc_domain_cb() 106 struct efc_domain *domain = container_of(arg, struct efc_domain, ref); _efc_domain_free() local 116 efc_domain_free(struct efc_domain * domain) efc_domain_free() argument 135 struct efc_domain *domain; efc_domain_alloc() local 174 struct efc_domain *domain = ctx->app; __efc_domain_common() local 196 struct efc_domain *domain = ctx->app; __efc_domain_common_shutdown() local 824 __efc_domain_attach_internal(struct efc_domain * domain,u32 s_id) __efc_domain_attach_internal() argument 834 efc_domain_attach(struct efc_domain * domain,u32 s_id) efc_domain_attach() argument 840 efc_domain_post_event(struct efc_domain * domain,enum efc_sm_event event,void * arg) efc_domain_post_event() argument 858 efct_domain_process_pending(struct efc_domain * domain) efct_domain_process_pending() argument 907 struct efc_domain *domain = efc->domain; efc_dispatch_frame() local 940 struct efc_domain *domain = (struct efc_domain *)arg; efc_domain_dispatch_frame() local [all...] |
| H A D | efc_cmds.c | 132 nport->indicator, nport->domain->indicator); in efc_nport_alloc_init_vpi() 207 struct efc_domain *domain, u8 *wwpn) in efc_cmd_nport_alloc() argument 227 if (domain) { in efc_cmd_nport_alloc() 237 /* domain NULL and wwpn non-NULL */ in efc_cmd_nport_alloc() 278 nport->domain->indicator, false); in efc_cmd_nport_attach() 317 efc_domain_get_mbox_status(struct efc_domain *domain, u8 *mqe, int status) in efc_domain_get_mbox_status() argument 319 struct efc *efc = domain->efc; in efc_domain_get_mbox_status() 325 domain->indicator, status, in efc_domain_get_mbox_status() 334 efc_domain_free_resources(struct efc_domain *domain, int evt, void *data) in efc_domain_free_resources() argument 336 struct efc *efc = domain in efc_domain_free_resources() 353 efc_domain_send_nport_evt(struct efc_domain * domain,int port_evt,int domain_evt,void * data) efc_domain_send_nport_evt() argument 369 struct efc_domain *domain = arg; efc_domain_alloc_read_sparm64_cb() local 383 efc_domain_alloc_read_sparm64(struct efc_domain * domain) efc_domain_alloc_read_sparm64() argument 410 struct efc_domain *domain = arg; efc_domain_alloc_init_vfi_cb() local 423 efc_domain_alloc_init_vfi(struct efc_domain * domain) efc_domain_alloc_init_vfi() argument 455 efc_cmd_domain_alloc(struct efc * efc,struct efc_domain * domain,u32 fcf) efc_cmd_domain_alloc() argument 499 struct efc_domain *domain = arg; efc_domain_attach_reg_vfi_cb() local 513 efc_cmd_domain_attach(struct efc * efc,struct efc_domain * domain,u32 fc_id) efc_cmd_domain_attach() argument 552 struct efc_domain *domain = arg; efc_domain_free_unreg_vfi_cb() local 567 efc_domain_free_unreg_vfi(struct efc_domain * domain) efc_domain_free_unreg_vfi() argument 594 efc_cmd_domain_free(struct efc * efc,struct efc_domain * domain) efc_cmd_domain_free() argument [all...] |
| /linux/drivers/pmdomain/imx/ |
| H A D | imx93-pd.c | 37 struct imx93_power_domain *domain = to_imx93_pd(genpd); in imx93_pd_on() local 38 void __iomem *addr = domain->addr; in imx93_pd_on() 42 ret = clk_bulk_prepare_enable(domain->num_clks, domain->clks); in imx93_pd_on() 44 dev_err(domain->dev, "failed to enable clocks for domain: %s\n", genpd->name); in imx93_pd_on() 55 dev_err(domain->dev, "pd_on timeout: name: %s, stat: %x\n", genpd->name, val); in imx93_pd_on() 64 struct imx93_power_domain *domain = to_imx93_pd(genpd); in imx93_pd_off() local 65 void __iomem *addr = domain->addr; in imx93_pd_off() 77 dev_err(domain in imx93_pd_off() 88 struct imx93_power_domain *domain = platform_get_drvdata(pdev); imx93_pd_remove() local 100 struct imx93_power_domain *domain; imx93_pd_probe() local [all...] |
| H A D | gpcv2.c | 318 struct imx_pgc_domain *domain = to_imx_pgc_domain(genpd); in imx_pgc_power_up() local 322 ret = pm_runtime_get_sync(domain->dev); in imx_pgc_power_up() 324 pm_runtime_put_noidle(domain->dev); in imx_pgc_power_up() 328 if (!IS_ERR(domain->regulator)) { in imx_pgc_power_up() 329 ret = regulator_enable(domain->regulator); in imx_pgc_power_up() 331 dev_err(domain->dev, in imx_pgc_power_up() 338 reset_control_assert(domain->reset); in imx_pgc_power_up() 340 /* Enable reset clocks for all devices in the domain */ in imx_pgc_power_up() 341 ret = clk_bulk_prepare_enable(domain->num_clks, domain in imx_pgc_power_up() 428 struct imx_pgc_domain *domain = to_imx_pgc_domain(genpd); imx_pgc_power_down() local 1326 struct imx_pgc_domain *domain = pdev->dev.platform_data; imx_pgc_domain_probe() local 1389 struct imx_pgc_domain *domain = pdev->dev.platform_data; imx_pgc_domain_remove() local 1486 struct imx_pgc_domain *domain; builtin_platform_driver() local [all...] |
| H A D | gpc.c | 71 /* Gate off domain when powered down */ in imx6_pm_domain_power_off() 75 /* Request GPC to power down domain */ in imx6_pm_domain_power_off() 103 /* Enable reset clocks for all devices in the domain */ in imx6_pm_domain_power_on() 107 /* Gate off domain when powered down */ in imx6_pm_domain_power_on() 111 /* Request GPC to power up domain */ in imx6_pm_domain_power_on() 119 pr_err("powerup request on domain %s timed out\n", genpd->name); in imx6_pm_domain_power_on() 124 /* Disable reset clocks for all devices in the domain */ in imx6_pm_domain_power_on() 131 static int imx_pgc_get_clocks(struct device *dev, struct imx_pm_domain *domain) in imx_pgc_get_clocks() argument 144 domain->clk[i] = clk; in imx_pgc_get_clocks() 146 domain in imx_pgc_get_clocks() 157 imx_pgc_put_clocks(struct imx_pm_domain * domain) imx_pgc_put_clocks() argument 165 imx_pgc_parse_dt(struct device * dev,struct imx_pm_domain * domain) imx_pgc_parse_dt() argument 182 struct imx_pm_domain *domain = pdev->dev.platform_data; imx_pgc_power_domain_probe() local 217 struct imx_pm_domain *domain = pdev->dev.platform_data; imx_pgc_power_domain_remove() local 362 struct imx_pm_domain *domain; imx_gpc_old_dt_init() local 456 struct imx_pm_domain *domain; imx_gpc_probe() local [all...] |
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_force_wake.c | 34 struct xe_force_wake_domain *domain = &fw->domains[id]; in init_domain() local 36 domain->id = id; in init_domain() 37 domain->reg_ctl = reg; in init_domain() 38 domain->reg_ack = ack; in init_domain() 39 domain->val = FORCEWAKE_MT(FORCEWAKE_KERNEL); in init_domain() 40 domain->mask = FORCEWAKE_MT_MASK(FORCEWAKE_KERNEL); in init_domain() 96 static void __domain_ctl(struct xe_gt *gt, struct xe_force_wake_domain *domain, bool wake) in __domain_ctl() argument 101 xe_mmio_write32(>->mmio, domain->reg_ctl, domain->mask | (wake ? domain in __domain_ctl() 104 __domain_wait(struct xe_gt * gt,struct xe_force_wake_domain * domain,bool wake) __domain_wait() argument 129 domain_wake(struct xe_gt * gt,struct xe_force_wake_domain * domain) domain_wake() argument 135 domain_wake_wait(struct xe_gt * gt,struct xe_force_wake_domain * domain) domain_wake_wait() argument 140 domain_sleep(struct xe_gt * gt,struct xe_force_wake_domain * domain) domain_sleep() argument 146 domain_sleep_wait(struct xe_gt * gt,struct xe_force_wake_domain * domain) domain_sleep_wait() argument 176 struct xe_force_wake_domain *domain; xe_force_wake_get() local 227 struct xe_force_wake_domain *domain; xe_force_wake_put() local [all...] |
| /linux/drivers/iommu/ |
| H A D | ipmmu-vmsa.c | 211 static u32 ipmmu_ctx_read_root(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_read_root() argument 214 return ipmmu_ctx_read(domain->mmu->root, domain->context_id, reg); in ipmmu_ctx_read_root() 217 static void ipmmu_ctx_write_root(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_write_root() argument 220 ipmmu_ctx_write(domain->mmu->root, domain->context_id, reg, data); in ipmmu_ctx_write_root() 223 static void ipmmu_ctx_write_all(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_write_all() argument 226 if (domain->mmu != domain->mmu->root) in ipmmu_ctx_write_all() 227 ipmmu_ctx_write(domain in ipmmu_ctx_write_all() 254 ipmmu_tlb_sync(struct ipmmu_vmsa_domain * domain) ipmmu_tlb_sync() argument 265 ipmmu_tlb_invalidate(struct ipmmu_vmsa_domain * domain) ipmmu_tlb_invalidate() argument 279 ipmmu_utlb_enable(struct ipmmu_vmsa_domain * domain,unsigned int utlb) ipmmu_utlb_enable() argument 300 ipmmu_utlb_disable(struct ipmmu_vmsa_domain * domain,unsigned int utlb) ipmmu_utlb_disable() argument 311 struct ipmmu_vmsa_domain *domain = cookie; ipmmu_tlb_flush_all() local 332 ipmmu_domain_allocate_context(struct ipmmu_vmsa_device * mmu,struct ipmmu_vmsa_domain * domain) ipmmu_domain_allocate_context() argument 364 ipmmu_domain_setup_context(struct ipmmu_vmsa_domain * domain) ipmmu_domain_setup_context() argument 417 ipmmu_domain_init_context(struct ipmmu_vmsa_domain * domain) ipmmu_domain_init_context() argument 467 ipmmu_domain_destroy_context(struct ipmmu_vmsa_domain * domain) ipmmu_domain_destroy_context() argument 487 ipmmu_domain_irq(struct ipmmu_vmsa_domain * domain) ipmmu_domain_irq() argument 567 struct ipmmu_vmsa_domain *domain; ipmmu_domain_alloc_paging() local 581 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); ipmmu_domain_free() local 597 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); ipmmu_attach_device() local 644 struct ipmmu_vmsa_domain *domain; ipmmu_iommu_identity_attach() local 673 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); ipmmu_map() local 683 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); ipmmu_unmap() local 690 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); ipmmu_flush_iotlb_all() local 705 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); ipmmu_iova_to_phys() local [all...] |
| H A D | iommu-sva.c | 73 struct iommu_domain *domain; in iommu_sva_bind_device() local 93 if (attach_handle->domain->mm != mm) { in iommu_sva_bind_device() 113 /* Search for an existing domain. */ in iommu_sva_bind_device() 114 list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) { in iommu_sva_bind_device() 115 ret = iommu_attach_device_pasid(domain, dev, iommu_mm->pasid, in iommu_sva_bind_device() 118 domain->users++; in iommu_sva_bind_device() 123 /* Allocate a new domain and set it on device pasid. */ in iommu_sva_bind_device() 124 domain = iommu_sva_domain_alloc(dev, mm); in iommu_sva_bind_device() 125 if (IS_ERR(domain)) { in iommu_sva_bind_device() 126 ret = PTR_ERR(domain); in iommu_sva_bind_device() 163 struct iommu_domain *domain = handle->handle.domain; iommu_sva_unbind_device() local 185 struct iommu_domain *domain = handle->handle.domain; iommu_sva_get_pasid() local 297 struct iommu_domain *domain; iommu_sva_domain_alloc() local [all...] |
| /linux/drivers/iommu/intel/ |
| H A D | cache.c | 42 /* Assign a cache tag with specified type to domain. */ 43 int cache_tag_assign(struct dmar_domain *domain, u16 did, struct device *dev, in cache_tag_assign() argument 67 spin_lock_irqsave(&domain->cache_lock, flags); in cache_tag_assign() 68 prev = &domain->cache_tags; in cache_tag_assign() 69 list_for_each_entry(temp, &domain->cache_tags, node) { in cache_tag_assign() 72 spin_unlock_irqrestore(&domain->cache_lock, flags); in cache_tag_assign() 86 spin_unlock_irqrestore(&domain->cache_lock, flags); in cache_tag_assign() 92 /* Unassign a cache tag with specified type from domain. */ 93 static void cache_tag_unassign(struct dmar_domain *domain, u16 did, in cache_tag_unassign() argument 102 spin_lock_irqsave(&domain in cache_tag_unassign() 117 domain_qi_batch_alloc(struct dmar_domain * domain) domain_qi_batch_alloc() argument 135 __cache_tag_assign_domain(struct dmar_domain * domain,u16 did,struct device * dev,ioasid_t pasid) __cache_tag_assign_domain() argument 156 __cache_tag_unassign_domain(struct dmar_domain * domain,u16 did,struct device * dev,ioasid_t pasid) __cache_tag_unassign_domain() argument 167 __cache_tag_assign_parent_domain(struct dmar_domain * domain,u16 did,struct device * dev,ioasid_t pasid) __cache_tag_assign_parent_domain() argument 188 __cache_tag_unassign_parent_domain(struct dmar_domain * domain,u16 did,struct device * dev,ioasid_t pasid) __cache_tag_unassign_parent_domain() argument 199 domain_get_id_for_dev(struct dmar_domain * domain,struct device * dev) domain_get_id_for_dev() argument 222 cache_tag_assign_domain(struct dmar_domain * domain,struct device * dev,ioasid_t pasid) cache_tag_assign_domain() argument 246 cache_tag_unassign_domain(struct dmar_domain * domain,struct device * dev,ioasid_t pasid) cache_tag_unassign_domain() argument 366 cache_tag_flush_iotlb(struct dmar_domain * domain,struct cache_tag * tag,unsigned long addr,unsigned long pages,unsigned long mask,int ih) cache_tag_flush_iotlb() argument 398 cache_tag_flush_devtlb_psi(struct dmar_domain * domain,struct cache_tag * tag,unsigned long addr,unsigned long mask) cache_tag_flush_devtlb_psi() argument 429 cache_tag_flush_range(struct dmar_domain * domain,unsigned long start,unsigned long end,int ih) cache_tag_flush_range() argument 482 cache_tag_flush_all(struct dmar_domain * domain) cache_tag_flush_all() argument 498 cache_tag_flush_range_np(struct dmar_domain * domain,unsigned long start,unsigned long end) cache_tag_flush_range_np() argument [all...] |
| H A D | svm.c | 55 struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier); in intel_arch_invalidate_secondary_tlbs() local 58 cache_tag_flush_all(domain); in intel_arch_invalidate_secondary_tlbs() 67 cache_tag_flush_range(domain, start, end - 1, 0); in intel_arch_invalidate_secondary_tlbs() 72 struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier); in intel_mm_release() local 89 spin_lock_irqsave(&domain->lock, flags); in intel_mm_release() 90 list_for_each_entry(dev_pasid, &domain->dev_pasids, link_domain) { in intel_mm_release() 95 spin_unlock_irqrestore(&domain->lock, flags); in intel_mm_release() 101 struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier); in intel_mm_free_notifier() local 103 kfree(domain->qi_batch); in intel_mm_free_notifier() 104 kfree(domain); in intel_mm_free_notifier() 148 intel_svm_set_dev_pasid(struct iommu_domain * domain,struct device * dev,ioasid_t pasid,struct iommu_domain * old) intel_svm_set_dev_pasid() argument 189 intel_svm_domain_free(struct iommu_domain * domain) intel_svm_domain_free() argument 205 struct dmar_domain *domain; intel_svm_domain_alloc() local [all...] |
| H A D | nested.c | 21 static int intel_nested_attach_dev(struct iommu_domain *domain, in intel_nested_attach_dev() argument 25 struct dmar_domain *dmar_domain = to_dmar_domain(domain); in intel_nested_attach_dev() 38 * Stage-1 domain cannot work alone, it is nested on a s2_domain. in intel_nested_attach_dev() 42 ret = paging_domain_compatible(&dmar_domain->s2_domain->domain, dev); in intel_nested_attach_dev() 44 dev_err_ratelimited(dev, "s2 domain is not compatible\n"); in intel_nested_attach_dev() 50 dev_err_ratelimited(dev, "Failed to attach domain to iommu\n"); in intel_nested_attach_dev() 58 ret = iopf_for_domain_set(domain, dev); in intel_nested_attach_dev() 67 info->domain = dmar_domain; in intel_nested_attach_dev() 75 iopf_for_domain_remove(domain, dev); in intel_nested_attach_dev() 84 static void intel_nested_domain_free(struct iommu_domain *domain) in intel_nested_domain_free() argument 96 intel_nested_cache_invalidate_user(struct iommu_domain * domain,struct iommu_user_data_array * array) intel_nested_cache_invalidate_user() argument 140 domain_setup_nested(struct intel_iommu * iommu,struct dmar_domain * domain,struct device * dev,ioasid_t pasid,struct iommu_domain * old) domain_setup_nested() argument 151 intel_nested_set_dev_pasid(struct iommu_domain * domain,struct device * dev,ioasid_t pasid,struct iommu_domain * old) intel_nested_set_dev_pasid() argument 210 struct dmar_domain *domain; intel_iommu_domain_alloc_nested() local [all...] |
| /linux/arch/arm/boot/dts/ti/keystone/ |
| H A D | keystone-k2hk-clocks.dtsi | 59 reg-names = "control", "domain"; 60 domain-id = <0>; 69 reg-names = "control", "domain"; 70 domain-id = <4>; 79 reg-names = "control", "domain"; 80 domain-id = <5>; 89 reg-names = "control", "domain"; 90 domain-id = <9>; 99 reg-names = "control", "domain"; 100 domain [all...] |
| H A D | keystone-k2l-clocks.dtsi | 49 reg-names = "control", "domain"; 51 domain-id = <0>; 60 reg-names = "control", "domain"; 61 domain-id = <4>; 70 reg-names = "control", "domain"; 71 domain-id = <9>; 80 reg-names = "control", "domain"; 81 domain-id = <10>; 90 reg-names = "control", "domain"; 91 domain [all...] |
| /linux/drivers/dca/ |
| H A D | dca-core.c | 45 struct dca_domain *domain; in dca_allocate_domain() local 47 domain = kzalloc(sizeof(*domain), GFP_NOWAIT); in dca_allocate_domain() 48 if (!domain) in dca_allocate_domain() 51 INIT_LIST_HEAD(&domain->dca_providers); in dca_allocate_domain() 52 domain->pci_rc = rc; in dca_allocate_domain() 54 return domain; in dca_allocate_domain() 57 static void dca_free_domain(struct dca_domain *domain) in dca_free_domain() argument 59 list_del(&domain->node); in dca_free_domain() 60 kfree(domain); in dca_free_domain() 82 struct dca_domain *domain; unregister_dca_providers() local 115 struct dca_domain *domain; dca_find_domain() local 127 struct dca_domain *domain; dca_get_domain() local 144 struct dca_domain *domain; dca_find_provider_by_dev() local 177 struct dca_domain *domain; dca_add_requester() local 343 struct dca_domain *domain, *newdomain = NULL; register_dca_provider() local 400 struct dca_domain *domain; unregister_dca_provider() local [all...] |
| /linux/include/linux/ |
| H A D | irqdomain.h | 5 * See Documentation/core-api/irq/irq-domain.rst for the details. 53 * @match: Match an interrupt controller device node to a domain, returns 113 * @name: Name of interrupt domain 119 * @mutex: Domain lock, hierarchical domains use root domain's lock 120 * @root: Pointer to root domain, or containing structure if non-hierarchical 135 * purposes related to the irq domain. 137 * @msi_parent_ops: Pointer to MSI parent domain methods for per device domain init 138 * @exit: Function called when the domain is destroyed 140 * Revmap data, used internally by the irq domain cod 455 irq_create_mapping(struct irq_domain * domain,irq_hw_number_t hwirq) irq_create_mapping() argument 471 irq_resolve_mapping(struct irq_domain * domain,irq_hw_number_t hwirq) irq_resolve_mapping() argument 484 irq_find_mapping(struct irq_domain * domain,irq_hw_number_t hwirq) irq_find_mapping() argument 581 irq_domain_alloc_irqs(struct irq_domain * domain,unsigned int nr_irqs,int node,void * arg) irq_domain_alloc_irqs() argument 605 irq_domain_is_hierarchy(struct irq_domain * domain) irq_domain_is_hierarchy() argument 610 irq_domain_is_ipi(struct irq_domain * domain) irq_domain_is_ipi() argument 615 irq_domain_is_ipi_per_cpu(struct irq_domain * domain) irq_domain_is_ipi_per_cpu() argument 620 irq_domain_is_ipi_single(struct irq_domain * domain) irq_domain_is_ipi_single() argument 625 irq_domain_is_msi(struct irq_domain * domain) irq_domain_is_msi() argument 630 irq_domain_is_msi_parent(struct irq_domain * domain) irq_domain_is_msi_parent() argument 635 irq_domain_is_msi_device(struct irq_domain * domain) irq_domain_is_msi_device() argument 640 irq_domain_is_msi_immutable(struct irq_domain * domain) irq_domain_is_msi_immutable() argument 645 irq_domain_alloc_irqs(struct irq_domain * domain,unsigned int nr_irqs,int node,void * arg) irq_domain_alloc_irqs() argument 653 irq_domain_is_hierarchy(struct irq_domain * domain) irq_domain_is_hierarchy() argument 658 irq_domain_is_ipi(struct irq_domain * domain) irq_domain_is_ipi() argument 663 irq_domain_is_ipi_per_cpu(struct irq_domain * domain) irq_domain_is_ipi_per_cpu() argument 668 irq_domain_is_ipi_single(struct irq_domain * domain) irq_domain_is_ipi_single() argument 673 irq_domain_is_msi(struct irq_domain * domain) irq_domain_is_msi() argument 678 irq_domain_is_msi_parent(struct irq_domain * domain) irq_domain_is_msi_parent() argument 683 irq_domain_is_msi_device(struct irq_domain * domain) irq_domain_is_msi_device() argument 694 msi_device_domain_alloc_wired(struct irq_domain * domain,unsigned int hwirq,unsigned int type) msi_device_domain_alloc_wired() argument 700 msi_device_domain_free_wired(struct irq_domain * domain,unsigned int virq) msi_device_domain_free_wired() argument [all...] |
| /linux/kernel/irq/ |
| H A D | msi.c | 41 * @domid: ID of the domain on which management operations should be done 56 /* The maximum domain size */ 62 static int msi_domain_prepare_irqs(struct irq_domain *domain, struct device *dev, 145 * @domid: The id of the interrupt domain to which the desriptor is added 186 (dev->msi.domain && in msi_ctrl_valid() 187 !dev->msi.data->__domains[ctrl->domid].domain))) in msi_ctrl_valid() 223 * @domid: Id of the domain to operate on 332 * If @dev::msi::domain is set and is a global MSI domain, copy the in msi_setup_device_data() 333 * pointer into the domain arra in msi_setup_device_data() 603 struct irq_domain *domain; msi_get_device_domain() local 623 struct irq_domain *domain; msi_domain_get_hwsize() local 640 msi_check_level(struct irq_domain * domain,struct msi_msg * msg) msi_check_level() argument 681 msi_domain_activate(struct irq_domain * domain,struct irq_data * irq_data,bool early) msi_domain_activate() argument 692 msi_domain_deactivate(struct irq_domain * domain,struct irq_data * irq_data) msi_domain_deactivate() argument 701 msi_domain_alloc(struct irq_domain * domain,unsigned int virq,unsigned int nr_irqs,void * arg) msi_domain_alloc() argument 733 msi_domain_free(struct irq_domain * domain,unsigned int virq,unsigned int nr_irqs) msi_domain_free() argument 746 msi_domain_translate(struct irq_domain * domain,struct irq_fwspec * fwspec,irq_hw_number_t * hwirq,unsigned int * type) msi_domain_translate() argument 792 msi_domain_ops_prepare(struct irq_domain * domain,struct device * dev,int nvec,msi_alloc_info_t * arg) msi_domain_ops_prepare() argument 799 msi_domain_ops_teardown(struct irq_domain * domain,msi_alloc_info_t * arg) msi_domain_ops_teardown() argument 809 msi_domain_ops_init(struct irq_domain * domain,struct msi_domain_info * info,unsigned int virq,irq_hw_number_t hwirq,msi_alloc_info_t * arg) msi_domain_ops_init() argument 870 struct irq_domain *domain; __msi_create_irq_domain() local 968 msi_parent_init_dev_msi_info(struct device * dev,struct irq_domain * domain,struct irq_domain * msi_parent_domain,struct msi_domain_info * msi_child_info) msi_parent_init_dev_msi_info() argument 1034 struct irq_domain *domain, *parent = dev->msi.domain; msi_create_device_irq_domain() local 1117 struct irq_domain *domain; msi_remove_device_irq_domain() local 1148 struct irq_domain *domain; msi_match_device_irq_domain() local 1159 msi_domain_prepare_irqs(struct irq_domain * domain,struct device * dev,int nvec,msi_alloc_info_t * arg) msi_domain_prepare_irqs() argument 1179 msi_check_reservation_mode(struct irq_domain * domain,struct msi_domain_info * info,struct device * dev) msi_check_reservation_mode() argument 1209 msi_handle_pci_fail(struct irq_domain * domain,struct msi_desc * desc,int allocated) msi_handle_pci_fail() argument 1235 msi_init_virq(struct irq_domain * domain,int virq,unsigned int vflags) msi_init_virq() argument 1274 populate_alloc_info(struct irq_domain * domain,struct device * dev,unsigned int nirqs,msi_alloc_info_t * arg) populate_alloc_info() argument 1292 __msi_domain_alloc_irqs(struct device * dev,struct irq_domain * domain,struct msi_ctrl * ctrl) __msi_domain_alloc_irqs() argument 1373 struct irq_domain *domain; __msi_domain_alloc_locked() local 1483 struct irq_domain *domain; __msi_domain_alloc_irq_at() local 1576 msi_device_domain_alloc_wired(struct irq_domain * domain,unsigned int hwirq,unsigned int type) msi_device_domain_alloc_wired() argument 1597 __msi_domain_free_irqs(struct device * dev,struct irq_domain * domain,struct msi_ctrl * ctrl) __msi_domain_free_irqs() argument 1630 struct irq_domain *domain; msi_domain_free_locked() local 1729 msi_device_domain_free_wired(struct irq_domain * domain,unsigned int virq) msi_device_domain_free_wired() argument 1750 msi_get_domain_info(struct irq_domain * domain) msi_get_domain_info() argument 1773 struct irq_domain *domain = dev_get_msi_domain(dev); msi_device_has_isolated_msi() local [all...] |
| /linux/drivers/soc/dove/ |
| H A D | pmu.c | 131 * This deals with the "old" Marvell sequence of bringing a power domain 143 static int pmu_domain_power_off(struct generic_pm_domain *domain) in pmu_domain_power_off() argument 145 struct pmu_domain *pmu_dom = to_pmu_domain(domain); in pmu_domain_power_off() 177 static int pmu_domain_power_on(struct generic_pm_domain *domain) in pmu_domain_power_on() argument 179 struct pmu_domain *pmu_dom = to_pmu_domain(domain); in pmu_domain_power_on() 211 static void __pmu_domain_register(struct pmu_domain *domain, in __pmu_domain_register() argument 214 unsigned int val = readl_relaxed(domain->pmu->pmu_base + PMU_PWR); in __pmu_domain_register() 216 domain->base.power_off = pmu_domain_power_off; in __pmu_domain_register() 217 domain->base.power_on = pmu_domain_power_on; in __pmu_domain_register() 219 pm_genpd_init(&domain in __pmu_domain_register() 230 struct irq_domain *domain = pmu->irq_domain; pmu_irq_handler() local 269 struct irq_domain *domain; dove_init_pmu_irq() local 325 struct pmu_domain *domain; dove_init_pmu_legacy() local 409 struct pmu_domain *domain; dove_init_pmu() local [all...] |
| /linux/drivers/irqchip/ |
| H A D | irq-atmel-aic5.c | 87 struct irq_domain *domain = d->domain; in aic5_mask() local 88 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_mask() 103 struct irq_domain *domain = d->domain; in aic5_unmask() local 104 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_unmask() 119 struct irq_domain *domain = d->domain; in aic5_retrigger() local 120 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_retrigger() 131 struct irq_domain *domain in aic5_set_type() local 150 struct irq_domain *domain = d->domain; aic5_suspend() local 179 struct irq_domain *domain = d->domain; aic5_resume() local 214 struct irq_domain *domain = d->domain; aic5_pm_shutdown() local 233 aic5_hw_init(struct irq_domain * domain) aic5_hw_init() argument 319 struct irq_domain *domain; aic5_of_init() local [all...] |
| /linux/arch/arm64/boot/dts/apple/ |
| H A D | t8103-pmgr.dtsi | 13 #power-domain-cells = <0>; 22 #power-domain-cells = <0>; 31 #power-domain-cells = <0>; 40 #power-domain-cells = <0>; 48 #power-domain-cells = <0>; 56 #power-domain-cells = <0>; 64 #power-domain-cells = <0>; 72 #power-domain-cells = <0>; 81 #power-domain-cells = <0>; 90 #power-domain [all...] |
| H A D | t8015-pmgr.dtsi | 12 #power-domain-cells = <0>; 21 #power-domain-cells = <0>; 30 #power-domain-cells = <0>; 39 #power-domain-cells = <0>; 48 #power-domain-cells = <0>; 57 #power-domain-cells = <0>; 66 #power-domain-cells = <0>; 75 #power-domain-cells = <0>; 83 #power-domain-cells = <0>; 92 #power-domain [all...] |
| H A D | t8112-pmgr.dtsi | 13 #power-domain-cells = <0>; 22 #power-domain-cells = <0>; 31 #power-domain-cells = <0>; 40 #power-domain-cells = <0>; 48 #power-domain-cells = <0>; 56 #power-domain-cells = <0>; 65 #power-domain-cells = <0>; 74 #power-domain-cells = <0>; 83 #power-domain-cells = <0>; 92 #power-domain [all...] |
| H A D | t8012-pmgr.dtsi | 12 #power-domain-cells = <0>; 21 #power-domain-cells = <0>; 30 #power-domain-cells = <0>; 39 #power-domain-cells = <0>; 47 #power-domain-cells = <0>; 56 #power-domain-cells = <0>; 64 #power-domain-cells = <0>; 73 #power-domain-cells = <0>; 82 #power-domain-cells = <0>; 90 #power-domain [all...] |
| H A D | s8001-pmgr.dtsi | 12 #power-domain-cells = <0>; 21 #power-domain-cells = <0>; 30 #power-domain-cells = <0>; 39 #power-domain-cells = <0>; 47 #power-domain-cells = <0>; 56 #power-domain-cells = <0>; 65 #power-domain-cells = <0>; 74 #power-domain-cells = <0>; 82 #power-domain-cells = <0>; 90 #power-domain [all...] |
| /linux/drivers/soc/qcom/ |
| H A D | qcom_pd_mapper.c | 25 const char *domain; member 71 struct qcom_pdm_domain *domain; in qcom_pdm_add_service_domain() local 75 list_for_each_entry(domain, &service->domains, list) { in qcom_pdm_add_service_domain() 76 if (!strcmp(domain->name, domain_name)) in qcom_pdm_add_service_domain() 90 domain = kzalloc(sizeof(*domain), GFP_KERNEL); in qcom_pdm_add_service_domain() 91 if (!domain) { in qcom_pdm_add_service_domain() 100 domain->name = domain_name; in qcom_pdm_add_service_domain() 101 domain->instance_id = instance_id; in qcom_pdm_add_service_domain() 102 list_add_tail(&domain in qcom_pdm_add_service_domain() 108 qcom_pdm_add_domain(struct qcom_pdm_data * data,const struct qcom_pdm_domain_data * domain) qcom_pdm_add_domain() argument 136 struct qcom_pdm_domain *domain, *tdomain; qcom_pdm_free_domains() local 180 struct qcom_pdm_domain *domain; qcom_pdm_get_domain_list() local [all...] |