| /linux/drivers/android/ |
| H A D | dbitmap.h | 31 static inline int dbitmap_enabled(struct dbitmap *dmap) in dbitmap_enabled() argument 33 return !!dmap->nbits; in dbitmap_enabled() 36 static inline void dbitmap_free(struct dbitmap *dmap) in dbitmap_free() argument 38 dmap->nbits = 0; in dbitmap_free() 39 kfree(dmap->map); in dbitmap_free() 40 dmap->map = NULL; in dbitmap_free() 44 static inline unsigned int dbitmap_shrink_nbits(struct dbitmap *dmap) in dbitmap_shrink_nbits() argument 48 if (dmap->nbits <= NBITS_MIN) in dbitmap_shrink_nbits() 57 bit = find_last_bit(dmap->map, dmap->nbits); in dbitmap_shrink_nbits() 58 if (bit < (dmap->nbits >> 2)) in dbitmap_shrink_nbits() [all …]
|
| H A D | binder_internal.h | 435 struct dbitmap dmap; member
|
| /linux/fs/fuse/ |
| H A D | dax.c | 130 struct fuse_dax_mapping *dmap; in alloc_dax_mapping() local 133 dmap = list_first_entry_or_null(&fcd->free_ranges, in alloc_dax_mapping() 135 if (dmap) { in alloc_dax_mapping() 136 list_del_init(&dmap->list); in alloc_dax_mapping() 143 return dmap; in alloc_dax_mapping() 148 struct fuse_dax_mapping *dmap) in __dmap_remove_busy_list() argument 150 list_del_init(&dmap->busy_list); in __dmap_remove_busy_list() 156 struct fuse_dax_mapping *dmap) in dmap_remove_busy_list() argument 159 __dmap_remove_busy_list(fcd, dmap); in dmap_remove_busy_list() 165 struct fuse_dax_mapping *dmap) in __dmap_add_to_free_pool() argument [all …]
|
| /linux/drivers/pmdomain/ti/ |
| H A D | omap_prm.c | 61 const struct omap_prm_domain_map *dmap; member 161 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_reton, 165 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact, 170 .pwrstctrl = 0, .pwrstst = 0x4, .dmap = &omap_prm_all, 174 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_alwon, 178 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_reton, 185 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact, 190 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_onoff_noauto, 194 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact 198 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_onoff_noauto [all …]
|
| /linux/drivers/md/ |
| H A D | dm-clone-metadata.c | 127 struct dirty_map dmap[2]; member 468 static int __dirty_map_init(struct dirty_map *dmap, unsigned long nr_words, in __dirty_map_init() argument 471 dmap->changed = 0; in __dirty_map_init() 473 dmap->dirty_words = kvzalloc(bitmap_size(nr_words), GFP_KERNEL); in __dirty_map_init() 474 if (!dmap->dirty_words) in __dirty_map_init() 477 dmap->dirty_regions = kvzalloc(bitmap_size(nr_regions), GFP_KERNEL); in __dirty_map_init() 478 if (!dmap->dirty_regions) { in __dirty_map_init() 479 kvfree(dmap->dirty_words); in __dirty_map_init() 486 static void __dirty_map_exit(struct dirty_map *dmap) in __dirty_map_exit() argument 488 kvfree(dmap->dirty_words); in __dirty_map_exit() [all …]
|
| H A D | dm-zoned-metadata.c | 1683 struct dmz_map *dmap; in dmz_load_mapping() local 1701 dmap = dmap_mblk->data; in dmz_load_mapping() 1707 dzone_id = le32_to_cpu(dmap[e].dzone_id); in dmz_load_mapping() 1735 bzone_id = le32_to_cpu(dmap[e].bzone_id); in dmz_load_mapping() 1832 struct dmz_map *dmap = dmap_mblk->data; in dmz_set_chunk_mapping() local 1835 dmap[map_idx].dzone_id = cpu_to_le32(dzone_id); in dmz_set_chunk_mapping() 1836 dmap[map_idx].bzone_id = cpu_to_le32(bzone_id); in dmz_set_chunk_mapping() 2045 struct dmz_map *dmap = dmap_mblk->data; in dmz_get_chunk_mapping() local 2055 dzone_id = le32_to_cpu(dmap[dmap_idx].dzone_id); in dmz_get_chunk_mapping()
|
| /linux/drivers/media/pci/intel/ipu6/ |
| H A D | ipu6-mmu.c | 424 iova = alloc_iova(&mmu->dmap->iovad, n_pages, in allocate_trash_buffer() 425 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in allocate_trash_buffer() 431 dma = dma_map_page(mmu->dmap->mmu_info->dev, mmu->trash_page, 0, in allocate_trash_buffer() 433 if (dma_mapping_error(mmu->dmap->mmu_info->dev, dma)) { in allocate_trash_buffer() 434 dev_err(mmu->dmap->mmu_info->dev, "Failed to map trash page\n"); in allocate_trash_buffer() 447 ret = ipu6_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr), in allocate_trash_buffer() 464 ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in allocate_trash_buffer() 466 dma_unmap_page(mmu->dmap->mmu_info->dev, mmu->pci_trash_page, in allocate_trash_buffer() 469 __free_iova(&mmu->dmap->iovad, iova); in allocate_trash_buffer() 479 mmu_info = mmu->dmap->mmu_info; in ipu6_mmu_hw_init() [all …]
|
| H A D | ipu6-dma.c | 174 iova = alloc_iova(&mmu->dmap->iovad, count, in ipu6_dma_alloc() 175 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu6_dma_alloc() 196 ret = ipu6_mmu_map(mmu->dmap->mmu_info, in ipu6_dma_alloc() 225 pci_dma_addr = ipu6_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu6_dma_alloc() 230 ipu6_mmu_unmap(mmu->dmap->mmu_info, ipu6_iova, PAGE_SIZE); in ipu6_dma_alloc() 236 __free_iova(&mmu->dmap->iovad, iova); in ipu6_dma_alloc() 249 struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle)); in ipu6_dma_free() 278 pci_dma_addr = ipu6_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu6_dma_free() 284 ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in ipu6_dma_free() 291 __free_iova(&mmu->dmap->iovad, iova); in ipu6_dma_free() [all …]
|
| H A D | ipu6-mmu.h | 35 struct ipu6_dma_mapping *dmap; member 48 struct ipu6_dma_mapping *dmap; member
|
| /linux/drivers/staging/media/ipu7/ |
| H A D | ipu7-mmu.c | 407 iova = alloc_iova(&mmu->dmap->iovad, n_pages, in allocate_trash_buffer() 408 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in allocate_trash_buffer() 414 dma = dma_map_page(mmu->dmap->mmu_info->dev, mmu->trash_page, 0, in allocate_trash_buffer() 416 if (dma_mapping_error(mmu->dmap->mmu_info->dev, dma)) { in allocate_trash_buffer() 417 dev_err(mmu->dmap->mmu_info->dev, "Failed to map trash page\n"); in allocate_trash_buffer() 430 ret = ipu7_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr), in allocate_trash_buffer() 447 ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in allocate_trash_buffer() 449 dma_unmap_page(mmu->dmap->mmu_info->dev, mmu->pci_trash_page, in allocate_trash_buffer() 452 __free_iova(&mmu->dmap->iovad, iova); in allocate_trash_buffer() 461 mmu_info = mmu->dmap->mmu_info; in __mmu_at_init() [all …]
|
| H A D | ipu7-dma.c | 174 iova = alloc_iova(&mmu->dmap->iovad, count, in ipu7_dma_alloc() 175 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu7_dma_alloc() 196 ret = ipu7_mmu_map(mmu->dmap->mmu_info, in ipu7_dma_alloc() 225 pci_dma_addr = ipu7_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu7_dma_alloc() 230 ipu7_mmu_unmap(mmu->dmap->mmu_info, ipu7_iova, PAGE_SIZE); in ipu7_dma_alloc() 236 __free_iova(&mmu->dmap->iovad, iova); in ipu7_dma_alloc() 249 struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle)); in ipu7_dma_free() 278 pci_dma_addr = ipu7_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu7_dma_free() 284 ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in ipu7_dma_free() 291 __free_iova(&mmu->dmap->iovad, iova); in ipu7_dma_free() [all …]
|
| H A D | ipu7-mmu.h | 376 struct ipu7_dma_mapping *dmap; member 389 struct ipu7_dma_mapping *dmap; member
|
| /linux/sound/soc/tegra/ |
| H A D | tegra_pcm.c | 80 struct snd_dmaengine_dai_dma_data *dmap; in tegra_pcm_open() local 88 dmap = snd_soc_dai_get_dma_data(cpu_dai, substream); in tegra_pcm_open() 101 chan = dma_request_chan(cpu_dai->dev, dmap->chan_name); in tegra_pcm_open() 105 dmap->chan_name); in tegra_pcm_open() 113 dmap->chan_name); in tegra_pcm_open() 146 struct snd_dmaengine_dai_dma_data *dmap; in tegra_pcm_hw_params() local 154 dmap = snd_soc_dai_get_dma_data(snd_soc_rtd_to_cpu(rtd, 0), substream); in tegra_pcm_hw_params() 155 if (!dmap) in tegra_pcm_hw_params() 169 slave_config.dst_addr = dmap->addr; in tegra_pcm_hw_params() 173 slave_config.src_addr = dmap->addr; in tegra_pcm_hw_params()
|
| /linux/fs/nilfs2/ |
| H A D | page.c | 243 int nilfs_copy_dirty_pages(struct address_space *dmap, in nilfs_copy_dirty_pages() argument 264 dfolio = filemap_grab_folio(dmap, folio->index); in nilfs_copy_dirty_pages() 298 void nilfs_copy_back_pages(struct address_space *dmap, in nilfs_copy_back_pages() argument 316 dfolio = filemap_lock_folio(dmap, index); in nilfs_copy_back_pages() 334 xa_lock_irq(&dmap->i_pages); in nilfs_copy_back_pages() 335 f = __xa_store(&dmap->i_pages, index, folio, GFP_NOFS); in nilfs_copy_back_pages() 341 folio->mapping = dmap; in nilfs_copy_back_pages() 342 dmap->nrpages++; in nilfs_copy_back_pages() 344 __xa_set_mark(&dmap->i_pages, index, in nilfs_copy_back_pages() 347 xa_unlock_irq(&dmap->i_pages); in nilfs_copy_back_pages()
|
| /linux/fs/xfs/ |
| H A D | xfs_reflink.c | 1315 struct xfs_bmbt_irec *dmap, in xfs_reflink_remap_extent() argument 1326 bool dmap_written = xfs_bmap_is_written_extent(dmap); in xfs_reflink_remap_extent() 1356 rblocks = dmap->br_blockcount; in xfs_reflink_remap_extent() 1358 dblocks = resblks + dmap->br_blockcount; in xfs_reflink_remap_extent() 1377 error = xfs_bmapi_read(ip, dmap->br_startoff, dmap->br_blockcount, in xfs_reflink_remap_extent() 1381 ASSERT(nimaps == 1 && smap.br_startoff == dmap->br_startoff); in xfs_reflink_remap_extent() 1388 dmap->br_blockcount = min(dmap->br_blockcount, smap.br_blockcount); in xfs_reflink_remap_extent() 1389 ASSERT(dmap->br_blockcount == smap.br_blockcount); in xfs_reflink_remap_extent() 1398 if (dmap->br_startblock == smap.br_startblock) { in xfs_reflink_remap_extent() 1399 if (dmap->br_state != smap.br_state) { in xfs_reflink_remap_extent() [all …]
|
| /linux/fs/jfs/ |
| H A D | jfs_dmap.c | 64 static void dbAllocBits(struct bmap * bmp, struct dmap * dp, s64 blkno, 73 static int dbAllocNext(struct bmap * bmp, struct dmap * dp, s64 blkno, 75 static int dbAllocNear(struct bmap * bmp, struct dmap * dp, s64 blkno, 78 static int dbAllocDmap(struct bmap * bmp, struct dmap * dp, s64 blkno, 80 static int dbAllocDmapLev(struct bmap * bmp, struct dmap * dp, int nblocks, 91 static int dbFreeBits(struct bmap * bmp, struct dmap * dp, s64 blkno, 93 static int dbFreeDmap(struct bmap * bmp, struct dmap * dp, s64 blkno, 101 static int dbAllocDmapBU(struct bmap * bmp, struct dmap * dp, s64 blkno, 103 static int dbInitDmap(struct dmap * dp, s64 blkno, int nblocks); 104 static int dbInitDmapTree(struct dmap * dp); [all …]
|
| /linux/include/drm/ttm/ |
| H A D | ttm_kmap_iter.h | 32 struct iosys_map *dmap, pgoff_t i); 42 struct iosys_map *dmap);
|
| /linux/drivers/misc/sgi-gru/ |
| H A D | grufault.c | 143 struct gru_tlb_fault_map *dmap) in get_clear_fault_map() argument 158 dmap->fault_bits[i] = k; in get_clear_fault_map() 522 struct gru_tlb_fault_map imap, dmap; in gru_intr() local 536 get_clear_fault_map(gru, &imap, &dmap); in gru_intr() 541 dmap.fault_bits[0], dmap.fault_bits[1]); in gru_intr() 543 for_each_cbr_in_tfm(cbrnum, dmap.fault_bits) { in gru_intr()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3.c | 77 &imgu_pipe->queues[i].dmap); in imgu_dummybufs_cleanup() 98 &imgu_pipe->queues[i].dmap, size)) { in imgu_dummybufs_preallocate() 138 &imgu_pipe->queues[i].dmap, in imgu_dummybufs_init() 146 imgu_pipe->queues[i].dmap.daddr); in imgu_dummybufs_init() 163 if (WARN_ON(!imgu_pipe->queues[queue].dmap.vaddr)) in imgu_dummybufs_get() 176 imgu_pipe->queues[queue].dmap.daddr); in imgu_dummybufs_get()
|
| H A D | ipu3.h | 105 struct imgu_css_map dmap; member
|
| /linux/drivers/edac/ |
| H A D | pnd2_edac.c | 418 static struct d_cr_dmap dmap[DNV_NUM_CHANNELS]; variable 474 RD_REGP(&dmap[i], d_cr_dmap, dnv_dports[i]) || in dnv_get_registers() 979 daddr->rank = dnv_get_bit(pmiaddr, dmap[pmiidx].rs0 + 13, 0); in dnv_pmi2mem() 981 daddr->rank |= dnv_get_bit(pmiaddr, dmap[pmiidx].rs1 + 13, 1); in dnv_pmi2mem() 989 daddr->bank = dnv_get_bit(pmiaddr, dmap[pmiidx].ba0 + 6, 0); in dnv_pmi2mem() 990 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].ba1 + 6, 1); in dnv_pmi2mem() 991 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].bg0 + 6, 2); in dnv_pmi2mem() 993 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].bg1 + 6, 3); in dnv_pmi2mem()
|
| /linux/drivers/iommu/ |
| H A D | tegra-smmu.c | 586 dma_addr_t *dmap) in tegra_smmu_pte_lookup() argument 596 *dmap = smmu_pde_to_dma(smmu, as->pd->val[pd_index]); in tegra_smmu_pte_lookup() 602 dma_addr_t *dmap, struct tegra_pt *pt) in as_get_pte() argument 629 *dmap = dma; in as_get_pte() 631 *dmap = smmu_pde_to_dma(smmu, as->pd->val[pde]); in as_get_pte()
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_tt.c | 488 struct iosys_map *dmap, in ttm_kmap_iter_tt_map_local() argument 494 iosys_map_set_vaddr(dmap, kmap_local_page_prot(iter_tt->tt->pages[i], in ttm_kmap_iter_tt_map_local()
|
| /linux/drivers/media/platform/xilinx/ |
| H A D | xilinx-vipp.c | 473 struct xvip_dma *dmap; in xvip_graph_cleanup() local 479 list_for_each_entry_safe(dma, dmap, &xdev->dmas, list) { in xvip_graph_cleanup()
|
| /linux/drivers/net/ethernet/apple/ |
| H A D | bmac.c | 177 dbdma_continue(volatile struct dbdma_regs __iomem *dmap) in dbdma_continue() argument 179 dbdma_st32(&dmap->control, in dbdma_continue() 185 dbdma_reset(volatile struct dbdma_regs __iomem *dmap) in dbdma_reset() argument 187 dbdma_st32(&dmap->control, in dbdma_reset() 190 while (dbdma_ld32(&dmap->status) & RUN) in dbdma_reset()
|