Lines Matching defs:attrs

438 static inline pgprot_t __get_dma_pgprot(unsigned long attrs, pgprot_t prot)
440 prot = (attrs & DMA_ATTR_WRITE_COMBINE) ?
535 unsigned long attrs, const void *caller)
548 .want_vaddr = ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) == 0),
606 dma_addr_t handle, unsigned long attrs,
616 .want_vaddr = ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) == 0),
733 static int __dma_info_to_prot(enum dma_data_direction dir, unsigned long attrs)
737 if (attrs & DMA_ATTR_PRIVILEGED)
853 gfp_t gfp, unsigned long attrs,
866 if (attrs & DMA_ATTR_FORCE_CONTIGUOUS)
885 if (attrs & DMA_ATTR_ALLOC_SINGLE_PAGES)
941 size_t size, unsigned long attrs)
946 if (attrs & DMA_ATTR_FORCE_CONTIGUOUS) {
963 unsigned long attrs)
988 __dma_info_to_prot(DMA_BIDIRECTIONAL, attrs),
1029 static struct page **__iommu_get_pages(void *cpu_addr, unsigned long attrs)
1034 if (attrs & DMA_ATTR_NO_KERNEL_MAPPING)
1042 unsigned long attrs)
1054 *handle = __iommu_create_mapping(dev, &page, size, attrs);
1076 dma_addr_t *handle, gfp_t gfp, unsigned long attrs)
1078 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL);
1088 coherent_flag, attrs);
1090 pages = __iommu_alloc_buffer(dev, size, gfp, attrs, coherent_flag);
1094 *handle = __iommu_create_mapping(dev, pages, size, attrs);
1098 if (attrs & DMA_ATTR_NO_KERNEL_MAPPING)
1111 __iommu_free_buffer(dev, pages, size, attrs);
1117 unsigned long attrs)
1119 struct page **pages = __iommu_get_pages(cpu_addr, attrs);
1130 vma->vm_page_prot = __get_dma_pgprot(attrs, vma->vm_page_prot);
1144 dma_addr_t handle, unsigned long attrs)
1155 pages = __iommu_get_pages(cpu_addr, attrs);
1161 if ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) == 0)
1165 __iommu_free_buffer(dev, pages, size, attrs);
1170 size_t size, unsigned long attrs)
1173 struct page **pages = __iommu_get_pages(cpu_addr, attrs);
1187 enum dma_data_direction dir, unsigned long attrs)
1207 if (!dev->dma_coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC))
1210 prot = __dma_info_to_prot(dir, attrs);
1241 int nents, enum dma_data_direction dir, unsigned long attrs)
1256 &dma->dma_address, dir, attrs);
1270 ret = __map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs);
1300 unsigned long attrs)
1309 if (!dev->dma_coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC))
1370 unsigned long attrs)
1376 if (!dev->dma_coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC))
1383 prot = __dma_info_to_prot(dir, attrs);
1406 size_t size, enum dma_data_direction dir, unsigned long attrs)
1417 if (!dev->dma_coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC)) {
1435 enum dma_data_direction dir, unsigned long attrs)
1448 prot = __dma_info_to_prot(dir, attrs) | IOMMU_MMIO;
1469 unsigned long attrs)
1812 gfp_t gfp, unsigned long attrs)
1815 __get_dma_pgprot(attrs, PAGE_KERNEL), false,
1816 attrs, __builtin_return_address(0));
1820 dma_addr_t dma_handle, unsigned long attrs)
1822 __arm_dma_free(dev, size, cpu_addr, dma_handle, attrs, false);