Lines Matching full:context
16 static void etnaviv_context_unmap(struct etnaviv_iommu_context *context, in etnaviv_context_unmap() argument
29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap()
39 static int etnaviv_context_map(struct etnaviv_iommu_context *context, in etnaviv_context_map() argument
55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map()
67 etnaviv_context_unmap(context, orig_iova, orig_size - size); in etnaviv_context_map()
72 static int etnaviv_iommu_map(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_map() argument
79 if (!context || !sgt) in etnaviv_iommu_map()
88 ret = etnaviv_context_map(context, da, pa, bytes, prot); in etnaviv_iommu_map()
98 etnaviv_context_unmap(context, iova, da - iova); in etnaviv_iommu_map()
102 static void etnaviv_iommu_unmap(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_unmap() argument
112 etnaviv_context_unmap(context, da, bytes); in etnaviv_iommu_unmap()
122 static void etnaviv_iommu_remove_mapping(struct etnaviv_iommu_context *context, in etnaviv_iommu_remove_mapping() argument
127 etnaviv_iommu_unmap(context, mapping->vram_node.start, in etnaviv_iommu_remove_mapping()
132 static int etnaviv_iommu_find_iova(struct etnaviv_iommu_context *context, in etnaviv_iommu_find_iova() argument
139 lockdep_assert_held(&context->lock); in etnaviv_iommu_find_iova()
147 ret = drm_mm_insert_node_in_range(&context->mm, node, in etnaviv_iommu_find_iova()
153 drm_mm_scan_init(&scan, &context->mm, size, 0, 0, mode); in etnaviv_iommu_find_iova()
157 list_for_each_entry(free, &context->mappings, mmu_node) { in etnaviv_iommu_find_iova()
199 etnaviv_iommu_remove_mapping(context, m); in etnaviv_iommu_find_iova()
200 m->context = NULL; in etnaviv_iommu_find_iova()
216 static int etnaviv_iommu_insert_exact(struct etnaviv_iommu_context *context, in etnaviv_iommu_insert_exact() argument
219 return drm_mm_insert_node_in_range(&context->mm, node, size, 0, 0, va, in etnaviv_iommu_insert_exact()
223 int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context, in etnaviv_iommu_map_gem() argument
233 mutex_lock(&context->lock); in etnaviv_iommu_map_gem()
236 if (context->global->version == ETNAVIV_IOMMU_V1 && in etnaviv_iommu_map_gem()
243 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem()
252 ret = etnaviv_iommu_insert_exact(context, node, in etnaviv_iommu_map_gem()
255 ret = etnaviv_iommu_find_iova(context, node, in etnaviv_iommu_map_gem()
261 ret = etnaviv_iommu_map(context, node->start, sgt, etnaviv_obj->base.size, in etnaviv_iommu_map_gem()
269 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem()
270 context->flush_seq++; in etnaviv_iommu_map_gem()
272 mutex_unlock(&context->lock); in etnaviv_iommu_map_gem()
277 void etnaviv_iommu_unmap_gem(struct etnaviv_iommu_context *context, in etnaviv_iommu_unmap_gem() argument
282 mutex_lock(&context->lock); in etnaviv_iommu_unmap_gem()
285 if (mapping->vram_node.mm == &context->mm) in etnaviv_iommu_unmap_gem()
286 etnaviv_iommu_remove_mapping(context, mapping); in etnaviv_iommu_unmap_gem()
289 context->flush_seq++; in etnaviv_iommu_unmap_gem()
290 mutex_unlock(&context->lock); in etnaviv_iommu_unmap_gem()
295 struct etnaviv_iommu_context *context = in etnaviv_iommu_context_free() local
298 etnaviv_cmdbuf_suballoc_unmap(context, &context->cmdbuf_mapping); in etnaviv_iommu_context_free()
300 context->global->ops->free(context); in etnaviv_iommu_context_free()
302 void etnaviv_iommu_context_put(struct etnaviv_iommu_context *context) in etnaviv_iommu_context_put() argument
304 kref_put(&context->refcount, etnaviv_iommu_context_free); in etnaviv_iommu_context_put()
344 struct etnaviv_iommu_context *context) in etnaviv_iommu_restore() argument
346 context->global->ops->restore(gpu, context); in etnaviv_iommu_restore()
349 int etnaviv_iommu_get_suballoc_va(struct etnaviv_iommu_context *context, in etnaviv_iommu_get_suballoc_va() argument
354 mutex_lock(&context->lock); in etnaviv_iommu_get_suballoc_va()
358 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
368 if (context->global->version == ETNAVIV_IOMMU_V1) { in etnaviv_iommu_get_suballoc_va()
374 ret = etnaviv_iommu_find_iova(context, node, size); in etnaviv_iommu_get_suballoc_va()
376 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
381 ret = etnaviv_context_map(context, node->start, paddr, size, in etnaviv_iommu_get_suballoc_va()
385 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
389 context->flush_seq++; in etnaviv_iommu_get_suballoc_va()
392 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_get_suballoc_va()
395 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
400 void etnaviv_iommu_put_suballoc_va(struct etnaviv_iommu_context *context, in etnaviv_iommu_put_suballoc_va() argument
405 mutex_lock(&context->lock); in etnaviv_iommu_put_suballoc_va()
408 if (mapping->use > 0 || context->global->version == ETNAVIV_IOMMU_V1) { in etnaviv_iommu_put_suballoc_va()
409 mutex_unlock(&context->lock); in etnaviv_iommu_put_suballoc_va()
413 etnaviv_context_unmap(context, node->start, node->size); in etnaviv_iommu_put_suballoc_va()
415 mutex_unlock(&context->lock); in etnaviv_iommu_put_suballoc_va()
418 size_t etnaviv_iommu_dump_size(struct etnaviv_iommu_context *context) in etnaviv_iommu_dump_size() argument
420 return context->global->ops->dump_size(context); in etnaviv_iommu_dump_size()
423 void etnaviv_iommu_dump(struct etnaviv_iommu_context *context, void *buf) in etnaviv_iommu_dump() argument
425 context->global->ops->dump(context, buf); in etnaviv_iommu_dump()