| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_kms.c | 60 struct vmw_surface *surf = vmw_user_object_surface(&vps->uo); in vmw_du_plane_unpin_surf() 280 vmw_user_object_ref(&vps->uo); in vmw_du_plane_duplicate_state() 329 vmw_user_object_unref(&vps->uo); in vmw_du_plane_destroy_state() 422 struct vmw_bo *bo = vmw_user_object_buffer(&vfbs->uo); in vmw_framebuffer_surface_destroy() 423 struct vmw_surface *surf = vmw_user_object_surface(&vfbs->uo); in vmw_framebuffer_surface_destroy() 437 vmw_user_object_unref(&vfbs->uo); in vmw_framebuffer_surface_destroy() 487 struct vmw_bo *bo = vmw_user_object_buffer(&vfbs->uo); in vmw_framebuffer_surface_create_handle() 501 struct vmw_user_object *uo, in vmw_kms_new_framebuffer_surface() argument 517 surface = vmw_user_object_surface(uo); in vmw_kms_new_framebuffer_surface() 553 memcpy(&vfbs->uo, uo, sizeof(vfbs->uo)); in vmw_kms_new_framebuffer_surface() [all …]
|
| H A D | vmwgfx_cursor_plane.c | 77 struct vmw_surface *surface = vmw_user_object_surface(&vps->uo); in vmw_cursor_plane_update_legacy() 95 struct vmw_surface *surface = vmw_user_object_surface(&vps->uo); in vmw_cursor_update_type() 115 struct vmw_bo *bo = vmw_user_object_buffer(&vps->uo); in vmw_cursor_update_mob() 473 if (!vmw_user_object_is_null(&vps->uo)) in vmw_cursor_plane_cleanup_fb() 474 vmw_user_object_unmap(&vps->uo); in vmw_cursor_plane_cleanup_fb() 480 vmw_user_object_unref(&vps->uo); in vmw_cursor_plane_cleanup_fb() 487 struct vmw_bo *new_bo = vmw_user_object_buffer(&new_vps->uo); in vmw_cursor_buffer_changed() 488 struct vmw_bo *old_bo = vmw_user_object_buffer(&old_vps->uo); in vmw_cursor_buffer_changed() 503 surf = vmw_user_object_surface(&new_vps->uo); in vmw_cursor_buffer_changed() 607 if (!vmw_user_object_is_null(&vps->uo)) { in vmw_cursor_plane_prepare_fb() [all …]
|
| H A D | vmwgfx_stdu.c | 740 srf = &vmw_user_object_surface(&vfbs->uo)->res; in vmw_kms_stdu_surface_dirty() 948 if (vmw_user_object_surface(&vps->uo)) in vmw_stdu_primary_plane_cleanup_fb() 986 if (vmw_user_object_surface(&vps->uo)) { in vmw_stdu_primary_plane_prepare_fb() 988 vmw_user_object_unref(&vps->uo); in vmw_stdu_primary_plane_prepare_fb() 998 vmw_user_object_surface(&new_vfbs->uo)->metadata.base_size.width == hdisplay && in vmw_stdu_primary_plane_prepare_fb() 999 vmw_user_object_surface(&new_vfbs->uo)->metadata.base_size.height == vdisplay) in vmw_stdu_primary_plane_prepare_fb() 1037 metadata = vmw_user_object_surface(&new_vfbs->uo)->metadata; in vmw_stdu_primary_plane_prepare_fb() 1044 if (vmw_user_object_surface(&vps->uo)) { in vmw_stdu_primary_plane_prepare_fb() 1046 vmw_user_object_surface(&vps->uo)->metadata.base_size; in vmw_stdu_primary_plane_prepare_fb() 1050 vmw_user_object_surface(&vps->uo)->metadata.format != metadata.format) { in vmw_stdu_primary_plane_prepare_fb() [all …]
|
| H A D | vmwgfx_resource.c | 295 struct vmw_user_object *uo) in vmw_user_object_lookup() argument 301 WARN_ON(uo->surface || uo->buffer); in vmw_user_object_lookup() 307 uo->surface = vmw_res_to_srf(res); in vmw_user_object_lookup() 311 uo->surface = NULL; in vmw_user_object_lookup() 312 ret = vmw_user_bo_lookup(filp, handle, &uo->buffer); in vmw_user_object_lookup() 313 if (!ret && !uo->buffer->is_dumb) { in vmw_user_object_lookup() 314 uo->surface = vmw_lookup_surface_for_buffer(dev_priv, in vmw_user_object_lookup() 315 uo->buffer, in vmw_user_object_lookup() 317 if (uo->surface) in vmw_user_object_lookup() 318 vmw_user_bo_unref(&uo->buffer); in vmw_user_object_lookup()
|
| H A D | vmwgfx_drv.h | 764 u32 handle, struct vmw_user_object *uo); 765 struct vmw_user_object *vmw_user_object_ref(struct vmw_user_object *uo); 766 void vmw_user_object_unref(struct vmw_user_object *uo); 767 bool vmw_user_object_is_null(struct vmw_user_object *uo); 768 struct vmw_surface *vmw_user_object_surface(struct vmw_user_object *uo); 769 struct vmw_bo *vmw_user_object_buffer(struct vmw_user_object *uo); 770 void *vmw_user_object_map(struct vmw_user_object *uo); 771 void *vmw_user_object_map_size(struct vmw_user_object *uo, size_t size); 772 void vmw_user_object_unmap(struct vmw_user_object *uo); 773 bool vmw_user_object_is_mapped(struct vmw_user_object *uo);
|
| H A D | vmwgfx_scrn.c | 244 sou->buffer = vmw_user_object_buffer(&vps->uo); in vmw_sou_crtc_mode_set_nofb() 380 struct vmw_bo *bo = vmw_user_object_buffer(&vps->uo); in vmw_sou_primary_plane_cleanup_fb() 384 vmw_user_object_unref(&vps->uo); in vmw_sou_primary_plane_cleanup_fb() 419 vmw_user_object_unref(&vps->uo); in vmw_sou_primary_plane_prepare_fb() 428 bo = vmw_user_object_buffer(&vps->uo); in vmw_sou_primary_plane_prepare_fb() 438 vmw_user_object_unref(&vps->uo); in vmw_sou_primary_plane_prepare_fb() 448 ret = vmw_bo_create(dev_priv, &bo_params, &vps->uo.buffer); in vmw_sou_primary_plane_prepare_fb() 459 return vmw_bo_pin_in_vram(dev_priv, vps->uo.buffer, true); in vmw_sou_primary_plane_prepare_fb() 594 surf = vmw_user_object_surface(&vfbs->uo); in vmw_sou_surface_pre_clip() 1111 srf = &vmw_user_object_surface(&vfbs->uo)->res; in vmw_kms_sou_do_surface_dirty()
|
| H A D | vmwgfx_kms.h | 203 struct vmw_user_object uo; member 247 struct vmw_user_object uo; member 401 struct vmw_user_object *uo,
|
| H A D | vmwgfx_ldu.c | 153 vmw_user_object_buffer(&vmw_framebuffer_to_vfbs(&vfb->base)->uo); in vmw_ldu_fb_pin() 176 vmw_user_object_buffer(&vmw_framebuffer_to_vfbs(&vfb->base)->uo); in vmw_ldu_fb_unpin()
|
| /linux/drivers/tee/qcomtee/ |
| H A D | user_obj.c | 229 struct qcomtee_user_object *uo = to_qcomtee_user_object(object); in qcomtee_user_object_dispatch() local 230 struct qcomtee_context_data *ctxdata = uo->ctx->data; in qcomtee_user_object_dispatch() 239 ureq->object_id = uo->object_id; in qcomtee_user_object_dispatch() 319 struct qcomtee_user_object *uo = to_qcomtee_user_object(object); in qcomtee_user_object_release() local 320 struct qcomtee_context_data *ctxdata = uo->ctx->data; in qcomtee_user_object_release() 326 if (!uo->notify) in qcomtee_user_object_release() 334 ureq->object_id = uo->object_id; in qcomtee_user_object_release() 346 teedev_ctx_put(uo->ctx); in qcomtee_user_object_release() 347 kfree(uo); in qcomtee_user_object_release() 400 struct qcomtee_user_object *uo; in qcomtee_user_param_from_object() local [all …]
|
| /linux/drivers/clk/bcm/ |
| H A D | clk-ns2.c | 26 #define VCO_CTRL_VAL(uo, lo) { .u_offset = uo, .l_offset = lo } argument
|
| H A D | clk-cygnus.c | 34 #define VCO_CTRL_VAL(uo, lo) { .u_offset = uo, .l_offset = lo } argument
|
| H A D | clk-sr.c | 28 #define VCO_CTRL_VAL(uo, lo) { .u_offset = uo, .l_offset = lo } argument
|
| /linux/drivers/net/ethernet/broadcom/genet/ |
| H A D | bcmgenet.h | 103 u32 uo; /* RO # of unknown op code pkt */ member
|
| H A D | bcmgenet.c | 1102 STAT_GENET_MIB_RX("rx_unknown", mib.rx.uo),
|
| /linux/drivers/net/ethernet/broadcom/ |
| H A D | bcmsysport.h | 520 u32 uo; /* RO # of unknown op code pkt */ member
|
| H A D | bcmsysport.c | 220 STAT_MIB_RX("rx_unknown", mib.rx.uo),
|
| /linux/include/linux/ |
| H A D | skbuff.h | 3793 #define __skb_frag_dma_map1(dev, frag, offset, uf, uo) ({ \ argument 3795 size_t uo = (offset); \ 3797 __skb_frag_dma_map(dev, uf, uo, skb_frag_size(uf) - uo, \
|