/linux/drivers/vfio/ |
H A D | group.c | 26 static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group, in vfio_device_get_from_name() argument 31 mutex_lock(&group->device_lock); in vfio_device_get_from_name() 32 list_for_each_entry(it, &group->device_list, group_next) { in vfio_device_get_from_name() 50 mutex_unlock(&group->device_lock); in vfio_device_get_from_name() 58 static bool vfio_group_has_iommu(struct vfio_group *group) in vfio_group_has_iommu() argument 60 lockdep_assert_held(&group->group_lock); in vfio_group_has_iommu() 65 WARN_ON(!group->container != !group->container_users); in vfio_group_has_iommu() 67 return group->container || group in vfio_group_has_iommu() 76 vfio_group_ioctl_unset_container(struct vfio_group * group) vfio_group_ioctl_unset_container() argument 102 vfio_group_ioctl_set_container(struct vfio_group * group,int __user * arg) vfio_group_ioctl_set_container() argument 298 vfio_group_ioctl_get_device_fd(struct vfio_group * group,char __user * arg) vfio_group_ioctl_get_device_fd() argument 338 vfio_group_ioctl_get_status(struct vfio_group * group,struct vfio_group_status __user * arg) vfio_group_ioctl_get_status() argument 380 struct vfio_group *group = filep->private_data; vfio_group_fops_unl_ioctl() local 399 struct vfio_group *group = device->group; vfio_device_block_group() local 417 struct vfio_group *group = device->group; vfio_device_unblock_group() local 426 struct vfio_group *group = vfio_group_fops_open() local 468 struct vfio_group *group = filep->private_data; vfio_group_fops_release() local 503 struct vfio_group *group; vfio_group_find_from_iommu() local 520 struct vfio_group *group = container_of(dev, struct vfio_group, dev); vfio_group_release() local 533 struct vfio_group *group; vfio_group_alloc() local 570 struct vfio_group *group; vfio_create_group() local 607 struct vfio_group *group; vfio_noiommu_group_alloc() local 638 vfio_group_has_device(struct vfio_group * group,struct device * dev) vfio_group_has_device() argument 656 struct vfio_group *group; vfio_group_find_or_alloc() local 697 struct vfio_group *group; vfio_device_set_group() local 714 struct vfio_group *group = device->group; vfio_device_remove_group() local 775 struct vfio_group *group = device->group; vfio_device_group_use_iommu() local 792 struct vfio_group *group = device->group; vfio_device_group_unuse_iommu() local 810 struct vfio_group *group = file->private_data; vfio_group_from_file() local 827 struct vfio_group *group = vfio_group_from_file(file); vfio_file_iommu_group() local 856 vfio_group_enforced_coherent(struct vfio_group * group) vfio_group_enforced_coherent() argument 879 vfio_group_set_kvm(struct vfio_group * group,struct kvm * kvm) vfio_group_set_kvm() argument 895 struct vfio_group *group = vfio_group_from_file(file); vfio_file_has_dev() local [all...] |
H A D | container.c | 143 * it's freed via kref. Must support container/group/device being 167 device->group->container->iommu_driver; in vfio_device_container_register() 171 device->group->container->iommu_data, device); in vfio_device_container_register() 177 device->group->container->iommu_driver; in vfio_device_container_unregister() 181 device->group->container->iommu_data, device); in vfio_device_container_unregister() 239 struct vfio_group *group; in __vfio_container_attach_groups() local 242 list_for_each_entry(group, &container->group_list, container_next) { in __vfio_container_attach_groups() 243 ret = driver->ops->attach_group(data, group->iommu_group, in __vfio_container_attach_groups() 244 group->type); in __vfio_container_attach_groups() 252 list_for_each_entry_continue_reverse(group, in __vfio_container_attach_groups() 417 vfio_container_attach_group(struct vfio_container * container,struct vfio_group * group) vfio_container_attach_group() argument 468 vfio_group_detach_container(struct vfio_group * group) vfio_group_detach_container() argument 503 vfio_group_use_container(struct vfio_group * group) vfio_group_use_container() argument 522 vfio_group_unuse_container(struct vfio_group * group) vfio_group_unuse_container() argument [all...] |
/linux/drivers/infiniband/hw/mlx4/ |
H A D | mcg.c | 50 #define mcg_warn_group(group, format, arg...) \ argument 52 (group)->name, group->demux->port, ## arg) 54 #define mcg_debug_group(group, format, arg...) \ argument 56 (group)->name, (group)->demux->port, ## arg) 58 #define mcg_error_group(group, format, arg...) \ argument 59 pr_err(" %16s: " format, (group)->name, ## arg) 136 struct mcast_group *group; member 144 mcg_warn_group(group, "di 166 struct mcast_group *group; mcast_find() local 184 mcast_insert(struct mlx4_ib_demux_ctx * ctx,struct mcast_group * group) mcast_insert() argument 252 send_join_to_wire(struct mcast_group * group,struct ib_sa_mad * sa_mad) send_join_to_wire() argument 279 send_leave_to_wire(struct mcast_group * group,u8 join_state) send_leave_to_wire() argument 318 send_reply_to_slave(int slave,struct mcast_group * group,struct ib_sa_mad * req_sa_mad,u16 status) send_reply_to_slave() argument 437 release_group(struct mcast_group * group,int from_timeout_handler) release_group() argument 474 adjust_membership(struct mcast_group * group,u8 join_state,int inc) adjust_membership() argument 483 get_leave_state(struct mcast_group * group) get_leave_state() argument 495 join_group(struct mcast_group * group,int slave,u8 join_mask) join_group() argument 511 leave_group(struct mcast_group * group,int slave,u8 leave_state) leave_group() argument 524 check_leave(struct mcast_group * group,int slave,u8 leave_mask) check_leave() argument 542 struct mcast_group *group; mlx4_ib_mcg_timeout_handler() local 584 handle_leave_req(struct mcast_group * group,u8 leave_mask,struct mcast_req * req) handle_leave_req() argument 605 handle_join_req(struct mcast_group * group,u8 join_mask,struct mcast_req * req) handle_join_req() argument 644 struct mcast_group *group; mlx4_ib_mcg_work_handler() local 750 struct mcast_group *group = NULL, *cur_group, *n; search_relocate_mgid0_group() local 813 struct mcast_group *group, *cur_group; acquire_group() local 871 struct mcast_group *group = req->group; queue_req() local 888 struct mcast_group *group; mlx4_ib_mcg_demux_handler() local 937 struct mcast_group *group; mlx4_ib_mcg_multiplex_handler() local 993 struct mcast_group *group = sysfs_show_group() local 1070 force_clean_group(struct mcast_group * group) force_clean_group() argument 1087 struct mcast_group *group; _mlx4_ib_mcg_port_cleanup() local 1173 clear_pending_reqs(struct mcast_group * group,int vf) clear_pending_reqs() argument 1206 push_deleteing_req(struct mcast_group * group,int slave) push_deleteing_req() argument 1237 struct mcast_group *group; clean_vf_mcast() local [all...] |
/linux/fs/notify/ |
H A D | group.c | 20 * Final freeing of a group 22 static void fsnotify_final_destroy_group(struct fsnotify_group *group) in fsnotify_final_destroy_group() argument 24 if (group->ops->free_group_priv) in fsnotify_final_destroy_group() 25 group->ops->free_group_priv(group); in fsnotify_final_destroy_group() 27 mem_cgroup_put(group->memcg); in fsnotify_final_destroy_group() 28 mutex_destroy(&group->mark_mutex); in fsnotify_final_destroy_group() 30 kfree(group); in fsnotify_final_destroy_group() 34 * Stop queueing new events for this group. Once this function returns 35 * fsnotify_add_event() will not add any new events to the group' 37 fsnotify_group_stop_queueing(struct fsnotify_group * group) fsnotify_group_stop_queueing() argument 50 fsnotify_destroy_group(struct fsnotify_group * group) fsnotify_destroy_group() argument 99 fsnotify_get_group(struct fsnotify_group * group) fsnotify_get_group() argument 107 fsnotify_put_group(struct fsnotify_group * group) fsnotify_put_group() argument 118 struct fsnotify_group *group; __fsnotify_alloc_group() local 157 struct fsnotify_group *group = file->private_data; fsnotify_fasync() local [all...] |
H A D | notification.c | 7 * Basic idea behind the notification queue: An fsnotify group (like inotify) 10 * event to the group notify queue. Since a single event might need to be on 11 * multiple group's notification queues we can't add the event directly to each 17 * another group a new event_holder (from fsnotify_event_holder_cachep) will be 50 void fsnotify_destroy_event(struct fsnotify_group *group, in fsnotify_destroy_event() argument 53 /* Overflow events are per-group and we don't want to free them */ in fsnotify_destroy_event() 54 if (!event || event == group->overflow_event) in fsnotify_destroy_event() 63 spin_lock(&group->notification_lock); in fsnotify_destroy_event() 65 spin_unlock(&group->notification_lock); in fsnotify_destroy_event() 67 group in fsnotify_destroy_event() 81 fsnotify_insert_event(struct fsnotify_group * group,struct fsnotify_event * event,int (* merge)(struct fsnotify_group *,struct fsnotify_event *),void (* insert)(struct fsnotify_group *,struct fsnotify_event *)) fsnotify_insert_event() argument 132 fsnotify_remove_queued_event(struct fsnotify_group * group,struct fsnotify_event * event) fsnotify_remove_queued_event() argument 148 fsnotify_peek_first_event(struct fsnotify_group * group) fsnotify_peek_first_event() argument 163 fsnotify_remove_first_event(struct fsnotify_group * group) fsnotify_remove_first_event() argument 181 fsnotify_flush_notify(struct fsnotify_group * group) fsnotify_flush_notify() argument [all...] |
/linux/drivers/iommu/ |
H A D | iommu.c | 49 /* Tags used with xa_tag_pointer() in group->pasid_array */ 77 #define for_each_group_device(group, pos) \ argument 78 list_for_each_entry(pos, &(group)->devices, list) 82 ssize_t (*show)(struct iommu_group *group, char *buf); 83 ssize_t (*store)(struct iommu_group *group, 105 struct iommu_group *group); 114 static int __iommu_device_set_domain(struct iommu_group *group, 118 static int __iommu_group_set_domain_internal(struct iommu_group *group, 121 static int __iommu_group_set_domain(struct iommu_group *group, in __iommu_group_set_domain() argument 124 return __iommu_group_set_domain_internal(group, new_domai in __iommu_group_set_domain() 126 __iommu_group_set_domain_nofail(struct iommu_group * group,struct iommu_domain * new_domain) __iommu_group_set_domain_nofail() argument 417 struct iommu_group *group; iommu_init_device() local 495 struct iommu_group *group = dev->iommu_group; iommu_deinit_device() local 561 struct iommu_group *group; __iommu_probe_device() local 658 __iommu_group_free_device(struct iommu_group * group,struct group_device * grp_dev) __iommu_group_free_device() argument 684 struct iommu_group *group = dev->iommu_group; __iommu_group_remove_device() local 711 struct iommu_group *group = dev->iommu_group; iommu_release_device() local 760 struct iommu_group *group = to_iommu_group(kobj); iommu_group_attr_show() local 773 struct iommu_group *group = to_iommu_group(kobj); iommu_group_attr_store() local 786 iommu_group_create_file(struct iommu_group * group,struct iommu_group_attribute * attr) iommu_group_create_file() argument 792 iommu_group_remove_file(struct iommu_group * group,struct iommu_group_attribute * attr) iommu_group_remove_file() argument 798 iommu_group_show_name(struct iommu_group * group,char * buf) iommu_group_show_name() argument 879 iommu_get_group_resv_regions(struct iommu_group * group,struct list_head * head) iommu_get_group_resv_regions() argument 908 iommu_group_show_resv_regions(struct iommu_group * group,char * buf) iommu_group_show_resv_regions() argument 930 iommu_group_show_type(struct iommu_group * group,char * buf) iommu_group_show_type() argument 970 struct iommu_group *group = to_iommu_group(kobj); iommu_group_release() local 1005 struct iommu_group *group; iommu_group_alloc() local 1072 iommu_group_get_iommudata(struct iommu_group * group) iommu_group_get_iommudata() argument 1088 iommu_group_set_iommudata(struct iommu_group * group,void * iommu_data,void (* release)(void * iommu_data)) iommu_group_set_iommudata() argument 1104 iommu_group_set_name(struct iommu_group * group,const char * name) iommu_group_set_name() argument 1194 iommu_group_alloc_device(struct iommu_group * group,struct device * dev) iommu_group_alloc_device() argument 1257 iommu_group_add_device(struct iommu_group * group,struct device * dev) iommu_group_add_device() argument 1284 struct iommu_group *group = dev->iommu_group; iommu_group_remove_device() local 1307 struct iommu_group *group = dev->iommu_group; iommu_group_mutex_assert() local 1314 iommu_group_first_dev(struct iommu_group * group) iommu_group_first_dev() argument 1331 iommu_group_for_each_dev(struct iommu_group * group,void * data,int (* fn)(struct device *,void *)) iommu_group_for_each_dev() argument 1359 struct iommu_group *group = dev->iommu_group; iommu_group_get() local 1375 iommu_group_ref_get(struct iommu_group * group) iommu_group_ref_get() argument 1389 iommu_group_put(struct iommu_group * group) iommu_group_put() argument 1402 iommu_group_id(struct iommu_group * group) iommu_group_id() argument 1431 struct iommu_group *group; get_pci_function_alias_group() local 1465 struct iommu_group *group; get_pci_alias_group() local 1499 struct iommu_group *group; global() member 1536 struct iommu_group *group; generic_single_device_group() local 1556 struct iommu_group *group = NULL; pci_device_group() local 1619 struct iommu_group *group; fsl_mc_device_group() local 1649 __iommu_group_alloc_default_domain(struct iommu_group * group,int req_type) __iommu_group_alloc_default_domain() argument 1686 iommu_group_alloc_default_domain(struct iommu_group * group,int req_type) iommu_group_alloc_default_domain() argument 1724 iommu_group_default_domain(struct iommu_group * group) iommu_group_default_domain() argument 1765 iommu_get_def_domain_type(struct iommu_group * group,struct device * dev,int cur_type) iommu_get_def_domain_type() argument 1807 iommu_get_default_domain_type(struct iommu_group * group,int target_type) iommu_get_default_domain_type() argument 1884 struct iommu_group *group, *next; bus_iommu_probe() local 1960 iommu_group_has_isolated_msi(struct iommu_group * group) iommu_group_has_isolated_msi() argument 2079 __iommu_group_set_core_domain(struct iommu_group * group) __iommu_group_set_core_domain() argument 2122 struct iommu_group *group = dev->iommu_group; iommu_attach_device() local 2156 struct iommu_group *group = dev->iommu_group; iommu_detach_device() local 2175 struct iommu_group *group = dev->iommu_group; iommu_get_domain_for_dev() local 2218 __iommu_attach_group(struct iommu_domain * domain,struct iommu_group * group) __iommu_attach_group() argument 2246 iommu_attach_group(struct iommu_domain * domain,struct iommu_group * group) iommu_attach_group() argument 2258 __iommu_device_set_domain(struct iommu_group * group,struct device * dev,struct iommu_domain * new_domain,unsigned int flags) __iommu_device_set_domain() argument 2317 __iommu_group_set_domain_internal(struct iommu_group * group,struct iommu_domain * new_domain,unsigned int flags) __iommu_group_set_domain_internal() argument 2383 iommu_detach_group(struct iommu_domain * domain,struct iommu_group * group) iommu_detach_group() argument 2954 iommu_setup_default_domain(struct iommu_group * group,int target_type) iommu_setup_default_domain() argument 3059 iommu_group_store_type(struct iommu_group * group,const char * buf,size_t count) iommu_group_store_type() argument 3125 struct iommu_group *group = dev->iommu_group; iommu_device_use_default_domain() local 3163 struct iommu_group *group = dev->iommu_group; iommu_device_unuse_default_domain() local 3175 __iommu_group_alloc_blocking_domain(struct iommu_group * group) __iommu_group_alloc_blocking_domain() argument 3200 __iommu_take_dma_ownership(struct iommu_group * group,void * owner) __iommu_take_dma_ownership() argument 3229 iommu_group_claim_dma_owner(struct iommu_group * group,void * owner) iommu_group_claim_dma_owner() argument 3262 struct iommu_group *group = dev->iommu_group; iommu_device_claim_dma_owner() local 3288 __iommu_release_dma_ownership(struct iommu_group * group) __iommu_release_dma_ownership() argument 3305 iommu_group_release_dma_owner(struct iommu_group * group) iommu_group_release_dma_owner() argument 3322 struct iommu_group *group = dev->iommu_group; iommu_device_release_dma_owner() local 3340 iommu_group_dma_owner_claimed(struct iommu_group * group) iommu_group_dma_owner_claimed() argument 3363 __iommu_set_group_pasid(struct iommu_domain * domain,struct iommu_group * group,ioasid_t pasid,struct iommu_domain * old) __iommu_set_group_pasid() argument 3401 __iommu_remove_group_pasid(struct iommu_group * group,ioasid_t pasid,struct iommu_domain * domain) __iommu_remove_group_pasid() argument 3430 struct iommu_group *group = dev->iommu_group; iommu_attach_device_pasid() local 3517 struct iommu_group *group = dev->iommu_group; iommu_replace_device_pasid() local 3599 struct iommu_group *group = dev->iommu_group; iommu_detach_device_pasid() local 3650 iommu_attach_handle_get(struct iommu_group * group,ioasid_t pasid,unsigned int type) iommu_attach_handle_get() argument 3686 iommu_attach_group_handle(struct iommu_domain * domain,struct iommu_group * group,struct iommu_attach_handle * handle) iommu_attach_group_handle() argument 3732 iommu_detach_group_handle(struct iommu_domain * domain,struct iommu_group * group) iommu_detach_group_handle() argument 3757 iommu_replace_group_handle(struct iommu_group * group,struct iommu_domain * new_domain,struct iommu_attach_handle * handle) iommu_replace_group_handle() argument 3806 struct iommu_group *group = dev->iommu_group; iommu_dma_prepare_msi() local [all...] |
/linux/drivers/gpu/drm/xe/ |
H A D | xe_hw_engine_group.c | 19 struct xe_hw_engine_group *group = container_of(w, struct xe_hw_engine_group, resume_work); in hw_engine_group_resume_lr_jobs_func() local 23 err = xe_hw_engine_group_get_mode(group, EXEC_MODE_LR, &previous_mode); in hw_engine_group_resume_lr_jobs_func() 30 list_for_each_entry(q, &group->exec_queue_list, hw_engine_group_link) { in hw_engine_group_resume_lr_jobs_func() 38 xe_hw_engine_group_put(group); in hw_engine_group_resume_lr_jobs_func() 44 struct xe_hw_engine_group *group; in hw_engine_group_alloc() local 47 group = drmm_kzalloc(&xe->drm, sizeof(*group), GFP_KERNEL); in hw_engine_group_alloc() 48 if (!group) in hw_engine_group_alloc() 51 group->resume_wq = alloc_workqueue("xe-resume-lr-jobs-wq", 0, 0); in hw_engine_group_alloc() 52 if (!group in hw_engine_group_alloc() 122 xe_hw_engine_group_add_exec_queue(struct xe_hw_engine_group * group,struct xe_exec_queue * q) xe_hw_engine_group_add_exec_queue() argument 163 xe_hw_engine_group_del_exec_queue(struct xe_hw_engine_group * group,struct xe_exec_queue * q) xe_hw_engine_group_del_exec_queue() argument 183 xe_hw_engine_group_resume_faulting_lr_jobs(struct xe_hw_engine_group * group) xe_hw_engine_group_resume_faulting_lr_jobs() argument 194 xe_hw_engine_group_suspend_faulting_lr_jobs(struct xe_hw_engine_group * group) xe_hw_engine_group_suspend_faulting_lr_jobs() argument 239 xe_hw_engine_group_wait_for_dma_fence_jobs(struct xe_hw_engine_group * group) xe_hw_engine_group_wait_for_dma_fence_jobs() argument 262 switch_mode(struct xe_hw_engine_group * group) switch_mode() argument 296 xe_hw_engine_group_get_mode(struct xe_hw_engine_group * group,enum xe_hw_engine_group_execution_mode new_mode,enum xe_hw_engine_group_execution_mode * previous_mode) xe_hw_engine_group_get_mode() argument 331 xe_hw_engine_group_put(struct xe_hw_engine_group * group) xe_hw_engine_group_put() argument [all...] |
/linux/arch/sparc/kernel/ |
H A D | hvapi.c | 19 unsigned long group; member 28 { .group = HV_GRP_SUN4V, .flags = FLAG_PRE_API }, 29 { .group = HV_GRP_CORE, .flags = FLAG_PRE_API }, 30 { .group = HV_GRP_INTR, }, 31 { .group = HV_GRP_SOFT_STATE, }, 32 { .group = HV_GRP_TM, }, 33 { .group = HV_GRP_PCI, .flags = FLAG_PRE_API }, 34 { .group = HV_GRP_LDOM, }, 35 { .group = HV_GRP_SVC_CHAN, .flags = FLAG_PRE_API }, 36 { .group 59 __get_info(unsigned long group) __get_info() argument 96 sun4v_hvapi_register(unsigned long group,unsigned long major,unsigned long * minor) sun4v_hvapi_register() argument 147 sun4v_hvapi_unregister(unsigned long group) sun4v_hvapi_unregister() argument 160 sun4v_hvapi_get(unsigned long group,unsigned long * major,unsigned long * minor) sun4v_hvapi_get() argument 184 unsigned long group, major, minor; sun4v_hvapi_init() local [all...] |
/linux/drivers/infiniband/core/ |
H A D | multicast.c | 117 struct mcast_group *group; member 133 struct mcast_group *group; in mcast_find() local 137 group = rb_entry(node, struct mcast_group, node); in mcast_find() 138 ret = memcmp(mgid->raw, group->rec.mgid.raw, sizeof *mgid); in mcast_find() 140 return group; in mcast_find() 151 struct mcast_group *group, in mcast_insert() argument 163 ret = memcmp(group->rec.mgid.raw, cur_group->rec.mgid.raw, in mcast_insert() 164 sizeof group->rec.mgid); in mcast_insert() 174 rb_link_node(&group->node, parent, link); in mcast_insert() 175 rb_insert_color(&group in mcast_insert() 185 release_group(struct mcast_group * group) release_group() argument 208 struct mcast_group *group = member->group; queue_join() local 228 adjust_membership(struct mcast_group * group,u8 join_state,int inc) adjust_membership() argument 243 get_leave_state(struct mcast_group * group) get_leave_state() argument 330 send_join(struct mcast_group * group,struct mcast_member * member) send_join() argument 345 send_leave(struct mcast_group * group,u8 leave_state) send_leave() argument 365 join_group(struct mcast_group * group,struct mcast_member * member,u8 join_state) join_group() argument 376 fail_join(struct mcast_group * group,struct mcast_member * member,int status) fail_join() argument 385 process_group_error(struct mcast_group * group) process_group_error() argument 426 struct mcast_group *group; mcast_work_handler() local 491 process_join_error(struct mcast_group * group,int status) process_join_error() argument 514 struct mcast_group *group = context; join_handler() local 548 struct mcast_group *group = context; leave_handler() local 560 struct mcast_group *group, *cur_group; acquire_group() local 665 struct mcast_group *group; ib_sa_free_multicast() local 698 struct mcast_group *group; ib_sa_get_mcmember_rec() local 774 struct mcast_group *group; mcast_groups_event() local [all...] |
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_sched.c | 42 * The scheduling happens at the scheduling group level, each group 51 * rotating the groups passed to the firmware so every group gets 67 * queue ring-buffer, and the group is scheduled for execution if it 70 * Kernel-side group scheduling is timeslice-based. When we have less 73 * groups than slots, we let each group a chance to execute stuff for 75 * to schedule. The group selection algorithm is based on 80 * group/queue state that would be based on information we wouldn't have 82 * reason we don't do 'cooperative' scheduling (encoding FW group slot 84 * drm_gpu_scheduler::prepare_job() hook, and treating group rotatio 106 struct panthor_group *group; global() member 723 group_queue_work(group,wname) global() argument 787 struct panthor_group *group; global() member 852 panthor_queue_get_syncwait_obj(struct panthor_group * group,struct panthor_queue * queue) panthor_queue_get_syncwait_obj() argument 884 group_free_queue(struct panthor_group * group,struct panthor_queue * queue) group_free_queue() argument 909 struct panthor_group *group = container_of(work, group_release_work() local 927 struct panthor_group *group = container_of(kref, group_release() local 939 group_put(struct panthor_group * group) group_put() argument 946 group_get(struct panthor_group * group) group_get() argument 962 group_bind_locked(struct panthor_group * group,u32 csg_id) group_bind_locked() argument 1004 group_unbind_locked(struct panthor_group * group) group_unbind_locked() argument 1093 struct panthor_group *group = ptdev->scheduler->csg_slots[csg_id].group; cs_slot_reset_locked() local 1146 struct panthor_group *group = ptdev->scheduler->csg_slots[csg_id].group; cs_slot_sync_queue_state_locked() local 1198 struct panthor_group *group = csg_slot->group; csg_slot_sync_queues_state_locked() local 1217 struct panthor_group *group; csg_slot_sync_state_locked() local 1290 struct panthor_group *group; csg_slot_prog_locked() local 1347 struct panthor_group *group = csg_slot->group; cs_slot_process_fatal_event_locked() local 1389 struct panthor_group *group = csg_slot->group; cs_slot_process_fault_event_locked() local 1431 group_process_tiler_oom(struct panthor_group * group,u32 cs_id) group_process_tiler_oom() argument 1519 struct panthor_group *group = group_tiler_oom_work() local 1539 struct panthor_group *group = csg_slot->group; cs_slot_process_tiler_oom_event_locked() local 1616 struct panthor_group *group = csg_slot->group; csg_slot_sync_update_locked() local 1631 struct panthor_group *group = csg_slot->group; csg_slot_process_progress_timer_event_locked() local 1892 group_is_idle(struct panthor_group * group) group_is_idle() argument 1905 group_can_run(struct panthor_group * group) group_can_run() argument 1920 struct panthor_group *group, *tmp; tick_ctx_pick_groups_from_list() local 1964 tick_ctx_insert_old_group(struct panthor_scheduler * sched,struct panthor_sched_tick_ctx * ctx,struct panthor_group * group,bool full_tick) tick_ctx_insert_old_group() argument 2017 struct panthor_group *group = csg_slot->group; tick_ctx_init() local 2051 group_term_post_processing(struct panthor_group * group) group_term_post_processing() argument 2101 struct panthor_group *group = group_term_work() local 2113 struct panthor_group *group, *tmp; tick_ctx_cleanup() local 2164 struct panthor_group *group, *tmp; tick_ctx_apply() local 2456 panthor_queue_eval_syncwait(struct panthor_group * group,u8 queue_idx) panthor_queue_eval_syncwait() argument 2490 struct panthor_group *group, *tmp; sync_upd_work() local 2530 group_schedule_locked(struct panthor_group * group,u32 queue_mask) group_schedule_locked() argument 2608 panthor_group_stop(struct panthor_group * group) panthor_group_stop() argument 2621 panthor_group_start(struct panthor_group * group) panthor_group_start() argument 2671 struct panthor_group *group; panthor_sched_suspend() local 2793 struct panthor_group *group, *group_tmp; panthor_sched_pre_reset() local 2828 struct panthor_group *group, *group_tmp; panthor_sched_post_reset() local 2857 struct panthor_group *group = job->group; update_fdinfo_stats() local 2874 struct panthor_group *group; panthor_fdinfo_gather_group_samples() local 2893 struct panthor_group *group = group_sync_upd_work() local 2987 struct panthor_group *group = job->group; get_job_cs_params() local 3116 struct panthor_group *group = job->group; queue_run_job() local 3216 struct panthor_group *group = job->group; queue_timedout_job() local 3287 group_create_queue(struct panthor_group * group,const struct drm_panthor_queue_create * args) group_create_queue() argument 3393 add_group_kbo_sizes(struct panthor_device * ptdev,struct panthor_group * group) add_group_kbo_sizes() argument 3425 struct panthor_group *group = NULL; panthor_group_create() local 3555 struct panthor_group *group; panthor_group_destroy() local 3589 struct panthor_group *group; group_from_handle() local 3604 struct panthor_group *group; panthor_group_get_state() local 3646 struct panthor_group *group; panthor_group_pool_destroy() local 3672 struct panthor_group *group; panthor_fdinfo_gather_group_mem_info() local [all...] |
/linux/kernel/sched/ |
H A D | psi.c | 205 static void group_init(struct psi_group *group) in group_init() argument 207 group->enabled = true; in group_init() 208 group->avg_last_update = sched_clock(); in group_init() 209 group->avg_next_update = group->avg_last_update + psi_period; in group_init() 210 mutex_init(&group->avgs_lock); in group_init() 213 INIT_LIST_HEAD(&group->avg_triggers); in group_init() 214 memset(group->avg_nr_triggers, 0, sizeof(group->avg_nr_triggers)); in group_init() 215 INIT_DELAYED_WORK(&group in group_init() 271 get_recent_times(struct psi_group * group,int cpu,enum psi_aggregators aggregator,u32 * times,u32 * pchanged_states) get_recent_times() argument 362 collect_percpu_times(struct psi_group * group,enum psi_aggregators aggregator,u32 * pchanged_states) collect_percpu_times() argument 463 update_triggers(struct psi_group * group,u64 now,enum psi_aggregators aggregator) update_triggers() argument 525 update_averages(struct psi_group * group,u64 now) update_averages() argument 581 struct psi_group *group; psi_avgs_work() local 613 init_rtpoll_triggers(struct psi_group * group,u64 now) init_rtpoll_triggers() argument 626 psi_schedule_rtpoll_work(struct psi_group * group,unsigned long delay,bool force) psi_schedule_rtpoll_work() argument 653 psi_rtpoll_work(struct psi_group * group) psi_rtpoll_work() argument 740 struct psi_group *group = (struct psi_group *)data; psi_rtpoll_worker() local 758 struct psi_group *group = timer_container_of(group, t, rtpoll_timer); poll_timer_fn() local 793 for_each_group(iter,group) global() argument 796 psi_group_change(struct psi_group * group,int cpu,unsigned int clear,unsigned int set,u64 now,bool wake_clock) psi_group_change() argument 1210 psi_cgroup_restart(struct psi_group * group) psi_cgroup_restart() argument 1245 psi_show(struct seq_file * m,struct psi_group * group,enum psi_res res) psi_show() argument 1295 psi_trigger_create(struct psi_group * group,char * buf,enum psi_res res,struct file * file,struct kernfs_open_file * of) psi_trigger_create() argument 1399 struct psi_group *group; psi_trigger_destroy() local [all...] |
/linux/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_sdlb.c | 60 static void sparx5_sdlb_group_disable(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_disable() argument 64 ANA_AC_SDLB_PUP_CTRL(group)); in sparx5_sdlb_group_disable() 67 static void sparx5_sdlb_group_enable(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_enable() argument 71 ANA_AC_SDLB_PUP_CTRL(group)); in sparx5_sdlb_group_enable() 74 static u32 sparx5_sdlb_group_get_first(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_get_first() argument 78 val = spx5_rd(sparx5, ANA_AC_SDLB_XLB_START(group)); in sparx5_sdlb_group_get_first() 83 static u32 sparx5_sdlb_group_get_next(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_get_next() argument 93 static bool sparx5_sdlb_group_is_first(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_is_first() argument 96 return lb == sparx5_sdlb_group_get_first(sparx5, group); in sparx5_sdlb_group_is_first() 99 static bool sparx5_sdlb_group_is_last(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_is_last() argument 105 sparx5_sdlb_group_is_empty(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_is_empty() argument 114 sparx5_sdlb_group_get_last(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_get_last() argument 129 sparx5_sdlb_group_is_singular(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_is_singular() argument 138 sparx5_sdlb_group_get_adjacent(struct sparx5 * sparx5,u32 group,u32 idx,u32 * prev,u32 * next,u32 * first) sparx5_sdlb_group_get_adjacent() argument 163 sparx5_sdlb_group_get_count(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_get_count() argument 183 const struct sparx5_sdlb_group *group; sparx5_sdlb_group_get_by_rate() local 208 sparx5_sdlb_group_get_by_index(struct sparx5 * sparx5,u32 idx,u32 * group) sparx5_sdlb_group_get_by_index() argument 236 sparx5_sdlb_group_link(struct sparx5 * sparx5,u32 group,u32 idx,u32 first,u32 next,bool empty) sparx5_sdlb_group_link() argument 260 sparx5_sdlb_group_add(struct sparx5 * sparx5,u32 group,u32 idx) sparx5_sdlb_group_add() argument 275 sparx5_sdlb_group_del(struct sparx5 * sparx5,u32 group,u32 idx) sparx5_sdlb_group_del() argument 310 struct sparx5_sdlb_group *group; sparx5_sdlb_group_init() local [all...] |
H A D | sparx5_qos.c | 84 static u32 sparx5_lg_get_leak_time(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_leak_time() argument 88 value = spx5_rd(sparx5, HSCH_HSCH_TIMER_CFG(layer, group)); in sparx5_lg_get_leak_time() 92 static void sparx5_lg_set_leak_time(struct sparx5 *sparx5, u32 layer, u32 group, in sparx5_lg_set_leak_time() argument 96 HSCH_HSCH_TIMER_CFG(layer, group)); in sparx5_lg_set_leak_time() 99 static u32 sparx5_lg_get_first(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_first() argument 103 value = spx5_rd(sparx5, HSCH_HSCH_LEAK_CFG(layer, group)); in sparx5_lg_get_first() 107 static u32 sparx5_lg_get_next(struct sparx5 *sparx5, u32 layer, u32 group, in sparx5_lg_get_next() argument 117 static u32 sparx5_lg_get_last(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_last() argument 121 itr = sparx5_lg_get_first(sparx5, layer, group); in sparx5_lg_get_last() 124 next = sparx5_lg_get_next(sparx5, layer, group, it in sparx5_lg_get_last() 132 sparx5_lg_is_last(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx) sparx5_lg_is_last() argument 138 sparx5_lg_is_first(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx) sparx5_lg_is_first() argument 144 sparx5_lg_is_empty(struct sparx5 * sparx5,u32 layer,u32 group) sparx5_lg_is_empty() argument 149 sparx5_lg_is_singular(struct sparx5 * sparx5,u32 layer,u32 group) sparx5_lg_is_singular() argument 158 sparx5_lg_enable(struct sparx5 * sparx5,u32 layer,u32 group,u32 leak_time) sparx5_lg_enable() argument 164 sparx5_lg_disable(struct sparx5 * sparx5,u32 layer,u32 group) sparx5_lg_disable() argument 170 sparx5_lg_get_group_by_index(struct sparx5 * sparx5,u32 layer,u32 idx,u32 * group) sparx5_lg_get_group_by_index() argument 198 sparx5_lg_get_group_by_rate(u32 layer,u32 rate,u32 * group) sparx5_lg_get_group_by_rate() argument 215 sparx5_lg_get_adjacent(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx,u32 * prev,u32 * next,u32 * first) sparx5_lg_get_adjacent() argument 241 sparx5_lg_conf_set(struct sparx5 * sparx5,u32 layer,u32 group,u32 se_first,u32 idx,u32 idx_next,bool empty) sparx5_lg_conf_set() argument 271 sparx5_lg_del(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx) sparx5_lg_del() argument 332 sparx5_shaper_conf_set(struct sparx5_port * port,const struct sparx5_shaper * sh,u32 layer,u32 idx,u32 group) sparx5_shaper_conf_set() argument 506 u32 group; sparx5_tc_tbf_add() local 535 u32 group; sparx5_tc_tbf_del() local [all...] |
/linux/drivers/gpio/ |
H A D | gpio-lpc32xx.c | 168 static inline u32 gpreg_read(struct lpc32xx_gpio_chip *group, unsigned long offset) in gpreg_read() argument 170 return __raw_readl(group->reg_base + offset); in gpreg_read() 173 static inline void gpreg_write(struct lpc32xx_gpio_chip *group, u32 val, unsigned long offset) in gpreg_write() argument 175 __raw_writel(val, group->reg_base + offset); in gpreg_write() 178 static void __set_gpio_dir_p012(struct lpc32xx_gpio_chip *group, in __set_gpio_dir_p012() argument 182 gpreg_write(group, GPIO012_PIN_TO_BIT(pin), in __set_gpio_dir_p012() 183 group->gpio_grp->dir_clr); in __set_gpio_dir_p012() 185 gpreg_write(group, GPIO012_PIN_TO_BIT(pin), in __set_gpio_dir_p012() 186 group->gpio_grp->dir_set); in __set_gpio_dir_p012() 189 static void __set_gpio_dir_p3(struct lpc32xx_gpio_chip *group, in __set_gpio_dir_p3() argument 200 __set_gpio_level_p012(struct lpc32xx_gpio_chip * group,unsigned pin,int high) __set_gpio_level_p012() argument 211 __set_gpio_level_p3(struct lpc32xx_gpio_chip * group,unsigned pin,int high) __set_gpio_level_p3() argument 222 __set_gpo_level_p3(struct lpc32xx_gpio_chip * group,unsigned pin,int high) __set_gpo_level_p3() argument 231 __get_gpio_state_p012(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpio_state_p012() argument 238 __get_gpio_state_p3(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpio_state_p3() argument 250 __get_gpi_state_p3(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpi_state_p3() argument 256 __get_gpo_state_p3(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpo_state_p3() argument 268 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_input_p012() local 278 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_input_p3() local 293 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_get_value_p012() local 300 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_get_value_p3() local 307 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpi_get_value() local 315 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_output_p012() local 326 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_output_p3() local 337 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_out_always() local 346 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_set_value_p012() local 356 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_set_value_p3() local 366 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpo_set_value() local 375 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpo_get_value() local [all...] |
/linux/kernel/time/ |
H A D | timer_migration.c | 22 * lowest level group contains CPUs, the next level groups of CPU groups 43 * Each group has a designated migrator CPU/group as long as a CPU/group is 44 * active in the group. This designated role is necessary to avoid that all 45 * active CPUs in a group try to migrate expired timers from other CPUs, 48 * When a CPU is awake, it checks in it's own timer tick the group 53 * If it finds expired timers in one of the group queues it pulls them over 55 * group and the parent groups if required. 60 * CPU does not queue an event in the LVL0 group 437 tmigr_check_migrator(struct tmigr_group * group,u8 childmask) tmigr_check_migrator() argument 449 tmigr_check_migrator_and_lonely(struct tmigr_group * group,u8 childmask) tmigr_check_migrator_and_lonely() argument 466 tmigr_check_lonely(struct tmigr_group * group) tmigr_check_lonely() argument 528 struct tmigr_group *child = NULL, *group = tmc->tmgroup; __walk_groups() local 560 tmigr_next_groupevt(struct tmigr_group * group) tmigr_next_groupevt() argument 593 tmigr_next_expired_groupevt(struct tmigr_group * group,u64 now) tmigr_next_expired_groupevt() argument 610 tmigr_next_groupevt_expires(struct tmigr_group * group) tmigr_next_groupevt_expires() argument 622 tmigr_active_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_active_up() argument 722 tmigr_update_events(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_update_events() argument 869 tmigr_new_timer_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_new_timer_up() argument 1002 tmigr_handle_remote_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_handle_remote_up() argument 1101 tmigr_requires_handle_remote_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_requires_handle_remote_up() argument 1244 tmigr_inactive_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_inactive_up() argument 1396 struct tmigr_group *group = tmc->tmgroup; tmigr_quick_check() local 1484 tmigr_init_group(struct tmigr_group * group,unsigned int lvl,int node) tmigr_init_group() argument 1526 struct tmigr_group *tmp, *group = NULL; tmigr_get_group() local 1644 struct tmigr_group *group, *child, **stack; tmigr_setup_groups() local [all...] |
/linux/drivers/clk/renesas/ |
H A D | clk-mstp.c | 32 * struct mstp_clock_group - MSTP gating clocks group 34 * @data: clock specifier translation for clocks in this group 39 * @clks: clocks in this group 54 * @group: MSTP clocks group 59 struct mstp_clock_group *group; member 64 static inline u32 cpg_mstp_read(struct mstp_clock_group *group, in cpg_mstp_read() argument 67 return group->width_8bit ? readb(reg) : readl(reg); in cpg_mstp_read() 70 static inline void cpg_mstp_write(struct mstp_clock_group *group, u32 val, in cpg_mstp_write() argument 73 group in cpg_mstp_write() 79 struct mstp_clock_group *group = clock->group; cpg_mstp_clock_endisable() local 128 struct mstp_clock_group *group = clock->group; cpg_mstp_clock_is_enabled() local 147 cpg_mstp_clock_register(const char * name,const char * parent_name,unsigned int index,struct mstp_clock_group * group) cpg_mstp_clock_register() argument 182 struct mstp_clock_group *group; cpg_mstp_clocks_init() local [all...] |
/linux/fs/notify/inotify/ |
H A D | inotify_user.c | 141 struct fsnotify_group *group = file->private_data; in inotify_poll() local 144 poll_wait(file, &group->notification_waitq, wait); in inotify_poll() 145 spin_lock(&group->notification_lock); in inotify_poll() 146 if (!fsnotify_notify_queue_is_empty(group)) in inotify_poll() 148 spin_unlock(&group->notification_lock); in inotify_poll() 168 * Called with the group->notification_lock held. 170 static struct fsnotify_event *get_one_event(struct fsnotify_group *group, in get_one_event() argument 176 event = fsnotify_peek_first_event(group); in get_one_event() 180 pr_debug("%s: group=%p event=%p\n", __func__, group, even in get_one_event() 199 copy_event_to_user(struct fsnotify_group * group,struct fsnotify_event * fsn_event,char __user * buf) copy_event_to_user() argument 252 struct fsnotify_group *group; inotify_read() local 303 struct fsnotify_group *group = file->private_data; inotify_release() local 316 struct fsnotify_group *group; inotify_ioctl() local 414 inotify_idr_find_locked(struct fsnotify_group * group,int wd) inotify_idr_find_locked() argument 435 inotify_idr_find(struct fsnotify_group * group,int wd) inotify_idr_find() argument 452 inotify_remove_from_idr(struct fsnotify_group * group,struct inotify_inode_mark * i_mark) inotify_remove_from_idr() argument 521 inotify_ignored_and_remove_idr(struct fsnotify_mark * fsn_mark,struct fsnotify_group * group) inotify_ignored_and_remove_idr() argument 536 inotify_update_existing_watch(struct fsnotify_group * group,struct inode * inode,u32 arg) inotify_update_existing_watch() argument 590 inotify_new_watch(struct fsnotify_group * group,struct inode * inode,u32 arg) inotify_new_watch() argument 638 inotify_update_watch(struct fsnotify_group * group,struct inode * inode,u32 arg) inotify_update_watch() argument 655 struct fsnotify_group *group; inotify_new_group() local 696 struct fsnotify_group *group; do_inotify_init() local 732 struct fsnotify_group *group; SYSCALL_DEFINE3() local 788 struct fsnotify_group *group; SYSCALL_DEFINE2() local [all...] |
H A D | inotify_fsnotify.c | 49 static int inotify_merge(struct fsnotify_group *group, in inotify_merge() argument 52 struct list_head *list = &group->notification_list; in inotify_merge() 66 struct fsnotify_group *group = inode_mark->group; in inotify_handle_inode_event() local 77 pr_debug("%s: group=%p mark=%p mask=%x\n", __func__, group, inode_mark, in inotify_handle_inode_event() 95 old_memcg = set_active_memcg(group->memcg); in inotify_handle_inode_event() 104 fsnotify_queue_overflow(group); in inotify_handle_inode_event() 126 ret = fsnotify_add_event(group, fsn_event, inotify_merge); in inotify_handle_inode_event() 129 fsnotify_destroy_event(group, fsn_even in inotify_handle_inode_event() 138 inotify_freeing_mark(struct fsnotify_mark * fsn_mark,struct fsnotify_group * group) inotify_freeing_mark() argument 178 inotify_free_group_priv(struct fsnotify_group * group) inotify_free_group_priv() argument 187 inotify_free_event(struct fsnotify_group * group,struct fsnotify_event * fsn_event) inotify_free_event() argument [all...] |
/linux/drivers/media/platform/renesas/rcar-vin/ |
H A D | rcar-core.c | 50 * than one group for the whole system be replaced with a linked 55 * be only one group for all instances. 62 static void rvin_group_cleanup(struct rvin_group *group) in rvin_group_cleanup() argument 64 media_device_cleanup(&group->mdev); in rvin_group_cleanup() 65 mutex_destroy(&group->lock); in rvin_group_cleanup() 68 static int rvin_group_init(struct rvin_group *group, struct rvin_dev *vin, in rvin_group_init() argument 72 struct media_device *mdev = &group->mdev; in rvin_group_init() 76 mutex_init(&group->lock); in rvin_group_init() 79 group->count = 0; in rvin_group_init() 82 group in rvin_group_init() 104 struct rvin_group *group = rvin_group_release() local 122 struct rvin_group *group; rvin_group_get() local 176 struct rvin_group *group = vin->group; rvin_group_put() local 194 rvin_group_entity_to_remote_id(struct rvin_group * group,struct media_entity * entity) rvin_group_entity_to_remote_id() argument 243 struct rvin_group *group = vin->group; rvin_group_notify_unbind() local 281 struct rvin_group *group = vin->group; rvin_group_notify_bound() local 552 struct rvin_group *group = container_of(link->graph_obj.mdev, rvin_csi2_link_notify() local 670 rvin_csi2_create_link(struct rvin_group * group,unsigned int id,const struct rvin_group_route * route) rvin_csi2_create_link() argument 695 rvin_parallel_setup_links(struct rvin_group * group) rvin_parallel_setup_links() argument 732 rvin_csi2_setup_links(struct rvin_group * group) rvin_csi2_setup_links() argument 788 rvin_isp_setup_links(struct rvin_group * group) rvin_isp_setup_links() argument [all...] |
/linux/drivers/pinctrl/aspeed/ |
H A D | pinmux-aspeed.h | 144 * pins in the function's group to disable the higher-priority signals such 221 * Conversely, failing to allocate all pins in a group indicates some bits (as 222 * well as pins) required for the group's configuration will already be in use, 224 * group. 244 * the group and the function. In this way copy/paste errors cause duplicate 248 * no override errors in the pin, group and function arrays. 513 #define SIG_DESC_LIST_SYM(sig, group) sig_descs_ ## sig ## _ ## group argument 514 #define SIG_DESC_LIST_DECL(sig, group, ...) \ argument 515 static const struct aspeed_sig_desc SIG_DESC_LIST_SYM(sig, group)[] 518 SIG_EXPR_SYM(sig,group) global() argument 519 SIG_EXPR_DECL_(sig,group,func) global() argument 545 SIG_EXPR_DECL(sig,group,func,...) global() argument 555 SIG_EXPR_PTR(sig,group) global() argument 557 SIG_EXPR_LIST_SYM(sig,group) global() argument 574 SIG_EXPR_LIST_DECL(sig,group,...) global() argument 594 SIG_EXPR_LIST_ALIAS(pin,sig,group) global() argument 628 SIG_EXPR_LIST_DECL_SEMG(pin,sig,group,func,...) global() argument 650 SIG_EXPR_LIST_PTR(sig,group) global() argument 741 GROUP_SYM(group) global() argument 742 GROUP_DECL(group,...) global() argument 749 FUNC_DECL_1(func,group) global() argument [all...] |
/linux/include/trace/events/ |
H A D | timer_migration.h | 14 TP_PROTO(struct tmigr_group *group), 16 TP_ARGS(group), 19 __field( void *, group ) 25 __entry->group = group; 26 __entry->lvl = group->level; 27 __entry->numa_node = group->numa_node; 30 TP_printk("group=%p lvl=%d numa=%d", 31 __entry->group, __entry->lvl, __entry->numa_node) 58 TP_printk("group [all...] |
/linux/fs/notify/fanotify/ |
H A D | fanotify_user.c | 35 * Legacy fanotify marks limits (8192) is per group and we introduced a tunable 37 * of fanotify marks per user is <max marks per group> * <max groups per user>. 174 * With group flag FAN_REPORT_NAME, if name was not recorded in in fanotify_event_len() 199 static void fanotify_unhash_event(struct fsnotify_group *group, in fanotify_unhash_event() argument 202 assert_spin_locked(&group->notification_lock); in fanotify_unhash_event() 204 pr_debug("%s: group=%p event=%p bucket=%u\n", __func__, in fanotify_unhash_event() 205 group, event, fanotify_event_hash_bucket(group, event)); in fanotify_unhash_event() 219 static struct fanotify_event *get_one_event(struct fsnotify_group *group, in get_one_event() argument 225 unsigned int info_mode = FAN_GROUP_FLAG(group, FANOTIFY_INFO_MODE in get_one_event() 256 create_fd(struct fsnotify_group * group,const struct path * path,struct file ** file) create_fd() argument 306 finish_permission_event(struct fsnotify_group * group,struct fanotify_perm_event * event,u32 response,struct fanotify_response_info_audit_rule * friar) finish_permission_event() argument 327 process_access_response(struct fsnotify_group * group,struct fanotify_response * response_struct,const char __user * info,size_t info_len) process_access_response() argument 739 copy_event_to_user(struct fsnotify_group * group,struct fanotify_event * event,char __user * buf,size_t count) copy_event_to_user() argument 883 struct fsnotify_group *group = file->private_data; fanotify_poll() local 898 struct fsnotify_group *group; fanotify_read() local 974 struct fsnotify_group *group; fanotify_write() local 1005 struct fsnotify_group *group = file->private_data; fanotify_release() local 1060 struct fsnotify_group *group; fanotify_ioctl() local 1172 fanotify_remove_mark(struct fsnotify_group * group,void * obj,unsigned int obj_type,__u32 mask,unsigned int flags,__u32 umask) fanotify_remove_mark() argument 1270 fanotify_set_mark_fsid(struct fsnotify_group * group,struct fsnotify_mark * mark,struct fan_fsid * fsid) fanotify_set_mark_fsid() argument 1321 fanotify_add_new_mark(struct fsnotify_group * group,void * obj,unsigned int obj_type,unsigned int fan_flags,struct fan_fsid * fsid) fanotify_add_new_mark() argument 1376 fanotify_group_init_error_pool(struct fsnotify_group * group) fanotify_group_init_error_pool() argument 1422 fanotify_add_mark(struct fsnotify_group * group,void * obj,unsigned int obj_type,__u32 mask,unsigned int fan_flags,struct fan_fsid * fsid) fanotify_add_mark() argument 1503 struct fsnotify_group *group; SYSCALL_DEFINE2() local 1742 fanotify_events_supported(struct fsnotify_group * group,const struct path * path,__u64 mask,unsigned int flags) fanotify_events_supported() argument 1806 struct fsnotify_group *group; do_fanotify_mark() local [all...] |
/linux/net/psample/ |
H A D | psample.c | 41 struct psample_group *group, in psample_group_nl_fill() argument 52 ret = nla_put_u32(msg, PSAMPLE_ATTR_SAMPLE_GROUP, group->group_num); in psample_group_nl_fill() 56 ret = nla_put_u32(msg, PSAMPLE_ATTR_GROUP_REFCOUNT, group->refcount); in psample_group_nl_fill() 60 ret = nla_put_u32(msg, PSAMPLE_ATTR_GROUP_SEQ, group->seq); in psample_group_nl_fill() 75 struct psample_group *group; in psample_nl_cmd_get_group_dumpit() local 81 list_for_each_entry(group, &psample_groups_list, list) { in psample_nl_cmd_get_group_dumpit() 82 if (!net_eq(group->net, sock_net(msg->sk))) in psample_nl_cmd_get_group_dumpit() 88 err = psample_group_nl_fill(msg, group, PSAMPLE_CMD_NEW_GROUP, in psample_nl_cmd_get_group_dumpit() 123 static void psample_group_notify(struct psample_group *group, in psample_group_notify() argument 133 err = psample_group_nl_fill(msg, group, cm in psample_group_notify() 144 struct psample_group *group; psample_group_create() local 158 psample_group_destroy(struct psample_group * group) psample_group_destroy() argument 168 struct psample_group *group; psample_group_lookup() local 178 struct psample_group *group; psample_group_get() local 196 psample_group_take(struct psample_group * group) psample_group_take() argument 204 psample_group_put(struct psample_group * group) psample_group_put() argument 363 psample_sample_packet(struct psample_group * group,const struct sk_buff * skb,u32 sample_rate,const struct psample_metadata * md) psample_sample_packet() argument [all...] |
/linux/Documentation/devicetree/bindings/pinctrl/ |
H A D | marvell,armada-37xx-pinctrl.txt | 32 group: jtag 36 group sdio0 40 group emmc_nb 44 group pwm0 48 group pwm1 52 group pwm2 56 group pwm3 60 group pmic1 64 group pmic0 68 group i2c [all...] |
/linux/Documentation/filesystems/ext4/ |
H A D | blockgroup.rst | 9 The layout of a standard block group is approximately as follows (each 33 For the special case of block group 0, the first 1024 bytes are unused, 40 The ext4 driver primarily works with the superblock and the group 41 descriptors that are found in block group 0. Redundant copies of the 42 superblock and group descriptors are written to some of the block groups 45 paragraph for more details). If the group does not have a redundant 46 copy, the block group begins with the data block bitmap. Note also that 48 GDT block” space after the block group descriptors and before the start 57 As for the ordering of items in a block group, it is generally 58 established that the super block and the group descripto [all...] |