Home
last modified time | relevance | path

Searched refs:group (Results 1 – 25 of 1421) sorted by relevance

12345678910>>...57

/linux/drivers/vfio/
H A Dgroup.c26 static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group, in vfio_device_get_from_name() argument
31 mutex_lock(&group->device_lock); in vfio_device_get_from_name()
32 list_for_each_entry(it, &group->device_list, group_next) { in vfio_device_get_from_name()
50 mutex_unlock(&group->device_lock); in vfio_device_get_from_name()
58 static bool vfio_group_has_iommu(struct vfio_group *group) in vfio_group_has_iommu() argument
60 lockdep_assert_held(&group->group_lock); in vfio_group_has_iommu()
65 WARN_ON(!group->container != !group->container_users); in vfio_group_has_iommu()
67 return group->container || group in vfio_group_has_iommu()
76 vfio_group_ioctl_unset_container(struct vfio_group * group) vfio_group_ioctl_unset_container() argument
102 vfio_group_ioctl_set_container(struct vfio_group * group,int __user * arg) vfio_group_ioctl_set_container() argument
298 vfio_group_ioctl_get_device_fd(struct vfio_group * group,char __user * arg) vfio_group_ioctl_get_device_fd() argument
338 vfio_group_ioctl_get_status(struct vfio_group * group,struct vfio_group_status __user * arg) vfio_group_ioctl_get_status() argument
380 struct vfio_group *group = filep->private_data; vfio_group_fops_unl_ioctl() local
399 struct vfio_group *group = device->group; vfio_device_block_group() local
417 struct vfio_group *group = device->group; vfio_device_unblock_group() local
426 struct vfio_group *group = vfio_group_fops_open() local
468 struct vfio_group *group = filep->private_data; vfio_group_fops_release() local
503 struct vfio_group *group; vfio_group_find_from_iommu() local
520 struct vfio_group *group = container_of(dev, struct vfio_group, dev); vfio_group_release() local
533 struct vfio_group *group; vfio_group_alloc() local
570 struct vfio_group *group; vfio_create_group() local
607 struct vfio_group *group; vfio_noiommu_group_alloc() local
638 vfio_group_has_device(struct vfio_group * group,struct device * dev) vfio_group_has_device() argument
656 struct vfio_group *group; vfio_group_find_or_alloc() local
697 struct vfio_group *group; vfio_device_set_group() local
714 struct vfio_group *group = device->group; vfio_device_remove_group() local
775 struct vfio_group *group = device->group; vfio_device_group_use_iommu() local
792 struct vfio_group *group = device->group; vfio_device_group_unuse_iommu() local
810 struct vfio_group *group = file->private_data; vfio_group_from_file() local
827 struct vfio_group *group = vfio_group_from_file(file); vfio_file_iommu_group() local
856 vfio_group_enforced_coherent(struct vfio_group * group) vfio_group_enforced_coherent() argument
879 vfio_group_set_kvm(struct vfio_group * group,struct kvm * kvm) vfio_group_set_kvm() argument
895 struct vfio_group *group = vfio_group_from_file(file); vfio_file_has_dev() local
[all...]
H A Dcontainer.c143 * it's freed via kref. Must support container/group/device being
167 device->group->container->iommu_driver; in vfio_device_container_register()
171 device->group->container->iommu_data, device); in vfio_device_container_register()
177 device->group->container->iommu_driver; in vfio_device_container_unregister()
181 device->group->container->iommu_data, device); in vfio_device_container_unregister()
239 struct vfio_group *group; in __vfio_container_attach_groups() local
242 list_for_each_entry(group, &container->group_list, container_next) { in __vfio_container_attach_groups()
243 ret = driver->ops->attach_group(data, group->iommu_group, in __vfio_container_attach_groups()
244 group->type); in __vfio_container_attach_groups()
252 list_for_each_entry_continue_reverse(group, in __vfio_container_attach_groups()
417 vfio_container_attach_group(struct vfio_container * container,struct vfio_group * group) vfio_container_attach_group() argument
468 vfio_group_detach_container(struct vfio_group * group) vfio_group_detach_container() argument
503 vfio_group_use_container(struct vfio_group * group) vfio_group_use_container() argument
522 vfio_group_unuse_container(struct vfio_group * group) vfio_group_unuse_container() argument
[all...]
/linux/drivers/infiniband/hw/mlx4/
H A Dmcg.c50 #define mcg_warn_group(group, format, arg...) \ argument
52 (group)->name, group->demux->port, ## arg)
54 #define mcg_debug_group(group, format, arg...) \ argument
56 (group)->name, (group)->demux->port, ## arg)
58 #define mcg_error_group(group, format, arg...) \ argument
59 pr_err(" %16s: " format, (group)->name, ## arg)
136 struct mcast_group *group; member
144 mcg_warn_group(group, "di
166 struct mcast_group *group; mcast_find() local
184 mcast_insert(struct mlx4_ib_demux_ctx * ctx,struct mcast_group * group) mcast_insert() argument
252 send_join_to_wire(struct mcast_group * group,struct ib_sa_mad * sa_mad) send_join_to_wire() argument
279 send_leave_to_wire(struct mcast_group * group,u8 join_state) send_leave_to_wire() argument
318 send_reply_to_slave(int slave,struct mcast_group * group,struct ib_sa_mad * req_sa_mad,u16 status) send_reply_to_slave() argument
437 release_group(struct mcast_group * group,int from_timeout_handler) release_group() argument
474 adjust_membership(struct mcast_group * group,u8 join_state,int inc) adjust_membership() argument
483 get_leave_state(struct mcast_group * group) get_leave_state() argument
495 join_group(struct mcast_group * group,int slave,u8 join_mask) join_group() argument
511 leave_group(struct mcast_group * group,int slave,u8 leave_state) leave_group() argument
524 check_leave(struct mcast_group * group,int slave,u8 leave_mask) check_leave() argument
542 struct mcast_group *group; mlx4_ib_mcg_timeout_handler() local
584 handle_leave_req(struct mcast_group * group,u8 leave_mask,struct mcast_req * req) handle_leave_req() argument
605 handle_join_req(struct mcast_group * group,u8 join_mask,struct mcast_req * req) handle_join_req() argument
644 struct mcast_group *group; mlx4_ib_mcg_work_handler() local
750 struct mcast_group *group = NULL, *cur_group, *n; search_relocate_mgid0_group() local
813 struct mcast_group *group, *cur_group; acquire_group() local
871 struct mcast_group *group = req->group; queue_req() local
888 struct mcast_group *group; mlx4_ib_mcg_demux_handler() local
937 struct mcast_group *group; mlx4_ib_mcg_multiplex_handler() local
993 struct mcast_group *group = sysfs_show_group() local
1070 force_clean_group(struct mcast_group * group) force_clean_group() argument
1087 struct mcast_group *group; _mlx4_ib_mcg_port_cleanup() local
1173 clear_pending_reqs(struct mcast_group * group,int vf) clear_pending_reqs() argument
1206 push_deleteing_req(struct mcast_group * group,int slave) push_deleteing_req() argument
1237 struct mcast_group *group; clean_vf_mcast() local
[all...]
/linux/fs/notify/
H A Dgroup.c20 * Final freeing of a group
22 static void fsnotify_final_destroy_group(struct fsnotify_group *group) in fsnotify_final_destroy_group() argument
24 if (group->ops->free_group_priv) in fsnotify_final_destroy_group()
25 group->ops->free_group_priv(group); in fsnotify_final_destroy_group()
27 mem_cgroup_put(group->memcg); in fsnotify_final_destroy_group()
28 mutex_destroy(&group->mark_mutex); in fsnotify_final_destroy_group()
30 kfree(group); in fsnotify_final_destroy_group()
34 * Stop queueing new events for this group. Once this function returns
35 * fsnotify_add_event() will not add any new events to the group'
37 fsnotify_group_stop_queueing(struct fsnotify_group * group) fsnotify_group_stop_queueing() argument
50 fsnotify_destroy_group(struct fsnotify_group * group) fsnotify_destroy_group() argument
99 fsnotify_get_group(struct fsnotify_group * group) fsnotify_get_group() argument
107 fsnotify_put_group(struct fsnotify_group * group) fsnotify_put_group() argument
118 struct fsnotify_group *group; __fsnotify_alloc_group() local
157 struct fsnotify_group *group = file->private_data; fsnotify_fasync() local
[all...]
H A Dnotification.c7 * Basic idea behind the notification queue: An fsnotify group (like inotify)
10 * event to the group notify queue. Since a single event might need to be on
11 * multiple group's notification queues we can't add the event directly to each
17 * another group a new event_holder (from fsnotify_event_holder_cachep) will be
50 void fsnotify_destroy_event(struct fsnotify_group *group, in fsnotify_destroy_event() argument
53 /* Overflow events are per-group and we don't want to free them */ in fsnotify_destroy_event()
54 if (!event || event == group->overflow_event) in fsnotify_destroy_event()
63 spin_lock(&group->notification_lock); in fsnotify_destroy_event()
65 spin_unlock(&group->notification_lock); in fsnotify_destroy_event()
67 group in fsnotify_destroy_event()
81 fsnotify_insert_event(struct fsnotify_group * group,struct fsnotify_event * event,int (* merge)(struct fsnotify_group *,struct fsnotify_event *),void (* insert)(struct fsnotify_group *,struct fsnotify_event *)) fsnotify_insert_event() argument
132 fsnotify_remove_queued_event(struct fsnotify_group * group,struct fsnotify_event * event) fsnotify_remove_queued_event() argument
148 fsnotify_peek_first_event(struct fsnotify_group * group) fsnotify_peek_first_event() argument
163 fsnotify_remove_first_event(struct fsnotify_group * group) fsnotify_remove_first_event() argument
181 fsnotify_flush_notify(struct fsnotify_group * group) fsnotify_flush_notify() argument
[all...]
/linux/drivers/iommu/
H A Diommu.c49 /* Tags used with xa_tag_pointer() in group->pasid_array */
77 #define for_each_group_device(group, pos) \ argument
78 list_for_each_entry(pos, &(group)->devices, list)
82 ssize_t (*show)(struct iommu_group *group, char *buf);
83 ssize_t (*store)(struct iommu_group *group,
105 struct iommu_group *group);
114 static int __iommu_device_set_domain(struct iommu_group *group,
118 static int __iommu_group_set_domain_internal(struct iommu_group *group,
121 static int __iommu_group_set_domain(struct iommu_group *group, in __iommu_group_set_domain() argument
124 return __iommu_group_set_domain_internal(group, new_domai in __iommu_group_set_domain()
126 __iommu_group_set_domain_nofail(struct iommu_group * group,struct iommu_domain * new_domain) __iommu_group_set_domain_nofail() argument
417 struct iommu_group *group; iommu_init_device() local
495 struct iommu_group *group = dev->iommu_group; iommu_deinit_device() local
561 struct iommu_group *group; __iommu_probe_device() local
658 __iommu_group_free_device(struct iommu_group * group,struct group_device * grp_dev) __iommu_group_free_device() argument
684 struct iommu_group *group = dev->iommu_group; __iommu_group_remove_device() local
711 struct iommu_group *group = dev->iommu_group; iommu_release_device() local
760 struct iommu_group *group = to_iommu_group(kobj); iommu_group_attr_show() local
773 struct iommu_group *group = to_iommu_group(kobj); iommu_group_attr_store() local
786 iommu_group_create_file(struct iommu_group * group,struct iommu_group_attribute * attr) iommu_group_create_file() argument
792 iommu_group_remove_file(struct iommu_group * group,struct iommu_group_attribute * attr) iommu_group_remove_file() argument
798 iommu_group_show_name(struct iommu_group * group,char * buf) iommu_group_show_name() argument
879 iommu_get_group_resv_regions(struct iommu_group * group,struct list_head * head) iommu_get_group_resv_regions() argument
908 iommu_group_show_resv_regions(struct iommu_group * group,char * buf) iommu_group_show_resv_regions() argument
930 iommu_group_show_type(struct iommu_group * group,char * buf) iommu_group_show_type() argument
970 struct iommu_group *group = to_iommu_group(kobj); iommu_group_release() local
1005 struct iommu_group *group; iommu_group_alloc() local
1072 iommu_group_get_iommudata(struct iommu_group * group) iommu_group_get_iommudata() argument
1088 iommu_group_set_iommudata(struct iommu_group * group,void * iommu_data,void (* release)(void * iommu_data)) iommu_group_set_iommudata() argument
1104 iommu_group_set_name(struct iommu_group * group,const char * name) iommu_group_set_name() argument
1194 iommu_group_alloc_device(struct iommu_group * group,struct device * dev) iommu_group_alloc_device() argument
1257 iommu_group_add_device(struct iommu_group * group,struct device * dev) iommu_group_add_device() argument
1284 struct iommu_group *group = dev->iommu_group; iommu_group_remove_device() local
1307 struct iommu_group *group = dev->iommu_group; iommu_group_mutex_assert() local
1314 iommu_group_first_dev(struct iommu_group * group) iommu_group_first_dev() argument
1331 iommu_group_for_each_dev(struct iommu_group * group,void * data,int (* fn)(struct device *,void *)) iommu_group_for_each_dev() argument
1359 struct iommu_group *group = dev->iommu_group; iommu_group_get() local
1375 iommu_group_ref_get(struct iommu_group * group) iommu_group_ref_get() argument
1389 iommu_group_put(struct iommu_group * group) iommu_group_put() argument
1402 iommu_group_id(struct iommu_group * group) iommu_group_id() argument
1431 struct iommu_group *group; get_pci_function_alias_group() local
1465 struct iommu_group *group; get_pci_alias_group() local
1499 struct iommu_group *group; global() member
1536 struct iommu_group *group; generic_single_device_group() local
1556 struct iommu_group *group = NULL; pci_device_group() local
1619 struct iommu_group *group; fsl_mc_device_group() local
1649 __iommu_group_alloc_default_domain(struct iommu_group * group,int req_type) __iommu_group_alloc_default_domain() argument
1686 iommu_group_alloc_default_domain(struct iommu_group * group,int req_type) iommu_group_alloc_default_domain() argument
1724 iommu_group_default_domain(struct iommu_group * group) iommu_group_default_domain() argument
1765 iommu_get_def_domain_type(struct iommu_group * group,struct device * dev,int cur_type) iommu_get_def_domain_type() argument
1807 iommu_get_default_domain_type(struct iommu_group * group,int target_type) iommu_get_default_domain_type() argument
1884 struct iommu_group *group, *next; bus_iommu_probe() local
1960 iommu_group_has_isolated_msi(struct iommu_group * group) iommu_group_has_isolated_msi() argument
2079 __iommu_group_set_core_domain(struct iommu_group * group) __iommu_group_set_core_domain() argument
2122 struct iommu_group *group = dev->iommu_group; iommu_attach_device() local
2156 struct iommu_group *group = dev->iommu_group; iommu_detach_device() local
2175 struct iommu_group *group = dev->iommu_group; iommu_get_domain_for_dev() local
2218 __iommu_attach_group(struct iommu_domain * domain,struct iommu_group * group) __iommu_attach_group() argument
2246 iommu_attach_group(struct iommu_domain * domain,struct iommu_group * group) iommu_attach_group() argument
2258 __iommu_device_set_domain(struct iommu_group * group,struct device * dev,struct iommu_domain * new_domain,unsigned int flags) __iommu_device_set_domain() argument
2317 __iommu_group_set_domain_internal(struct iommu_group * group,struct iommu_domain * new_domain,unsigned int flags) __iommu_group_set_domain_internal() argument
2383 iommu_detach_group(struct iommu_domain * domain,struct iommu_group * group) iommu_detach_group() argument
2954 iommu_setup_default_domain(struct iommu_group * group,int target_type) iommu_setup_default_domain() argument
3059 iommu_group_store_type(struct iommu_group * group,const char * buf,size_t count) iommu_group_store_type() argument
3125 struct iommu_group *group = dev->iommu_group; iommu_device_use_default_domain() local
3163 struct iommu_group *group = dev->iommu_group; iommu_device_unuse_default_domain() local
3175 __iommu_group_alloc_blocking_domain(struct iommu_group * group) __iommu_group_alloc_blocking_domain() argument
3200 __iommu_take_dma_ownership(struct iommu_group * group,void * owner) __iommu_take_dma_ownership() argument
3229 iommu_group_claim_dma_owner(struct iommu_group * group,void * owner) iommu_group_claim_dma_owner() argument
3262 struct iommu_group *group = dev->iommu_group; iommu_device_claim_dma_owner() local
3288 __iommu_release_dma_ownership(struct iommu_group * group) __iommu_release_dma_ownership() argument
3305 iommu_group_release_dma_owner(struct iommu_group * group) iommu_group_release_dma_owner() argument
3322 struct iommu_group *group = dev->iommu_group; iommu_device_release_dma_owner() local
3340 iommu_group_dma_owner_claimed(struct iommu_group * group) iommu_group_dma_owner_claimed() argument
3363 __iommu_set_group_pasid(struct iommu_domain * domain,struct iommu_group * group,ioasid_t pasid,struct iommu_domain * old) __iommu_set_group_pasid() argument
3401 __iommu_remove_group_pasid(struct iommu_group * group,ioasid_t pasid,struct iommu_domain * domain) __iommu_remove_group_pasid() argument
3430 struct iommu_group *group = dev->iommu_group; iommu_attach_device_pasid() local
3517 struct iommu_group *group = dev->iommu_group; iommu_replace_device_pasid() local
3599 struct iommu_group *group = dev->iommu_group; iommu_detach_device_pasid() local
3650 iommu_attach_handle_get(struct iommu_group * group,ioasid_t pasid,unsigned int type) iommu_attach_handle_get() argument
3686 iommu_attach_group_handle(struct iommu_domain * domain,struct iommu_group * group,struct iommu_attach_handle * handle) iommu_attach_group_handle() argument
3732 iommu_detach_group_handle(struct iommu_domain * domain,struct iommu_group * group) iommu_detach_group_handle() argument
3757 iommu_replace_group_handle(struct iommu_group * group,struct iommu_domain * new_domain,struct iommu_attach_handle * handle) iommu_replace_group_handle() argument
3806 struct iommu_group *group = dev->iommu_group; iommu_dma_prepare_msi() local
[all...]
/linux/drivers/gpu/drm/xe/
H A Dxe_hw_engine_group.c19 struct xe_hw_engine_group *group = container_of(w, struct xe_hw_engine_group, resume_work); in hw_engine_group_resume_lr_jobs_func() local
23 err = xe_hw_engine_group_get_mode(group, EXEC_MODE_LR, &previous_mode); in hw_engine_group_resume_lr_jobs_func()
30 list_for_each_entry(q, &group->exec_queue_list, hw_engine_group_link) { in hw_engine_group_resume_lr_jobs_func()
38 xe_hw_engine_group_put(group); in hw_engine_group_resume_lr_jobs_func()
44 struct xe_hw_engine_group *group; in hw_engine_group_alloc() local
47 group = drmm_kzalloc(&xe->drm, sizeof(*group), GFP_KERNEL); in hw_engine_group_alloc()
48 if (!group) in hw_engine_group_alloc()
51 group->resume_wq = alloc_workqueue("xe-resume-lr-jobs-wq", 0, 0); in hw_engine_group_alloc()
52 if (!group in hw_engine_group_alloc()
122 xe_hw_engine_group_add_exec_queue(struct xe_hw_engine_group * group,struct xe_exec_queue * q) xe_hw_engine_group_add_exec_queue() argument
163 xe_hw_engine_group_del_exec_queue(struct xe_hw_engine_group * group,struct xe_exec_queue * q) xe_hw_engine_group_del_exec_queue() argument
183 xe_hw_engine_group_resume_faulting_lr_jobs(struct xe_hw_engine_group * group) xe_hw_engine_group_resume_faulting_lr_jobs() argument
194 xe_hw_engine_group_suspend_faulting_lr_jobs(struct xe_hw_engine_group * group) xe_hw_engine_group_suspend_faulting_lr_jobs() argument
239 xe_hw_engine_group_wait_for_dma_fence_jobs(struct xe_hw_engine_group * group) xe_hw_engine_group_wait_for_dma_fence_jobs() argument
262 switch_mode(struct xe_hw_engine_group * group) switch_mode() argument
296 xe_hw_engine_group_get_mode(struct xe_hw_engine_group * group,enum xe_hw_engine_group_execution_mode new_mode,enum xe_hw_engine_group_execution_mode * previous_mode) xe_hw_engine_group_get_mode() argument
331 xe_hw_engine_group_put(struct xe_hw_engine_group * group) xe_hw_engine_group_put() argument
[all...]
/linux/arch/sparc/kernel/
H A Dhvapi.c19 unsigned long group; member
28 { .group = HV_GRP_SUN4V, .flags = FLAG_PRE_API },
29 { .group = HV_GRP_CORE, .flags = FLAG_PRE_API },
30 { .group = HV_GRP_INTR, },
31 { .group = HV_GRP_SOFT_STATE, },
32 { .group = HV_GRP_TM, },
33 { .group = HV_GRP_PCI, .flags = FLAG_PRE_API },
34 { .group = HV_GRP_LDOM, },
35 { .group = HV_GRP_SVC_CHAN, .flags = FLAG_PRE_API },
36 { .group
59 __get_info(unsigned long group) __get_info() argument
96 sun4v_hvapi_register(unsigned long group,unsigned long major,unsigned long * minor) sun4v_hvapi_register() argument
147 sun4v_hvapi_unregister(unsigned long group) sun4v_hvapi_unregister() argument
160 sun4v_hvapi_get(unsigned long group,unsigned long * major,unsigned long * minor) sun4v_hvapi_get() argument
184 unsigned long group, major, minor; sun4v_hvapi_init() local
[all...]
/linux/drivers/infiniband/core/
H A Dmulticast.c117 struct mcast_group *group; member
133 struct mcast_group *group; in mcast_find() local
137 group = rb_entry(node, struct mcast_group, node); in mcast_find()
138 ret = memcmp(mgid->raw, group->rec.mgid.raw, sizeof *mgid); in mcast_find()
140 return group; in mcast_find()
151 struct mcast_group *group, in mcast_insert() argument
163 ret = memcmp(group->rec.mgid.raw, cur_group->rec.mgid.raw, in mcast_insert()
164 sizeof group->rec.mgid); in mcast_insert()
174 rb_link_node(&group->node, parent, link); in mcast_insert()
175 rb_insert_color(&group in mcast_insert()
185 release_group(struct mcast_group * group) release_group() argument
208 struct mcast_group *group = member->group; queue_join() local
228 adjust_membership(struct mcast_group * group,u8 join_state,int inc) adjust_membership() argument
243 get_leave_state(struct mcast_group * group) get_leave_state() argument
330 send_join(struct mcast_group * group,struct mcast_member * member) send_join() argument
345 send_leave(struct mcast_group * group,u8 leave_state) send_leave() argument
365 join_group(struct mcast_group * group,struct mcast_member * member,u8 join_state) join_group() argument
376 fail_join(struct mcast_group * group,struct mcast_member * member,int status) fail_join() argument
385 process_group_error(struct mcast_group * group) process_group_error() argument
426 struct mcast_group *group; mcast_work_handler() local
491 process_join_error(struct mcast_group * group,int status) process_join_error() argument
514 struct mcast_group *group = context; join_handler() local
548 struct mcast_group *group = context; leave_handler() local
560 struct mcast_group *group, *cur_group; acquire_group() local
665 struct mcast_group *group; ib_sa_free_multicast() local
698 struct mcast_group *group; ib_sa_get_mcmember_rec() local
774 struct mcast_group *group; mcast_groups_event() local
[all...]
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_sched.c42 * The scheduling happens at the scheduling group level, each group
51 * rotating the groups passed to the firmware so every group gets
67 * queue ring-buffer, and the group is scheduled for execution if it
70 * Kernel-side group scheduling is timeslice-based. When we have less
73 * groups than slots, we let each group a chance to execute stuff for
75 * to schedule. The group selection algorithm is based on
80 * group/queue state that would be based on information we wouldn't have
82 * reason we don't do 'cooperative' scheduling (encoding FW group slot
84 * drm_gpu_scheduler::prepare_job() hook, and treating group rotatio
106 struct panthor_group *group; global() member
723 group_queue_work(group,wname) global() argument
787 struct panthor_group *group; global() member
852 panthor_queue_get_syncwait_obj(struct panthor_group * group,struct panthor_queue * queue) panthor_queue_get_syncwait_obj() argument
884 group_free_queue(struct panthor_group * group,struct panthor_queue * queue) group_free_queue() argument
909 struct panthor_group *group = container_of(work, group_release_work() local
927 struct panthor_group *group = container_of(kref, group_release() local
939 group_put(struct panthor_group * group) group_put() argument
946 group_get(struct panthor_group * group) group_get() argument
962 group_bind_locked(struct panthor_group * group,u32 csg_id) group_bind_locked() argument
1004 group_unbind_locked(struct panthor_group * group) group_unbind_locked() argument
1093 struct panthor_group *group = ptdev->scheduler->csg_slots[csg_id].group; cs_slot_reset_locked() local
1146 struct panthor_group *group = ptdev->scheduler->csg_slots[csg_id].group; cs_slot_sync_queue_state_locked() local
1198 struct panthor_group *group = csg_slot->group; csg_slot_sync_queues_state_locked() local
1217 struct panthor_group *group; csg_slot_sync_state_locked() local
1290 struct panthor_group *group; csg_slot_prog_locked() local
1347 struct panthor_group *group = csg_slot->group; cs_slot_process_fatal_event_locked() local
1389 struct panthor_group *group = csg_slot->group; cs_slot_process_fault_event_locked() local
1431 group_process_tiler_oom(struct panthor_group * group,u32 cs_id) group_process_tiler_oom() argument
1519 struct panthor_group *group = group_tiler_oom_work() local
1539 struct panthor_group *group = csg_slot->group; cs_slot_process_tiler_oom_event_locked() local
1616 struct panthor_group *group = csg_slot->group; csg_slot_sync_update_locked() local
1631 struct panthor_group *group = csg_slot->group; csg_slot_process_progress_timer_event_locked() local
1892 group_is_idle(struct panthor_group * group) group_is_idle() argument
1905 group_can_run(struct panthor_group * group) group_can_run() argument
1920 struct panthor_group *group, *tmp; tick_ctx_pick_groups_from_list() local
1964 tick_ctx_insert_old_group(struct panthor_scheduler * sched,struct panthor_sched_tick_ctx * ctx,struct panthor_group * group,bool full_tick) tick_ctx_insert_old_group() argument
2017 struct panthor_group *group = csg_slot->group; tick_ctx_init() local
2051 group_term_post_processing(struct panthor_group * group) group_term_post_processing() argument
2101 struct panthor_group *group = group_term_work() local
2113 struct panthor_group *group, *tmp; tick_ctx_cleanup() local
2164 struct panthor_group *group, *tmp; tick_ctx_apply() local
2456 panthor_queue_eval_syncwait(struct panthor_group * group,u8 queue_idx) panthor_queue_eval_syncwait() argument
2490 struct panthor_group *group, *tmp; sync_upd_work() local
2530 group_schedule_locked(struct panthor_group * group,u32 queue_mask) group_schedule_locked() argument
2608 panthor_group_stop(struct panthor_group * group) panthor_group_stop() argument
2621 panthor_group_start(struct panthor_group * group) panthor_group_start() argument
2671 struct panthor_group *group; panthor_sched_suspend() local
2793 struct panthor_group *group, *group_tmp; panthor_sched_pre_reset() local
2828 struct panthor_group *group, *group_tmp; panthor_sched_post_reset() local
2857 struct panthor_group *group = job->group; update_fdinfo_stats() local
2874 struct panthor_group *group; panthor_fdinfo_gather_group_samples() local
2893 struct panthor_group *group = group_sync_upd_work() local
2987 struct panthor_group *group = job->group; get_job_cs_params() local
3116 struct panthor_group *group = job->group; queue_run_job() local
3216 struct panthor_group *group = job->group; queue_timedout_job() local
3287 group_create_queue(struct panthor_group * group,const struct drm_panthor_queue_create * args) group_create_queue() argument
3393 add_group_kbo_sizes(struct panthor_device * ptdev,struct panthor_group * group) add_group_kbo_sizes() argument
3425 struct panthor_group *group = NULL; panthor_group_create() local
3555 struct panthor_group *group; panthor_group_destroy() local
3589 struct panthor_group *group; group_from_handle() local
3604 struct panthor_group *group; panthor_group_get_state() local
3646 struct panthor_group *group; panthor_group_pool_destroy() local
3672 struct panthor_group *group; panthor_fdinfo_gather_group_mem_info() local
[all...]
/linux/kernel/sched/
H A Dpsi.c205 static void group_init(struct psi_group *group) in group_init() argument
207 group->enabled = true; in group_init()
208 group->avg_last_update = sched_clock(); in group_init()
209 group->avg_next_update = group->avg_last_update + psi_period; in group_init()
210 mutex_init(&group->avgs_lock); in group_init()
213 INIT_LIST_HEAD(&group->avg_triggers); in group_init()
214 memset(group->avg_nr_triggers, 0, sizeof(group->avg_nr_triggers)); in group_init()
215 INIT_DELAYED_WORK(&group in group_init()
271 get_recent_times(struct psi_group * group,int cpu,enum psi_aggregators aggregator,u32 * times,u32 * pchanged_states) get_recent_times() argument
362 collect_percpu_times(struct psi_group * group,enum psi_aggregators aggregator,u32 * pchanged_states) collect_percpu_times() argument
463 update_triggers(struct psi_group * group,u64 now,enum psi_aggregators aggregator) update_triggers() argument
525 update_averages(struct psi_group * group,u64 now) update_averages() argument
581 struct psi_group *group; psi_avgs_work() local
613 init_rtpoll_triggers(struct psi_group * group,u64 now) init_rtpoll_triggers() argument
626 psi_schedule_rtpoll_work(struct psi_group * group,unsigned long delay,bool force) psi_schedule_rtpoll_work() argument
653 psi_rtpoll_work(struct psi_group * group) psi_rtpoll_work() argument
740 struct psi_group *group = (struct psi_group *)data; psi_rtpoll_worker() local
758 struct psi_group *group = timer_container_of(group, t, rtpoll_timer); poll_timer_fn() local
793 for_each_group(iter,group) global() argument
796 psi_group_change(struct psi_group * group,int cpu,unsigned int clear,unsigned int set,u64 now,bool wake_clock) psi_group_change() argument
1210 psi_cgroup_restart(struct psi_group * group) psi_cgroup_restart() argument
1245 psi_show(struct seq_file * m,struct psi_group * group,enum psi_res res) psi_show() argument
1295 psi_trigger_create(struct psi_group * group,char * buf,enum psi_res res,struct file * file,struct kernfs_open_file * of) psi_trigger_create() argument
1399 struct psi_group *group; psi_trigger_destroy() local
[all...]
/linux/drivers/net/ethernet/microchip/sparx5/
H A Dsparx5_sdlb.c60 static void sparx5_sdlb_group_disable(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_disable() argument
64 ANA_AC_SDLB_PUP_CTRL(group)); in sparx5_sdlb_group_disable()
67 static void sparx5_sdlb_group_enable(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_enable() argument
71 ANA_AC_SDLB_PUP_CTRL(group)); in sparx5_sdlb_group_enable()
74 static u32 sparx5_sdlb_group_get_first(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_get_first() argument
78 val = spx5_rd(sparx5, ANA_AC_SDLB_XLB_START(group)); in sparx5_sdlb_group_get_first()
83 static u32 sparx5_sdlb_group_get_next(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_get_next() argument
93 static bool sparx5_sdlb_group_is_first(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_is_first() argument
96 return lb == sparx5_sdlb_group_get_first(sparx5, group); in sparx5_sdlb_group_is_first()
99 static bool sparx5_sdlb_group_is_last(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_is_last() argument
105 sparx5_sdlb_group_is_empty(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_is_empty() argument
114 sparx5_sdlb_group_get_last(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_get_last() argument
129 sparx5_sdlb_group_is_singular(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_is_singular() argument
138 sparx5_sdlb_group_get_adjacent(struct sparx5 * sparx5,u32 group,u32 idx,u32 * prev,u32 * next,u32 * first) sparx5_sdlb_group_get_adjacent() argument
163 sparx5_sdlb_group_get_count(struct sparx5 * sparx5,u32 group) sparx5_sdlb_group_get_count() argument
183 const struct sparx5_sdlb_group *group; sparx5_sdlb_group_get_by_rate() local
208 sparx5_sdlb_group_get_by_index(struct sparx5 * sparx5,u32 idx,u32 * group) sparx5_sdlb_group_get_by_index() argument
236 sparx5_sdlb_group_link(struct sparx5 * sparx5,u32 group,u32 idx,u32 first,u32 next,bool empty) sparx5_sdlb_group_link() argument
260 sparx5_sdlb_group_add(struct sparx5 * sparx5,u32 group,u32 idx) sparx5_sdlb_group_add() argument
275 sparx5_sdlb_group_del(struct sparx5 * sparx5,u32 group,u32 idx) sparx5_sdlb_group_del() argument
310 struct sparx5_sdlb_group *group; sparx5_sdlb_group_init() local
[all...]
H A Dsparx5_qos.c84 static u32 sparx5_lg_get_leak_time(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_leak_time() argument
88 value = spx5_rd(sparx5, HSCH_HSCH_TIMER_CFG(layer, group)); in sparx5_lg_get_leak_time()
92 static void sparx5_lg_set_leak_time(struct sparx5 *sparx5, u32 layer, u32 group, in sparx5_lg_set_leak_time() argument
96 HSCH_HSCH_TIMER_CFG(layer, group)); in sparx5_lg_set_leak_time()
99 static u32 sparx5_lg_get_first(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_first() argument
103 value = spx5_rd(sparx5, HSCH_HSCH_LEAK_CFG(layer, group)); in sparx5_lg_get_first()
107 static u32 sparx5_lg_get_next(struct sparx5 *sparx5, u32 layer, u32 group, in sparx5_lg_get_next() argument
117 static u32 sparx5_lg_get_last(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_last() argument
121 itr = sparx5_lg_get_first(sparx5, layer, group); in sparx5_lg_get_last()
124 next = sparx5_lg_get_next(sparx5, layer, group, it in sparx5_lg_get_last()
132 sparx5_lg_is_last(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx) sparx5_lg_is_last() argument
138 sparx5_lg_is_first(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx) sparx5_lg_is_first() argument
144 sparx5_lg_is_empty(struct sparx5 * sparx5,u32 layer,u32 group) sparx5_lg_is_empty() argument
149 sparx5_lg_is_singular(struct sparx5 * sparx5,u32 layer,u32 group) sparx5_lg_is_singular() argument
158 sparx5_lg_enable(struct sparx5 * sparx5,u32 layer,u32 group,u32 leak_time) sparx5_lg_enable() argument
164 sparx5_lg_disable(struct sparx5 * sparx5,u32 layer,u32 group) sparx5_lg_disable() argument
170 sparx5_lg_get_group_by_index(struct sparx5 * sparx5,u32 layer,u32 idx,u32 * group) sparx5_lg_get_group_by_index() argument
198 sparx5_lg_get_group_by_rate(u32 layer,u32 rate,u32 * group) sparx5_lg_get_group_by_rate() argument
215 sparx5_lg_get_adjacent(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx,u32 * prev,u32 * next,u32 * first) sparx5_lg_get_adjacent() argument
241 sparx5_lg_conf_set(struct sparx5 * sparx5,u32 layer,u32 group,u32 se_first,u32 idx,u32 idx_next,bool empty) sparx5_lg_conf_set() argument
271 sparx5_lg_del(struct sparx5 * sparx5,u32 layer,u32 group,u32 idx) sparx5_lg_del() argument
332 sparx5_shaper_conf_set(struct sparx5_port * port,const struct sparx5_shaper * sh,u32 layer,u32 idx,u32 group) sparx5_shaper_conf_set() argument
506 u32 group; sparx5_tc_tbf_add() local
535 u32 group; sparx5_tc_tbf_del() local
[all...]
/linux/drivers/gpio/
H A Dgpio-lpc32xx.c168 static inline u32 gpreg_read(struct lpc32xx_gpio_chip *group, unsigned long offset) in gpreg_read() argument
170 return __raw_readl(group->reg_base + offset); in gpreg_read()
173 static inline void gpreg_write(struct lpc32xx_gpio_chip *group, u32 val, unsigned long offset) in gpreg_write() argument
175 __raw_writel(val, group->reg_base + offset); in gpreg_write()
178 static void __set_gpio_dir_p012(struct lpc32xx_gpio_chip *group, in __set_gpio_dir_p012() argument
182 gpreg_write(group, GPIO012_PIN_TO_BIT(pin), in __set_gpio_dir_p012()
183 group->gpio_grp->dir_clr); in __set_gpio_dir_p012()
185 gpreg_write(group, GPIO012_PIN_TO_BIT(pin), in __set_gpio_dir_p012()
186 group->gpio_grp->dir_set); in __set_gpio_dir_p012()
189 static void __set_gpio_dir_p3(struct lpc32xx_gpio_chip *group, in __set_gpio_dir_p3() argument
200 __set_gpio_level_p012(struct lpc32xx_gpio_chip * group,unsigned pin,int high) __set_gpio_level_p012() argument
211 __set_gpio_level_p3(struct lpc32xx_gpio_chip * group,unsigned pin,int high) __set_gpio_level_p3() argument
222 __set_gpo_level_p3(struct lpc32xx_gpio_chip * group,unsigned pin,int high) __set_gpo_level_p3() argument
231 __get_gpio_state_p012(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpio_state_p012() argument
238 __get_gpio_state_p3(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpio_state_p3() argument
250 __get_gpi_state_p3(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpi_state_p3() argument
256 __get_gpo_state_p3(struct lpc32xx_gpio_chip * group,unsigned pin) __get_gpo_state_p3() argument
268 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_input_p012() local
278 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_input_p3() local
293 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_get_value_p012() local
300 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_get_value_p3() local
307 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpi_get_value() local
315 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_output_p012() local
326 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_output_p3() local
337 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_dir_out_always() local
346 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_set_value_p012() local
356 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpio_set_value_p3() local
366 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpo_set_value() local
375 struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip); lpc32xx_gpo_get_value() local
[all...]
/linux/kernel/time/
H A Dtimer_migration.c22 * lowest level group contains CPUs, the next level groups of CPU groups
43 * Each group has a designated migrator CPU/group as long as a CPU/group is
44 * active in the group. This designated role is necessary to avoid that all
45 * active CPUs in a group try to migrate expired timers from other CPUs,
48 * When a CPU is awake, it checks in it's own timer tick the group
53 * If it finds expired timers in one of the group queues it pulls them over
55 * group and the parent groups if required.
60 * CPU does not queue an event in the LVL0 group
437 tmigr_check_migrator(struct tmigr_group * group,u8 childmask) tmigr_check_migrator() argument
449 tmigr_check_migrator_and_lonely(struct tmigr_group * group,u8 childmask) tmigr_check_migrator_and_lonely() argument
466 tmigr_check_lonely(struct tmigr_group * group) tmigr_check_lonely() argument
528 struct tmigr_group *child = NULL, *group = tmc->tmgroup; __walk_groups() local
560 tmigr_next_groupevt(struct tmigr_group * group) tmigr_next_groupevt() argument
593 tmigr_next_expired_groupevt(struct tmigr_group * group,u64 now) tmigr_next_expired_groupevt() argument
610 tmigr_next_groupevt_expires(struct tmigr_group * group) tmigr_next_groupevt_expires() argument
622 tmigr_active_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_active_up() argument
722 tmigr_update_events(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_update_events() argument
869 tmigr_new_timer_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_new_timer_up() argument
1002 tmigr_handle_remote_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_handle_remote_up() argument
1101 tmigr_requires_handle_remote_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_requires_handle_remote_up() argument
1244 tmigr_inactive_up(struct tmigr_group * group,struct tmigr_group * child,struct tmigr_walk * data) tmigr_inactive_up() argument
1396 struct tmigr_group *group = tmc->tmgroup; tmigr_quick_check() local
1484 tmigr_init_group(struct tmigr_group * group,unsigned int lvl,int node) tmigr_init_group() argument
1526 struct tmigr_group *tmp, *group = NULL; tmigr_get_group() local
1644 struct tmigr_group *group, *child, **stack; tmigr_setup_groups() local
[all...]
/linux/drivers/clk/renesas/
H A Dclk-mstp.c32 * struct mstp_clock_group - MSTP gating clocks group
34 * @data: clock specifier translation for clocks in this group
39 * @clks: clocks in this group
54 * @group: MSTP clocks group
59 struct mstp_clock_group *group; member
64 static inline u32 cpg_mstp_read(struct mstp_clock_group *group, in cpg_mstp_read() argument
67 return group->width_8bit ? readb(reg) : readl(reg); in cpg_mstp_read()
70 static inline void cpg_mstp_write(struct mstp_clock_group *group, u32 val, in cpg_mstp_write() argument
73 group in cpg_mstp_write()
79 struct mstp_clock_group *group = clock->group; cpg_mstp_clock_endisable() local
128 struct mstp_clock_group *group = clock->group; cpg_mstp_clock_is_enabled() local
147 cpg_mstp_clock_register(const char * name,const char * parent_name,unsigned int index,struct mstp_clock_group * group) cpg_mstp_clock_register() argument
182 struct mstp_clock_group *group; cpg_mstp_clocks_init() local
[all...]
/linux/fs/notify/inotify/
H A Dinotify_user.c141 struct fsnotify_group *group = file->private_data; in inotify_poll() local
144 poll_wait(file, &group->notification_waitq, wait); in inotify_poll()
145 spin_lock(&group->notification_lock); in inotify_poll()
146 if (!fsnotify_notify_queue_is_empty(group)) in inotify_poll()
148 spin_unlock(&group->notification_lock); in inotify_poll()
168 * Called with the group->notification_lock held.
170 static struct fsnotify_event *get_one_event(struct fsnotify_group *group, in get_one_event() argument
176 event = fsnotify_peek_first_event(group); in get_one_event()
180 pr_debug("%s: group=%p event=%p\n", __func__, group, even in get_one_event()
199 copy_event_to_user(struct fsnotify_group * group,struct fsnotify_event * fsn_event,char __user * buf) copy_event_to_user() argument
252 struct fsnotify_group *group; inotify_read() local
303 struct fsnotify_group *group = file->private_data; inotify_release() local
316 struct fsnotify_group *group; inotify_ioctl() local
414 inotify_idr_find_locked(struct fsnotify_group * group,int wd) inotify_idr_find_locked() argument
435 inotify_idr_find(struct fsnotify_group * group,int wd) inotify_idr_find() argument
452 inotify_remove_from_idr(struct fsnotify_group * group,struct inotify_inode_mark * i_mark) inotify_remove_from_idr() argument
521 inotify_ignored_and_remove_idr(struct fsnotify_mark * fsn_mark,struct fsnotify_group * group) inotify_ignored_and_remove_idr() argument
536 inotify_update_existing_watch(struct fsnotify_group * group,struct inode * inode,u32 arg) inotify_update_existing_watch() argument
590 inotify_new_watch(struct fsnotify_group * group,struct inode * inode,u32 arg) inotify_new_watch() argument
638 inotify_update_watch(struct fsnotify_group * group,struct inode * inode,u32 arg) inotify_update_watch() argument
655 struct fsnotify_group *group; inotify_new_group() local
696 struct fsnotify_group *group; do_inotify_init() local
732 struct fsnotify_group *group; SYSCALL_DEFINE3() local
788 struct fsnotify_group *group; SYSCALL_DEFINE2() local
[all...]
H A Dinotify_fsnotify.c49 static int inotify_merge(struct fsnotify_group *group, in inotify_merge() argument
52 struct list_head *list = &group->notification_list; in inotify_merge()
66 struct fsnotify_group *group = inode_mark->group; in inotify_handle_inode_event() local
77 pr_debug("%s: group=%p mark=%p mask=%x\n", __func__, group, inode_mark, in inotify_handle_inode_event()
95 old_memcg = set_active_memcg(group->memcg); in inotify_handle_inode_event()
104 fsnotify_queue_overflow(group); in inotify_handle_inode_event()
126 ret = fsnotify_add_event(group, fsn_event, inotify_merge); in inotify_handle_inode_event()
129 fsnotify_destroy_event(group, fsn_even in inotify_handle_inode_event()
138 inotify_freeing_mark(struct fsnotify_mark * fsn_mark,struct fsnotify_group * group) inotify_freeing_mark() argument
178 inotify_free_group_priv(struct fsnotify_group * group) inotify_free_group_priv() argument
187 inotify_free_event(struct fsnotify_group * group,struct fsnotify_event * fsn_event) inotify_free_event() argument
[all...]
/linux/drivers/media/platform/renesas/rcar-vin/
H A Drcar-core.c50 * than one group for the whole system be replaced with a linked
55 * be only one group for all instances.
62 static void rvin_group_cleanup(struct rvin_group *group) in rvin_group_cleanup() argument
64 media_device_cleanup(&group->mdev); in rvin_group_cleanup()
65 mutex_destroy(&group->lock); in rvin_group_cleanup()
68 static int rvin_group_init(struct rvin_group *group, struct rvin_dev *vin, in rvin_group_init() argument
72 struct media_device *mdev = &group->mdev; in rvin_group_init()
76 mutex_init(&group->lock); in rvin_group_init()
79 group->count = 0; in rvin_group_init()
82 group in rvin_group_init()
104 struct rvin_group *group = rvin_group_release() local
122 struct rvin_group *group; rvin_group_get() local
176 struct rvin_group *group = vin->group; rvin_group_put() local
194 rvin_group_entity_to_remote_id(struct rvin_group * group,struct media_entity * entity) rvin_group_entity_to_remote_id() argument
243 struct rvin_group *group = vin->group; rvin_group_notify_unbind() local
281 struct rvin_group *group = vin->group; rvin_group_notify_bound() local
552 struct rvin_group *group = container_of(link->graph_obj.mdev, rvin_csi2_link_notify() local
670 rvin_csi2_create_link(struct rvin_group * group,unsigned int id,const struct rvin_group_route * route) rvin_csi2_create_link() argument
695 rvin_parallel_setup_links(struct rvin_group * group) rvin_parallel_setup_links() argument
732 rvin_csi2_setup_links(struct rvin_group * group) rvin_csi2_setup_links() argument
788 rvin_isp_setup_links(struct rvin_group * group) rvin_isp_setup_links() argument
[all...]
/linux/drivers/pinctrl/aspeed/
H A Dpinmux-aspeed.h144 * pins in the function's group to disable the higher-priority signals such
221 * Conversely, failing to allocate all pins in a group indicates some bits (as
222 * well as pins) required for the group's configuration will already be in use,
224 * group.
244 * the group and the function. In this way copy/paste errors cause duplicate
248 * no override errors in the pin, group and function arrays.
513 #define SIG_DESC_LIST_SYM(sig, group) sig_descs_ ## sig ## _ ## group argument
514 #define SIG_DESC_LIST_DECL(sig, group, ...) \ argument
515 static const struct aspeed_sig_desc SIG_DESC_LIST_SYM(sig, group)[]
518 SIG_EXPR_SYM(sig,group) global() argument
519 SIG_EXPR_DECL_(sig,group,func) global() argument
545 SIG_EXPR_DECL(sig,group,func,...) global() argument
555 SIG_EXPR_PTR(sig,group) global() argument
557 SIG_EXPR_LIST_SYM(sig,group) global() argument
574 SIG_EXPR_LIST_DECL(sig,group,...) global() argument
594 SIG_EXPR_LIST_ALIAS(pin,sig,group) global() argument
628 SIG_EXPR_LIST_DECL_SEMG(pin,sig,group,func,...) global() argument
650 SIG_EXPR_LIST_PTR(sig,group) global() argument
741 GROUP_SYM(group) global() argument
742 GROUP_DECL(group,...) global() argument
749 FUNC_DECL_1(func,group) global() argument
[all...]
/linux/include/trace/events/
H A Dtimer_migration.h14 TP_PROTO(struct tmigr_group *group),
16 TP_ARGS(group),
19 __field( void *, group )
25 __entry->group = group;
26 __entry->lvl = group->level;
27 __entry->numa_node = group->numa_node;
30 TP_printk("group=%p lvl=%d numa=%d",
31 __entry->group, __entry->lvl, __entry->numa_node)
58 TP_printk("group
[all...]
/linux/fs/notify/fanotify/
H A Dfanotify_user.c35 * Legacy fanotify marks limits (8192) is per group and we introduced a tunable
37 * of fanotify marks per user is <max marks per group> * <max groups per user>.
174 * With group flag FAN_REPORT_NAME, if name was not recorded in in fanotify_event_len()
199 static void fanotify_unhash_event(struct fsnotify_group *group, in fanotify_unhash_event() argument
202 assert_spin_locked(&group->notification_lock); in fanotify_unhash_event()
204 pr_debug("%s: group=%p event=%p bucket=%u\n", __func__, in fanotify_unhash_event()
205 group, event, fanotify_event_hash_bucket(group, event)); in fanotify_unhash_event()
219 static struct fanotify_event *get_one_event(struct fsnotify_group *group, in get_one_event() argument
225 unsigned int info_mode = FAN_GROUP_FLAG(group, FANOTIFY_INFO_MODE in get_one_event()
256 create_fd(struct fsnotify_group * group,const struct path * path,struct file ** file) create_fd() argument
306 finish_permission_event(struct fsnotify_group * group,struct fanotify_perm_event * event,u32 response,struct fanotify_response_info_audit_rule * friar) finish_permission_event() argument
327 process_access_response(struct fsnotify_group * group,struct fanotify_response * response_struct,const char __user * info,size_t info_len) process_access_response() argument
739 copy_event_to_user(struct fsnotify_group * group,struct fanotify_event * event,char __user * buf,size_t count) copy_event_to_user() argument
883 struct fsnotify_group *group = file->private_data; fanotify_poll() local
898 struct fsnotify_group *group; fanotify_read() local
974 struct fsnotify_group *group; fanotify_write() local
1005 struct fsnotify_group *group = file->private_data; fanotify_release() local
1060 struct fsnotify_group *group; fanotify_ioctl() local
1172 fanotify_remove_mark(struct fsnotify_group * group,void * obj,unsigned int obj_type,__u32 mask,unsigned int flags,__u32 umask) fanotify_remove_mark() argument
1270 fanotify_set_mark_fsid(struct fsnotify_group * group,struct fsnotify_mark * mark,struct fan_fsid * fsid) fanotify_set_mark_fsid() argument
1321 fanotify_add_new_mark(struct fsnotify_group * group,void * obj,unsigned int obj_type,unsigned int fan_flags,struct fan_fsid * fsid) fanotify_add_new_mark() argument
1376 fanotify_group_init_error_pool(struct fsnotify_group * group) fanotify_group_init_error_pool() argument
1422 fanotify_add_mark(struct fsnotify_group * group,void * obj,unsigned int obj_type,__u32 mask,unsigned int fan_flags,struct fan_fsid * fsid) fanotify_add_mark() argument
1503 struct fsnotify_group *group; SYSCALL_DEFINE2() local
1742 fanotify_events_supported(struct fsnotify_group * group,const struct path * path,__u64 mask,unsigned int flags) fanotify_events_supported() argument
1806 struct fsnotify_group *group; do_fanotify_mark() local
[all...]
/linux/net/psample/
H A Dpsample.c41 struct psample_group *group, in psample_group_nl_fill() argument
52 ret = nla_put_u32(msg, PSAMPLE_ATTR_SAMPLE_GROUP, group->group_num); in psample_group_nl_fill()
56 ret = nla_put_u32(msg, PSAMPLE_ATTR_GROUP_REFCOUNT, group->refcount); in psample_group_nl_fill()
60 ret = nla_put_u32(msg, PSAMPLE_ATTR_GROUP_SEQ, group->seq); in psample_group_nl_fill()
75 struct psample_group *group; in psample_nl_cmd_get_group_dumpit() local
81 list_for_each_entry(group, &psample_groups_list, list) { in psample_nl_cmd_get_group_dumpit()
82 if (!net_eq(group->net, sock_net(msg->sk))) in psample_nl_cmd_get_group_dumpit()
88 err = psample_group_nl_fill(msg, group, PSAMPLE_CMD_NEW_GROUP, in psample_nl_cmd_get_group_dumpit()
123 static void psample_group_notify(struct psample_group *group, in psample_group_notify() argument
133 err = psample_group_nl_fill(msg, group, cm in psample_group_notify()
144 struct psample_group *group; psample_group_create() local
158 psample_group_destroy(struct psample_group * group) psample_group_destroy() argument
168 struct psample_group *group; psample_group_lookup() local
178 struct psample_group *group; psample_group_get() local
196 psample_group_take(struct psample_group * group) psample_group_take() argument
204 psample_group_put(struct psample_group * group) psample_group_put() argument
363 psample_sample_packet(struct psample_group * group,const struct sk_buff * skb,u32 sample_rate,const struct psample_metadata * md) psample_sample_packet() argument
[all...]
/linux/Documentation/devicetree/bindings/pinctrl/
H A Dmarvell,armada-37xx-pinctrl.txt32 group: jtag
36 group sdio0
40 group emmc_nb
44 group pwm0
48 group pwm1
52 group pwm2
56 group pwm3
60 group pmic1
64 group pmic0
68 group i2c
[all...]
/linux/Documentation/filesystems/ext4/
H A Dblockgroup.rst9 The layout of a standard block group is approximately as follows (each
33 For the special case of block group 0, the first 1024 bytes are unused,
40 The ext4 driver primarily works with the superblock and the group
41 descriptors that are found in block group 0. Redundant copies of the
42 superblock and group descriptors are written to some of the block groups
45 paragraph for more details). If the group does not have a redundant
46 copy, the block group begins with the data block bitmap. Note also that
48 GDT block” space after the block group descriptors and before the start
57 As for the ordering of items in a block group, it is generally
58 established that the super block and the group descripto
[all...]

12345678910>>...57