Lines Matching full:work
29 /* List head pointing to ordered work list */
55 struct btrfs_fs_info * __pure btrfs_work_owner(const struct btrfs_work *work) in btrfs_work_owner() argument
57 return work->wq->fs_info; in btrfs_work_owner()
163 * Hook for threshold which will be called before executing the work,
213 struct btrfs_work *work; in run_ordered_work() local
222 work = list_first_entry(list, struct btrfs_work, ordered_list); in run_ordered_work()
223 if (!test_bit(WORK_DONE_BIT, &work->flags)) in run_ordered_work()
229 * updates from ordinary work function. in run_ordered_work()
235 * we leave the work item on the list as a barrier so in run_ordered_work()
236 * that later work items that are done don't have their in run_ordered_work()
239 if (test_and_set_bit(WORK_ORDER_DONE_BIT, &work->flags)) in run_ordered_work()
241 trace_btrfs_ordered_sched(work); in run_ordered_work()
243 work->ordered_func(work, false); in run_ordered_work()
247 list_del(&work->ordered_list); in run_ordered_work()
250 if (work == self) { in run_ordered_work()
252 * This is the work item that the worker is currently in run_ordered_work()
256 * of work items. I.e., if a work item with the same in run_ordered_work()
257 * address and work function is queued twice, the second in run_ordered_work()
259 * work item may be freed and recycled with the same in run_ordered_work()
260 * work function; the workqueue code assumes that the in run_ordered_work()
261 * original work item cannot depend on the recycled work in run_ordered_work()
264 * Note that different types of Btrfs work can depend on in run_ordered_work()
265 * each other, and one type of work on one Btrfs in run_ordered_work()
266 * filesystem may even depend on the same type of work in run_ordered_work()
268 * Therefore, we must not allow the current work item to in run_ordered_work()
278 work->ordered_func(work, true); in run_ordered_work()
279 /* NB: work must not be dereferenced past this point. */ in run_ordered_work()
280 trace_btrfs_all_work_done(wq->fs_info, work); in run_ordered_work()
294 struct btrfs_work *work = container_of(normal_work, struct btrfs_work, in btrfs_work_helper() local
296 struct btrfs_workqueue *wq = work->wq; in btrfs_work_helper()
300 * We should not touch things inside work in the following cases: in btrfs_work_helper()
301 * 1) after work->func() if it has no ordered_func(..., true) to free in btrfs_work_helper()
302 * Since the struct is freed in work->func(). in btrfs_work_helper()
304 * The work may be freed in other threads almost instantly. in btrfs_work_helper()
307 if (work->ordered_func) in btrfs_work_helper()
310 trace_btrfs_work_sched(work); in btrfs_work_helper()
312 work->func(work); in btrfs_work_helper()
315 * Ensures all memory accesses done in the work function are in btrfs_work_helper()
317 * which is going to executed the ordered work sees them. in btrfs_work_helper()
321 set_bit(WORK_DONE_BIT, &work->flags); in btrfs_work_helper()
322 run_ordered_work(wq, work); in btrfs_work_helper()
324 /* NB: work must not be dereferenced past this point. */ in btrfs_work_helper()
325 trace_btrfs_all_work_done(wq->fs_info, work); in btrfs_work_helper()
329 void btrfs_init_work(struct btrfs_work *work, btrfs_func_t func, in btrfs_init_work() argument
332 work->func = func; in btrfs_init_work()
333 work->ordered_func = ordered_func; in btrfs_init_work()
334 INIT_WORK(&work->normal_work, btrfs_work_helper); in btrfs_init_work()
335 INIT_LIST_HEAD(&work->ordered_list); in btrfs_init_work()
336 work->flags = 0; in btrfs_init_work()
339 void btrfs_queue_work(struct btrfs_workqueue *wq, struct btrfs_work *work) in btrfs_queue_work() argument
343 work->wq = wq; in btrfs_queue_work()
345 if (work->ordered_func) { in btrfs_queue_work()
347 list_add_tail(&work->ordered_list, &wq->ordered_list); in btrfs_queue_work()
350 trace_btrfs_work_queued(work); in btrfs_queue_work()
351 queue_work(wq->normal_wq, &work->normal_work); in btrfs_queue_work()