Lines Matching refs:ds
287 struct dm_deferred_set *ds;
302 struct dm_deferred_set *ds;
304 ds = kmalloc(sizeof(*ds), GFP_KERNEL);
305 if (!ds)
308 spin_lock_init(&ds->lock);
309 ds->current_entry = 0;
310 ds->sweeper = 0;
312 ds->entries[i].ds = ds;
313 ds->entries[i].count = 0;
314 INIT_LIST_HEAD(&ds->entries[i].work_items);
317 return ds;
321 void dm_deferred_set_destroy(struct dm_deferred_set *ds)
323 kfree(ds);
327 struct dm_deferred_entry *dm_deferred_entry_inc(struct dm_deferred_set *ds)
332 spin_lock_irqsave(&ds->lock, flags);
333 entry = ds->entries + ds->current_entry;
335 spin_unlock_irqrestore(&ds->lock, flags);
346 static void __sweep(struct dm_deferred_set *ds, struct list_head *head)
348 while ((ds->sweeper != ds->current_entry) &&
349 !ds->entries[ds->sweeper].count) {
350 list_splice_init(&ds->entries[ds->sweeper].work_items, head);
351 ds->sweeper = ds_next(ds->sweeper);
354 if ((ds->sweeper == ds->current_entry) && !ds->entries[ds->sweeper].count)
355 list_splice_init(&ds->entries[ds->sweeper].work_items, head);
362 spin_lock_irqsave(&entry->ds->lock, flags);
365 __sweep(entry->ds, head);
366 spin_unlock_irqrestore(&entry->ds->lock, flags);
373 int dm_deferred_set_add_work(struct dm_deferred_set *ds, struct list_head *work)
378 spin_lock_irq(&ds->lock);
379 if ((ds->sweeper == ds->current_entry) &&
380 !ds->entries[ds->current_entry].count)
383 list_add(work, &ds->entries[ds->current_entry].work_items);
384 next_entry = ds_next(ds->current_entry);
385 if (!ds->entries[next_entry].count)
386 ds->current_entry = next_entry;
388 spin_unlock_irq(&ds->lock);