Lines Matching refs:trees

389 	 * We spread entries across multiple trees to reduce contention
394 struct buffer_tree trees[];
407 read_lock_bh(&bc->trees[cache_index(block, bc->num_locks)].u.spinlock);
409 down_read(&bc->trees[cache_index(block, bc->num_locks)].u.lock);
415 read_unlock_bh(&bc->trees[cache_index(block, bc->num_locks)].u.spinlock);
417 up_read(&bc->trees[cache_index(block, bc->num_locks)].u.lock);
423 write_lock_bh(&bc->trees[cache_index(block, bc->num_locks)].u.spinlock);
425 down_write(&bc->trees[cache_index(block, bc->num_locks)].u.lock);
431 write_unlock_bh(&bc->trees[cache_index(block, bc->num_locks)].u.spinlock);
433 up_write(&bc->trees[cache_index(block, bc->num_locks)].u.lock);
459 write_lock_bh(&lh->cache->trees[index].u.spinlock);
461 down_write(&lh->cache->trees[index].u.lock);
464 read_lock_bh(&lh->cache->trees[index].u.spinlock);
466 down_read(&lh->cache->trees[index].u.lock);
474 write_unlock_bh(&lh->cache->trees[index].u.spinlock);
476 up_write(&lh->cache->trees[index].u.lock);
479 read_unlock_bh(&lh->cache->trees[index].u.spinlock);
481 up_read(&lh->cache->trees[index].u.lock);
537 rwlock_init(&bc->trees[i].u.spinlock);
539 init_rwsem(&bc->trees[i].u.lock);
540 bc->trees[i].root = RB_ROOT;
552 WARN_ON_ONCE(!RB_EMPTY_ROOT(&bc->trees[i].root));
610 b = __cache_get(&bc->trees[cache_index(block, bc->num_locks)].root, block);
684 rb_erase(&b->node, &bc->trees[cache_index(b->block, bc->num_locks)].root);
850 r = __cache_insert(&bc->trees[cache_index(b->block, bc->num_locks)].root, b);
876 rb_erase(&b->node, &bc->trees[cache_index(b->block, bc->num_locks)].root);
947 down_write(&bc->trees[i].u.lock);
948 __remove_range(bc, &bc->trees[i].root, begin, end, pred, release);
949 up_write(&bc->trees[i].u.lock);