Lines Matching refs:osd

49 static void link_request(struct ceph_osd *osd, struct ceph_osd_request *req);
50 static void unlink_request(struct ceph_osd *osd, struct ceph_osd_request *req);
51 static void link_linger(struct ceph_osd *osd,
53 static void unlink_linger(struct ceph_osd *osd,
55 static void clear_backoffs(struct ceph_osd *osd);
77 static inline void verify_osd_locked(struct ceph_osd *osd) in verify_osd_locked() argument
79 struct ceph_osd_client *osdc = osd->o_osdc; in verify_osd_locked()
81 WARN_ON(!(mutex_is_locked(&osd->lock) && in verify_osd_locked()
92 static inline void verify_osd_locked(struct ceph_osd *osd) { } in verify_osd_locked() argument
432 t->osd = CEPH_HOMELESS_OSD; in target_init()
461 dest->osd = src->osd; in target_copy()
1180 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in DEFINE_RB_FUNCS() local
1182 for (p = rb_first(&osd->o_requests); p; ) { in DEFINE_RB_FUNCS()
1202 static bool osd_homeless(struct ceph_osd *osd) in osd_homeless() argument
1204 return osd->o_osd == CEPH_HOMELESS_OSD; in osd_homeless()
1207 static bool osd_registered(struct ceph_osd *osd) in osd_registered() argument
1209 verify_osdc_locked(osd->o_osdc); in osd_registered()
1211 return !RB_EMPTY_NODE(&osd->o_node); in osd_registered()
1217 static void osd_init(struct ceph_osd *osd) in osd_init() argument
1219 refcount_set(&osd->o_ref, 1); in osd_init()
1220 RB_CLEAR_NODE(&osd->o_node); in osd_init()
1221 spin_lock_init(&osd->o_requests_lock); in osd_init()
1222 osd->o_requests = RB_ROOT; in osd_init()
1223 osd->o_linger_requests = RB_ROOT; in osd_init()
1224 osd->o_backoff_mappings = RB_ROOT; in osd_init()
1225 osd->o_backoffs_by_id = RB_ROOT; in osd_init()
1226 INIT_LIST_HEAD(&osd->o_osd_lru); in osd_init()
1227 INIT_LIST_HEAD(&osd->o_keepalive_item); in osd_init()
1228 osd->o_incarnation = 1; in osd_init()
1229 mutex_init(&osd->lock); in osd_init()
1239 static void osd_cleanup(struct ceph_osd *osd) in osd_cleanup() argument
1241 WARN_ON(!RB_EMPTY_NODE(&osd->o_node)); in osd_cleanup()
1242 WARN_ON(!RB_EMPTY_ROOT(&osd->o_requests)); in osd_cleanup()
1243 WARN_ON(!RB_EMPTY_ROOT(&osd->o_linger_requests)); in osd_cleanup()
1244 WARN_ON(!RB_EMPTY_ROOT(&osd->o_backoff_mappings)); in osd_cleanup()
1245 WARN_ON(!RB_EMPTY_ROOT(&osd->o_backoffs_by_id)); in osd_cleanup()
1246 WARN_ON(!list_empty(&osd->o_osd_lru)); in osd_cleanup()
1247 WARN_ON(!list_empty(&osd->o_keepalive_item)); in osd_cleanup()
1249 ceph_init_sparse_read(&osd->o_sparse_read); in osd_cleanup()
1251 if (osd->o_auth.authorizer) { in osd_cleanup()
1252 WARN_ON(osd_homeless(osd)); in osd_cleanup()
1253 ceph_auth_destroy_authorizer(osd->o_auth.authorizer); in osd_cleanup()
1262 struct ceph_osd *osd; in create_osd() local
1266 osd = kzalloc_obj(*osd, GFP_NOIO | __GFP_NOFAIL); in create_osd()
1267 osd_init(osd); in create_osd()
1268 osd->o_osdc = osdc; in create_osd()
1269 osd->o_osd = onum; in create_osd()
1270 osd->o_sparse_op_idx = -1; in create_osd()
1272 ceph_init_sparse_read(&osd->o_sparse_read); in create_osd()
1274 ceph_con_init(&osd->o_con, osd, &osd_con_ops, &osdc->client->msgr); in create_osd()
1276 return osd; in create_osd()
1279 static struct ceph_osd *get_osd(struct ceph_osd *osd) in get_osd() argument
1281 if (refcount_inc_not_zero(&osd->o_ref)) { in get_osd()
1282 dout("get_osd %p -> %d\n", osd, refcount_read(&osd->o_ref)); in get_osd()
1283 return osd; in get_osd()
1285 dout("get_osd %p FAIL\n", osd); in get_osd()
1290 static void put_osd(struct ceph_osd *osd) in put_osd() argument
1292 dout("put_osd %p -> %d\n", osd, refcount_read(&osd->o_ref) - 1); in put_osd()
1293 if (refcount_dec_and_test(&osd->o_ref)) { in put_osd()
1294 osd_cleanup(osd); in put_osd()
1295 kfree(osd); in put_osd()
1299 DEFINE_RB_FUNCS(osd, struct ceph_osd, o_osd, o_node) in DEFINE_RB_FUNCS() argument
1301 static void __move_osd_to_lru(struct ceph_osd *osd) in DEFINE_RB_FUNCS()
1303 struct ceph_osd_client *osdc = osd->o_osdc; in DEFINE_RB_FUNCS()
1305 dout("%s osd %p osd%d\n", __func__, osd, osd->o_osd); in DEFINE_RB_FUNCS()
1306 BUG_ON(!list_empty(&osd->o_osd_lru)); in DEFINE_RB_FUNCS()
1309 list_add_tail(&osd->o_osd_lru, &osdc->osd_lru); in DEFINE_RB_FUNCS()
1312 osd->lru_ttl = jiffies + osdc->client->options->osd_idle_ttl; in DEFINE_RB_FUNCS()
1315 static void maybe_move_osd_to_lru(struct ceph_osd *osd) in maybe_move_osd_to_lru() argument
1317 if (RB_EMPTY_ROOT(&osd->o_requests) && in maybe_move_osd_to_lru()
1318 RB_EMPTY_ROOT(&osd->o_linger_requests)) in maybe_move_osd_to_lru()
1319 __move_osd_to_lru(osd); in maybe_move_osd_to_lru()
1322 static void __remove_osd_from_lru(struct ceph_osd *osd) in __remove_osd_from_lru() argument
1324 struct ceph_osd_client *osdc = osd->o_osdc; in __remove_osd_from_lru()
1326 dout("%s osd %p osd%d\n", __func__, osd, osd->o_osd); in __remove_osd_from_lru()
1329 if (!list_empty(&osd->o_osd_lru)) in __remove_osd_from_lru()
1330 list_del_init(&osd->o_osd_lru); in __remove_osd_from_lru()
1338 static void close_osd(struct ceph_osd *osd) in close_osd() argument
1340 struct ceph_osd_client *osdc = osd->o_osdc; in close_osd()
1344 dout("%s osd %p osd%d\n", __func__, osd, osd->o_osd); in close_osd()
1346 ceph_con_close(&osd->o_con); in close_osd()
1348 for (n = rb_first(&osd->o_requests); n; ) { in close_osd()
1355 unlink_request(osd, req); in close_osd()
1358 for (n = rb_first(&osd->o_linger_requests); n; ) { in close_osd()
1366 unlink_linger(osd, lreq); in close_osd()
1369 clear_backoffs(osd); in close_osd()
1371 __remove_osd_from_lru(osd); in close_osd()
1372 erase_osd(&osdc->osds, osd); in close_osd()
1373 put_osd(osd); in close_osd()
1379 static int reopen_osd(struct ceph_osd *osd) in reopen_osd() argument
1383 dout("%s osd %p osd%d\n", __func__, osd, osd->o_osd); in reopen_osd()
1385 if (RB_EMPTY_ROOT(&osd->o_requests) && in reopen_osd()
1386 RB_EMPTY_ROOT(&osd->o_linger_requests)) { in reopen_osd()
1387 close_osd(osd); in reopen_osd()
1391 peer_addr = &osd->o_osdc->osdmap->osd_addr[osd->o_osd]; in reopen_osd()
1392 if (!memcmp(peer_addr, &osd->o_con.peer_addr, sizeof (*peer_addr)) && in reopen_osd()
1393 !ceph_con_opened(&osd->o_con)) { in reopen_osd()
1399 for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) { in reopen_osd()
1408 ceph_con_close(&osd->o_con); in reopen_osd()
1409 ceph_con_open(&osd->o_con, CEPH_ENTITY_TYPE_OSD, osd->o_osd, peer_addr); in reopen_osd()
1410 osd->o_incarnation++; in reopen_osd()
1418 struct ceph_osd *osd; in lookup_create_osd() local
1426 osd = lookup_osd(&osdc->osds, o); in lookup_create_osd()
1428 osd = &osdc->homeless_osd; in lookup_create_osd()
1429 if (!osd) { in lookup_create_osd()
1433 osd = create_osd(osdc, o); in lookup_create_osd()
1434 insert_osd(&osdc->osds, osd); in lookup_create_osd()
1435 ceph_con_open(&osd->o_con, CEPH_ENTITY_TYPE_OSD, osd->o_osd, in lookup_create_osd()
1436 &osdc->osdmap->osd_addr[osd->o_osd]); in lookup_create_osd()
1439 dout("%s osdc %p osd%d -> osd %p\n", __func__, osdc, o, osd); in lookup_create_osd()
1440 return osd; in lookup_create_osd()
1448 static void link_request(struct ceph_osd *osd, struct ceph_osd_request *req) in link_request() argument
1450 verify_osd_locked(osd); in link_request()
1452 dout("%s osd %p osd%d req %p tid %llu\n", __func__, osd, osd->o_osd, in link_request()
1455 if (!osd_homeless(osd)) in link_request()
1456 __remove_osd_from_lru(osd); in link_request()
1458 atomic_inc(&osd->o_osdc->num_homeless); in link_request()
1460 get_osd(osd); in link_request()
1461 spin_lock(&osd->o_requests_lock); in link_request()
1462 insert_request(&osd->o_requests, req); in link_request()
1463 spin_unlock(&osd->o_requests_lock); in link_request()
1464 req->r_osd = osd; in link_request()
1467 static void unlink_request(struct ceph_osd *osd, struct ceph_osd_request *req) in unlink_request() argument
1469 verify_osd_locked(osd); in unlink_request()
1470 WARN_ON(req->r_osd != osd); in unlink_request()
1471 dout("%s osd %p osd%d req %p tid %llu\n", __func__, osd, osd->o_osd, in unlink_request()
1475 spin_lock(&osd->o_requests_lock); in unlink_request()
1476 erase_request(&osd->o_requests, req); in unlink_request()
1477 spin_unlock(&osd->o_requests_lock); in unlink_request()
1478 put_osd(osd); in unlink_request()
1480 if (!osd_homeless(osd)) in unlink_request()
1481 maybe_move_osd_to_lru(osd); in unlink_request()
1483 atomic_dec(&osd->o_osdc->num_homeless); in unlink_request()
1603 t->osd = CEPH_HOMELESS_OSD; in calc_target()
1628 t->osd = CEPH_HOMELESS_OSD; in calc_target()
1697 t->osd = acting.osds[pos]; in calc_target()
1700 t->osd = acting.primary; in calc_target()
1712 legacy_change, force_resend, split, ct_res, t->osd); in calc_target()
1971 static void clear_backoffs(struct ceph_osd *osd) in DEFINE_RB_FUNCS()
1973 while (!RB_EMPTY_ROOT(&osd->o_backoff_mappings)) { in DEFINE_RB_FUNCS()
1975 rb_entry(rb_first(&osd->o_backoff_mappings), in DEFINE_RB_FUNCS()
1984 erase_backoff_by_id(&osd->o_backoffs_by_id, backoff); in DEFINE_RB_FUNCS()
1987 erase_spg_mapping(&osd->o_backoff_mappings, spg); in DEFINE_RB_FUNCS()
2018 struct ceph_osd *osd = req->r_osd; in should_plug_request() local
2023 spg = lookup_spg_mapping(&osd->o_backoff_mappings, &req->r_t.spgid); in should_plug_request()
2033 __func__, req, req->r_tid, osd->o_osd, backoff->spgid.pgid.pool, in should_plug_request()
2318 struct ceph_osd *osd = req->r_osd; in send_request() local
2320 verify_osd_locked(osd); in send_request()
2321 WARN_ON(osd->o_osd != req->r_t.osd); in send_request()
2345 req->r_t.spgid.shard, osd->o_osd, req->r_t.epoch, req->r_flags, in send_request()
2352 req->r_sent = osd->o_incarnation; in send_request()
2354 ceph_con_send(&osd->o_con, ceph_msg_get(req->r_request)); in send_request()
2384 struct ceph_osd *osd; in __submit_request() local
2398 osd = lookup_create_osd(osdc, req->r_t.osd, wrlocked); in __submit_request()
2399 if (IS_ERR(osd)) { in __submit_request()
2400 WARN_ON(PTR_ERR(osd) != -EAGAIN || wrlocked); in __submit_request()
2439 } else if (!osd_homeless(osd)) { in __submit_request()
2445 mutex_lock(&osd->lock); in __submit_request()
2452 link_request(osd, req); in __submit_request()
2457 mutex_unlock(&osd->lock); in __submit_request()
2777 WARN_ON(lreq->osd); in linger_release()
2838 static void link_linger(struct ceph_osd *osd, in DEFINE_RB_INSDEL_FUNCS()
2841 verify_osd_locked(osd); in DEFINE_RB_INSDEL_FUNCS()
2842 WARN_ON(!lreq->linger_id || lreq->osd); in DEFINE_RB_INSDEL_FUNCS()
2843 dout("%s osd %p osd%d lreq %p linger_id %llu\n", __func__, osd, in DEFINE_RB_INSDEL_FUNCS()
2844 osd->o_osd, lreq, lreq->linger_id); in DEFINE_RB_INSDEL_FUNCS()
2846 if (!osd_homeless(osd)) in DEFINE_RB_INSDEL_FUNCS()
2847 __remove_osd_from_lru(osd); in DEFINE_RB_INSDEL_FUNCS()
2849 atomic_inc(&osd->o_osdc->num_homeless); in DEFINE_RB_INSDEL_FUNCS()
2851 get_osd(osd); in DEFINE_RB_INSDEL_FUNCS()
2852 insert_linger(&osd->o_linger_requests, lreq); in DEFINE_RB_INSDEL_FUNCS()
2853 lreq->osd = osd; in DEFINE_RB_INSDEL_FUNCS()
2856 static void unlink_linger(struct ceph_osd *osd, in unlink_linger() argument
2859 verify_osd_locked(osd); in unlink_linger()
2860 WARN_ON(lreq->osd != osd); in unlink_linger()
2861 dout("%s osd %p osd%d lreq %p linger_id %llu\n", __func__, osd, in unlink_linger()
2862 osd->o_osd, lreq, lreq->linger_id); in unlink_linger()
2864 lreq->osd = NULL; in unlink_linger()
2865 erase_linger(&osd->o_linger_requests, lreq); in unlink_linger()
2866 put_osd(osd); in unlink_linger()
2868 if (!osd_homeless(osd)) in unlink_linger()
2869 maybe_move_osd_to_lru(osd); in unlink_linger()
2871 atomic_dec(&osd->o_osdc->num_homeless); in unlink_linger()
3254 link_request(lreq->osd, req); in send_linger_ping()
3261 struct ceph_osd *osd; in linger_submit() local
3267 osd = lookup_create_osd(osdc, lreq->t.osd, true); in linger_submit()
3268 link_linger(osd, lreq); in linger_submit()
3301 unlink_linger(lreq->osd, lreq); in __linger_cancel()
3448 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in handle_timeout() local
3451 for (p = rb_first(&osd->o_requests); p; ) { in handle_timeout()
3459 req, req->r_tid, osd->o_osd); in handle_timeout()
3465 req->r_tid, osd->o_osd); in handle_timeout()
3469 for (p = rb_first(&osd->o_linger_requests); p; p = rb_next(p)) { in handle_timeout()
3474 lreq, lreq->linger_id, osd->o_osd); in handle_timeout()
3484 list_move_tail(&osd->o_keepalive_item, &slow_osds); in handle_timeout()
3506 struct ceph_osd *osd = list_first_entry(&slow_osds, in handle_timeout() local
3509 list_del_init(&osd->o_keepalive_item); in handle_timeout()
3510 ceph_con_keepalive(&osd->o_con); in handle_timeout()
3524 struct ceph_osd *osd, *nosd; in handle_osds_timeout() local
3528 list_for_each_entry_safe(osd, nosd, &osdc->osd_lru, o_osd_lru) { in handle_osds_timeout()
3529 if (time_before(jiffies, osd->lru_ttl)) in handle_osds_timeout()
3532 WARN_ON(!RB_EMPTY_ROOT(&osd->o_requests)); in handle_osds_timeout()
3533 WARN_ON(!RB_EMPTY_ROOT(&osd->o_linger_requests)); in handle_osds_timeout()
3534 close_osd(osd); in handle_osds_timeout()
3749 static void handle_reply(struct ceph_osd *osd, struct ceph_msg *msg) in handle_reply() argument
3751 struct ceph_osd_client *osdc = osd->o_osdc; in handle_reply()
3762 if (!osd_registered(osd)) { in handle_reply()
3763 dout("%s osd%d unknown\n", __func__, osd->o_osd); in handle_reply()
3766 WARN_ON(osd->o_osd != le64_to_cpu(msg->hdr.src.num)); in handle_reply()
3768 mutex_lock(&osd->lock); in handle_reply()
3769 req = lookup_request(&osd->o_requests, tid); in handle_reply()
3771 dout("%s osd%d tid %llu unknown\n", __func__, osd->o_osd, tid); in handle_reply()
3803 unlink_request(osd, req); in handle_reply()
3804 mutex_unlock(&osd->lock); in handle_reply()
3821 unlink_request(osd, req); in handle_reply()
3822 mutex_unlock(&osd->lock); in handle_reply()
3867 mutex_unlock(&osd->lock); in handle_reply()
3876 mutex_unlock(&osd->lock); in handle_reply()
3912 struct ceph_osd *osd; in recalc_linger_target() local
3914 osd = lookup_create_osd(osdc, lreq->t.osd, true); in recalc_linger_target()
3915 if (osd != lreq->osd) { in recalc_linger_target()
3916 unlink_linger(lreq->osd, lreq); in recalc_linger_target()
3917 link_linger(osd, lreq); in recalc_linger_target()
3927 static void scan_requests(struct ceph_osd *osd, in scan_requests() argument
3934 struct ceph_osd_client *osdc = osd->o_osdc; in scan_requests()
3938 for (n = rb_first(&osd->o_linger_requests); n; ) { in scan_requests()
3974 for (n = rb_first(&osd->o_requests); n; ) { in scan_requests()
3996 unlink_request(osd, req); in scan_requests()
4061 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in handle_one_map() local
4065 scan_requests(osd, skipped_map, was_full, true, need_resend, in handle_one_map()
4067 if (!ceph_osd_is_up(osdc->osdmap, osd->o_osd) || in handle_one_map()
4068 memcmp(&osd->o_con.peer_addr, in handle_one_map()
4069 ceph_osd_addr(osdc->osdmap, osd->o_osd), in handle_one_map()
4071 close_osd(osd); in handle_one_map()
4104 struct ceph_osd *osd; in kick_requests() local
4109 osd = lookup_create_osd(osdc, req->r_t.osd, true); in kick_requests()
4110 link_request(osd, req); in kick_requests()
4112 if (!osd_homeless(osd) && !req->r_t.paused) in kick_requests()
4120 if (!osd_homeless(lreq->osd)) in kick_requests()
4247 static void kick_osd_requests(struct ceph_osd *osd) in kick_osd_requests() argument
4251 clear_backoffs(osd); in kick_osd_requests()
4253 for (n = rb_first(&osd->o_requests); n; ) { in kick_osd_requests()
4266 for (n = rb_first(&osd->o_linger_requests); n; n = rb_next(n)) { in kick_osd_requests()
4279 struct ceph_osd *osd = con->private; in osd_fault() local
4280 struct ceph_osd_client *osdc = osd->o_osdc; in osd_fault()
4282 dout("%s osd %p osd%d\n", __func__, osd, osd->o_osd); in osd_fault()
4285 if (!osd_registered(osd)) { in osd_fault()
4286 dout("%s osd%d unknown\n", __func__, osd->o_osd); in osd_fault()
4290 osd->o_sparse_op_idx = -1; in osd_fault()
4291 ceph_init_sparse_read(&osd->o_sparse_read); in osd_fault()
4293 if (!reopen_osd(osd)) in osd_fault()
4294 kick_osd_requests(osd); in osd_fault()
4398 static void handle_backoff_block(struct ceph_osd *osd, struct MOSDBackoff *m) in handle_backoff_block() argument
4404 dout("%s osd%d spgid %llu.%xs%d id %llu\n", __func__, osd->o_osd, in handle_backoff_block()
4407 spg = lookup_spg_mapping(&osd->o_backoff_mappings, &m->spgid); in handle_backoff_block()
4415 insert_spg_mapping(&osd->o_backoff_mappings, spg); in handle_backoff_block()
4431 insert_backoff_by_id(&osd->o_backoffs_by_id, backoff); in handle_backoff_block()
4442 ceph_con_send(&osd->o_con, msg); in handle_backoff_block()
4457 static void handle_backoff_unblock(struct ceph_osd *osd, in handle_backoff_unblock() argument
4464 dout("%s osd%d spgid %llu.%xs%d id %llu\n", __func__, osd->o_osd, in handle_backoff_unblock()
4467 backoff = lookup_backoff_by_id(&osd->o_backoffs_by_id, m->id); in handle_backoff_unblock()
4470 __func__, osd->o_osd, m->spgid.pgid.pool, in handle_backoff_unblock()
4478 __func__, osd->o_osd, m->spgid.pgid.pool, in handle_backoff_unblock()
4483 spg = lookup_spg_mapping(&osd->o_backoff_mappings, &backoff->spgid); in handle_backoff_unblock()
4487 erase_backoff_by_id(&osd->o_backoffs_by_id, backoff); in handle_backoff_unblock()
4491 erase_spg_mapping(&osd->o_backoff_mappings, spg); in handle_backoff_unblock()
4495 for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) { in handle_backoff_unblock()
4515 static void handle_backoff(struct ceph_osd *osd, struct ceph_msg *msg) in handle_backoff() argument
4517 struct ceph_osd_client *osdc = osd->o_osdc; in handle_backoff()
4522 if (!osd_registered(osd)) { in handle_backoff()
4523 dout("%s osd%d unknown\n", __func__, osd->o_osd); in handle_backoff()
4527 WARN_ON(osd->o_osd != le64_to_cpu(msg->hdr.src.num)); in handle_backoff()
4529 mutex_lock(&osd->lock); in handle_backoff()
4539 handle_backoff_block(osd, &m); in handle_backoff()
4542 handle_backoff_unblock(osd, &m); in handle_backoff()
4545 pr_err("%s osd%d unknown op %d\n", __func__, osd->o_osd, m.op); in handle_backoff()
4552 mutex_unlock(&osd->lock); in handle_backoff()
4729 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in ceph_osdc_sync() local
4731 mutex_lock(&osd->lock); in ceph_osdc_sync()
4732 for (p = rb_first(&osd->o_requests); p; p = rb_next(p)) { in ceph_osdc_sync()
4743 mutex_unlock(&osd->lock); in ceph_osdc_sync()
4752 mutex_unlock(&osd->lock); in ceph_osdc_sync()
5187 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in ceph_osdc_reopen_osds() local
5190 if (!reopen_osd(osd)) in ceph_osdc_reopen_osds()
5191 kick_osd_requests(osd); in ceph_osdc_reopen_osds()
5278 struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds), in ceph_osdc_stop() local
5280 close_osd(osd); in ceph_osdc_stop()
5361 struct ceph_osd *osd = con->private; in osd_dispatch() local
5362 struct ceph_osd_client *osdc = osd->o_osdc; in osd_dispatch()
5370 handle_reply(osd, msg); in osd_dispatch()
5373 handle_backoff(osd, msg); in osd_dispatch()
5414 struct ceph_osd *osd = con->private; in get_reply() local
5415 struct ceph_osd_client *osdc = osd->o_osdc; in get_reply()
5424 if (!osd_registered(osd)) { in get_reply()
5425 dout("%s osd%d unknown, skipping\n", __func__, osd->o_osd); in get_reply()
5429 WARN_ON(osd->o_osd != le64_to_cpu(hdr->src.num)); in get_reply()
5431 mutex_lock(&osd->lock); in get_reply()
5432 req = lookup_request(&osd->o_requests, tid); in get_reply()
5435 osd->o_osd, tid); in get_reply()
5444 __func__, osd->o_osd, req->r_tid, front_len, in get_reply()
5457 __func__, osd->o_osd, req->r_tid, data_len, in get_reply()
5470 mutex_unlock(&osd->lock); in get_reply()
5507 struct ceph_osd *osd = con->private; in osd_alloc_msg() local
5520 osd->o_osd, type); in osd_alloc_msg()
5531 struct ceph_osd *osd = con->private; in osd_get_con() local
5532 if (get_osd(osd)) in osd_get_con()
5539 struct ceph_osd *osd = con->private; in osd_put_con() local
5540 put_osd(osd); in osd_put_con()