Lines Matching refs:nsock
71 struct nbd_sock *nsock; member
305 static void nbd_mark_nsock_dead(struct nbd_device *nbd, struct nbd_sock *nsock, in nbd_mark_nsock_dead() argument
308 if (!nsock->dead && notify && !nbd_disconnected(nbd->config)) { in nbd_mark_nsock_dead()
317 if (!nsock->dead) { in nbd_mark_nsock_dead()
318 kernel_sock_shutdown(nsock->sock, SHUT_RDWR); in nbd_mark_nsock_dead()
329 nsock->dead = true; in nbd_mark_nsock_dead()
330 nsock->pending = NULL; in nbd_mark_nsock_dead()
331 nsock->sent = 0; in nbd_mark_nsock_dead()
409 struct nbd_sock *nsock = config->socks[i]; in sock_shutdown() local
410 mutex_lock(&nsock->tx_lock); in sock_shutdown()
411 nbd_mark_nsock_dead(nbd, nsock, 0); in sock_shutdown()
412 mutex_unlock(&nsock->tx_lock); in sock_shutdown()
493 struct nbd_sock *nsock = in nbd_xmit_timeout() local
495 mutex_lock(&nsock->tx_lock); in nbd_xmit_timeout()
502 if (cmd->cookie == nsock->cookie) in nbd_xmit_timeout()
503 nbd_mark_nsock_dead(nbd, nsock, 1); in nbd_xmit_timeout()
504 mutex_unlock(&nsock->tx_lock); in nbd_xmit_timeout()
518 struct nbd_sock *nsock = config->socks[cmd->index]; in nbd_xmit_timeout() local
525 mutex_lock(&nsock->tx_lock); in nbd_xmit_timeout()
526 if (cmd->cookie != nsock->cookie) { in nbd_xmit_timeout()
528 mutex_unlock(&nsock->tx_lock); in nbd_xmit_timeout()
533 mutex_unlock(&nsock->tx_lock); in nbd_xmit_timeout()
626 struct nbd_sock *nsock, in nbd_sched_pending_work() argument
634 nsock->pending = req; in nbd_sched_pending_work()
635 nsock->sent = sent; in nbd_sched_pending_work()
638 schedule_work(&nsock->work); in nbd_sched_pending_work()
650 struct nbd_sock *nsock = config->socks[index]; in nbd_send_cmd() local
659 int sent = nsock->sent, skip = 0; in nbd_send_cmd()
662 lockdep_assert_held(&nsock->tx_lock); in nbd_send_cmd()
700 cmd->cookie = nsock->cookie; in nbd_send_cmd()
726 nbd_sched_pending_work(nbd, nsock, cmd, sent); in nbd_send_cmd()
764 nbd_sched_pending_work(nbd, nsock, cmd, sent); in nbd_send_cmd()
785 nsock->pending = NULL; in nbd_send_cmd()
786 nsock->sent = 0; in nbd_send_cmd()
802 nbd_mark_nsock_dead(nbd, nsock, 1); in nbd_send_cmd()
810 struct nbd_sock *nsock = container_of(work, struct nbd_sock, work); in nbd_pending_cmd_work() local
811 struct request *req = nsock->pending; in nbd_pending_cmd_work()
823 mutex_lock(&nsock->tx_lock); in nbd_pending_cmd_work()
826 if (!nsock->pending) in nbd_pending_cmd_work()
838 mutex_unlock(&nsock->tx_lock); in nbd_pending_cmd_work()
978 struct nbd_sock *nsock = args->nsock; in recv_work() local
985 if (nbd_read_reply(nbd, nsock->sock, &reply)) in recv_work()
1020 mutex_lock(&nsock->tx_lock); in recv_work()
1021 nbd_mark_nsock_dead(nbd, nsock, 1); in recv_work()
1022 mutex_unlock(&nsock->tx_lock); in recv_work()
1062 struct nbd_sock *nsock = config->socks[index]; in find_fallback() local
1063 int fallback = nsock->fallback_index; in find_fallback()
1078 if (nsock->fallback_index < 0 || in find_fallback()
1079 nsock->fallback_index >= config->num_connections || in find_fallback()
1080 config->socks[nsock->fallback_index]->dead) { in find_fallback()
1090 nsock->fallback_index = new_index; in find_fallback()
1097 new_index = nsock->fallback_index; in find_fallback()
1122 struct nbd_sock *nsock; in nbd_handle_cmd() local
1142 nsock = config->socks[index]; in nbd_handle_cmd()
1143 mutex_lock(&nsock->tx_lock); in nbd_handle_cmd()
1144 if (nsock->dead) { in nbd_handle_cmd()
1147 mutex_unlock(&nsock->tx_lock); in nbd_handle_cmd()
1172 if (unlikely(nsock->pending && nsock->pending != req)) { in nbd_handle_cmd()
1179 mutex_unlock(&nsock->tx_lock); in nbd_handle_cmd()
1247 struct nbd_sock *nsock; in nbd_add_socket() local
1277 nsock = kzalloc_obj(*nsock); in nbd_add_socket()
1278 if (!nsock) { in nbd_add_socket()
1286 kfree(nsock); in nbd_add_socket()
1293 nsock->fallback_index = -1; in nbd_add_socket()
1294 nsock->dead = false; in nbd_add_socket()
1295 mutex_init(&nsock->tx_lock); in nbd_add_socket()
1296 nsock->sock = sock; in nbd_add_socket()
1297 nsock->pending = NULL; in nbd_add_socket()
1298 nsock->sent = 0; in nbd_add_socket()
1299 nsock->cookie = 0; in nbd_add_socket()
1300 INIT_WORK(&nsock->work, nbd_pending_cmd_work); in nbd_add_socket()
1301 socks[config->num_connections++] = nsock; in nbd_add_socket()
1332 struct nbd_sock *nsock = config->socks[i]; in nbd_reconnect_socket() local
1334 if (!nsock->dead) in nbd_reconnect_socket()
1337 mutex_lock(&nsock->tx_lock); in nbd_reconnect_socket()
1338 if (!nsock->dead) { in nbd_reconnect_socket()
1339 mutex_unlock(&nsock->tx_lock); in nbd_reconnect_socket()
1347 old = nsock->sock; in nbd_reconnect_socket()
1348 nsock->fallback_index = -1; in nbd_reconnect_socket()
1349 nsock->sock = sock; in nbd_reconnect_socket()
1350 nsock->dead = false; in nbd_reconnect_socket()
1354 args->nsock = nsock; in nbd_reconnect_socket()
1355 nsock->cookie++; in nbd_reconnect_socket()
1356 mutex_unlock(&nsock->tx_lock); in nbd_reconnect_socket()
1402 struct nbd_sock *nsock = config->socks[i]; in send_disconnects() local
1405 mutex_lock(&nsock->tx_lock); in send_disconnects()
1410 mutex_unlock(&nsock->tx_lock); in send_disconnects()
1536 args->nsock = config->socks[i]; in nbd_start_device()