Searched refs:c_path (Results 1 – 13 of 13) sorted by relevance
| /linux/net/rds/ |
| H A D | rds_single_path.h | 5 #define c_xmit_rm c_path[0].cp_xmit_rm 6 #define c_xmit_sg c_path[0].cp_xmit_sg 7 #define c_xmit_hdr_off c_path[0].cp_xmit_hdr_off 8 #define c_xmit_data_off c_path[0].cp_xmit_data_off 9 #define c_xmit_atomic_sent c_path[0].cp_xmit_atomic_sent 10 #define c_xmit_rdma_sent c_path[0].cp_xmit_rdma_sent 11 #define c_xmit_data_sent c_path[0].cp_xmit_data_sent 12 #define c_lock c_path[0].cp_lock 13 #define c_next_tx_seq c_path[0].cp_next_tx_seq 14 #define c_send_queue c_path[0].cp_send_queue [all …]
|
| H A D | connection.c | 200 conn->c_path = kzalloc_objs(struct rds_conn_path, npaths, gfp); in __rds_conn_create() 201 if (!conn->c_path) { in __rds_conn_create() 230 kfree(conn->c_path); in __rds_conn_create() 258 kfree(conn->c_path); in __rds_conn_create() 270 __rds_conn_path_init(conn, &conn->c_path[i], in __rds_conn_create() 272 conn->c_path[i].cp_index = i; in __rds_conn_create() 273 conn->c_path[i].cp_wq = in __rds_conn_create() 276 if (!conn->c_path[i].cp_wq) in __rds_conn_create() 277 conn->c_path[i].cp_wq = rds_wq; in __rds_conn_create() 286 free_cp = conn->c_path; in __rds_conn_create() [all …]
|
| H A D | tcp_listen.c | 98 struct rds_conn_path *cp = &conn->c_path[i]; in rds_tcp_accept_one_path() 118 tc = conn->c_path->cp_transport_data; in rds_tcp_conn_slots_available() 135 rds_conn_path_drop(conn->c_path, 0); in rds_tcp_conn_slots_available() 271 rds_conn_path_connect_if_down(&conn->c_path[0]); in rds_tcp_accept_one()
|
| H A D | send.c | 130 struct rds_conn_path *cp0 = conn->c_path; in rds_mprds_cp0_catchup() 777 rds_send_path_drop_acked(&conn->c_path[0], ack, is_acked); in rds_send_drop_acked() 818 cp = &conn->c_path[0]; in rds_send_drop_to() 1337 cpath = &conn->c_path[RDS_MPATH_HASH(rs, conn->c_npaths ? : 1)]; in rds_sendmsg() 1339 cpath = &conn->c_path[0]; in rds_sendmsg() 1348 !rds_conn_path_up(&conn->c_path[0])) { in rds_sendmsg() 1354 &conn->c_path[0].cp_flags)) in rds_sendmsg() 1355 queue_delayed_work(conn->c_path[0].cp_wq, in rds_sendmsg() 1356 &conn->c_path[0].cp_conn_w, 0); in rds_sendmsg() 1551 struct rds_conn_path *cp = &conn->c_path[cp_index]; in rds_send_ping()
|
| H A D | tcp.c | 389 conn->c_path[i].cp_transport_data = tc; in rds_tcp_conn_alloc() 390 tc->t_cpath = &conn->c_path[i]; in rds_tcp_conn_alloc() 394 conn->c_path[i].cp_transport_data); in rds_tcp_conn_alloc() 398 tc = conn->c_path[i].cp_transport_data; in rds_tcp_conn_alloc() 406 rds_tcp_conn_free(conn->c_path[j].cp_transport_data); in rds_tcp_conn_alloc()
|
| H A D | rds.h | 168 struct rds_conn_path *c_path; member 838 return rds_conn_path_transition(&conn->c_path[0], old, new); in rds_conn_transition() 851 return rds_conn_path_state(&conn->c_path[0]); in rds_conn_state() 870 return rds_conn_path_up(&conn->c_path[0]); in rds_conn_up() 883 return rds_conn_path_connecting(&conn->c_path[0]); in rds_conn_connecting()
|
| H A D | recv.c | 147 cp = &conn->c_path[i]; in rds_conn_peer_gen_update() 246 struct rds_conn_path *cp0 = conn->c_path; in rds_recv_hs_exthdrs() 294 cp = &conn->c_path[i]; in rds_start_mprds() 330 cp = &conn->c_path[0]; in rds_recv_incoming()
|
| H A D | ib.c | 90 rds_conn_path_drop(&ic->conn->c_path[0], true); in rds_ib_dev_shutdown() 504 struct rds_conn_path *cp = &conn->c_path[0]; in rds_ib_is_unloading()
|
| H A D | threads.c | 102 rds_connect_path_complete(&conn->c_path[0], RDS_CONN_CONNECTING); in rds_connect_complete()
|
| H A D | cong.c | 225 struct rds_conn_path *cp = &conn->c_path[0]; in rds_cong_queue_updates()
|
| H A D | ib_send.c | 300 queue_delayed_work(conn->c_path->cp_wq, &conn->c_send_w, 0); in rds_ib_send_cqe_handler() 422 queue_delayed_work(conn->c_path->cp_wq, &conn->c_send_w, 0); in rds_ib_send_add_credits()
|
| H A D | ib_recv.c | 460 queue_delayed_work(conn->c_path->cp_wq, &conn->c_recv_w, 1); in rds_ib_recv_refill()
|
| H A D | ib_cm.c | 300 rds_send_xmit(&ic->conn->c_path[0]); in rds_ib_tasklet_fn_send()
|