Lines Matching defs:msk

445 	struct mptcp_sock *msk = mptcp_sk(subflow->conn);
477 opts->csum_reqd = READ_ONCE(msk->csum_enabled);
558 struct mptcp_sock *msk = mptcp_sk(subflow->conn);
565 opts->csum_reqd = READ_ONCE(msk->csum_enabled);
585 /* passive sockets msk will set the 'can_ack' after accept(), even
589 if (!READ_ONCE(msk->can_ack)) {
594 ack_seq = READ_ONCE(msk->ack_seq);
595 if (READ_ONCE(msk->use_64bit_ack)) {
650 struct mptcp_sock *msk = mptcp_sk(subflow->conn);
660 if (!mptcp_pm_should_add_signal(msk) ||
662 !mptcp_pm_add_addr_signal(msk, skb, opt_size, remaining, &addr,
696 opts->ahmac = add_addr_generate_hmac(READ_ONCE(msk->local_key),
697 READ_ONCE(msk->remote_key),
714 struct mptcp_sock *msk = mptcp_sk(subflow->conn);
718 if (!mptcp_pm_should_rm_signal(msk) ||
719 !(mptcp_pm_rm_addr_signal(msk, remaining, &rm_list)))
789 struct mptcp_sock *msk = mptcp_sk(subflow->conn);
799 opts->rcvr_key = READ_ONCE(msk->remote_key);
834 struct mptcp_sock *msk = mptcp_sk(subflow->conn);
841 if (unlikely(__mptcp_check_fallback(msk) && !mptcp_check_infinite_map(skb)))
858 snd_data_fin = mptcp_data_fin_enabled(msk);
929 static bool check_fully_established(struct mptcp_sock *msk, struct sock *ssk,
939 * note that msk is always set by subflow_syn_recv_sock()
989 WRITE_ONCE(msk->pm.remote_deny_join_id0, true);
991 if (unlikely(!READ_ONCE(msk->pm.server_side)))
995 mptcp_data_lock((struct sock *)msk);
996 __mptcp_subflow_fully_established(msk, subflow, mp_opt);
997 mptcp_data_unlock((struct sock *)msk);
1011 mptcp_pm_subflow_established(msk);
1013 mptcp_pm_fully_established(msk, ssk);
1038 static void __mptcp_snd_una_update(struct mptcp_sock *msk, u64 new_snd_una)
1040 msk->bytes_acked += new_snd_una - msk->snd_una;
1041 WRITE_ONCE(msk->snd_una, new_snd_una);
1044 static void ack_update_msk(struct mptcp_sock *msk,
1048 u64 new_wnd_end, new_snd_una, snd_nxt = READ_ONCE(msk->snd_nxt);
1049 struct sock *sk = (struct sock *)msk;
1058 old_snd_una = msk->snd_una;
1067 if (after64(new_wnd_end, msk->wnd_end))
1068 WRITE_ONCE(msk->wnd_end, new_wnd_end);
1071 if (after64(msk->wnd_end, snd_nxt))
1075 __mptcp_snd_una_update(msk, new_snd_una);
1078 msk->last_ack_recv = tcp_jiffies32;
1083 new_wnd_end, READ_ONCE(msk->wnd_end));
1086 bool mptcp_update_rcv_data_fin(struct mptcp_sock *msk, u64 data_fin_seq, bool use_64bit)
1093 if (READ_ONCE(msk->rcv_data_fin))
1096 WRITE_ONCE(msk->rcv_data_fin_seq,
1097 mptcp_expand_seq(READ_ONCE(msk->ack_seq), data_fin_seq, use_64bit));
1098 WRITE_ONCE(msk->rcv_data_fin, 1);
1103 static bool add_addr_hmac_valid(struct mptcp_sock *msk,
1111 hmac = add_addr_generate_hmac(READ_ONCE(msk->remote_key),
1112 READ_ONCE(msk->local_key),
1115 pr_debug("msk=%p, ahmac=%llu, mp_opt->ahmac=%llu\n",
1116 msk, hmac, mp_opt->ahmac);
1125 struct mptcp_sock *msk = mptcp_sk(subflow->conn);
1129 if (__mptcp_check_fallback(msk)) {
1139 /* on fallback we just need to ignore the msk-level snd_una, as
1142 __mptcp_snd_una_update(msk, READ_ONCE(msk->snd_nxt));
1154 if (!check_fully_established(msk, sk, subflow, skb, &mp_opt))
1159 READ_ONCE(msk->local_key) == mp_opt.rcvr_key) {
1160 WRITE_ONCE(msk->rcv_fastclose, true);
1161 mptcp_schedule_work((struct sock *)msk);
1166 add_addr_hmac_valid(msk, &mp_opt)) {
1171 mptcp_pm_add_addr_echoed(msk, &mp_opt.addr);
1172 mptcp_pm_del_add_timer(msk, &mp_opt.addr, true);
1181 mptcp_pm_rm_addr_received(msk, &mp_opt.rm_list);
1208 ack_update_msk(msk, sk, &mp_opt);
1217 mptcp_update_rcv_data_fin(msk, mp_opt.data_seq, mp_opt.dsn64))
1218 mptcp_schedule_work((struct sock *)msk);
1263 struct mptcp_sock *msk;
1268 msk = mptcp_sk(subflow->conn);
1270 ack_seq = READ_ONCE(msk->ack_seq);
1273 rcv_wnd_old = atomic64_read(&msk->rcv_wnd_sent);
1278 rcv_wnd = atomic64_cmpxchg(&msk->rcv_wnd_sent, rcv_wnd_old, rcv_wnd_new);
1317 WRITE_ONCE(msk->old_wspace, tp->rcv_wnd);