mptcp_sk
struct mptcp_sock *msk = mptcp_sk(sk);
return mptcp_sk(mptcp_subflow_ctx(sk)->conn);
mptcp_early_fallback(mptcp_sk(subflow->conn), subflow,
mptcp_sk(sk)->bytes_received += skb->len;
struct mptcp_sock *msk = mptcp_sk(sk);
READ_ONCE(mptcp_sk(sk)->backlog_len);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
msk = mptcp_sk(subflow->conn);
msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
u64 data_fin_tx_seq = READ_ONCE(mptcp_sk(subflow->conn)->write_seq) - 1;
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
ssk = __mptcp_nmpc_sk(mptcp_sk(newsk));
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_sk(sk)->cb_flags |= BIT(MPTCP_PUSH_PENDING);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
const struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
const struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_event(MPTCP_EVENT_SUB_CLOSED, mptcp_sk(sk), ssk, GFP_KERNEL);
mptcp_pm_subflow_check_next(mptcp_sk(sk), subflow);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
strscpy(mptcp_sk(sk)->ca_name, icsk->icsk_ca_ops->name,
sizeof(mptcp_sk(sk)->ca_name));
ret = mptcp_init_sched(mptcp_sk(sk),
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
if (__mptcp_check_fallback(mptcp_sk(sk))) {
WRITE_ONCE(mptcp_sk(sk)->snd_una, mptcp_sk(sk)->snd_nxt);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
ssk = mptcp_sk(sk)->first;
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp6_sock *msk6 = container_of(mptcp_sk(sk), struct mptcp6_sock, msk);
msk = mptcp_sk(nsk);
msk->setsockopt_seq = mptcp_sk(sk)->setsockopt_seq;
mptcp_init_sched(msk, mptcp_sk(sk)->sched);
struct mptcp_sock *msk = mptcp_sk(sk);
__set_bit(MPTCP_CLEAN_UNA, &mptcp_sk(sk)->cb_flags);
__set_bit(MPTCP_PUSH_PENDING, &mptcp_sk(sk)->cb_flags);
struct mptcp_sock *msk = mptcp_sk(sk);
__set_bit(MPTCP_PUSH_PENDING, &mptcp_sk(sk)->cb_flags);
__set_bit(MPTCP_SYNC_SNDBUF, &mptcp_sk(sk)->cb_flags);
struct mptcp_sock *msk = mptcp_sk(sk);
msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sock->sk);
struct mptcp_sock *msk = mptcp_sk(sock->sk);
mptcp_sk(sk)->timer_ival = 0;
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sock->sk);
msk = mptcp_sk(newsk);
msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_rps_record_subflows(mptcp_sk(sk));
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_sk(sk)->timer_ival = TCP_RTO_MIN << retransmits;
mptcp_sk(sk)->timer_ival = tout > 0 ? tout : TCP_RTO_MIN;
mptcp_for_each_subflow(mptcp_sk(sk), subflow)
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
if (sk->sk_state != TCP_SYN_SENT && !__mptcp_check_fallback(mptcp_sk(sk)))
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
tout = mptcp_sk(sk)->timer_ival;
if (schedule_work(&mptcp_sk(sk)->work))
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
msk = mptcp_sk(sk);
return __tcp_win_from_space(mptcp_sk(sk)->scaling_ratio, space);
return __tcp_space_from_win(mptcp_sk(sk)->scaling_ratio, win);
READ_ONCE(mptcp_sk(sk)->backlog_len) -
const struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
const struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
u64 data_avail = mptcp_data_avail(mptcp_sk(sk));
mptcp_reset_tout_timer(mptcp_sk(sk), 0);
READ_ONCE(mptcp_sk(sk)->fully_established);
val = READ_ONCE(mptcp_sk(sk)->notsent_lowat);
const struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_for_each_subflow(mptcp_sk(sk), subflow) {
struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_for_each_subflow(mptcp_sk(sk), subflow) {
struct mptcp_sock *msk = mptcp_sk(sk);
WRITE_ONCE(mptcp_sk(subflow->conn)->use_64bit_ack, !!mpext->dsn64);
msk = mptcp_sk(subflow->conn);
__set_bit(MPTCP_ERROR_REPORT, &mptcp_sk(sk)->cb_flags);
msk = mptcp_sk(parent);
struct mptcp_sock *msk = mptcp_sk(sk);
mptcp_sockopt_sync_locked(mptcp_sk(sk), sf->sk);
subflow_sched_work_if_closed(mptcp_sk(parent), sk);
&mptcp_sk(sk)->flags))
if (!test_and_set_bit(MPTCP_WORK_CLOSE_SUBFLOW, &mptcp_sk(sk)->flags))
struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_sock *msk = mptcp_sk(sk);
msk = mptcp_sk(parent);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
owner = mptcp_sk(ctx->conn);
struct mptcp_sock *msk = mptcp_sk(subflow->conn);
msk = mptcp_sk(sk);
msk = mptcp_sk(sk);
ret = mptcp_sk(sk);
if (mptcp_sk(sk)->token == token)
return mptcp_sk(sk);