drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1281
struct inet_connection_sock *icsk = inet_csk(sk);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1551
reqsk_queue_removed(&inet_csk(lsk)->icsk_accept_queue, oreq);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
282
if (inet_csk(sk)->icsk_bind_hash)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
615
reqsk_queue_removed(&inet_csk(parent)->icsk_accept_queue, req);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
987
inet_csk(sk)->icsk_pmtu_cookie = pmtu;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.h
89
#define ACCEPT_QUEUE(sk) (&inet_csk(sk)->icsk_accept_queue.rskq_accept_head)
drivers/xen/pvcalls-back.c
794
icsk = inet_csk(mappass->sock->sk);
include/linux/tcp.h
592
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
include/net/espintcp.h
35
const struct inet_connection_sock *icsk = inet_csk(sk);
include/net/inet_connection_sock.h
156
return (void *)inet_csk(sk)->icsk_ca_priv;
include/net/inet_connection_sock.h
181
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED;
include/net/inet_connection_sock.h
186
return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED;
include/net/inet_connection_sock.h
191
memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack));
include/net/inet_connection_sock.h
207
struct inet_connection_sock *icsk = inet_csk(sk);
include/net/inet_connection_sock.h
232
struct inet_connection_sock *icsk = inet_csk(sk);
include/net/inet_connection_sock.h
283
reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue);
include/net/inet_connection_sock.h
288
return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue);
include/net/inet_connection_sock.h
308
return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ?
include/net/inet_connection_sock.h
324
inet_csk(sk)->icsk_ack.pingpong =
include/net/inet_connection_sock.h
330
inet_csk(sk)->icsk_ack.pingpong = 0;
include/net/inet_connection_sock.h
335
return inet_csk(sk)->icsk_ack.pingpong >=
include/net/inet_connection_sock.h
341
struct inet_connection_sock *icsk = inet_csk(sk);
include/net/inet_connection_sock.h
349
return inet_test_bit(IS_ICSK, sk) && !!inet_csk(sk)->icsk_ulp_ops;
include/net/inet_connection_sock.h
354
struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
1374
const struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
1381
const struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
1388
const struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
1395
const struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
1402
const struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
1484
(1 << inet_csk(sk)->icsk_ca_state);
include/net/tcp.h
1590
return max_t(unsigned long, inet_csk(sk)->icsk_rto, TCP_RTO_MIN);
include/net/tcp.h
1598
inet_csk(sk)->icsk_backoff);
include/net/tcp.h
1606
if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending)
include/net/tcp.h
1661
const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
include/net/tcp.h
1669
if (delta > inet_csk(sk)->icsk_rto)
include/net/tcp.h
1775
(tcp_receive_window(tp) <= inet_csk(sk)->icsk_ack.rcv_mss);
include/net/tcp.h
1834
const int rto = inet_csk(sk)->icsk_rto;
include/net/tcp.h
2125
ctx = rcu_dereference(inet_csk(sk)->icsk_accept_queue.fastopenq.ctx);
include/net/tcp.h
2615
inet_csk(sk)->icsk_ca_state,
include/net/tcp.h
2617
inet_csk(sk)->icsk_pmtu_cookie);
include/net/tcp.h
2624
u32 rto = inet_csk(sk)->icsk_rto;
include/net/tcp.h
401
struct inet_connection_sock *icsk = inet_csk(sk);
include/net/tcp.h
855
return READ_ONCE(inet_csk(sk)->icsk_rto_max);
include/net/tcp.h
860
inet_csk(sk)->icsk_rto = min(inet_csk(sk)->icsk_rto, tcp_rto_max(sk));
include/net/tcp.h
882
u32 rto_min = READ_ONCE(inet_csk(sk)->icsk_rto_min);
include/net/tcp_ecn.h
326
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
include/net/tcp_ecn.h
61
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
include/net/tls.h
371
const struct inet_connection_sock *icsk = inet_csk(sk);
net/core/devmem.c
384
if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk)) {
net/core/filter.c
5357
int rto_min_us = jiffies_to_usecs(inet_csk(sk)->icsk_rto_min);
net/core/filter.c
5363
int delack_max_us = jiffies_to_usecs(inet_csk(sk)->icsk_delack_max);
net/core/filter.c
5405
inet_csk(sk)->icsk_delack_max = timeout;
net/core/filter.c
5412
inet_csk(sk)->icsk_rto_min = timeout;
net/core/filter.c
5436
if (!inet_csk(sk)->icsk_ca_ops)
net/core/sock.c
2635
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/af_inet.c
221
!inet_csk(sk)->icsk_accept_queue.fastopenq.max_qlen) {
net/ipv4/cipso_ipv4.c
1904
sk_conn = inet_csk(sk);
net/ipv4/cipso_ipv4.c
2062
struct inet_connection_sock *sk_conn = inet_csk(sk);
net/ipv4/inet_connection_sock.c
1008
reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req);
net/ipv4/inet_connection_sock.c
1055
reqsk_queue_migrated(&inet_csk(nsk)->icsk_accept_queue, req);
net/ipv4/inet_connection_sock.c
1061
icsk = inet_csk(sk_listener);
net/ipv4/inet_connection_sock.c
1120
reqsk_queue_removed(&inet_csk(oreq->rsk_listener)->icsk_accept_queue, oreq);
net/ipv4/inet_connection_sock.c
1176
struct inet_connection_sock *icsk = inet_csk(newsk);
net/ipv4/inet_connection_sock.c
1204
newicsk = inet_csk(newsk);
net/ipv4/inet_connection_sock.c
1268
WARN_ON(inet_sk(sk)->inet_num && !inet_csk(sk)->icsk_bind_hash);
net/ipv4/inet_connection_sock.c
1305
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
1324
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
1386
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
net/ipv4/inet_connection_sock.c
1412
reqsk_queue_removed(&inet_csk(req->rsk_listener)->icsk_accept_queue, req);
net/ipv4/inet_connection_sock.c
1458
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
578
if (!inet_csk(sk)->icsk_bind_hash)
net/ipv4/inet_connection_sock.c
580
WARN_ON(inet_csk(sk)->icsk_bind_hash != tb);
net/ipv4/inet_connection_sock.c
581
WARN_ON(inet_csk(sk)->icsk_bind2_hash != tb2);
net/ipv4/inet_connection_sock.c
604
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
652
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
725
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
735
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_connection_sock.c
747
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/inet_hashtables.c
204
inet_csk(sk)->icsk_bind_hash = tb;
net/ipv4/inet_hashtables.c
205
inet_csk(sk)->icsk_bind2_hash = tb2;
net/ipv4/inet_hashtables.c
225
tb = inet_csk(sk)->icsk_bind_hash;
net/ipv4/inet_hashtables.c
226
inet_csk(sk)->icsk_bind_hash = NULL;
net/ipv4/inet_hashtables.c
231
if (inet_csk(sk)->icsk_bind2_hash) {
net/ipv4/inet_hashtables.c
232
struct inet_bind2_bucket *tb2 = inet_csk(sk)->icsk_bind2_hash;
net/ipv4/inet_hashtables.c
235
inet_csk(sk)->icsk_bind2_hash = NULL;
net/ipv4/inet_hashtables.c
270
tb = inet_csk(sk)->icsk_bind_hash;
net/ipv4/inet_hashtables.c
271
tb2 = inet_csk(sk)->icsk_bind2_hash;
net/ipv4/inet_hashtables.c
761
struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash;
net/ipv4/inet_hashtables.c
771
inet_csk(sk2)->icsk_bind_hash == tb &&
net/ipv4/inet_hashtables.c
937
if (!inet_csk(sk)->icsk_bind2_hash) {
net/ipv4/inet_hashtables.c
977
inet_bind2_bucket_destroy(hinfo->bind2_bucket_cachep, inet_csk(sk)->icsk_bind2_hash);
net/ipv4/inet_hashtables.c
991
inet_bind2_bucket_init(tb2, net, head2, inet_csk(sk)->icsk_bind_hash, sk);
net/ipv4/inet_hashtables.c
997
inet_csk(sk)->icsk_bind2_hash = tb2;
net/ipv4/inet_timewait_sock.c
109
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/ip_sockglue.c
1093
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/syncookies.c
201
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
1572
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
3052
if (inet_csk(sk)->icsk_bind_hash &&
net/ipv4/tcp.c
3397
inet_csk(sk)->icsk_backoff = 0;
net/ipv4/tcp.c
3403
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
3735
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_PUSHED;
net/ipv4/tcp.c
3755
WRITE_ONCE(inet_csk(sk)->icsk_syn_retries, val);
net/ipv4/tcp.c
3768
WRITE_ONCE(inet_csk(sk)->icsk_user_timeout, val);
net/ipv4/tcp.c
3880
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
3981
WRITE_ONCE(inet_csk(sk)->icsk_rto_max, msecs_to_jiffies(val));
net/ipv4/tcp.c
3988
WRITE_ONCE(inet_csk(sk)->icsk_rto_min, rto_min);
net/ipv4/tcp.c
3996
WRITE_ONCE(inet_csk(sk)->icsk_delack_max, delack_max);
net/ipv4/tcp.c
4214
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
4247
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
431
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
4481
READ_ONCE(inet_csk(sk)->icsk_retransmits));
net/ipv4/tcp.c
4488
nla_put_u8(stats, TCP_NLA_CA_STATE, inet_csk(sk)->icsk_ca_state);
net/ipv4/tcp.c
4513
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp.c
4859
val = jiffies_to_usecs(READ_ONCE(inet_csk(sk)->icsk_rto_min));
net/ipv4/tcp.c
4862
val = jiffies_to_usecs(READ_ONCE(inet_csk(sk)->icsk_delack_max));
net/ipv4/tcp.c
4890
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_bbr.c
485
u8 prev_state = bbr->prev_ca_state, state = inet_csk(sk)->icsk_ca_state;
net/ipv4/tcp_bic.c
192
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_cong.c
219
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_cong.c
238
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_cong.c
253
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_cong.c
272
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_cong.c
40
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_cong.c
415
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_dctcp.c
114
inet_csk(sk)->icsk_ca_ops = &dctcp_reno;
net/ipv4/tcp_dctcp.c
186
new_state != inet_csk(sk)->icsk_ca_state)
net/ipv4/tcp_dctcp.c
227
if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) {
net/ipv4/tcp_dctcp.h
29
if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) {
net/ipv4/tcp_dctcp.h
33
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
net/ipv4/tcp_diag.c
118
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_diag.c
148
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_diag.c
284
return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, r, nlmsg_flags,
net/ipv4/tcp_diag.c
369
if (inet_sk_diag_fill(sk, inet_csk(sk), skb,
net/ipv4/tcp_fastopen.c
130
inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1);
net/ipv4/tcp_fastopen.c
169
q = &inet_csk(sk)->icsk_accept_queue.fastopenq;
net/ipv4/tcp_fastopen.c
331
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
net/ipv4/tcp_fastopen.c
335
child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL,
net/ipv4/tcp_fastopen.c
401
fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq;
net/ipv4/tcp_fastopen.c
57
fastopenq = &inet_csk(lsk)->icsk_accept_queue.fastopenq;
net/ipv4/tcp_fastopen.c
676
u32 timeouts = inet_csk(sk)->icsk_retransmits;
net/ipv4/tcp_htcp.c
104
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_htcp.c
84
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
1019
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
1189
inet_csk(sk)->icsk_rto = __tcp_set_rto(tp);
net/ipv4/tcp_input.c
1291
tp->rx_opt.sack_ok, inet_csk(sk)->icsk_ca_state,
net/ipv4/tcp_input.c
1784
inet_csk(sk)->icsk_ca_state,
net/ipv4/tcp_input.c
229
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
2400
if (inet_csk(sk)->icsk_ca_state != TCP_CA_Loss || tp->undo_marker)
net/ipv4/tcp_input.c
2555
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
2597
!inet_csk(sk)->icsk_mtup.probe_size;
net/ipv4/tcp_input.c
2857
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
2902
DBGUNDO(sk, inet_csk(sk)->icsk_ca_state == TCP_CA_Loss ? "loss" : "retrans");
net/ipv4/tcp_input.c
2904
if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss)
net/ipv4/tcp_input.c
2949
WRITE_ONCE(inet_csk(sk)->icsk_retransmits, 0);
net/ipv4/tcp_input.c
2980
tp->snd_ssthresh = inet_csk(sk)->icsk_ca_ops->ssthresh(sk);
net/ipv4/tcp_input.c
3016
if (inet_csk(sk)->icsk_ca_ops->cong_control)
net/ipv4/tcp_input.c
3021
(inet_csk(sk)->icsk_ca_state == TCP_CA_CWR || tp->undo_marker)) {
net/ipv4/tcp_input.c
3034
if (inet_csk(sk)->icsk_ca_state < TCP_CA_CWR) {
net/ipv4/tcp_input.c
3050
if (inet_csk(sk)->icsk_ca_state != state) {
net/ipv4/tcp_input.c
3068
if (inet_csk(sk)->icsk_ca_state != TCP_CA_CWR) {
net/ipv4/tcp_input.c
3075
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3085
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3117
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
313
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
325
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3331
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
338
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3496
inet_csk(sk)->icsk_backoff = 0;
net/ipv4/tcp_input.c
3515
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3526
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3538
u32 rto = inet_csk(sk)->icsk_rto;
net/ipv4/tcp_input.c
3606
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3786
icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3809
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3834
inet_csk(sk)->icsk_ca_state != TCP_CA_Open;
net/ipv4/tcp_input.c
3861
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
3979
tcp_sync_mss(sk, inet_csk(sk)->icsk_pmtu_cookie);
net/ipv4/tcp_input.c
4144
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
4248
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
4803
return inet_csk(sk)->icsk_rto * (USEC_PER_SEC / HZ);
net/ipv4/tcp_input.c
4808
return inet_csk(sk)->icsk_rto * 1200 / HZ;
net/ipv4/tcp_input.c
5072
if (inet_csk(sk)->icsk_ca_state != TCP_CA_Loss &&
net/ipv4/tcp_input.c
538
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
net/ipv4/tcp_input.c
555
((1 << inet_csk(sk)->icsk_ca_state) &
net/ipv4/tcp_input.c
5677
inet_csk(sk)->icsk_ack.pending |=
net/ipv4/tcp_input.c
5703
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
net/ipv4/tcp_input.c
608
const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
net/ipv4/tcp_input.c
6142
if (((tp->rcv_nxt - tp->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss &&
net/ipv4/tcp_input.c
6153
inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOW) {
net/ipv4/tcp_input.c
6531
inet_csk(sk)->icsk_af_ops->sk_rx_dst_set(sk, skb);
net/ipv4/tcp_input.c
6727
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
673
return 2 * inet_csk(sk)->icsk_ack.rcv_mss;
net/ipv4/tcp_input.c
6756
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
6875
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
7129
if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss && !tp->packets_out)
net/ipv4/tcp_input.c
7133
WRITE_ONCE(inet_csk(sk)->icsk_retransmits, 0);
net/ipv4/tcp_input.c
7173
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
727
inet_csk(sk)->icsk_ack.quick |= 1;
net/ipv4/tcp_input.c
7302
if (!inet_csk(sk)->icsk_ca_ops->cong_control)
net/ipv4/tcp_input.c
7547
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
net/ipv4/tcp_input.c
781
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_input.c
815
inet_csk(sk)->icsk_ack.rcv_mss = hint;
net/ipv4/tcp_input.c
904
TCP_SKB_CB(skb)->seq >= inet_csk(sk)->icsk_ack.rcv_mss) {
net/ipv4/tcp_input.c
999
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_ipv4.c
1740
inet_csk(newsk)->icsk_ext_hdr_len = 0;
net/ipv4/tcp_ipv4.c
1742
inet_csk(newsk)->icsk_ext_hdr_len = inet_opt->opt.optlen;
net/ipv4/tcp_ipv4.c
2470
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_ipv4.c
2521
if (inet_csk(sk)->icsk_bind_hash)
net/ipv4/tcp_ipv4.c
2851
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_ipv4.c
295
inet_csk(sk)->icsk_ext_hdr_len = psp_sk_overhead(sk);
net/ipv4/tcp_ipv4.c
297
inet_csk(sk)->icsk_ext_hdr_len += inet_opt->opt.optlen;
net/ipv4/tcp_ipv4.c
394
inet_csk(sk)->icsk_pmtu_cookie > dmtu) {
net/ipv4/tcp_ipv4.c
444
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_metrics.c
342
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_metrics.c
527
inet_csk(sk)->icsk_rto = crtt + max(2 * crtt, tcp_rto_min(sk));
net/ipv4/tcp_metrics.c
538
inet_csk(sk)->icsk_rto = TCP_TIMEOUT_FALLBACK;
net/ipv4/tcp_minisocks.c
329
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_minisocks.c
501
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_minisocks.c
564
newicsk = inet_csk(newsk);
net/ipv4/tcp_minisocks.c
914
if (req->num_timeout < READ_ONCE(inet_csk(sk)->icsk_accept_queue.rskq_defer_accept) &&
net/ipv4/tcp_minisocks.c
927
child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL,
net/ipv4/tcp_minisocks.c
937
reqsk_queue_removed(&inet_csk(req->rsk_listener)->icsk_accept_queue, req);
net/ipv4/tcp_nv.c
241
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
1045
inet_csk(sk)->icsk_retransmits < 2 &&
net/ipv4/tcp_output.c
1355
inet_csk(sk)->icsk_af_ops->mtu_reduced(sk);
net/ipv4/tcp_output.c
1515
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
177
while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd)
net/ipv4/tcp_output.c
188
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
1976
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2010
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2023
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2060
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2095
if (mtu != inet_csk(sk)->icsk_pmtu_cookie)
net/ipv4/tcp_output.c
2121
if (inet_csk(sk)->icsk_ca_state == TCP_CA_Open &&
net/ipv4/tcp_output.c
2137
const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
net/ipv4/tcp_output.c
2166
(s32)(tcp_jiffies32 - tp->snd_cwnd_stamp) >= inet_csk(sk)->icsk_rto &&
net/ipv4/tcp_output.c
2257
const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
net/ipv4/tcp_output.c
2436
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2546
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2664
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
2681
inet_csk(sk)->icsk_ca_state != TCP_CA_Open ||
net/ipv4/tcp_output.c
292
if (unlikely(inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOMEM)) {
net/ipv4/tcp_output.c
3103
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
3141
jiffies_to_usecs(inet_csk(sk)->icsk_rto) :
net/ipv4/tcp_output.c
3195
smp_store_release(&inet_csk(sk)->icsk_pending, 0);
net/ipv4/tcp_output.c
3228
smp_store_release(&inet_csk(sk)->icsk_pending, 0);
net/ipv4/tcp_output.c
3318
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
3554
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
3587
if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk)) {
net/ipv4/tcp_output.c
3729
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
3797
inet_csk(sk)->icsk_rto, true);
net/ipv4/tcp_output.c
4086
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
4174
inet_csk(sk)->icsk_rto = tcp_timeout_init(sk);
net/ipv4/tcp_output.c
4175
WRITE_ONCE(inet_csk(sk)->icsk_retransmits, 0);
net/ipv4/tcp_output.c
4201
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
4351
if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk))
net/ipv4/tcp_output.c
4395
inet_csk(sk)->icsk_rto, false);
net/ipv4/tcp_output.c
4404
return min(READ_ONCE(inet_csk(sk)->icsk_delack_max), delack_from_rto_min);
net/ipv4/tcp_output.c
4413
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
4479
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
4604
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_output.c
91
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_recovery.c
109
timeout, inet_csk(sk)->icsk_rto);
net/ipv4/tcp_recovery.c
126
if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) {
net/ipv4/tcp_recovery.c
128
if (!inet_csk(sk)->icsk_ca_ops->cong_control)
net/ipv4/tcp_recovery.c
13
if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery)
net/ipv4/tcp_recovery.c
133
if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS)
net/ipv4/tcp_recovery.c
144
const u8 state = inet_csk(sk)->icsk_ca_state;
net/ipv4/tcp_timer.c
221
if (!inet_csk(sk)->icsk_retransmits)
net/ipv4/tcp_timer.c
244
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
30
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
310
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
390
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
441
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
458
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
497
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
52
const struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
538
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
696
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_timer.c
734
if (!smp_load_acquire(&inet_csk(sk)->icsk_pending))
net/ipv4/tcp_timer.c
759
sk_reset_timer(sk, &inet_csk(sk)->icsk_keepalive_timer, jiffies + len);
net/ipv4/tcp_timer.c
764
sk_stop_timer(sk, &inet_csk(sk)->icsk_keepalive_timer);
net/ipv4/tcp_ulp.c
106
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_ulp.c
114
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv4/tcp_ulp.c
132
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv6/ipv6_sockglue.c
110
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv6/ipv6_sockglue.c
597
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv6/tcp_ipv6.c
1326
inet_csk(newsk)->icsk_af_ops = &ipv6_mapped;
net/ipv6/tcp_ipv6.c
137
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv6/tcp_ipv6.c
1448
inet_csk(newsk)->icsk_ext_hdr_len = 0;
net/ipv6/tcp_ipv6.c
1450
inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen +
net/ipv6/tcp_ipv6.c
2080
struct inet_connection_sock *icsk = inet_csk(sk);
net/ipv6/tcp_ipv6.c
2136
const struct inet_connection_sock *icsk = inet_csk(sp);
net/ipv6/tcp_ipv6.c
372
if (inet_csk(sk)->icsk_pmtu_cookie > dmtu) {
net/mptcp/ctrl.c
537
timeouts = inet_csk(ssk)->icsk_retransmits;
net/mptcp/diag.c
32
sf = rcu_dereference(inet_csk(sk)->icsk_ulp_data);
net/mptcp/mptcp_diag.c
104
if (!ctx || strcmp(inet_csk(sk)->icsk_ulp_ops->name, "mptcp"))
net/mptcp/mptcp_diag.c
23
return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, req, NLM_F_MULTI,
net/mptcp/mptcp_diag.c
52
err = inet_sk_diag_fill(sk, inet_csk(sk), rep, cb, req, 0,
net/mptcp/options.c
442
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/pm.c
312
struct inet_connection_sock *icsk = inet_csk(ssk);
net/mptcp/protocol.c
2577
if (!inet_csk(ssk)->icsk_ulp_ops) {
net/mptcp/protocol.c
2686
if (!inet_csk(sk)->icsk_mtup.probe_timestamp ||
net/mptcp/protocol.c
2691
inet_csk(sk)->icsk_mtup.probe_timestamp + mptcp_close_timeout(sk));
net/mptcp/protocol.c
2760
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/protocol.c
2856
if (!fail_tout && !inet_csk(sk)->icsk_mtup.probe_timestamp)
net/mptcp/protocol.c
2859
close_timeout = (unsigned long)inet_csk(sk)->icsk_mtup.probe_timestamp -
net/mptcp/protocol.c
2865
timeout = inet_csk(sk)->icsk_mtup.probe_timestamp ? close_timeout : fail_tout;
net/mptcp/protocol.c
2867
sk_reset_timer(sk, &inet_csk(sk)->mptcp_tout_timer, timeout);
net/mptcp/protocol.c
2931
inet_csk(ssk)->icsk_ack.rcv_mss = TCP_MIN_MSS;
net/mptcp/protocol.c
3005
inet_csk(sk)->icsk_sync_mss = mptcp_sync_mss;
net/mptcp/protocol.c
3025
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/protocol.c
3223
sk_stop_timer(sk, &inet_csk(sk)->mptcp_tout_timer);
net/mptcp/protocol.c
3685
struct inet_connection_sock *icsk = inet_csk(ssk);
net/mptcp/protocol.c
497
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/protocol.c
515
return inet_csk(ssk)->icsk_pending && !subflow->stale_count ?
net/mptcp/protocol.c
570
const struct inet_connection_sock *icsk = inet_csk(ssk);
net/mptcp/protocol.h
598
const struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/protocol.h
872
inet_csk(sk)->icsk_af_ops = ctx->icsk_af_ops;
net/mptcp/protocol.h
894
if (!inet_csk(sk)->icsk_mtup.probe_timestamp)
net/mptcp/protocol.h
897
sk_stop_timer(sk, &inet_csk(sk)->mptcp_tout_timer);
net/mptcp/protocol.h
898
inet_csk(sk)->icsk_mtup.probe_timestamp = 0;
net/mptcp/protocol.h
904
inet_csk(sk)->icsk_mtup.probe_timestamp = tout ? : 1;
net/mptcp/sockopt.c
1576
if (inet_csk(sk)->icsk_ca_ops != inet_csk(ssk)->icsk_ca_ops)
net/mptcp/sockopt.c
999
info->mptcpi_retransmits = inet_csk(sk)->icsk_retransmits;
net/mptcp/subflow.c
1520
if (reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue))
net/mptcp/subflow.c
1537
(tcp_sk(sk)->rcv_nxt - tcp_sk(sk)->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss)
net/mptcp/subflow.c
1538
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
net/mptcp/subflow.c
1566
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/subflow.c
1844
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/subflow.c
1895
struct request_sock_queue *queue = &inet_csk(listener_ssk)->icsk_accept_queue;
net/mptcp/subflow.c
1966
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/subflow.c
770
struct inet_connection_sock *icsk = inet_csk(sk);
net/mptcp/subflow.c
788
if (inet_csk(ssk)->icsk_ulp_ops) {
net/psp/psp_sock.c
238
icsk = inet_csk(sk);
net/smc/af_smc.c
160
if (inet_csk(child)->icsk_af_ops == inet_csk(sk)->icsk_af_ops)
net/smc/af_smc.c
161
inet_csk(child)->icsk_af_ops = smc->ori_af_ops;
net/smc/af_smc.c
2685
smc->ori_af_ops = inet_csk(smc->clcsock->sk)->icsk_af_ops;
net/smc/af_smc.c
2690
inet_csk(smc->clcsock->sk)->icsk_af_ops = &smc->af_ops;
net/tls/tls_main.c
1132
ctx = rcu_dereference(inet_csk(sk)->icsk_ulp_data);
net/tls/tls_main.c
367
struct inet_connection_sock *icsk = inet_csk(sk);
net/tls/tls_main.c
915
struct inet_connection_sock *icsk = inet_csk(sk);
net/tls/tls_toe.c
48
struct inet_connection_sock *icsk = inet_csk(sk);
net/xfrm/espintcp.c
456
struct inet_connection_sock *icsk = inet_csk(sk);
tools/testing/selftests/bpf/progs/bpf_cc_cubic.c
123
(1 << inet_csk(sk)->icsk_ca_state)) {
tools/testing/selftests/bpf/progs/bpf_cc_cubic.c
130
inet_csk(sk)->icsk_ca_state == TCP_CA_CWR) {
tools/testing/selftests/bpf/progs/bpf_dctcp.c
161
new_state != BPF_CORE_READ_BITFIELD(inet_csk(sk), icsk_ca_state))
tools/testing/selftests/bpf/progs/bpf_dctcp.c
193
if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) {
tools/testing/selftests/bpf/progs/bpf_dctcp.c
197
inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
tools/testing/selftests/bpf/progs/bpf_tracing_net.h
170
return (void *)inet_csk(sk)->icsk_ca_priv;