Symbol: udp_sk
drivers/net/gtp.c
861
WRITE_ONCE(udp_sk(sk)->encap_type, 0);
drivers/net/gtp.c
913
switch (READ_ONCE(udp_sk(sk)->encap_type)) {
drivers/net/ovpn/socket.c
154
u8 type = READ_ONCE(udp_sk(sk)->encap_type);
drivers/net/ovpn/udp.c
38
if (unlikely(READ_ONCE(udp_sk(sk)->encap_type) != UDP_ENCAP_OVPNINUDP))
drivers/net/ovpn/udp.c
416
if ((READ_ONCE(udp_sk(ovpn_sock->sk)->encap_type) == UDP_ENCAP_OVPNINUDP) &&
drivers/net/ovpn/udp.c
444
WRITE_ONCE(udp_sk(sk)->encap_type, 0);
drivers/net/ovpn/udp.c
445
WRITE_ONCE(udp_sk(sk)->encap_rcv, NULL);
drivers/net/ovpn/udp.c
446
WRITE_ONCE(udp_sk(sk)->encap_destroy, NULL);
include/linux/udp.h
123
test_bit(UDP_FLAGS_##nr, &udp_sk(sk)->udp_flags)
include/linux/udp.h
125
set_bit(UDP_FLAGS_##nr, &udp_sk(sk)->udp_flags)
include/linux/udp.h
127
test_and_set_bit(UDP_FLAGS_##nr, &udp_sk(sk)->udp_flags)
include/linux/udp.h
129
clear_bit(UDP_FLAGS_##nr, &udp_sk(sk)->udp_flags)
include/linux/udp.h
131
assign_bit(UDP_FLAGS_##nr, &udp_sk(sk)->udp_flags, val)
include/linux/udp.h
140
WRITE_ONCE(udp_sk(sk)->peeking_with_offset, val >= 0);
include/linux/udp.h
211
READ_ONCE(udp_sk(sk)->encap_rcv) &&
include/net/udp.h
182
return !hlist_nulls_unhashed(&udp_sk(sk)->udp_lrpa_node);
include/net/udp.h
289
struct udp_sock *up = udp_sk(sk);
include/net/udp.h
307
numa_drop_add(&udp_sk(sk)->drop_counters, 1);
include/net/udp.h
384
return sk_rmem_alloc_get(sk) - READ_ONCE(udp_sk(sk)->forward_deficit);
include/net/udplite.h
74
u16 pcslen = READ_ONCE(udp_sk(sk)->pcslen);
net/core/sock.c
4413
!skb_queue_empty_lockless(&udp_sk(sk)->reader_queue))
net/ipv4/udp.c
1008
if (udp_sk(sk)->encap_err_rcv)
net/ipv4/udp.c
1009
udp_sk(sk)->encap_err_rcv(sk, skb, err, uh->dest, info,
net/ipv4/udp.c
1035
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
1214
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
1275
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
152
(bitmap || udp_sk(sk2)->udp_port_hash == num) &&
net/ipv4/udp.c
1555
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
1633
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
165
__set_bit(udp_sk(sk2)->udp_port_hash >> log,
net/ipv4/udp.c
1717
udp_prod_queue = &udp_sk(sk)->udp_prod_queue[numa_node_id()];
net/ipv4/udp.c
1810
numa_drop_add(&udp_sk(sk)->drop_counters, nb);
net/ipv4/udp.c
1838
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
1869
if (unlikely(READ_ONCE(udp_sk(sk)->peeking_with_offset)))
net/ipv4/udp.c
189
(udp_sk(sk2)->udp_port_hash == num) &&
net/ipv4/udp.c
1923
struct sk_buff_head *rcvq = &udp_sk(sk)->reader_queue;
net/ipv4/udp.c
1981
queue = &udp_sk(sk)->reader_queue;
net/ipv4/udp.c
2169
if (!__sk_queue_drop_skb(sk, &udp_sk(sk)->reader_queue, skb, flags,
net/ipv4/udp.c
219
(udp_sk(sk2)->udp_port_hash == udp_sk(sk)->udp_port_hash) &&
net/ipv4/udp.c
2254
udp_sk(sk)->udp_port_hash);
net/ipv4/udp.c
2255
hslot2 = udp_hashslot2(udptable, udp_sk(sk)->udp_portaddr_hash);
net/ipv4/udp.c
2266
hlist_del_init_rcu(&udp_sk(sk)->udp_portaddr_node);
net/ipv4/udp.c
2287
udp_sk(sk)->udp_port_hash);
net/ipv4/udp.c
2288
hslot2 = udp_hashslot2(udptable, udp_sk(sk)->udp_portaddr_hash);
net/ipv4/udp.c
2300
hlist_del_init_rcu(&udp_sk(sk)->udp_portaddr_node);
net/ipv4/udp.c
2305
hlist_add_head_rcu(&udp_sk(sk)->udp_portaddr_node,
net/ipv4/udp.c
2341
udp_sk(sk)->udp_portaddr_hash = newhash;
net/ipv4/udp.c
2405
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
292
unsigned int slot2 = udp_sk(sk)->udp_portaddr_hash ^ snum;
net/ipv4/udp.c
2941
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
2978
if (udp_sk(sk)->gro_receive != new_gro_receive) {
net/ipv4/udp.c
2983
if (udp_sk(sk)->gro_receive)
net/ipv4/udp.c
2986
WRITE_ONCE(udp_sk(sk)->gro_receive, new_gro_receive);
net/ipv4/udp.c
3000
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
3142
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp.c
318
udp_sk(sk)->udp_port_hash = snum;
net/ipv4/udp.c
319
udp_sk(sk)->udp_portaddr_hash ^= snum;
net/ipv4/udp.c
3226
if (!skb_queue_empty_lockless(&udp_sk(sk)->reader_queue))
net/ipv4/udp.c
324
udp_sk(sk)->udp_port_hash = 0;
net/ipv4/udp.c
325
udp_sk(sk)->udp_portaddr_hash ^= snum;
net/ipv4/udp.c
335
hslot2 = udp_hashslot2(udptable, udp_sk(sk)->udp_portaddr_hash);
net/ipv4/udp.c
339
hlist_add_tail_rcu(&udp_sk(sk)->udp_portaddr_node,
net/ipv4/udp.c
342
hlist_add_head_rcu(&udp_sk(sk)->udp_portaddr_node,
net/ipv4/udp.c
3466
__bpf_md_ptr(struct udp_sock *, udp_sk);
net/ipv4/udp.c
364
udp_sk(sk)->udp_portaddr_hash = hash2_partial;
net/ipv4/udp.c
3653
struct udp_sock *udp_sk, uid_t uid, int bucket)
net/ipv4/udp.c
3659
ctx.udp_sk = udp_sk;
net/ipv4/udp.c
378
udp_sk(sk)->udp_port_hash != hnum ||
net/ipv4/udp.c
3973
struct udp_sock *udp_sk, uid_t uid, int bucket)
net/ipv4/udp.c
4035
{ offsetof(struct bpf_iter__udp, udp_sk),
net/ipv4/udp.c
584
hslot4 = udp_hashslot4(udptable, udp_sk(sk)->udp_lrpa_hash);
net/ipv4/udp.c
586
udp_sk(sk)->udp_lrpa_hash = newhash4;
net/ipv4/udp.c
590
hlist_nulls_del_init_rcu(&udp_sk(sk)->udp_lrpa_node);
net/ipv4/udp.c
595
hlist_nulls_add_head_rcu(&udp_sk(sk)->udp_lrpa_node,
net/ipv4/udp.c
607
hslot2 = udp_hashslot2(udptable, udp_sk(sk)->udp_portaddr_hash);
net/ipv4/udp.c
608
hslot4 = udp_hashslot4(udptable, udp_sk(sk)->udp_lrpa_hash);
net/ipv4/udp.c
611
hlist_nulls_del_init_rcu(&udp_sk(sk)->udp_lrpa_node);
net/ipv4/udp.c
634
hslot = udp_hashslot(udptable, net, udp_sk(sk)->udp_port_hash);
net/ipv4/udp.c
635
hslot2 = udp_hashslot2(udptable, udp_sk(sk)->udp_portaddr_hash);
net/ipv4/udp.c
637
udp_sk(sk)->udp_lrpa_hash = hash;
net/ipv4/udp.c
644
hlist_nulls_add_head_rcu(&udp_sk(sk)->udp_lrpa_node,
net/ipv4/udp.c
798
udp_sk(sk)->udp_port_hash != hnum ||
net/ipv4/udp.c
890
up = udp_sk(sk);
net/ipv4/udp.c
903
up = udp_sk(sk);
net/ipv4/udp.c
948
if (!sk || READ_ONCE(udp_sk(sk)->encap_type)) {
net/ipv4/udp_bpf.c
26
return !skb_queue_empty(&udp_sk(sk)->reader_queue) ||
net/ipv4/udp_offload.c
158
return call_gro_receive_sk(udp_sk(sk)->gro_receive, sk, head, skb);
net/ipv4/udp_offload.c
167
return call_gro_receive_sk(udp_sk(sk)->gro_receive, sk, head, skb);
net/ipv4/udp_offload.c
55
struct udp_sock *tup, *up = udp_sk(sk);
net/ipv4/udp_offload.c
802
if (!sk || !udp_sk(sk)->gro_receive) {
net/ipv4/udp_offload.c
82
struct udp_sock *up = udp_sk(sk);
net/ipv4/udp_offload.c
937
if (sk && udp_sk(sk)->gro_complete) {
net/ipv4/udp_offload.c
950
err = udp_sk(sk)->gro_complete(sk, skb,
net/ipv4/udp_tunnel_core.c
84
udp_sk(sk)->encap_type = cfg->encap_type;
net/ipv4/udp_tunnel_core.c
85
udp_sk(sk)->encap_rcv = cfg->encap_rcv;
net/ipv4/udp_tunnel_core.c
86
udp_sk(sk)->encap_err_rcv = cfg->encap_err_rcv;
net/ipv4/udp_tunnel_core.c
87
udp_sk(sk)->encap_err_lookup = cfg->encap_err_lookup;
net/ipv4/udp_tunnel_core.c
88
udp_sk(sk)->encap_destroy = cfg->encap_destroy;
net/ipv4/udp_tunnel_core.c
89
udp_sk(sk)->gro_receive = cfg->gro_receive;
net/ipv4/udp_tunnel_core.c
90
udp_sk(sk)->gro_complete = cfg->gro_complete;
net/ipv4/xfrm4_input.c
171
udp_sk(sk)->encap_type);
net/ipv4/xfrm4_input.c
86
struct udp_sock *up = udp_sk(sk);
net/ipv6/ipv6_sockglue.c
568
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
106
udp_sk(sk)->udp_portaddr_hash = hash2_partial;
net/ipv6/udp.c
1275
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
140
udp_sk(sk)->udp_port_hash != hnum ||
net/ipv6/udp.c
1459
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
1480
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
1800
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
1813
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
586
if (!__sk_queue_drop_skb(sk, &udp_sk(sk)->reader_queue, skb, flags,
net/ipv6/udp.c
671
up = udp_sk(sk);
net/ipv6/udp.c
684
up = udp_sk(sk);
net/ipv6/udp.c
721
if (!sk || READ_ONCE(udp_sk(sk)->encap_type)) {
net/ipv6/udp.c
764
if (udp_sk(sk)->encap_err_rcv)
net/ipv6/udp.c
765
udp_sk(sk)->encap_err_rcv(sk, skb, err, uh->dest,
net/ipv6/udp.c
830
struct udp_sock *up = udp_sk(sk);
net/ipv6/udp.c
950
if (udp_sk(sk)->udp_port_hash != hnum ||
net/ipv6/xfrm6_input.c
169
udp_sk(sk)->encap_type);
net/ipv6/xfrm6_input.c
81
struct udp_sock *up = udp_sk(sk);
net/l2tp/l2tp_core.c
154
WRITE_ONCE(udp_sk(sk)->encap_type, 0);
net/l2tp/l2tp_core.c
155
udp_sk(sk)->encap_rcv = NULL;
net/l2tp/l2tp_core.c
156
udp_sk(sk)->encap_destroy = NULL;
net/netfilter/ipvs/ip_vs_sync.c
1706
struct udp_sock *up = udp_sk(sk);
net/rxrpc/io_thread.c
26
int rxrpc_encap_rcv(struct sock *udp_sk, struct sk_buff *skb)
net/rxrpc/io_thread.c
29
struct rxrpc_local *local = rcu_dereference_sk_user_data(udp_sk);
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
28
struct udp_sock *udp_sk = ctx->udp_sk;
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
35
if (udp_sk == (void *)0)
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
46
inet = &udp_sk->inet;
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
50
inet = &udp_sk->inet;
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
55
rqueue = inet->sk.sk_rmem_alloc.counter - udp_sk->forward_deficit;
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
66
inet->sk.sk_refcnt.refs.counter, udp_sk,
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
67
udp_sk->drop_counters.drops0.counter +
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
68
udp_sk->drop_counters.drops1.counter);
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
35
struct udp_sock *udp_sk = ctx->udp_sk;
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
43
if (udp_sk == (void *)0)
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
50
udp6_sk = bpf_skc_to_udp6_sock(udp_sk);
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
54
inet = &udp_sk->inet;
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
57
rqueue = inet->sk.sk_rmem_alloc.counter - udp_sk->forward_deficit;
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
74
inet->sk.sk_refcnt.refs.counter, udp_sk,
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
75
udp_sk->drop_counters.drops0.counter +
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
76
udp_sk->drop_counters.drops1.counter);
tools/testing/selftests/bpf/progs/sock_destroy_prog.c
107
struct udp_sock *udp_sk = ctx->udp_sk;
tools/testing/selftests/bpf/progs/sock_destroy_prog.c
108
struct sock *sk = (struct sock *) udp_sk;
tools/testing/selftests/bpf/progs/sock_destroy_prog.c
129
struct udp_sock *udp_sk = ctx->udp_sk;
tools/testing/selftests/bpf/progs/sock_destroy_prog.c
130
struct sock *sk = (struct sock *) udp_sk;
tools/testing/selftests/bpf/progs/sock_destroy_prog.c
137
inet = &udp_sk->inet;
tools/testing/selftests/bpf/progs/sock_iter_batch.c
126
bucket[idx] = udp_sk(sk)->udp_portaddr_hash & udptable->mask;
tools/testing/selftests/bpf/progs/sock_iter_batch.c
99
struct sock *sk = (struct sock *)ctx->udp_sk;