crypto/cast5_generic.c
406
z[0] = x[0] ^ s5[xi(13)] ^ s6[xi(15)] ^ s7[xi(12)] ^ sb8[xi(14)] ^
crypto/cast5_generic.c
407
s7[xi(8)];
crypto/cast5_generic.c
409
sb8[xi(10)];
crypto/cast5_generic.c
411
s5[xi(9)];
crypto/cast5_generic.c
413
s6[xi(11)];
crypto/cast5_generic.c
424
x[1] = z[0] ^ s5[xi(0)] ^ s6[xi(2)] ^ s7[xi(1)] ^ sb8[xi(3)] ^
crypto/cast5_generic.c
426
x[2] = z[1] ^ s5[xi(7)] ^ s6[xi(6)] ^ s7[xi(5)] ^ sb8[xi(4)] ^
crypto/cast5_generic.c
428
x[3] = z[3] ^ s5[xi(10)] ^ s6[xi(9)] ^ s7[xi(11)] ^ sb8[xi(8)] ^
crypto/cast5_generic.c
430
k[4] = s5[xi(3)] ^ s6[xi(2)] ^ s7[xi(12)] ^ sb8[xi(13)] ^
crypto/cast5_generic.c
431
s5[xi(8)];
crypto/cast5_generic.c
432
k[5] = s5[xi(1)] ^ s6[xi(0)] ^ s7[xi(14)] ^ sb8[xi(15)] ^
crypto/cast5_generic.c
433
s6[xi(13)];
crypto/cast5_generic.c
434
k[6] = s5[xi(7)] ^ s6[xi(6)] ^ s7[xi(8)] ^ sb8[xi(9)] ^ s7[xi(3)];
crypto/cast5_generic.c
435
k[7] = s5[xi(5)] ^ s6[xi(4)] ^ s7[xi(10)] ^ sb8[xi(11)] ^
crypto/cast5_generic.c
436
sb8[xi(7)];
crypto/cast5_generic.c
438
z[0] = x[0] ^ s5[xi(13)] ^ s6[xi(15)] ^ s7[xi(12)] ^ sb8[xi(14)] ^
crypto/cast5_generic.c
439
s7[xi(8)];
crypto/cast5_generic.c
441
sb8[xi(10)];
crypto/cast5_generic.c
443
s5[xi(9)];
crypto/cast5_generic.c
445
s6[xi(11)];
crypto/cast5_generic.c
456
x[1] = z[0] ^ s5[xi(0)] ^ s6[xi(2)] ^ s7[xi(1)] ^ sb8[xi(3)] ^
crypto/cast5_generic.c
458
x[2] = z[1] ^ s5[xi(7)] ^ s6[xi(6)] ^ s7[xi(5)] ^ sb8[xi(4)] ^
crypto/cast5_generic.c
460
x[3] = z[3] ^ s5[xi(10)] ^ s6[xi(9)] ^ s7[xi(11)] ^ sb8[xi(8)] ^
crypto/cast5_generic.c
462
k[12] = s5[xi(8)] ^ s6[xi(9)] ^ s7[xi(7)] ^ sb8[xi(6)] ^ s5[xi(3)];
crypto/cast5_generic.c
463
k[13] = s5[xi(10)] ^ s6[xi(11)] ^ s7[xi(5)] ^ sb8[xi(4)] ^
crypto/cast5_generic.c
464
s6[xi(7)];
crypto/cast5_generic.c
465
k[14] = s5[xi(12)] ^ s6[xi(13)] ^ s7[xi(3)] ^ sb8[xi(2)] ^
crypto/cast5_generic.c
466
s7[xi(8)];
crypto/cast5_generic.c
467
k[15] = s5[xi(14)] ^ s6[xi(15)] ^ s7[xi(1)] ^ sb8[xi(0)] ^
crypto/cast5_generic.c
468
sb8[xi(13)];
drivers/gpu/drm/vc4/vc4_render_cl.c
266
uint8_t xi, yi;
drivers/gpu/drm/vc4/vc4_render_cl.c
369
for (xi = 0; xi < xtiles; xi++) {
drivers/gpu/drm/vc4/vc4_render_cl.c
370
int x = positive_x ? min_x_tile + xi : max_x_tile - xi;
drivers/gpu/drm/vc4/vc4_render_cl.c
371
bool first = (xi == 0 && yi == 0);
drivers/gpu/drm/vc4/vc4_render_cl.c
372
bool last = (xi == xtiles - 1 && yi == ytiles - 1);
drivers/net/ethernet/mellanox/mlx5/core/en.h
469
union mlx5e_xdp_info *xi;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
228
union mlx5e_xdp_info xi)
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
232
fifo->xi[i] = xi;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
238
return fifo->xi[(*fifo->cc)++ & fifo->mask];
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1433
kvfree(sq->db.xdpi_fifo.xi);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1448
size = array_size(sizeof(*xdpi_fifo->xi), entries);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1449
xdpi_fifo->xi = kvzalloc_node(size, GFP_KERNEL, numa);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1450
if (!xdpi_fifo->xi)
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2685
struct xdp_rxq_info *xi;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2690
xi = q->xdp_rxq_info;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2693
xdp_rxq_info_unreg(xi);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2694
kfree(xi);
drivers/scsi/ncr53c8xx.c
662
int xi = 0;
drivers/scsi/ncr53c8xx.c
755
if (xi < SCSI_NCR_MAX_EXCLUDES)
drivers/scsi/ncr53c8xx.c
756
driver_setup.excludes[xi++] = val;
drivers/scsi/sym53c8xx_2/sym_glue.c
84
int xi = 0;
drivers/scsi/sym53c8xx_2/sym_glue.c
86
while (p && (xi < 8)) {
drivers/scsi/sym53c8xx_2/sym_glue.c
89
sym_driver_setup.excludes[xi++] = val;
fs/ocfs2/xattr.c
150
struct ocfs2_xattr_info *xi);
fs/ocfs2/xattr.c
1514
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
1516
return loc->xl_ops->xlo_can_reuse(loc, xi);
fs/ocfs2/xattr.c
1521
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
1523
return loc->xl_ops->xlo_check_space(loc, xi);
fs/ocfs2/xattr.c
1539
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
154
struct ocfs2_xattr_info *xi);
fs/ocfs2/xattr.c
1541
int size = namevalue_size_xi(xi);
fs/ocfs2/xattr.c
1546
loc->xl_entry->xe_value_size = cpu_to_le64(xi->xi_value_len);
fs/ocfs2/xattr.c
1547
loc->xl_entry->xe_name_len = xi->xi_name_len;
fs/ocfs2/xattr.c
1548
ocfs2_xattr_set_type(loc->xl_entry, xi->xi_name_index);
fs/ocfs2/xattr.c
1550
xi->xi_value_len <= OCFS2_XATTR_INLINE_SIZE);
fs/ocfs2/xattr.c
1555
memcpy(nameval_buf, xi->xi_name, xi->xi_name_len);
fs/ocfs2/xattr.c
1606
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
1613
namevalue_size_xi(xi);
fs/ocfs2/xattr.c
1632
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
1636
int needed_space = ocfs2_xi_entry_usage(xi);
fs/ocfs2/xattr.c
1645
if (ocfs2_xa_can_reuse_entry(loc, xi))
fs/ocfs2/xattr.c
1767
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
1770
namevalue_size_xi(xi);
fs/ocfs2/xattr.c
1794
struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
1799
int needed_space = ocfs2_xi_entry_usage(xi);
fs/ocfs2/xattr.c
1800
int size = namevalue_size_xi(xi);
fs/ocfs2/xattr.c
1811
if (ocfs2_xa_can_reuse_entry(loc, xi))
fs/ocfs2/xattr.c
2096
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
2100
int name_size = OCFS2_XATTR_SIZE(xi->xi_name_len);
fs/ocfs2/xattr.c
2104
int xi_local = xi->xi_value_len <= OCFS2_XATTR_INLINE_SIZE;
fs/ocfs2/xattr.c
2127
xi->xi_value_len) {
fs/ocfs2/xattr.c
2128
rc = ocfs2_xa_value_truncate(loc, xi->xi_value_len,
fs/ocfs2/xattr.c
2141
loc->xl_entry->xe_value_size = cpu_to_le64(xi->xi_value_len);
fs/ocfs2/xattr.c
2157
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
2165
rc = ocfs2_xa_check_space(loc, xi);
fs/ocfs2/xattr.c
2170
if (ocfs2_xa_can_reuse_entry(loc, xi)) {
fs/ocfs2/xattr.c
2172
rc = ocfs2_xa_reuse_entry(loc, xi, ctxt);
fs/ocfs2/xattr.c
2197
ocfs2_xa_add_namevalue(loc, xi);
fs/ocfs2/xattr.c
2198
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE)
fs/ocfs2/xattr.c
2202
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE) {
fs/ocfs2/xattr.c
2204
rc = ocfs2_xa_value_truncate(loc, xi->xi_value_len, ctxt);
fs/ocfs2/xattr.c
2233
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
2238
int name_size = OCFS2_XATTR_SIZE(xi->xi_name_len);
fs/ocfs2/xattr.c
2243
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE) {
fs/ocfs2/xattr.c
2247
xi->xi_value,
fs/ocfs2/xattr.c
2248
xi->xi_value_len);
fs/ocfs2/xattr.c
2250
memcpy(nameval_buf + name_size, xi->xi_value, xi->xi_value_len);
fs/ocfs2/xattr.c
2256
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
226
static int namevalue_size_xi(struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
2260
u32 name_hash = ocfs2_xattr_name_hash(loc->xl_inode, xi->xi_name,
fs/ocfs2/xattr.c
2261
xi->xi_name_len);
fs/ocfs2/xattr.c
2277
if (!xi->xi_value) {
fs/ocfs2/xattr.c
228
return namevalue_size(xi->xi_name_len, xi->xi_value_len);
fs/ocfs2/xattr.c
2282
ret = ocfs2_xa_prepare_entry(loc, xi, name_hash, ctxt);
fs/ocfs2/xattr.c
2289
ret = ocfs2_xa_store_value(loc, xi, ctxt);
fs/ocfs2/xattr.c
267
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
2803
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
2826
ret = ocfs2_xa_set(&loc, xi, ctxt);
fs/ocfs2/xattr.c
294
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
2986
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
3016
ret = ocfs2_xa_set(&loc, xi, ctxt);
fs/ocfs2/xattr.c
3029
ret = ocfs2_xattr_set_entry_index_block(inode, xi, xs, ctxt);
fs/ocfs2/xattr.c
3037
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
3062
if (free >= (sizeof(struct ocfs2_xattr_entry) + namevalue_size_xi(xi)))
fs/ocfs2/xattr.c
3070
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
3086
xi->xi_value_len);
fs/ocfs2/xattr.c
3094
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE)
fs/ocfs2/xattr.c
3101
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE) {
fs/ocfs2/xattr.c
3145
if (!xi->xi_value) {
fs/ocfs2/xattr.c
3162
if (ocfs2_xattr_can_be_in_inode(inode, xi, xis)) {
fs/ocfs2/xattr.c
3174
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE) {
fs/ocfs2/xattr.c
3206
(value_size >= xi->xi_value_len)) ||
fs/ocfs2/xattr.c
3208
OCFS2_XATTR_ROOT_SIZE >= xi->xi_value_len))
fs/ocfs2/xattr.c
3260
if (xi->xi_value_len > OCFS2_XATTR_INLINE_SIZE) {
fs/ocfs2/xattr.c
3282
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
3296
ret = ocfs2_calc_xattr_set_need(inode, di, xi, xis, xbs,
fs/ocfs2/xattr.c
3304
trace_ocfs2_init_xattr_set_ctxt(xi->xi_name, meta_add,
fs/ocfs2/xattr.c
3338
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
3345
if (!xi->xi_value) {
fs/ocfs2/xattr.c
3348
ret = ocfs2_xattr_ibody_set(inode, xi, xis, ctxt);
fs/ocfs2/xattr.c
3350
ret = ocfs2_xattr_block_set(inode, xi, xbs, ctxt);
fs/ocfs2/xattr.c
3353
ret = ocfs2_xattr_ibody_set(inode, xi, xis, ctxt);
fs/ocfs2/xattr.c
3359
xi->xi_value = NULL;
fs/ocfs2/xattr.c
3360
xi->xi_value_len = 0;
fs/ocfs2/xattr.c
3366
xi,
fs/ocfs2/xattr.c
3383
ret = ocfs2_xattr_block_set(inode, xi, xbs, ctxt);
fs/ocfs2/xattr.c
3387
xi->xi_name_index,
fs/ocfs2/xattr.c
3388
xi->xi_name, xbs);
fs/ocfs2/xattr.c
3396
xi,
fs/ocfs2/xattr.c
3418
ret = ocfs2_xattr_block_set(inode, xi, xbs, ctxt);
fs/ocfs2/xattr.c
3426
xi->xi_value = NULL;
fs/ocfs2/xattr.c
3427
xi->xi_value_len = 0;
fs/ocfs2/xattr.c
3431
xi,
fs/ocfs2/xattr.c
3447
ret = ocfs2_xattr_ibody_set(inode, xi,
fs/ocfs2/xattr.c
3491
struct ocfs2_xattr_info xi = {
fs/ocfs2/xattr.c
3543
ret = __ocfs2_xattr_set_handle(inode, di, &xi, &xis, &xbs, &ctxt);
fs/ocfs2/xattr.c
3576
struct ocfs2_xattr_info xi = {
fs/ocfs2/xattr.c
3644
ret = ocfs2_prepare_refcount_xattr(inode, di, &xi,
fs/ocfs2/xattr.c
3665
ret = ocfs2_init_xattr_set_ctxt(inode, di, &xi, &xis,
fs/ocfs2/xattr.c
3681
ret = __ocfs2_xattr_set_handle(inode, di, &xi, &xis, &xbs, &ctxt);
fs/ocfs2/xattr.c
5569
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
5576
trace_ocfs2_xattr_set_entry_bucket(xi->xi_name);
fs/ocfs2/xattr.c
5580
ret = ocfs2_xa_set(&loc, xi, ctxt);
fs/ocfs2/xattr.c
5598
ret = ocfs2_xa_set(&loc, xi, ctxt);
fs/ocfs2/xattr.c
5612
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
5618
trace_ocfs2_xattr_set_entry_index_block(xi->xi_name);
fs/ocfs2/xattr.c
5620
ret = ocfs2_xattr_set_entry_bucket(inode, xi, xs, ctxt);
fs/ocfs2/xattr.c
5638
xi->xi_name);
fs/ocfs2/xattr.c
564
static int ocfs2_xi_entry_usage(struct ocfs2_xattr_info *xi)
fs/ocfs2/xattr.c
566
return namevalue_size_xi(xi) +
fs/ocfs2/xattr.c
5662
xi->xi_name_index,
fs/ocfs2/xattr.c
5663
xi->xi_name, xs);
fs/ocfs2/xattr.c
5669
ret = ocfs2_xattr_set_entry_bucket(inode, xi, xs, ctxt);
fs/ocfs2/xattr.c
5785
struct ocfs2_xattr_info *xi,
fs/ocfs2/xattr.c
5887
if (!xi->xi_value || xi->xi_value_len <= OCFS2_XATTR_INLINE_SIZE) {
lib/bch.c
1142
unsigned int sum, x, y, remaining, ak = 0, xi[BCH_MAX_M];
lib/bch.c
1156
memset(xi, 0, sizeof(xi));
lib/bch.c
1162
if (y && (r < m) && !xi[r]) {
lib/bch.c
1164
xi[r] = 1;
lib/crypto/curve25519-hacl64.c
369
u64 xi = a[0];
lib/crypto/curve25519-hacl64.c
371
a[0] = yi - xi;
lib/crypto/curve25519-hacl64.c
374
u64 xi = a[1];
lib/crypto/curve25519-hacl64.c
376
a[1] = yi - xi;
lib/crypto/curve25519-hacl64.c
379
u64 xi = a[2];
lib/crypto/curve25519-hacl64.c
381
a[2] = yi - xi;
lib/crypto/curve25519-hacl64.c
384
u64 xi = a[3];
lib/crypto/curve25519-hacl64.c
386
a[3] = yi - xi;
lib/crypto/curve25519-hacl64.c
389
u64 xi = a[4];
lib/crypto/curve25519-hacl64.c
391
a[4] = yi - xi;
lib/crypto/curve25519-hacl64.c
403
u64 xi = b[0];
lib/crypto/curve25519-hacl64.c
404
tmp[0] = ((u128)(xi) * (s));
lib/crypto/curve25519-hacl64.c
407
u64 xi = b[1];
lib/crypto/curve25519-hacl64.c
408
tmp[1] = ((u128)(xi) * (s));
lib/crypto/curve25519-hacl64.c
411
u64 xi = b[2];
lib/crypto/curve25519-hacl64.c
412
tmp[2] = ((u128)(xi) * (s));
lib/crypto/curve25519-hacl64.c
415
u64 xi = b[3];
lib/crypto/curve25519-hacl64.c
416
tmp[3] = ((u128)(xi) * (s));
lib/crypto/curve25519-hacl64.c
419
u64 xi = b[4];
lib/crypto/curve25519-hacl64.c
420
tmp[4] = ((u128)(xi) * (s));
lib/crypto/curve25519-hacl64.c
54
u128 xi = input[0];
lib/crypto/curve25519-hacl64.c
55
output[0] = ((u64)(xi));
lib/crypto/curve25519-hacl64.c
58
u128 xi = input[1];
lib/crypto/curve25519-hacl64.c
59
output[1] = ((u64)(xi));
lib/crypto/curve25519-hacl64.c
62
u128 xi = input[2];
lib/crypto/curve25519-hacl64.c
63
output[2] = ((u64)(xi));
lib/crypto/curve25519-hacl64.c
66
u128 xi = input[3];
lib/crypto/curve25519-hacl64.c
67
output[3] = ((u64)(xi));
lib/crypto/curve25519-hacl64.c
70
u128 xi = input[4];
lib/crypto/curve25519-hacl64.c
71
output[4] = ((u64)(xi));
net/ceph/auth_none.c
133
struct ceph_auth_none_info *xi;
net/ceph/auth_none.c
136
xi = kzalloc_obj(*xi, GFP_NOFS);
net/ceph/auth_none.c
137
if (!xi)
net/ceph/auth_none.c
140
xi->starting = true;
net/ceph/auth_none.c
143
ac->private = xi;
net/ceph/auth_none.c
17
struct ceph_auth_none_info *xi = ac->private;
net/ceph/auth_none.c
19
xi->starting = true;
net/ceph/auth_none.c
30
struct ceph_auth_none_info *xi = ac->private;
net/ceph/auth_none.c
32
return !xi->starting;
net/ceph/auth_none.c
37
struct ceph_auth_none_info *xi = ac->private;
net/ceph/auth_none.c
39
return xi->starting;
net/ceph/auth_none.c
77
struct ceph_auth_none_info *xi = ac->private;
net/ceph/auth_none.c
79
xi->starting = false;
net/ceph/auth_x.c
1173
struct ceph_x_info *xi;
net/ceph/auth_x.c
1177
xi = kzalloc_obj(*xi, GFP_NOFS);
net/ceph/auth_x.c
1178
if (!xi)
net/ceph/auth_x.c
1187
ret = ceph_crypto_key_clone(&xi->secret, ac->key);
net/ceph/auth_x.c
1193
ret = ceph_crypto_key_prepare(&xi->secret, client_key_usages,
net/ceph/auth_x.c
1200
xi->starting = true;
net/ceph/auth_x.c
1201
xi->ticket_handlers = RB_ROOT;
net/ceph/auth_x.c
1204
ac->private = xi;
net/ceph/auth_x.c
1209
ceph_crypto_key_destroy(&xi->secret);
net/ceph/auth_x.c
1211
kfree(xi);
net/ceph/auth_x.c
154
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
155
struct rb_node *parent = NULL, **p = &xi->ticket_handlers.rb_node;
net/ceph/auth_x.c
174
rb_insert_color(&th->node, &xi->ticket_handlers);
net/ceph/auth_x.c
181
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
184
rb_erase(&th->node, &xi->ticket_handlers);
net/ceph/auth_x.c
195
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
300
xi->have_keys |= th->service;
net/ceph/auth_x.c
40
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
45
missing = ac->want_keys & ~xi->have_keys;
net/ceph/auth_x.c
48
ac->want_keys, xi->have_keys, missing, !missing);
net/ceph/auth_x.c
508
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
511
*pneed = ac->want_keys & ~(xi->have_keys);
net/ceph/auth_x.c
531
xi->have_keys &= ~service;
net/ceph/auth_x.c
538
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
54
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
551
xi->have_keys, need);
net/ceph/auth_x.c
555
void *enc_buf = xi->auth_authorizer.enc_buf;
net/ceph/auth_x.c
566
if (xi->secret.type == CEPH_CRYPTO_AES) {
net/ceph/auth_x.c
567
blob = enc_buf + ceph_x_encrypt_offset(&xi->secret);
net/ceph/auth_x.c
576
blob->server_challenge = cpu_to_le64(xi->server_challenge);
net/ceph/auth_x.c
578
if (xi->secret.type == CEPH_CRYPTO_AES) {
net/ceph/auth_x.c
579
ret = ceph_x_encrypt(&xi->secret, 0 /* dummy */,
net/ceph/auth_x.c
585
ceph_hmac_sha256(&xi->secret, blob, sizeof(*blob),
net/ceph/auth_x.c
59
ac->want_keys, xi->have_keys, need, !!need);
net/ceph/auth_x.c
595
xi->server_challenge, le64_to_cpu(auth->client_challenge),
net/ceph/auth_x.c
612
ret = ceph_x_build_authorizer(ac, th, &xi->auth_authorizer);
net/ceph/auth_x.c
620
xi->auth_authorizer.buf->vec.iov_base,
net/ceph/auth_x.c
621
xi->auth_authorizer.buf->vec.iov_len, e_range);
net/ceph/auth_x.c
666
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
673
ret = ceph_x_proc_ticket_reply(ac, &xi->secret, p, end);
net/ceph/auth_x.c
735
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
743
if (xi->starting) {
net/ceph/auth_x.c
749
xi->server_challenge = le64_to_cpu(sc->server_challenge);
net/ceph/auth_x.c
751
xi->server_challenge);
net/ceph/auth_x.c
752
xi->starting = false;
net/ceph/auth_x.c
753
xi->have_keys &= ~CEPH_ENTITY_TYPE_AUTH;
net/ceph/auth_x.c
783
if (ac->want_keys == xi->have_keys)
net/ceph/auth_x.c
974
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
977
xi->starting = true;
net/ceph/auth_x.c
978
xi->server_challenge = 0;
net/ceph/auth_x.c
983
struct ceph_x_info *xi = ac->private;
net/ceph/auth_x.c
987
ceph_crypto_key_destroy(&xi->secret);
net/ceph/auth_x.c
989
while ((p = rb_first(&xi->ticket_handlers)) != NULL) {
net/ceph/auth_x.c
995
ceph_x_authorizer_cleanup(&xi->auth_authorizer);
net/xfrm/xfrm_interface_core.c
146
#define for_each_xfrmi_rcu(start, xi) \
net/xfrm/xfrm_interface_core.c
147
for (xi = rcu_dereference(start); xi; xi = rcu_dereference(xi->next))
net/xfrm/xfrm_interface_core.c
157
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
159
for_each_xfrmi_rcu(xfrmn->xfrmi[xfrmi_hash(x->if_id)], xi) {
net/xfrm/xfrm_interface_core.c
160
if (x->if_id == xi->p.if_id &&
net/xfrm/xfrm_interface_core.c
161
(xi->dev->flags & IFF_UP))
net/xfrm/xfrm_interface_core.c
162
return xi;
net/xfrm/xfrm_interface_core.c
165
xi = rcu_dereference(xfrmn->collect_md_xfrmi);
net/xfrm/xfrm_interface_core.c
166
if (xi && (xi->dev->flags & IFF_UP))
net/xfrm/xfrm_interface_core.c
167
return xi;
net/xfrm/xfrm_interface_core.c
177
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
205
xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
206
res->net = xi->net;
net/xfrm/xfrm_interface_core.c
208
if (xi->p.collect_md)
net/xfrm/xfrm_interface_core.c
211
res->if_id = xi->p.if_id;
net/xfrm/xfrm_interface_core.c
215
static void xfrmi_link(struct xfrmi_net *xfrmn, struct xfrm_if *xi)
net/xfrm/xfrm_interface_core.c
217
struct xfrm_if __rcu **xip = &xfrmn->xfrmi[xfrmi_hash(xi->p.if_id)];
net/xfrm/xfrm_interface_core.c
219
rcu_assign_pointer(xi->next , rtnl_dereference(*xip));
net/xfrm/xfrm_interface_core.c
220
rcu_assign_pointer(*xip, xi);
net/xfrm/xfrm_interface_core.c
223
static void xfrmi_unlink(struct xfrmi_net *xfrmn, struct xfrm_if *xi)
net/xfrm/xfrm_interface_core.c
228
for (xip = &xfrmn->xfrmi[xfrmi_hash(xi->p.if_id)];
net/xfrm/xfrm_interface_core.c
231
if (xi == iter) {
net/xfrm/xfrm_interface_core.c
232
rcu_assign_pointer(*xip, xi->next);
net/xfrm/xfrm_interface_core.c
240
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
242
gro_cells_destroy(&xi->gro_cells);
net/xfrm/xfrm_interface_core.c
247
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
256
if (xi->p.collect_md)
net/xfrm/xfrm_interface_core.c
257
rcu_assign_pointer(xfrmn->collect_md_xfrmi, xi);
net/xfrm/xfrm_interface_core.c
259
xfrmi_link(xfrmn, xi);
net/xfrm/xfrm_interface_core.c
270
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
274
(xi = rtnl_dereference(*xip)) != NULL;
net/xfrm/xfrm_interface_core.c
275
xip = &xi->next)
net/xfrm/xfrm_interface_core.c
276
if (xi->p.if_id == p->if_id)
net/xfrm/xfrm_interface_core.c
277
return xi;
net/xfrm/xfrm_interface_core.c
284
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
285
struct xfrmi_net *xfrmn = net_generic(xi->net, xfrmi_net_id);
net/xfrm/xfrm_interface_core.c
287
if (xi->p.collect_md)
net/xfrm/xfrm_interface_core.c
290
xfrmi_unlink(xfrmn, xi);
net/xfrm/xfrm_interface_core.c
363
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
372
xi = xfrmi_lookup(xs_net(x), x);
net/xfrm/xfrm_interface_core.c
373
if (!xi)
net/xfrm/xfrm_interface_core.c
377
dev = xi->dev;
net/xfrm/xfrm_interface_core.c
387
xnet = !net_eq(xi->net, dev_net(skb->dev));
net/xfrm/xfrm_interface_core.c
407
if (xi->p.collect_md) {
net/xfrm/xfrm_interface_core.c
426
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
435
if (xi->p.collect_md) {
net/xfrm/xfrm_interface_core.c
452
if_id = xi->p.if_id;
net/xfrm/xfrm_interface_core.c
456
dst = xfrm_lookup_with_ifid(xi->net, dst, fl, NULL, 0, if_id);
net/xfrm/xfrm_interface_core.c
504
xfrmi_scrub_packet(skb, !net_eq(xi->net, dev_net(dev)));
net/xfrm/xfrm_interface_core.c
508
err = dst_output(xi->net, skb_to_full_sk(skb), skb);
net/xfrm/xfrm_interface_core.c
527
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
570
fl.flowi_oif = xi->p.link;
net/xfrm/xfrm_interface_core.c
594
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
630
xi = xfrmi_lookup(net, x);
net/xfrm/xfrm_interface_core.c
631
if (!xi) {
net/xfrm/xfrm_interface_core.c
655
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
684
xi = xfrmi_lookup(net, x);
net/xfrm/xfrm_interface_core.c
685
if (!xi) {
net/xfrm/xfrm_interface_core.c
700
static int xfrmi_change(struct xfrm_if *xi, const struct xfrm_if_parms *p)
net/xfrm/xfrm_interface_core.c
702
if (xi->p.link != p->link)
net/xfrm/xfrm_interface_core.c
705
xi->p.if_id = p->if_id;
net/xfrm/xfrm_interface_core.c
710
static int xfrmi_update(struct xfrm_if *xi, struct xfrm_if_parms *p)
net/xfrm/xfrm_interface_core.c
712
struct net *net = xi->net;
net/xfrm/xfrm_interface_core.c
716
xfrmi_unlink(xfrmn, xi);
net/xfrm/xfrm_interface_core.c
718
err = xfrmi_change(xi, p);
net/xfrm/xfrm_interface_core.c
719
xfrmi_link(xfrmn, xi);
net/xfrm/xfrm_interface_core.c
720
netdev_state_change(xi->dev);
net/xfrm/xfrm_interface_core.c
726
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
728
return READ_ONCE(xi->p.link);
net/xfrm/xfrm_interface_core.c
763
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
764
struct net_device *phydev = __dev_get_by_index(xi->net, xi->p.link);
net/xfrm/xfrm_interface_core.c
767
err = gro_cells_init(&xi->gro_cells, dev);
net/xfrm/xfrm_interface_core.c
822
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
845
xi = xfrmi_locate(net, &p);
net/xfrm/xfrm_interface_core.c
846
if (xi)
net/xfrm/xfrm_interface_core.c
850
xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
851
xi->p = p;
net/xfrm/xfrm_interface_core.c
852
xi->net = net;
net/xfrm/xfrm_interface_core.c
853
xi->dev = dev;
net/xfrm/xfrm_interface_core.c
868
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
869
struct net *net = xi->net;
net/xfrm/xfrm_interface_core.c
878
if (p.collect_md || xi->p.collect_md) {
net/xfrm/xfrm_interface_core.c
883
xi = xfrmi_locate(net, &p);
net/xfrm/xfrm_interface_core.c
884
if (!xi) {
net/xfrm/xfrm_interface_core.c
885
xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
887
if (xi->dev != dev)
net/xfrm/xfrm_interface_core.c
891
return xfrmi_update(xi, &p);
net/xfrm/xfrm_interface_core.c
908
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
909
struct xfrm_if_parms *parm = &xi->p;
net/xfrm/xfrm_interface_core.c
913
(xi->p.collect_md && nla_put_flag(skb, IFLA_XFRM_COLLECT_METADATA)))
net/xfrm/xfrm_interface_core.c
923
struct xfrm_if *xi = netdev_priv(dev);
net/xfrm/xfrm_interface_core.c
925
return READ_ONCE(xi->net);
net/xfrm/xfrm_interface_core.c
955
struct xfrm_if *xi;
net/xfrm/xfrm_interface_core.c
960
(xi = rtnl_net_dereference(net, *xip)) != NULL;
net/xfrm/xfrm_interface_core.c
961
xip = &xi->next)
net/xfrm/xfrm_interface_core.c
962
unregister_netdevice_queue(xi->dev, dev_to_kill);
net/xfrm/xfrm_interface_core.c
965
xi = rtnl_net_dereference(net, xfrmn->collect_md_xfrmi);
net/xfrm/xfrm_interface_core.c
966
if (xi)
net/xfrm/xfrm_interface_core.c
967
unregister_netdevice_queue(xi->dev, dev_to_kill);