Symbol: prng
arch/s390/boot/kaslr.c
48
struct prng_parm prng = {
arch/s390/boot/kaslr.c
76
*(unsigned long *) prng.parm_block ^= seed;
arch/s390/boot/kaslr.c
78
cpacf_kmc(CPACF_KMC_PRNG, prng.parm_block,
arch/s390/boot/kaslr.c
81
memcpy(prng.parm_block, entropy, sizeof(entropy));
arch/s390/boot/kaslr.c
84
cpacf_kmc(CPACF_KMC_PRNG, prng.parm_block, (u8 *) &random,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
115
.prng = CE_ALG_PRNG_V2,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
132
.prng = CE_ALG_PRNG,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
151
.prng = CE_ALG_PRNG,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
168
.prng = CE_ALG_PRNG,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
52
.prng = CE_ALG_PRNG,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
69
.prng = CE_ALG_PRNG,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
91
.prng = CE_ALG_PRNG_V2,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
929
if (ce->variant->prng == CE_ID_NOTSUPP) {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-prng.c
122
common = ce->variant->prng | CE_COMM_INT;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce.h
159
unsigned char prng;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1198
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1248
order = i915_random_order(count * count, &prng);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1351
static u32 igt_random_size(struct rnd_state *prng,
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1364
size = prandom_u32_state(prng) & mask;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1375
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1399
size = igt_random_size(&prng, min, rounddown_pow_of_two(size));
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1622
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1657
sz = i915_prandom_u32_max_state(SZ_4M, &prng);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1706
order = i915_random_order(count * count, &prng);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1716
u32 rnd = i915_prandom_u32_max_state(UINT_MAX, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
300
struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
329
i915_prandom_u32_max_state(CLIENT_NUM_TILING_TYPES, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
420
struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
426
x = i915_prandom_u32_max_state(t->width, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
427
y = i915_prandom_u32_max_state(t->height, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
536
tiled_blits_create(struct intel_engine_cs *engine, struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
579
err = tiled_blits_create_buffers(t, WIDTH, HEIGHT, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
601
struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
614
fill_scratch(t, map, prandom_u32_state(prng));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
615
GEM_BUG_ON(verify_buffer(t, &t->scratch, prng));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
621
err = verify_buffer(t, &t->buffers[i], prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
632
static int tiled_blits_bounce(struct tiled_blits *t, struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
655
err = verify_buffer(t, &t->buffers[2], prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
663
struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
668
t = tiled_blits_create(engine, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
672
err = tiled_blits_prepare(t, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
676
err = tiled_blits_bounce(t, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
708
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
728
err = __igt_client_tiled_blits(engine, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
289
random_engine(struct drm_i915_private *i915, struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
298
count = i915_prandom_u32_max_state(count, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
313
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
332
ctx.engine = random_engine(i915, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
368
i915_random_reorder(offsets, ncachelines, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
370
values[n] = prandom_u32_state(&prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1378
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1432
if (prandom_u32_state(&prng) & 1)
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1779
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1859
offset = igt_random_offset(&prng,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
128
page = i915_prandom_u32_max_state(npages, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
454
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
494
i915_prandom_u32_max_state(I915_TILING_Y + 1, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
520
i915_prandom_u32_max_state(max_pitch, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
526
err = check_partial_mapping(obj, &tile, &prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
96
struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_execlists.c
3478
struct rnd_state prng;
drivers/gpu/drm/i915/gt/selftest_execlists.c
3486
&smoke->prng)];
drivers/gpu/drm/i915/gt/selftest_execlists.c
3624
ctx, random_priority(&smoke->prng),
drivers/gpu/drm/i915/gt/selftest_execlists.c
3642
.prng = I915_RND_STATE_INITIALIZER(i915_selftest.random_seed),
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
893
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
932
i915_prandom_u32_max_state(512, &prng),
drivers/gpu/drm/i915/gt/selftest_lrc.c
1770
struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_lrc.c
1779
prandom_bytes_state(prng,
drivers/gpu/drm/i915/gt/selftest_lrc.c
1799
static int __lrc_garbage(struct intel_engine_cs *engine, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_lrc.c
1809
hang = garbage(ce, prng);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1862
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1870
err = __lrc_garbage(engine, &prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
116
int x = i * 1024 + i915_prandom_u32_max_state(1024, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
261
u32 sz, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_migrate.c
335
i915_prandom_u32_max_state(1024, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
377
ccs_bytes_left), prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
42
u32 sz, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_migrate.c
439
migrate_copy(struct intel_migrate *migrate, u32 sz, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_migrate.c
441
return copy(migrate, __migrate_copy, sz, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
445
global_copy(struct intel_migrate *migrate, u32 sz, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_migrate.c
447
return copy(migrate, __global_copy, sz, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
477
migrate_clear(struct intel_migrate *migrate, u32 sz, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_migrate.c
479
return clear(migrate, __migrate_clear, sz, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
483
global_clear(struct intel_migrate *migrate, u32 sz, struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_migrate.c
485
return clear(migrate, __global_clear, sz, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
493
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
499
err = migrate_copy(migrate, sizes[i], &prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
501
err = global_copy(migrate, sizes[i], &prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
515
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
521
err = migrate_clear(migrate, sizes[i], &prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
523
err = global_clear(migrate, sizes[i], &prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
679
struct rnd_state prng;
drivers/gpu/drm/i915/gt/selftest_migrate.c
688
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
700
thread[i].prng =
drivers/gpu/drm/i915/gt/selftest_migrate.c
701
I915_RND_STATE_INITIALIZER(prandom_u32_state(&prng));
drivers/gpu/drm/i915/gt/selftest_migrate.c
742
return migrate_copy(tm->migrate, 2 * CHUNK_SZ, &tm->prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
757
return global_copy(tm->migrate, 2 * CHUNK_SZ, &tm->prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
772
return migrate_clear(tm->migrate, 2 * CHUNK_SZ, &tm->prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
779
return global_clear(tm->migrate, 2 * CHUNK_SZ, &tm->prng);
drivers/gpu/drm/i915/gt/selftest_rc6.c
191
struct rnd_state *prng,
drivers/gpu/drm/i915/gt/selftest_rc6.c
212
i915_prandom_shuffle(engines, sizeof(*engines), n, prng);
drivers/gpu/drm/i915/gt/selftest_rc6.c
223
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_rc6.c
230
engines = randomised_engines(gt, &prng, &count);
drivers/gpu/drm/i915/gt/selftest_timeline.c
130
&state->prng);
drivers/gpu/drm/i915/gt/selftest_timeline.c
133
&state->prng);
drivers/gpu/drm/i915/gt/selftest_timeline.c
162
state.prng = I915_RND_STATE_INITIALIZER(i915_selftest.random_seed);
drivers/gpu/drm/i915/gt/selftest_timeline.c
289
struct rnd_state prng;
drivers/gpu/drm/i915/gt/selftest_timeline.c
303
prandom_seed_state(&prng, i915_selftest.random_seed);
drivers/gpu/drm/i915/gt/selftest_timeline.c
311
WRITE_ONCE(x, prandom_u32_state(&prng));
drivers/gpu/drm/i915/gt/selftest_timeline.c
321
prandom_seed_state(&prng, i915_selftest.random_seed);
drivers/gpu/drm/i915/gt/selftest_timeline.c
326
u64 id = i915_prandom_u64_state(&prng);
drivers/gpu/drm/i915/gt/selftest_timeline.c
337
prandom_seed_state(&prng, i915_selftest.random_seed);
drivers/gpu/drm/i915/gt/selftest_timeline.c
341
u64 id = i915_prandom_u64_state(&prng);
drivers/gpu/drm/i915/gt/selftest_timeline.c
390
prandom_seed_state(&prng, i915_selftest.random_seed);
drivers/gpu/drm/i915/gt/selftest_timeline.c
395
u32 id = random_engine(&prng);
drivers/gpu/drm/i915/gt/selftest_timeline.c
396
u32 seqno = prandom_u32_state(&prng);
drivers/gpu/drm/i915/gt/selftest_timeline.c
70
struct rnd_state prng;
drivers/gpu/drm/i915/gt/selftest_tlb.c
238
I915_RND_STATE(prng);
drivers/gpu/drm/i915/gt/selftest_tlb.c
327
&prng);
drivers/gpu/drm/i915/gt/selftest_tlb.c
336
&prng);
drivers/gpu/drm/i915/gt/selftest_tlb.c
38
struct rnd_state *prng)
drivers/gpu/drm/i915/gt/selftest_tlb.c
67
addr = igt_random_offset(prng, addr, min(ce->vm->total, BIT_ULL(48)),
drivers/gpu/drm/i915/gt/selftest_tlb.c
91
addr = igt_random_offset(prng, addr, addr + align, 8, 8);
drivers/gpu/drm/i915/selftests/i915_gem.c
48
struct rnd_state prng;
drivers/gpu/drm/i915/selftests/i915_gem.c
55
prandom_seed_state(&prng, 0x12345678);
drivers/gpu/drm/i915/selftests/i915_gem.c
69
iowrite32(prandom_u32_state(&prng), &s[x]);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1351
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1395
order = i915_random_order(count, &prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1411
i915_random_reorder(order, count, &prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1542
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1661
offset = igt_random_offset(&prng,
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
260
I915_RND_SUBSTATE(prng, seed_prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
277
order = i915_random_order(count, &prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
369
i915_random_reorder(order, count, &prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
797
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
828
order = i915_random_order(count, &prng);
drivers/gpu/drm/i915/selftests/i915_request.c
319
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_request.c
338
order = i915_random_order(total, &prng);
drivers/gpu/drm/i915/selftests/i915_request.c
362
i915_random_reorder(order, total, &prng);
drivers/gpu/drm/i915/selftests/i915_request.c
363
count = 1 + i915_prandom_u32_max_state(max_batch, &prng);
drivers/gpu/drm/i915/selftests/i915_syncmap.c
204
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_syncmap.c
217
u64 context = i915_prandom_u64_state(&prng);
drivers/gpu/drm/i915/selftests/i915_syncmap.c
226
prandom_u32_state(&prng));
drivers/gpu/drm/i915/selftests/i915_syncmap.c
402
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_syncmap.c
414
u64 context = i915_prandom_u64_state(&prng) & ~MASK;
drivers/gpu/drm/i915/selftests/i915_syncmap.c
544
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/i915_syncmap.c
562
u64 context = i915_prandom_u64_state(&prng);
drivers/gpu/drm/i915/selftests/i915_syncmap.c
578
seqno = prandom_u32_state(&prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1006
count = i915_prandom_u32_max_state(count, prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1014
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1035
engine = random_engine_class(i915, I915_ENGINE_CLASS_COPY, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1041
sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1083
order = i915_random_order(count * count, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1090
bytes[0] = igt_random_offset(&prng, 64, PAGE_SIZE - 64, 0, sizeof(u32));
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1109
offset = igt_random_offset(&prng, 0, obj->base.size,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1112
val = prandom_u32_state(&prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1120
dword = igt_random_offset(&prng, offset,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
165
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
170
order = i915_random_order(count, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
184
u64 size = i915_prandom_u32_max_state(chunk_size, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
191
offset = igt_random_offset(&prng, 0, chunk_size, size,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
208
u32 size = i915_prandom_u32_max_state(cur_avail, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
243
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
279
target = i915_prandom_u64_state(&prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
454
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
470
if (i915_prandom_u64_state(&prng) & 1)
drivers/gpu/drm/i915/selftests/intel_memory_region.c
557
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
562
if (i915_prandom_u64_state(&prng) & 1)
drivers/gpu/drm/i915/selftests/intel_memory_region.c
565
div64_u64_rem(i915_prandom_u64_state(&prng), SZ_8G, &total);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
569
div64_u64_rem(i915_prandom_u64_state(&prng), total - ps, &io_size);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
585
div64_u64_rem(i915_prandom_u64_state(&prng), rem, &size);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
624
div64_u64_rem(i915_prandom_u64_state(&prng), rem, &size);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
712
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
736
order = i915_random_order(count * count, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
753
u32 rng = prandom_u32_state(&prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
868
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
875
size = max_t(u32, PAGE_SIZE, i915_prandom_u32_max_state(SZ_32M, &prng));
drivers/gpu/drm/i915/selftests/intel_memory_region.c
907
&prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
924
val = prandom_u32_state(&prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
953
I915_RND_STATE(prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
967
sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
994
struct rnd_state *prng)
drivers/gpu/drm/i915/selftests/scatterlist.c
282
struct rnd_state prng;
drivers/gpu/drm/i915/selftests/scatterlist.c
297
prandom_seed_state(&prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
299
err = alloc_table(&pt, sz, sz, *npages, &prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
306
prandom_seed_state(&prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
308
err = expect_pfn_sgtable(&pt, *npages, &prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
338
struct rnd_state prng;
drivers/gpu/drm/i915/selftests/scatterlist.c
340
prandom_seed_state(&prng, i915_selftest.random_seed);
drivers/gpu/drm/i915/selftests/scatterlist.c
341
err = alloc_table(&pt, prime, max, *npages, &prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
355
prandom_seed_state(&prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
358
*npages, &prng,
drivers/gpu/drm/tests/drm_buddy_test.c
131
DRM_RND_STATE(prng, random_seed);
drivers/gpu/drm/tests/drm_buddy_test.c
139
ps = roundup_pow_of_two(prandom_u32_state(&prng) % bias_size);
drivers/gpu/drm/tests/drm_buddy_test.c
149
order = drm_random_order(count, &prng);
drivers/gpu/drm/tests/drm_buddy_test.c
225
size = max(round_up(prandom_u32_state(&prng) % bias_rem, ps), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
247
size = max(round_up(prandom_u32_state(&prng) % bias_rem, ps), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
284
bias_start = round_up(prandom_u32_state(&prng) % (mm_size - ps), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
285
bias_end = round_up(bias_start + prandom_u32_state(&prng) % (mm_size - bias_start), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
290
u32 size = max(round_up(prandom_u32_state(&prng) % bias_rem, ps), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
310
bias_start -= round_up(prandom_u32_state(&prng) % bias_start, ps);
drivers/gpu/drm/tests/drm_buddy_test.c
312
bias_end += round_up(prandom_u32_state(&prng) % (mm_size - bias_end), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
342
bias_start = round_up(prandom_u32_state(&prng) % (mm_size - ps), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
343
bias_end = round_up(bias_start + prandom_u32_state(&prng) % (mm_size - bias_start), ps);
drivers/gpu/drm/tests/drm_buddy_test.c
348
size = max(round_up(prandom_u32_state(&prng) % bias_rem, ps), ps);
drivers/gpu/drm/xe/tests/xe_bo.c
449
struct rnd_state prng;
drivers/gpu/drm/xe/tests/xe_bo.c
454
prandom_seed_state(&prng, rand_seed);
drivers/gpu/drm/xe/tests/xe_bo.c
522
int ret = shrink_test_fill_random(bo, &prng, link);
drivers/gpu/drm/xe/tests/xe_bo.c
554
prandom_seed_state(&prng, rand_seed);
drivers/gpu/drm/xe/tests/xe_bo.c
576
failed = shrink_test_verify(test, bo, count, &prng, link);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
305
struct rnd_state prng;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
306
prandom_seed_state(&prng, dev->tv_freq ^ 0x55);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
308
prandom_u32_state(&prng) & 0x3f);
net/sched/sch_netem.c
1116
q->prng.seed = nla_get_u64(tb[TCA_NETEM_PRNG_SEED]);
net/sched/sch_netem.c
1118
q->prng.seed = get_random_u64();
net/sched/sch_netem.c
1119
prandom_seed_state(&q->prng.prng_state, q->prng.seed);
net/sched/sch_netem.c
114
} prng;
net/sched/sch_netem.c
1284
if (nla_put_u64_64bit(skb, TCA_NETEM_PRNG_SEED, q->prng.seed,
net/sched/sch_netem.c
190
static u32 get_crandom(struct crndstate *state, struct prng *p)
net/sched/sch_netem.c
213
u32 rnd = prandom_u32_state(&q->prng.prng_state);
net/sched/sch_netem.c
278
struct rnd_state *s = &q->prng.prng_state;
net/sched/sch_netem.c
302
return q->loss && q->loss >= get_crandom(&q->loss_cor, &q->prng);
net/sched/sch_netem.c
331
struct prng *prng,
net/sched/sch_netem.c
341
rnd = get_crandom(state, prng);
net/sched/sch_netem.c
464
if (q->duplicate && q->duplicate >= get_crandom(&q->dup_cor, &q->prng))
net/sched/sch_netem.c
499
if (q->corrupt && q->corrupt >= get_crandom(&q->corrupt_cor, &q->prng)) {
net/sched/sch_netem.c
556
q->reorder < get_crandom(&q->reorder_cor, &q->prng)) {
net/sched/sch_netem.c
561
&q->delay_cor, &q->prng, q->delay_dist);
net/sched/sch_netem.c
668
NULL, &q->prng, q->slot_dist);