lib/libexpat/lib/xmlparse.c
3396
ENTITY *entity;
lib/libexpat/lib/xmlparse.c
3418
entity = (ENTITY *)lookup(parser, &dtd->generalEntities, name, 0);
lib/libexpat/lib/xmlparse.c
3425
if (! entity)
lib/libexpat/lib/xmlparse.c
3427
else if (! entity->is_internal)
lib/libexpat/lib/xmlparse.c
3429
} else if (! entity) {
lib/libexpat/lib/xmlparse.c
3436
if (entity->open)
lib/libexpat/lib/xmlparse.c
3438
if (entity->notation)
lib/libexpat/lib/xmlparse.c
3440
if (entity->textPtr) {
lib/libexpat/lib/xmlparse.c
3444
parser->m_skippedEntityHandler(parser->m_handlerArg, entity->name,
lib/libexpat/lib/xmlparse.c
3450
result = processEntity(parser, entity, XML_FALSE, ENTITY_INTERNAL);
lib/libexpat/lib/xmlparse.c
3455
entity->open = XML_TRUE;
lib/libexpat/lib/xmlparse.c
3457
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
3461
parser->m_externalEntityRefHandlerArg, context, entity->base,
lib/libexpat/lib/xmlparse.c
3462
entity->systemId, entity->publicId))
lib/libexpat/lib/xmlparse.c
432
ENTITY *entity;
lib/libexpat/lib/xmlparse.c
503
static enum XML_Error processEntity(XML_Parser parser, ENTITY *entity,
lib/libexpat/lib/xmlparse.c
5425
ENTITY *entity = (ENTITY *)lookup(parser, &dtd->paramEntities,
lib/libexpat/lib/xmlparse.c
5427
if (! entity) {
lib/libexpat/lib/xmlparse.c
5436
entity->base = parser->m_curBase;
lib/libexpat/lib/xmlparse.c
5439
parser->m_externalEntityRefHandlerArg, 0, entity->base,
lib/libexpat/lib/xmlparse.c
5440
entity->systemId, entity->publicId))
lib/libexpat/lib/xmlparse.c
5472
ENTITY *entity = (ENTITY *)lookup(parser, &dtd->paramEntities,
lib/libexpat/lib/xmlparse.c
5474
if (! entity)
lib/libexpat/lib/xmlparse.c
5476
entity->base = parser->m_curBase;
lib/libexpat/lib/xmlparse.c
5479
parser->m_externalEntityRefHandlerArg, 0, entity->base,
lib/libexpat/lib/xmlparse.c
5480
entity->systemId, entity->publicId))
lib/libexpat/lib/xmlparse.c
552
static enum XML_Error storeSelfEntityValue(XML_Parser parser, ENTITY *entity);
lib/libexpat/lib/xmlparse.c
6004
ENTITY *entity;
lib/libexpat/lib/xmlparse.c
6009
entity = (ENTITY *)lookup(parser, &dtd->paramEntities, name, 0);
lib/libexpat/lib/xmlparse.c
6018
if (! entity)
lib/libexpat/lib/xmlparse.c
6020
else if (! entity->is_internal) {
lib/libexpat/lib/xmlparse.c
6043
} else if (! entity) {
lib/libexpat/lib/xmlparse.c
6053
if (entity->open)
lib/libexpat/lib/xmlparse.c
6055
if (entity->textPtr) {
lib/libexpat/lib/xmlparse.c
6059
result = processEntity(parser, entity, betweenDecl, ENTITY_INTERNAL);
lib/libexpat/lib/xmlparse.c
6067
entity->open = XML_TRUE;
lib/libexpat/lib/xmlparse.c
6068
entityTrackingOnOpen(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6070
parser->m_externalEntityRefHandlerArg, 0, entity->base,
lib/libexpat/lib/xmlparse.c
6071
entity->systemId, entity->publicId)) {
lib/libexpat/lib/xmlparse.c
6072
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6073
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6076
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6077
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
631
static void entityTrackingReportStats(XML_Parser parser, ENTITY *entity,
lib/libexpat/lib/xmlparse.c
633
static void entityTrackingOnOpen(XML_Parser parser, ENTITY *entity,
lib/libexpat/lib/xmlparse.c
635
static void entityTrackingOnClose(XML_Parser parser, ENTITY *entity,
lib/libexpat/lib/xmlparse.c
6360
processEntity(XML_Parser parser, ENTITY *entity, XML_Bool betweenDecl,
lib/libexpat/lib/xmlparse.c
6397
entity->open = XML_TRUE;
lib/libexpat/lib/xmlparse.c
6398
entity->hasMore = XML_TRUE;
lib/libexpat/lib/xmlparse.c
6400
entityTrackingOnOpen(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6402
entity->processed = 0;
lib/libexpat/lib/xmlparse.c
6405
openEntity->entity = entity;
lib/libexpat/lib/xmlparse.c
6426
ENTITY *entity;
lib/libexpat/lib/xmlparse.c
6434
entity = openEntity->entity;
lib/libexpat/lib/xmlparse.c
6437
if (entity->hasMore) {
lib/libexpat/lib/xmlparse.c
6438
textStart = ((const char *)entity->textPtr) + entity->processed;
lib/libexpat/lib/xmlparse.c
6439
textEnd = (const char *)(entity->textPtr + entity->textLen);
lib/libexpat/lib/xmlparse.c
6443
if (entity->is_param) {
lib/libexpat/lib/xmlparse.c
6463
entity->processed = (int)(next - (const char *)entity->textPtr);
lib/libexpat/lib/xmlparse.c
6470
entity->hasMore = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6471
if (! entity->is_param
lib/libexpat/lib/xmlparse.c
6481
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6488
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6496
parser->m_processor = entity->is_param ? prologProcessor : contentProcessor;
lib/libexpat/lib/xmlparse.c
6527
ENTITY *const entity = openEntity->entity;
lib/libexpat/lib/xmlparse.c
6529
= ((const char *)entity->textPtr) + entity->processed;
lib/libexpat/lib/xmlparse.c
6531
= (const char *)(entity->textPtr + entity->textLen);
lib/libexpat/lib/xmlparse.c
6534
if (entity->hasMore) {
lib/libexpat/lib/xmlparse.c
6544
entity->processed
lib/libexpat/lib/xmlparse.c
6545
= (int)(nextInEntity - (const char *)entity->textPtr);
lib/libexpat/lib/xmlparse.c
6552
entity->hasMore = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6558
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6565
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6664
ENTITY *entity;
lib/libexpat/lib/xmlparse.c
6686
entity = (ENTITY *)lookup(parser, &dtd->generalEntities, name, 0);
lib/libexpat/lib/xmlparse.c
6701
if (! entity)
lib/libexpat/lib/xmlparse.c
6703
else if (! entity->is_internal)
lib/libexpat/lib/xmlparse.c
6705
} else if (! entity) {
lib/libexpat/lib/xmlparse.c
6718
if (entity->open) {
lib/libexpat/lib/xmlparse.c
6741
if (entity->notation) {
lib/libexpat/lib/xmlparse.c
6746
if (! entity->textPtr) {
lib/libexpat/lib/xmlparse.c
6752
result = processEntity(parser, entity, XML_FALSE, ENTITY_ATTRIBUTE);
lib/libexpat/lib/xmlparse.c
6828
ENTITY *entity;
lib/libexpat/lib/xmlparse.c
6836
entity = (ENTITY *)lookup(parser, &dtd->paramEntities, name, 0);
lib/libexpat/lib/xmlparse.c
6838
if (! entity) {
lib/libexpat/lib/xmlparse.c
6848
if (entity->open || (entity == parser->m_declEntity)) {
lib/libexpat/lib/xmlparse.c
6854
if (entity->systemId) {
lib/libexpat/lib/xmlparse.c
6857
entity->open = XML_TRUE;
lib/libexpat/lib/xmlparse.c
6858
entityTrackingOnOpen(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6860
parser->m_externalEntityRefHandlerArg, 0, entity->base,
lib/libexpat/lib/xmlparse.c
6861
entity->systemId, entity->publicId)) {
lib/libexpat/lib/xmlparse.c
6862
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6863
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6867
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
6868
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
6874
result = processEntity(parser, entity, XML_FALSE, ENTITY_VALUE);
lib/libexpat/lib/xmlparse.c
6985
ENTITY *const entity = openEntity->entity;
lib/libexpat/lib/xmlparse.c
6987
= ((const char *)entity->textPtr) + entity->processed;
lib/libexpat/lib/xmlparse.c
6989
= (const char *)(entity->textPtr + entity->textLen);
lib/libexpat/lib/xmlparse.c
6992
if (entity->hasMore) {
lib/libexpat/lib/xmlparse.c
7002
entity->processed
lib/libexpat/lib/xmlparse.c
7003
= (int)(nextInEntity - (const char *)entity->textPtr);
lib/libexpat/lib/xmlparse.c
7010
entity->hasMore = XML_FALSE;
lib/libexpat/lib/xmlparse.c
7016
entityTrackingOnClose(parser, entity, __LINE__);
lib/libexpat/lib/xmlparse.c
7023
entity->open = XML_FALSE;
lib/libexpat/lib/xmlparse.c
7044
storeSelfEntityValue(XML_Parser parser, ENTITY *entity) {
lib/libexpat/lib/xmlparse.c
7052
|| ! poolAppendString(pool, entity->name)
lib/libexpat/lib/xmlparse.c
7058
entity->textPtr = poolStart(pool);
lib/libexpat/lib/xmlparse.c
7059
entity->textLen = (int)(poolLength(pool));
lib/libexpat/lib/xmlparse.c
8658
entityTrackingReportStats(XML_Parser rootParser, ENTITY *entity,
lib/libexpat/lib/xmlparse.c
8667
const char *const entityName = entity->name;
lib/libexpat/lib/xmlparse.c
8677
entity->is_param ? "%" : "&", entityName, action, entity->textLen,
lib/libexpat/lib/xmlparse.c
8682
entityTrackingOnOpen(XML_Parser originParser, ENTITY *entity, int sourceLine) {
lib/libexpat/lib/xmlparse.c
8693
entityTrackingReportStats(rootParser, entity, "OPEN ", sourceLine);
lib/libexpat/lib/xmlparse.c
8697
entityTrackingOnClose(XML_Parser originParser, ENTITY *entity, int sourceLine) {
lib/libexpat/lib/xmlparse.c
8701
entityTrackingReportStats(rootParser, entity, "CLOSE", sourceLine);
sys/dev/ipmi.c
1327
ipmi_sensor_type(int type, int ext_type, int units2, int entity)
sys/dev/ipmi.c
1349
if (entity == IPMI_ENTITY_PWRSUPPLY)
sys/dev/ipmi.c
1399
int entity, const char *name)
sys/dev/ipmi.c
1405
typ = ipmi_sensor_type(sensor_type, ext_type, s1->units2, entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
103
p->entities[i] = entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1109
struct drm_sched_entity *entity = p->entities[i];
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1110
struct drm_gpu_scheduler *sched = entity->rq->sched;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1241
sched = p->gang_leader->base.entity->rq->sched;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1494
struct drm_sched_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1504
wait->in.ring, &entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1510
fence = amdgpu_ctx_get_fence(ctx, entity, wait->in.handle);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1542
struct drm_sched_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1552
user->ring, &entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1558
fence = amdgpu_ctx_get_fence(ctx, entity, user->seq_no);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
416
struct drm_sched_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
425
deps[i].ring, &entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
431
fence = amdgpu_ctx_get_fence(ctx, entity, deps[i].handle);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
77
struct drm_sched_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
83
chunk_ib->ring, &entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
91
if (entity->rq == NULL)
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
96
if (p->entities[i] == entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
209
struct amdgpu_ctx_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
215
entity = kzalloc(struct_size(entity, fences, amdgpu_sched_jobs),
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
217
if (!entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
222
entity->hw_ip = hw_ip;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
223
entity->sequence = 1;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
252
r = drm_sched_entity_init(&entity->entity, drm_prio, scheds, num_scheds,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
258
if (cmpxchg(&ctx->entities[hw_ip][ring], NULL, entity))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
264
drm_sched_entity_fini(&entity->entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
267
kfree(entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
273
struct amdgpu_ctx_entity *entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
278
if (!entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
282
res = ktime_add(res, amdgpu_ctx_fence_time(entity->fences[i]));
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
283
dma_fence_put(entity->fences[i]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
286
amdgpu_xcp_release_sched(adev, entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
288
kfree(entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
33
container_of((e), struct amdgpu_ctx_entity, entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
435
u32 ring, struct drm_sched_entity **entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
462
ctx_entity = &ctx->entities[hw_ip][ring]->entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
469
*entity = ctx_entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
517
drm_sched_entity_destroy(&ctx->entities[i][j]->entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
758
struct drm_sched_entity *entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
761
struct amdgpu_ctx_entity *centity = to_amdgpu_ctx_entity(entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
785
struct drm_sched_entity *entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
788
struct amdgpu_ctx_entity *centity = to_amdgpu_ctx_entity(entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
824
drm_sched_entity_set_priority(&aentity->entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
833
drm_sched_entity_modify_sched(&aentity->entity, scheds,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
860
struct drm_sched_entity *entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
862
struct amdgpu_ctx_entity *centity = to_amdgpu_ctx_entity(entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
908
struct drm_sched_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
913
entity = &ctx->entities[i][j]->entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
914
timeout = drm_sched_entity_flush(entity, timeout);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
938
struct drm_sched_entity *entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
943
entity = &ctx->entities[i][j]->entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.c
944
drm_sched_entity_fini(entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.h
41
struct drm_sched_entity entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.h
77
u32 ring, struct drm_sched_entity **entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.h
79
struct drm_sched_entity *entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.h
82
struct drm_sched_entity *entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ctx.h
91
struct drm_sched_entity *entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
1461
struct drm_sched_entity entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
1470
r = drm_sched_entity_init(&entity, DRM_SCHED_PRIORITY_NORMAL,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
1484
r = amdgpu_job_alloc_with_ib(ring->adev, &entity, owner,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
1508
drm_sched_entity_destroy(&entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
184
struct drm_sched_entity *entity, void *owner,
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
201
if (!entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
204
return drm_sched_job_init(&(*job)->base, entity, 1, owner,
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
209
struct drm_sched_entity *entity, void *owner,
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
215
r = amdgpu_job_alloc(adev, NULL, entity, owner, 1, job,
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
223
if (entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
298
if (job->base.entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
423
drm_sched_entity_queue_pop(struct drm_sched_entity *entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.c
427
node = spsc_queue_pop(&entity->job_queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.h
105
return to_amdgpu_ring(job->base.entity->rq->sched);
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.h
109
struct drm_sched_entity *entity, void *owner,
sys/dev/pci/drm/amd/amdgpu/amdgpu_job.h
113
struct drm_sched_entity *entity, void *owner,
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
156
__entry->ring = to_amdgpu_ring(job->base.entity->rq->sched)->idx;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
159
to_amdgpu_ring(job->base.entity->rq->sched));
sys/dev/pci/drm/amd/amdgpu/amdgpu_ttm.c
2325
struct drm_sched_entity *entity = delayed ? &adev->mman.low_pr :
sys/dev/pci/drm/amd/amdgpu/amdgpu_ttm.c
2327
r = amdgpu_job_alloc_with_ib(adev, entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
1136
r = amdgpu_job_alloc_with_ib(ring->adev, &adev->uvd.entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
375
drm_sched_entity_destroy(&adev->uvd.entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
411
r = drm_sched_entity_init(&adev->uvd.entity, DRM_SCHED_PRIORITY_NORMAL,
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.h
65
struct drm_sched_entity entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
215
drm_sched_entity_destroy(&adev->vce.entity);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
243
r = drm_sched_entity_init(&adev->vce.entity, DRM_SCHED_PRIORITY_NORMAL,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
449
r = amdgpu_job_alloc_with_ib(ring->adev, &ring->adev->vce.entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
539
r = amdgpu_job_alloc_with_ib(ring->adev, &ring->adev->vce.entity,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.h
51
struct drm_sched_entity entity;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_sdma.c
47
struct drm_sched_entity *entity = p->immediate ? &p->vm->immediate
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_sdma.c
58
r = amdgpu_job_alloc_with_ib(p->adev, entity, AMDGPU_FENCE_OWNER_VM,
sys/dev/pci/drm/amd/amdgpu/amdgpu_xcp.c
466
struct amdgpu_ctx_entity *entity)
sys/dev/pci/drm/amd/amdgpu/amdgpu_xcp.c
474
sched = entity->entity.rq->sched;
sys/dev/pci/drm/amd/amdgpu/amdgpu_xcp.c
476
ring = to_amdgpu_ring(entity->entity.rq->sched);
sys/dev/pci/drm/amd/amdgpu/amdgpu_xcp.h
177
struct amdgpu_ctx_entity *entity);
sys/dev/pci/drm/amd/amdgpu/vcn_v3_0.c
1894
fence = amdgpu_ctx_get_fence(p->ctx, job->base.entity, ~0ull);
sys/dev/pci/drm/amd/amdgpu/vcn_v3_0.c
1902
drm_sched_entity_modify_sched(job->base.entity, scheds, 1);
sys/dev/pci/drm/amd/amdgpu/vcn_v4_0.c
1811
fence = amdgpu_ctx_get_fence(p->ctx, job->base.entity, ~0ull);
sys/dev/pci/drm/amd/amdgpu/vcn_v4_0.c
1819
drm_sched_entity_modify_sched(job->base.entity, scheds, 1);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
362
struct drm_sched_entity *entity;
sys/dev/pci/drm/include/drm/gpu_scheduler.h
660
struct drm_sched_entity *entity,
sys/dev/pci/drm/include/drm/gpu_scheduler.h
690
int drm_sched_entity_init(struct drm_sched_entity *entity,
sys/dev/pci/drm/include/drm/gpu_scheduler.h
695
long drm_sched_entity_flush(struct drm_sched_entity *entity, long timeout);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
696
void drm_sched_entity_fini(struct drm_sched_entity *entity);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
697
void drm_sched_entity_destroy(struct drm_sched_entity *entity);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
698
void drm_sched_entity_set_priority(struct drm_sched_entity *entity,
sys/dev/pci/drm/include/drm/gpu_scheduler.h
700
int drm_sched_entity_error(struct drm_sched_entity *entity);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
701
void drm_sched_entity_modify_sched(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
58
TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
59
TP_ARGS(sched_job, entity),
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
72
__entry->job_count = spsc_queue_count(&entity->job_queue);
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
87
TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
88
TP_ARGS(sched_job, entity)
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
92
TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
sys/dev/pci/drm/scheduler/gpu_scheduler_trace.h
93
TP_ARGS(sched_job, entity)
sys/dev/pci/drm/scheduler/sched_entity.c
100
entity->priority, sched_list[0]->num_rqs);
sys/dev/pci/drm/scheduler/sched_entity.c
101
entity->priority = max_t(s32, (s32) sched_list[0]->num_rqs - 1,
sys/dev/pci/drm/scheduler/sched_entity.c
104
entity->rq = sched_list[0]->sched_rq[entity->priority];
sys/dev/pci/drm/scheduler/sched_entity.c
107
init_completion(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_entity.c
110
complete_all(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_entity.c
112
mtx_init(&entity->lock, IPL_NONE);
sys/dev/pci/drm/scheduler/sched_entity.c
113
spsc_queue_init(&entity->job_queue);
sys/dev/pci/drm/scheduler/sched_entity.c
115
atomic_set(&entity->fence_seq, 0);
sys/dev/pci/drm/scheduler/sched_entity.c
116
entity->fence_context = dma_fence_context_alloc(2);
sys/dev/pci/drm/scheduler/sched_entity.c
134
void drm_sched_entity_modify_sched(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_entity.c
140
spin_lock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
141
entity->sched_list = sched_list;
sys/dev/pci/drm/scheduler/sched_entity.c
142
entity->num_sched_list = num_sched_list;
sys/dev/pci/drm/scheduler/sched_entity.c
143
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
147
static bool drm_sched_entity_is_idle(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
151
if (list_empty(&entity->list) ||
sys/dev/pci/drm/scheduler/sched_entity.c
152
spsc_queue_count(&entity->job_queue) == 0 ||
sys/dev/pci/drm/scheduler/sched_entity.c
153
entity->stopped)
sys/dev/pci/drm/scheduler/sched_entity.c
166
int drm_sched_entity_error(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
172
fence = rcu_dereference(entity->last_scheduled);
sys/dev/pci/drm/scheduler/sched_entity.c
236
static void drm_sched_entity_kill(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
241
if (!entity->rq)
sys/dev/pci/drm/scheduler/sched_entity.c
244
spin_lock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
245
entity->stopped = true;
sys/dev/pci/drm/scheduler/sched_entity.c
246
drm_sched_rq_remove_entity(entity->rq, entity);
sys/dev/pci/drm/scheduler/sched_entity.c
247
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
250
wait_for_completion(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_entity.c
253
prev = rcu_dereference_check(entity->last_scheduled, true);
sys/dev/pci/drm/scheduler/sched_entity.c
255
while ((job = drm_sched_entity_queue_pop(entity))) {
sys/dev/pci/drm/scheduler/sched_entity.c
287
long drm_sched_entity_flush(struct drm_sched_entity *entity, long timeout)
sys/dev/pci/drm/scheduler/sched_entity.c
297
if (!entity->rq)
sys/dev/pci/drm/scheduler/sched_entity.c
300
sched = entity->rq->sched;
sys/dev/pci/drm/scheduler/sched_entity.c
314
drm_sched_entity_is_idle(entity),
sys/dev/pci/drm/scheduler/sched_entity.c
318
drm_sched_entity_is_idle(entity));
sys/dev/pci/drm/scheduler/sched_entity.c
323
last_user = cmpxchg(&entity->last_user, current->group_leader, NULL);
sys/dev/pci/drm/scheduler/sched_entity.c
327
last_user = cmpxchg(&entity->last_user, curpr, NULL);
sys/dev/pci/drm/scheduler/sched_entity.c
332
drm_sched_entity_kill(entity);
sys/dev/pci/drm/scheduler/sched_entity.c
349
void drm_sched_entity_fini(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
356
drm_sched_entity_kill(entity);
sys/dev/pci/drm/scheduler/sched_entity.c
358
if (entity->dependency) {
sys/dev/pci/drm/scheduler/sched_entity.c
359
dma_fence_remove_callback(entity->dependency, &entity->cb);
sys/dev/pci/drm/scheduler/sched_entity.c
360
dma_fence_put(entity->dependency);
sys/dev/pci/drm/scheduler/sched_entity.c
361
entity->dependency = NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
364
dma_fence_put(rcu_dereference_check(entity->last_scheduled, true));
sys/dev/pci/drm/scheduler/sched_entity.c
365
RCU_INIT_POINTER(entity->last_scheduled, NULL);
sys/dev/pci/drm/scheduler/sched_entity.c
376
void drm_sched_entity_destroy(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
378
drm_sched_entity_flush(entity, MAX_WAIT_SCHED_ENTITY_Q_EMPTY);
sys/dev/pci/drm/scheduler/sched_entity.c
379
drm_sched_entity_fini(entity);
sys/dev/pci/drm/scheduler/sched_entity.c
390
struct drm_sched_entity *entity =
sys/dev/pci/drm/scheduler/sched_entity.c
393
entity->dependency = NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
395
drm_sched_wakeup(entity->rq->sched);
sys/dev/pci/drm/scheduler/sched_entity.c
406
void drm_sched_entity_set_priority(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_entity.c
409
spin_lock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
410
entity->priority = priority;
sys/dev/pci/drm/scheduler/sched_entity.c
411
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
419
static bool drm_sched_entity_add_dependency_cb(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_entity.c
422
struct drm_gpu_scheduler *sched = entity->rq->sched;
sys/dev/pci/drm/scheduler/sched_entity.c
423
struct dma_fence *fence = entity->dependency;
sys/dev/pci/drm/scheduler/sched_entity.c
426
if (fence->context == entity->fence_context ||
sys/dev/pci/drm/scheduler/sched_entity.c
427
fence->context == entity->fence_context + 1) {
sys/dev/pci/drm/scheduler/sched_entity.c
433
dma_fence_put(entity->dependency);
sys/dev/pci/drm/scheduler/sched_entity.c
446
dma_fence_put(entity->dependency);
sys/dev/pci/drm/scheduler/sched_entity.c
447
entity->dependency = fence;
sys/dev/pci/drm/scheduler/sched_entity.c
451
!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &entity->dependency->flags))
sys/dev/pci/drm/scheduler/sched_entity.c
452
trace_drm_sched_job_unschedulable(sched_job, entity->dependency);
sys/dev/pci/drm/scheduler/sched_entity.c
454
if (!dma_fence_add_callback(entity->dependency, &entity->cb,
sys/dev/pci/drm/scheduler/sched_entity.c
458
dma_fence_put(entity->dependency);
sys/dev/pci/drm/scheduler/sched_entity.c
464
struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
479
return job->sched->ops->prepare_job(job, entity);
sys/dev/pci/drm/scheduler/sched_entity.c
484
struct drm_sched_job *drm_sched_entity_pop_job(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
488
sched_job = drm_sched_entity_queue_peek(entity);
sys/dev/pci/drm/scheduler/sched_entity.c
492
while ((entity->dependency =
sys/dev/pci/drm/scheduler/sched_entity.c
493
drm_sched_job_dependency(sched_job, entity))) {
sys/dev/pci/drm/scheduler/sched_entity.c
494
if (drm_sched_entity_add_dependency_cb(entity, sched_job))
sys/dev/pci/drm/scheduler/sched_entity.c
499
if (entity->guilty && atomic_read(entity->guilty))
sys/dev/pci/drm/scheduler/sched_entity.c
502
dma_fence_put(rcu_dereference_check(entity->last_scheduled, true));
sys/dev/pci/drm/scheduler/sched_entity.c
503
rcu_assign_pointer(entity->last_scheduled,
sys/dev/pci/drm/scheduler/sched_entity.c
513
spsc_queue_pop(&entity->job_queue);
sys/dev/pci/drm/scheduler/sched_entity.c
522
next = drm_sched_entity_queue_peek(entity);
sys/dev/pci/drm/scheduler/sched_entity.c
526
spin_lock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
527
rq = entity->rq;
sys/dev/pci/drm/scheduler/sched_entity.c
529
drm_sched_rq_update_fifo_locked(entity, rq,
sys/dev/pci/drm/scheduler/sched_entity.c
532
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
540
sched_job->entity = NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
545
void drm_sched_entity_select_rq(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_entity.c
552
if (!entity->sched_list)
sys/dev/pci/drm/scheduler/sched_entity.c
556
if (spsc_queue_count(&entity->job_queue))
sys/dev/pci/drm/scheduler/sched_entity.c
567
fence = rcu_dereference_check(entity->last_scheduled, true);
sys/dev/pci/drm/scheduler/sched_entity.c
573
spin_lock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
574
sched = drm_sched_pick_best(entity->sched_list, entity->num_sched_list);
sys/dev/pci/drm/scheduler/sched_entity.c
575
rq = sched ? sched->sched_rq[entity->priority] : NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
576
if (rq != entity->rq) {
sys/dev/pci/drm/scheduler/sched_entity.c
577
drm_sched_rq_remove_entity(entity->rq, entity);
sys/dev/pci/drm/scheduler/sched_entity.c
578
entity->rq = rq;
sys/dev/pci/drm/scheduler/sched_entity.c
58
int drm_sched_entity_init(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_entity.c
581
if (entity->num_sched_list == 1)
sys/dev/pci/drm/scheduler/sched_entity.c
582
entity->sched_list = NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
584
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
598
struct drm_sched_entity *entity = sched_job->entity;
sys/dev/pci/drm/scheduler/sched_entity.c
602
trace_drm_sched_job_queue(sched_job, entity);
sys/dev/pci/drm/scheduler/sched_entity.c
611
atomic_inc(entity->rq->sched->score);
sys/dev/pci/drm/scheduler/sched_entity.c
613
WRITE_ONCE(entity->last_user, current->group_leader);
sys/dev/pci/drm/scheduler/sched_entity.c
615
WRITE_ONCE(entity->last_user, curproc->p_p);
sys/dev/pci/drm/scheduler/sched_entity.c
624
first = spsc_queue_push(&entity->job_queue, &sched_job->queue_node);
sys/dev/pci/drm/scheduler/sched_entity.c
632
spin_lock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
633
if (entity->stopped) {
sys/dev/pci/drm/scheduler/sched_entity.c
634
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
64
if (!(entity && sched_list && (num_sched_list == 0 || sched_list[0])))
sys/dev/pci/drm/scheduler/sched_entity.c
640
rq = entity->rq;
sys/dev/pci/drm/scheduler/sched_entity.c
644
drm_sched_rq_add_entity(rq, entity);
sys/dev/pci/drm/scheduler/sched_entity.c
647
drm_sched_rq_update_fifo_locked(entity, rq, submit_ts);
sys/dev/pci/drm/scheduler/sched_entity.c
650
spin_unlock(&entity->lock);
sys/dev/pci/drm/scheduler/sched_entity.c
67
memset(entity, 0, sizeof(struct drm_sched_entity));
sys/dev/pci/drm/scheduler/sched_entity.c
68
INIT_LIST_HEAD(&entity->list);
sys/dev/pci/drm/scheduler/sched_entity.c
69
entity->rq = NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
70
entity->guilty = guilty;
sys/dev/pci/drm/scheduler/sched_entity.c
71
entity->num_sched_list = num_sched_list;
sys/dev/pci/drm/scheduler/sched_entity.c
72
entity->priority = priority;
sys/dev/pci/drm/scheduler/sched_entity.c
74
entity->last_user = current->group_leader;
sys/dev/pci/drm/scheduler/sched_entity.c
76
entity->last_user = curproc->p_p;
sys/dev/pci/drm/scheduler/sched_entity.c
83
entity->sched_list = num_sched_list > 1 ? sched_list : NULL;
sys/dev/pci/drm/scheduler/sched_entity.c
84
RCU_INIT_POINTER(entity->last_scheduled, NULL);
sys/dev/pci/drm/scheduler/sched_entity.c
85
RB_CLEAR_NODE(&entity->rb_tree_node);
sys/dev/pci/drm/scheduler/sched_entity.c
98
if (entity->priority >= sched_list[0]->num_rqs) {
sys/dev/pci/drm/scheduler/sched_fence.c
229
struct drm_sched_fence *drm_sched_fence_alloc(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_fence.c
251
struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_fence.c
255
fence->sched = entity->rq->sched;
sys/dev/pci/drm/scheduler/sched_fence.c
256
seq = atomic_inc_return(&entity->fence_seq);
sys/dev/pci/drm/scheduler/sched_fence.c
258
&fence->lock, entity->fence_context, seq);
sys/dev/pci/drm/scheduler/sched_fence.c
260
&fence->lock, entity->fence_context + 1, seq);
sys/dev/pci/drm/scheduler/sched_internal.h
16
struct drm_sched_entity *entity);
sys/dev/pci/drm/scheduler/sched_internal.h
18
struct drm_sched_entity *entity);
sys/dev/pci/drm/scheduler/sched_internal.h
20
void drm_sched_rq_update_fifo_locked(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_internal.h
23
void drm_sched_entity_select_rq(struct drm_sched_entity *entity);
sys/dev/pci/drm/scheduler/sched_internal.h
24
struct drm_sched_job *drm_sched_entity_pop_job(struct drm_sched_entity *entity);
sys/dev/pci/drm/scheduler/sched_internal.h
29
struct drm_sched_entity *entity);
sys/dev/pci/drm/scheduler/sched_internal.h
46
drm_sched_entity_queue_pop(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_internal.h
50
node = spsc_queue_pop(&entity->job_queue);
sys/dev/pci/drm/scheduler/sched_internal.h
67
drm_sched_entity_queue_peek(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_internal.h
71
node = spsc_queue_peek(&entity->job_queue);
sys/dev/pci/drm/scheduler/sched_internal.h
80
drm_sched_entity_is_ready(struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_internal.h
82
if (!spsc_queue_count(&entity->job_queue))
sys/dev/pci/drm/scheduler/sched_internal.h
85
if (READ_ONCE(entity->dependency))
sys/dev/pci/drm/scheduler/sched_main.c
1104
struct drm_sched_entity *entity;
sys/dev/pci/drm/scheduler/sched_main.c
1110
entity = drm_sched_policy == DRM_SCHED_POLICY_FIFO ?
sys/dev/pci/drm/scheduler/sched_main.c
1113
if (entity)
sys/dev/pci/drm/scheduler/sched_main.c
1117
return IS_ERR(entity) ? NULL : entity;
sys/dev/pci/drm/scheduler/sched_main.c
118
struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_main.c
122
s_job = drm_sched_entity_queue_peek(entity);
sys/dev/pci/drm/scheduler/sched_main.c
1239
struct drm_sched_entity *entity;
sys/dev/pci/drm/scheduler/sched_main.c
1246
entity = drm_sched_select_entity(sched);
sys/dev/pci/drm/scheduler/sched_main.c
1247
if (!entity)
sys/dev/pci/drm/scheduler/sched_main.c
1250
sched_job = drm_sched_entity_pop_job(entity);
sys/dev/pci/drm/scheduler/sched_main.c
1252
complete_all(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_main.c
1262
trace_drm_sched_job_run(sched_job, entity);
sys/dev/pci/drm/scheduler/sched_main.c
1268
complete_all(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_main.c
147
static void drm_sched_rq_remove_fifo_locked(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_main.c
1491
struct drm_sched_entity *entity;
sys/dev/pci/drm/scheduler/sched_main.c
150
if (!RB_EMPTY_NODE(&entity->rb_tree_node)) {
sys/dev/pci/drm/scheduler/sched_main.c
1505
list_for_each_entry_safe(entity, tmp, &rq->entities, list) {
sys/dev/pci/drm/scheduler/sched_main.c
1507
entity->fence_context) {
sys/dev/pci/drm/scheduler/sched_main.c
1508
if (entity->guilty)
sys/dev/pci/drm/scheduler/sched_main.c
1509
atomic_set(entity->guilty, 1);
sys/dev/pci/drm/scheduler/sched_main.c
151
rb_erase_cached(&entity->rb_tree_node, &rq->rb_tree_root);
sys/dev/pci/drm/scheduler/sched_main.c
1514
if (&entity->list != &rq->entities)
sys/dev/pci/drm/scheduler/sched_main.c
152
RB_CLEAR_NODE(&entity->rb_tree_node);
sys/dev/pci/drm/scheduler/sched_main.c
156
void drm_sched_rq_update_fifo_locked(struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_main.c
165
lockdep_assert_held(&entity->lock);
sys/dev/pci/drm/scheduler/sched_main.c
168
drm_sched_rq_remove_fifo_locked(entity, rq);
sys/dev/pci/drm/scheduler/sched_main.c
170
entity->oldest_job_waiting = ts;
sys/dev/pci/drm/scheduler/sched_main.c
172
rb_add_cached(&entity->rb_tree_node, &rq->rb_tree_root,
sys/dev/pci/drm/scheduler/sched_main.c
203
struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_main.c
205
lockdep_assert_held(&entity->lock);
sys/dev/pci/drm/scheduler/sched_main.c
208
if (!list_empty(&entity->list))
sys/dev/pci/drm/scheduler/sched_main.c
212
list_add_tail(&entity->list, &rq->entities);
sys/dev/pci/drm/scheduler/sched_main.c
224
struct drm_sched_entity *entity)
sys/dev/pci/drm/scheduler/sched_main.c
226
lockdep_assert_held(&entity->lock);
sys/dev/pci/drm/scheduler/sched_main.c
228
if (list_empty(&entity->list))
sys/dev/pci/drm/scheduler/sched_main.c
234
list_del_init(&entity->list);
sys/dev/pci/drm/scheduler/sched_main.c
236
if (rq->current_entity == entity)
sys/dev/pci/drm/scheduler/sched_main.c
240
drm_sched_rq_remove_fifo_locked(entity, rq);
sys/dev/pci/drm/scheduler/sched_main.c
261
struct drm_sched_entity *entity;
sys/dev/pci/drm/scheduler/sched_main.c
265
entity = rq->current_entity;
sys/dev/pci/drm/scheduler/sched_main.c
266
if (entity) {
sys/dev/pci/drm/scheduler/sched_main.c
267
list_for_each_entry_continue(entity, &rq->entities, list) {
sys/dev/pci/drm/scheduler/sched_main.c
268
if (drm_sched_entity_is_ready(entity))
sys/dev/pci/drm/scheduler/sched_main.c
273
list_for_each_entry(entity, &rq->entities, list) {
sys/dev/pci/drm/scheduler/sched_main.c
274
if (drm_sched_entity_is_ready(entity))
sys/dev/pci/drm/scheduler/sched_main.c
277
if (entity == rq->current_entity)
sys/dev/pci/drm/scheduler/sched_main.c
286
if (!drm_sched_can_queue(sched, entity)) {
sys/dev/pci/drm/scheduler/sched_main.c
291
entity = ERR_PTR(-ENOSPC);
sys/dev/pci/drm/scheduler/sched_main.c
293
rq->current_entity = entity;
sys/dev/pci/drm/scheduler/sched_main.c
294
reinit_completion(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_main.c
299
return entity;
sys/dev/pci/drm/scheduler/sched_main.c
322
struct drm_sched_entity *entity;
sys/dev/pci/drm/scheduler/sched_main.c
324
entity = rb_entry(rb, struct drm_sched_entity, rb_tree_node);
sys/dev/pci/drm/scheduler/sched_main.c
325
if (drm_sched_entity_is_ready(entity)) {
sys/dev/pci/drm/scheduler/sched_main.c
329
if (!drm_sched_can_queue(sched, entity)) {
sys/dev/pci/drm/scheduler/sched_main.c
334
reinit_completion(&entity->entity_idle);
sys/dev/pci/drm/scheduler/sched_main.c
805
struct drm_sched_entity *entity,
sys/dev/pci/drm/scheduler/sched_main.c
809
if (!entity->rq) {
sys/dev/pci/drm/scheduler/sched_main.c
831
job->entity = entity;
sys/dev/pci/drm/scheduler/sched_main.c
833
job->s_fence = drm_sched_fence_alloc(entity, owner, drm_client_id);
sys/dev/pci/drm/scheduler/sched_main.c
865
struct drm_sched_entity *entity = job->entity;
sys/dev/pci/drm/scheduler/sched_main.c
867
BUG_ON(!entity);
sys/dev/pci/drm/scheduler/sched_main.c
868
drm_sched_entity_select_rq(entity);
sys/dev/pci/drm/scheduler/sched_main.c
869
sched = entity->rq->sched;
sys/dev/pci/drm/scheduler/sched_main.c
872
job->s_priority = entity->priority;
sys/dev/pci/drm/scheduler/sched_main.c
874
drm_sched_fence_init(job->s_fence, job->entity);
usr.sbin/rpki-client/extern.h
579
TAILQ_ENTRY(entity) entries;
usr.sbin/rpki-client/extern.h
591
TAILQ_HEAD(entityq, entity);
usr.sbin/rpki-client/extern.h
854
void entity_free(struct entity *);
usr.sbin/rpki-client/extern.h
855
void entity_read_req(struct ibuf *, struct entity *);
usr.sbin/rpki-client/extern.h
892
int repo_queued(struct repo *, struct entity *);
usr.sbin/rpki-client/filemode.c
742
struct entity *entp;
usr.sbin/rpki-client/filemode.c
791
struct entity *entp;
usr.sbin/rpki-client/filemode.c
838
entp = calloc(1, sizeof(struct entity));
usr.sbin/rpki-client/main.c
135
entity_free(struct entity *ent)
usr.sbin/rpki-client/main.c
154
entity_read_req(struct ibuf *b, struct entity *ent)
usr.sbin/rpki-client/main.c
178
entity_write_req(const struct entity *ent)
usr.sbin/rpki-client/main.c
230
struct entity *p, *np;
usr.sbin/rpki-client/main.c
249
struct entity *p;
usr.sbin/rpki-client/main.c
251
if ((p = calloc(1, sizeof(struct entity))) == NULL)
usr.sbin/rpki-client/parser.c
1028
struct entity *entp;
usr.sbin/rpki-client/parser.c
1187
struct entity *entp;
usr.sbin/rpki-client/parser.c
1242
entp = calloc(1, sizeof(struct entity));
usr.sbin/rpki-client/parser.c
186
const struct entity *entp, X509_STORE_CTX *ctx)
usr.sbin/rpki-client/parser.c
227
const struct entity *entp, X509_STORE_CTX *ctx)
usr.sbin/rpki-client/parser.c
312
parse_load_crl_from_mft(struct entity *entp, struct mft *mft, enum location loc,
usr.sbin/rpki-client/parser.c
367
proc_parser_mft_pre(struct entity *entp, char *file, struct crl **crl,
usr.sbin/rpki-client/parser.c
499
proc_parser_mft(struct entity *entp, struct mft **mp, char **crlfile,
usr.sbin/rpki-client/parser.c
579
const struct entity *entp, X509_STORE_CTX *ctx)
usr.sbin/rpki-client/parser.c
684
proc_parser_root_cert(struct entity *entp, struct cert **out_cert)
usr.sbin/rpki-client/parser.c
742
const struct entity *entp, X509_STORE_CTX *ctx)
usr.sbin/rpki-client/parser.c
782
const struct entity *entp, X509_STORE_CTX *ctx)
usr.sbin/rpki-client/parser.c
825
parse_load_file(struct entity *entp, unsigned char **f, size_t *flen)
usr.sbin/rpki-client/parser.c
848
struct entity *entp;
usr.sbin/rpki-client/repo.c
1424
repo_queued(struct repo *rp, struct entity *p)