Symbol: obj
games/adventure/hdr.h
70
extern int verb, obj, spk;
games/adventure/io.c
442
if ((obj = rnum()) < 0)
games/adventure/io.c
444
plac[obj] = rnum(); /* initial loc for this obj */
games/adventure/io.c
446
fixd[obj] = rnum();
games/adventure/io.c
448
fixd[obj] = 0;
games/adventure/main.c
182
obj = i;
games/adventure/main.c
183
if (obj > 100)
games/adventure/main.c
184
obj -= 100;
games/adventure/main.c
185
if (obj == steps && toting(nugget))
games/adventure/main.c
187
if (prop[obj] < 0) {
games/adventure/main.c
190
prop[obj] = 0;
games/adventure/main.c
191
if (obj == rug || obj == chain)
games/adventure/main.c
192
prop[obj] = 1;
games/adventure/main.c
198
ll = prop[obj]; /* 2006 */
games/adventure/main.c
199
if (obj == steps && loc == fixed[steps])
games/adventure/main.c
201
pspeak(obj, ll);
games/adventure/main.c
209
obj = 0;
games/adventure/main.c
341
obj = wd2[0];
games/adventure/main.c
342
if (obj != 0)
games/adventure/main.c
352
obj = atloc[loc];
games/adventure/main.c
359
obj = 0;
games/adventure/main.c
364
obj = clam;
games/adventure/main.c
366
obj = oyster;
games/adventure/main.c
368
obj = door;
games/adventure/main.c
370
obj = grate;
games/adventure/main.c
371
if (obj != 0 && here(chain))
games/adventure/main.c
374
obj = chain;
games/adventure/main.c
375
if (obj == 0)
games/adventure/main.c
454
obj = magzin;
games/adventure/main.c
456
obj = obj * 100 + tablet;
games/adventure/main.c
458
obj = obj * 100 + messag;
games/adventure/main.c
460
obj = oyster;
games/adventure/main.c
461
if (obj > 100 || obj == 0 || dark())
games/adventure/main.c
53
int verb, obj, spk;
games/adventure/main.c
537
if ((!toting(obj)) && (obj != rod || !toting(rod2)))
games/adventure/main.c
539
if (obj != rod || !at(fissur)||!toting(obj) || closng)
games/adventure/main.c
558
if (obj == bottle || obj == 0)
games/adventure/main.c
559
obj = liq();
games/adventure/main.c
560
if (obj == 0)
games/adventure/main.c
562
if (!toting(obj))
games/adventure/main.c
565
if (obj != oil && obj != water)
games/adventure/main.c
568
place[obj] = 0;
games/adventure/main.c
574
if (obj == oil)
games/adventure/main.c
580
if (obj != water)
games/adventure/main.c
588
if (obj == food)
games/adventure/main.c
590
if (obj == bird || obj == snake || obj == clam || obj == oyster
games/adventure/main.c
591
|| obj == dwarf || obj == dragon || obj == troll
games/adventure/main.c
592
|| obj == bear) spk = 71;
games/adventure/main.c
595
if (obj == 0 && liqloc(loc) != water && (liq() != water
games/adventure/main.c
598
if (obj != 0 && obj != water)
games/adventure/main.c
607
if (obj != lamp)
games/adventure/main.c
620
if (at(obj) || (liq() == obj && at(bottle))
games/adventure/main.c
624
if (dloc[i] == loc && dflag >= 2 && obj == dwarf)
games/adventure/main.c
628
if (toting(obj))
games/adventure/main.c
656
if (obj == magzin)
games/adventure/main.c
658
if (obj == tablet)
games/adventure/main.c
660
if (obj == messag)
games/adventure/main.c
662
if (obj == oyster && hinted[2] && toting(oyster))
games/adventure/main.c
664
if (obj != oyster || hinted[2] || !toting(oyster)
games/adventure/main.c
669
if (obj == mirror)
games/adventure/main.c
671
if (obj == vase && prop[vase] == 0) {
games/adventure/main.c
679
if (obj != mirror||!closed)
games/adventure/main.c
684
if (obj != dwarf||!closed)
games/adventure/main.c
693
obj = k;
games/adventure/main.c
717
if (obj != plant || !at(plant2) || prop[plant2] == 0)
games/adventure/main.c
719
obj = plant2;
games/adventure/main.c
721
l5130: if (obj != knife || knfloc != loc)
games/adventure/main.c
726
l5140: if (obj != rod || !here(rod2))
games/adventure/main.c
728
obj = rod2;
games/adventure/save.c
84
{&obj, sizeof(obj)},
games/adventure/subr.c
502
if (here(bird) && toting(rod) && obj == bird)
games/adventure/subr.c
544
obj = 0;
games/adventure/subr.c
555
if (toting(obj))
games/adventure/subr.c
558
if (obj == plant && prop[plant] <= 0)
games/adventure/subr.c
560
if (obj == bear && prop[bear] == 1)
games/adventure/subr.c
562
if (obj == chain && prop[bear] != 0)
games/adventure/subr.c
564
if (fixed[obj] != 0)
games/adventure/subr.c
566
if (obj == water || obj == oil) {
games/adventure/subr.c
567
if (here(bottle) && liq() == obj) {
games/adventure/subr.c
568
obj = bottle;
games/adventure/subr.c
571
obj = bottle;
games/adventure/subr.c
584
if (obj == bird) {
games/adventure/subr.c
597
l9014: if ((obj == bird || obj == cage) && prop[bird] != 0)
games/adventure/subr.c
598
carry(bird + cage - obj, loc);
games/adventure/subr.c
599
carry(obj, loc);
games/adventure/subr.c
601
if (obj == bottle && k != 0)
games/adventure/subr.c
611
if (k == obj)
games/adventure/subr.c
612
obj = bottle;
games/adventure/subr.c
613
if (obj == bottle && k != 0)
games/adventure/subr.c
615
if (obj == cage && prop[bird] != 0)
games/adventure/subr.c
617
if (obj == bird)
games/adventure/subr.c
619
drop(obj, loc);
games/adventure/subr.c
626
if (toting(rod2) && obj == rod && !toting(rod))
games/adventure/subr.c
627
obj = rod2;
games/adventure/subr.c
628
if (!toting(obj))
games/adventure/subr.c
630
if (obj == bird && here(snake)) {
games/adventure/subr.c
638
if (obj == coins && here(vend)) { /* 9024 */
games/adventure/subr.c
644
if (obj == bird && at(dragon) && prop[dragon] == 0) { /* 9025 */
games/adventure/subr.c
652
if (obj == bear && at(troll)) { /* 9026 */
games/adventure/subr.c
662
if (obj != vase || loc == plac[pillow]) { /* 9027 */
games/adventure/subr.c
679
if (obj == clam || obj == oyster) {
games/adventure/subr.c
681
if (obj == oyster)
games/adventure/subr.c
684
if (toting(obj))
games/adventure/subr.c
697
if (obj == door)
games/adventure/subr.c
699
if (obj == door && prop[door] == 1)
games/adventure/subr.c
701
if (obj == cage)
games/adventure/subr.c
703
if (obj == keys)
games/adventure/subr.c
705
if (obj == grate || obj == chain)
games/adventure/subr.c
709
if (obj == chain) {
games/adventure/subr.c
764
if (obj == 0) { /* 9122 */
games/adventure/subr.c
766
obj = dwarf;
games/adventure/subr.c
768
obj = obj * 100 + snake;
games/adventure/subr.c
770
obj = obj * 100 + dragon;
games/adventure/subr.c
772
obj = obj * 100 + troll;
games/adventure/subr.c
774
obj = obj * 100 + bear;
games/adventure/subr.c
775
if (obj > 100)
games/adventure/subr.c
777
if (obj == 0) {
games/adventure/subr.c
779
obj = bird;
games/adventure/subr.c
781
obj = 100 * obj + clam;
games/adventure/subr.c
782
if (obj > 100)
games/adventure/subr.c
786
if (obj == bird) { /* 9124 */
games/adventure/subr.c
796
if (obj == 0)
games/adventure/subr.c
798
if (obj == clam || obj == oyster)
games/adventure/subr.c
800
if (obj == snake)
games/adventure/subr.c
802
if (obj == dwarf)
games/adventure/subr.c
804
if (obj == dwarf && closed)
games/adventure/subr.c
806
if (obj == dragon)
games/adventure/subr.c
808
if (obj == troll)
games/adventure/subr.c
810
if (obj == bear)
games/adventure/subr.c
812
if (obj != dragon || prop[dragon] != 0)
games/adventure/subr.c
816
obj = 0;
games/adventure/subr.c
828
for (obj = 1; obj <= 100; obj++)
games/adventure/subr.c
829
if (place[obj] == plac[dragon] || place[obj] == fixd[dragon])
games/adventure/subr.c
830
move(obj, k);
games/adventure/subr.c
842
if (toting(rod2) && obj == rod && !toting(rod))
games/adventure/subr.c
843
obj = rod2;
games/adventure/subr.c
844
if (!toting(obj))
games/adventure/subr.c
846
if (obj >= 50 && obj <= maxtrs && at(troll)) {
games/adventure/subr.c
848
drop(obj, 0);
games/adventure/subr.c
856
if (obj == food && here(bear)) {
games/adventure/subr.c
857
obj = bear; /* 9177 */
games/adventure/subr.c
860
if (obj != axe)
games/adventure/subr.c
895
obj = 0;
games/adventure/subr.c
903
if (obj == bird) {
games/adventure/subr.c
907
if (obj == snake || obj == dragon || obj == troll) {
games/adventure/subr.c
909
if (obj == dragon && prop[dragon] != 0)
games/adventure/subr.c
911
if (obj == troll)
games/adventure/subr.c
913
if (obj != snake || closed || !here(bird))
games/adventure/subr.c
921
if (obj == dwarf) {
games/adventure/subr.c
928
if (obj == bear) {
games/adventure/subr.c
950
if (obj == vase) {
games/adventure/subr.c
961
if (obj != 0 && obj != bottle)
games/adventure/subr.c
963
if (obj == 0 && !here(bottle))
games/battlestar/command1.c
80
SetBit(nightfile[p->room].objects, p->obj);
games/battlestar/command1.c
87
ClearBit(dayfile[p->room].objects, p->obj);
games/battlestar/command5.c
268
int obj = -1, result = -1, person = 0, firstnumber, last1, last2;
games/battlestar/command5.c
279
obj = wordvalue[wordnumber];
games/battlestar/command5.c
280
if (obj == EVERYTHING)
games/battlestar/command5.c
286
wordvalue[wordnumber] == obj) && wordnumber <= wordcount)
games/battlestar/command5.c
307
if (person == NORMGOD && godready < 2 && !(obj == RING || obj == BRACELET))
games/battlestar/command5.c
316
if (result != -1 && (TestBit(location[position].objects, obj) || obj == AMULET || obj == MEDALION || obj == TALISMAN)) {
games/battlestar/command5.c
317
ClearBit(location[position].objects, obj);
games/battlestar/command5.c
326
if (obj == RING || obj == BRACELET) {
games/battlestar/command5.c
332
if (obj == AMULET || obj == MEDALION || obj == TALISMAN) {
games/battlestar/command5.c
352
if (obj == COINS) {
games/battlestar/extern.h
311
short obj;
games/battlestar/init.c
67
SetBit(location[p->room].objects, p->obj);
games/hack/def.monst.h
101
struct obj *minvent;
games/hack/def.obj.h
105
extern struct obj *fobj;
games/hack/def.obj.h
107
#define newobj(xl) (struct obj *) alloc((unsigned)(xl) + sizeof(struct obj))
games/hack/def.obj.h
66
struct obj *nobj;
games/hack/def.objclass.h
81
#define ARM_BONUS(obj) ((10 - objects[obj->otyp].a_ac) + obj->spe)
games/hack/hack.Decl.c
80
struct obj *fobj = 0, *fcobj = 0, *invent = 0, *uwep = 0, *uarm = 0,
games/hack/hack.apply.c
103
use_magic_whistle(obj);
games/hack/hack.apply.c
108
use_whistle(obj);
games/hack/hack.apply.c
118
if(obj != uwep)
games/hack/hack.apply.c
133
use_camera(struct obj *obj)
games/hack/hack.apply.c
179
struct obj *current_ice_box; /* a local variable of use_ice_box, to be
games/hack/hack.apply.c
182
in_ice_box(struct obj *obj)
games/hack/hack.apply.c
184
if(obj == current_ice_box ||
games/hack/hack.apply.c
185
(Punished && (obj == uball || obj == uchain))){
games/hack/hack.apply.c
189
if(obj->owornmask & (W_ARMOR | W_RING)) {
games/hack/hack.apply.c
193
if(obj->owt + current_ice_box->owt > 70) {
games/hack/hack.apply.c
197
if(obj == uwep) {
games/hack/hack.apply.c
202
setuwep((struct obj *) 0);
games/hack/hack.apply.c
204
current_ice_box->owt += obj->owt;
games/hack/hack.apply.c
205
freeinv(obj);
games/hack/hack.apply.c
206
obj->o_cnt_id = current_ice_box->o_id;
games/hack/hack.apply.c
207
obj->nobj = fcobj;
games/hack/hack.apply.c
208
fcobj = obj;
games/hack/hack.apply.c
209
obj->age = moves - obj->age; /* actual age */
games/hack/hack.apply.c
214
ck_ice_box(struct obj *obj)
games/hack/hack.apply.c
216
return(obj->o_cnt_id == current_ice_box->o_id);
games/hack/hack.apply.c
220
out_ice_box(struct obj *obj)
games/hack/hack.apply.c
222
struct obj *otmp;
games/hack/hack.apply.c
224
if (obj == fcobj)
games/hack/hack.apply.c
227
for(otmp = fcobj; otmp->nobj != obj; otmp = otmp->nobj)
games/hack/hack.apply.c
229
otmp->nobj = obj->nobj;
games/hack/hack.apply.c
231
current_ice_box->owt -= obj->owt;
games/hack/hack.apply.c
232
obj->age = moves - obj->age; /* simulated point of time */
games/hack/hack.apply.c
233
(void) addinv(obj);
games/hack/hack.apply.c
238
use_ice_box(struct obj *obj)
games/hack/hack.apply.c
241
struct obj *otmp;
games/hack/hack.apply.c
242
current_ice_box = obj; /* for use by in/out_ice_box */
games/hack/hack.apply.c
244
if(otmp->o_cnt_id == obj->o_id)
games/hack/hack.apply.c
288
use_whistle(struct obj *obj)
games/hack/hack.apply.c
304
use_magic_whistle(struct obj *obj)
games/hack/hack.apply.c
358
struct obj *obj;
games/hack/hack.apply.c
361
if ((obj = sobj_at(ENORMOUS_ROCK, dpx, dpy))) {
games/hack/hack.apply.c
362
fracture_rock(obj);
games/hack/hack.apply.c
426
use_pick_axe(struct obj *obj)
games/hack/hack.apply.c
435
if(obj != uwep) {
games/hack/hack.apply.c
442
pline("You now wield %s.", doname(obj));
games/hack/hack.apply.c
443
setuwep(obj);
games/hack/hack.apply.c
479
aobjnam(obj, "clang"));
games/hack/hack.apply.c
484
aobjnam(obj, NULL));
games/hack/hack.apply.c
72
static void use_camera(struct obj *);
games/hack/hack.apply.c
73
static void use_ice_box(struct obj *);
games/hack/hack.apply.c
74
int in_ice_box(struct obj *);
games/hack/hack.apply.c
75
int ck_ice_box(struct obj *);
games/hack/hack.apply.c
76
int out_ice_box(struct obj *);
games/hack/hack.apply.c
78
static void use_whistle(struct obj *);
games/hack/hack.apply.c
79
static void use_magic_whistle(struct obj *);
games/hack/hack.apply.c
81
static int use_pick_axe(struct obj *);
games/hack/hack.apply.c
86
struct obj *obj;
games/hack/hack.apply.c
89
obj = getobj("(", "use or apply");
games/hack/hack.apply.c
90
if(!obj) return(0);
games/hack/hack.apply.c
92
switch(obj->otyp){
games/hack/hack.apply.c
94
use_camera(obj); break;
games/hack/hack.apply.c
96
use_ice_box(obj); break;
games/hack/hack.apply.c
98
res = use_pick_axe(obj);
games/hack/hack.bones.c
78
struct obj *otmp;
games/hack/hack.c
154
struct obj *otmp;
games/hack/hack.c
380
movobj(struct obj *obj, int ox, int oy)
games/hack/hack.c
383
freeobj(obj);
games/hack/hack.c
384
unpobj(obj);
games/hack/hack.c
385
obj->nobj = fobj;
games/hack/hack.c
386
fobj = obj;
games/hack/hack.c
387
obj->ox = ox;
games/hack/hack.c
388
obj->oy = oy;
games/hack/hack.c
410
struct obj *obj, *obj2;
games/hack/hack.c
428
for(obj = fobj; obj; obj = obj->nobj)
games/hack/hack.c
429
if(obj->ox == u.ux && obj->oy == u.uy)
games/hack/hack.c
430
if(!Punished || obj != uchain)
games/hack/hack.c
438
for(obj = fobj; obj; obj = obj2) {
games/hack/hack.c
439
obj2 = obj->nobj; /* perhaps obj will be picked up */
games/hack/hack.c
440
if(obj->ox == u.ux && obj->oy == u.uy) {
games/hack/hack.c
444
if(Punished && obj == uchain)
games/hack/hack.c
450
pline("Pick up %s ? [ynaq]", doname(obj));
games/hack/hack.c
458
if(obj->otyp == DEAD_COCKATRICE && !uarmg){
games/hack/hack.c
465
if(obj->otyp == SCR_SCARE_MONSTER){
games/hack/hack.c
466
if(!obj->spe) obj->spe = 1;
games/hack/hack.c
473
delobj(obj);
games/hack/hack.c
478
wt = inv_weight() + obj->owt;
games/hack/hack.c
480
if(obj->quan > 1) {
games/hack/hack.c
482
int savequan = obj->quan;
games/hack/hack.c
486
obj->quan = qq;
games/hack/hack.c
487
if(iw + weight(obj) > 0)
games/hack/hack.c
490
obj->quan = savequan;
games/hack/hack.c
496
doname(obj));
games/hack/hack.c
497
(void) splitobj(obj, qq);
games/hack/hack.c
507
(obj->quan == 1) ? "is" : "are",
games/hack/hack.c
508
doname(obj),
games/hack/hack.c
519
freeobj(obj);
games/hack/hack.c
521
addtobill(obj); /* sets obj->unpaid if necessary */
games/hack/hack.c
522
{ int pickquan = obj->quan;
games/hack/hack.c
524
if(!Blind) obj->dknown = 1; /* this is done by prinv(),
games/hack/hack.c
526
obj = addinv(obj); /* might merge it with other objects */
games/hack/hack.c
527
mergquan = obj->quan;
games/hack/hack.c
528
obj->quan = pickquan; /* to fool prinv() */
games/hack/hack.c
529
prinv(obj);
games/hack/hack.c
530
obj->quan = mergquan;
games/hack/hack.c
73
static void movobj(struct obj *, int, int);
games/hack/hack.c
858
struct obj *otmp = invent;
games/hack/hack.c
880
struct obj *otmp = invent;
games/hack/hack.do.c
101
if(obj->owornmask & (W_ARMOR | W_RING)){
games/hack/hack.do.c
105
if(obj == uwep) {
games/hack/hack.do.c
110
setuwep((struct obj *) 0);
games/hack/hack.do.c
112
pline("You dropped %s.", doname(obj));
games/hack/hack.do.c
113
dropx(obj);
games/hack/hack.do.c
119
dropx(struct obj *obj)
games/hack/hack.do.c
121
freeinv(obj);
games/hack/hack.do.c
122
dropy(obj);
games/hack/hack.do.c
126
dropy(struct obj *obj)
games/hack/hack.do.c
128
if(obj->otyp == CRYSKNIFE)
games/hack/hack.do.c
129
obj->otyp = WORM_TOOTH;
games/hack/hack.do.c
130
obj->ox = u.ux;
games/hack/hack.do.c
131
obj->oy = u.uy;
games/hack/hack.do.c
132
obj->nobj = fobj;
games/hack/hack.do.c
133
fobj = obj;
games/hack/hack.do.c
135
subfrombill(obj);
games/hack/hack.do.c
136
stackobj(obj);
games/hack/hack.do.c
324
struct obj *obj;
games/hack/hack.do.c
328
obj = getobj("#)", "throw"); /* it is also possible to throw food */
games/hack/hack.do.c
330
if(!obj || !getdir(1)) /* ask "in what direction?" */
games/hack/hack.do.c
332
if(obj->owornmask & (W_ARMOR | W_RING)){
games/hack/hack.do.c
339
if(obj == uwep){
games/hack/hack.do.c
340
if(obj->cursed){
games/hack/hack.do.c
344
if(obj->quan > 1)
games/hack/hack.do.c
345
setuwep(splitobj(obj, 1));
games/hack/hack.do.c
347
setuwep((struct obj *) 0);
games/hack/hack.do.c
349
else if(obj->quan > 1)
games/hack/hack.do.c
350
(void) splitobj(obj, 1);
games/hack/hack.do.c
351
freeinv(obj);
games/hack/hack.do.c
359
Doname(obj)); /* note: obj->quan == 1 */
games/hack/hack.do.c
360
if(obj->olet == POTION_SYM)
games/hack/hack.do.c
361
potionhit(&youmonst, obj);
games/hack/hack.do.c
364
losehp(uarmh ? 1 : rnd((int)(obj->owt)), "falling object");
games/hack/hack.do.c
365
dropy(obj);
games/hack/hack.do.c
368
pline("%s hits the floor.", Doname(obj));
games/hack/hack.do.c
369
if(obj->otyp == EXPENSIVE_CAMERA) {
games/hack/hack.do.c
371
obfree(obj, Null(obj));
games/hack/hack.do.c
372
} else if(obj->otyp == EGG) {
games/hack/hack.do.c
374
obfree(obj, Null(obj));
games/hack/hack.do.c
375
} else if(obj->olet == POTION_SYM) {
games/hack/hack.do.c
377
potionbreathe(obj);
games/hack/hack.do.c
378
obfree(obj, Null(obj));
games/hack/hack.do.c
380
dropy(obj);
games/hack/hack.do.c
384
} else if(obj->otyp == BOOMERANG) {
games/hack/hack.do.c
387
(void) addinv(obj);
games/hack/hack.do.c
391
if(obj->otyp == PICK_AXE && shkcatch(obj))
games/hack/hack.do.c
394
mon = bhit(u.dx, u.dy, (obj->otyp == ICE_BOX) ? 1 :
games/hack/hack.do.c
395
(!Punished || obj != uball) ? 8 : !u.ustuck ? 5 : 1,
games/hack/hack.do.c
396
obj->olet, NULL, NULL, obj);
games/hack/hack.do.c
402
if(obj->olet == WEAPON_SYM) {
games/hack/hack.do.c
404
if(obj->otyp < ROCK) {
games/hack/hack.do.c
406
uwep->otyp != obj->otyp+(BOW-ARROW))
games/hack/hack.do.c
412
if(obj->otyp == BOOMERANG) tmp += 4;
games/hack/hack.do.c
413
tmp += obj->spe;
games/hack/hack.do.c
415
if(hmon(mon,obj,1) == TRUE){
games/hack/hack.do.c
418
cutworm(mon,bhitpos.x,bhitpos.y,obj->otyp);
games/hack/hack.do.c
422
if(obj->otyp < BOOMERANG && rn2(3)) {
games/hack/hack.do.c
424
obfree(obj, (struct obj *) 0);
games/hack/hack.do.c
427
} else miss(objects[obj->otyp].oc_name, mon);
games/hack/hack.do.c
428
} else if(obj->otyp == HEAVY_IRON_BALL) {
games/hack/hack.do.c
430
if(!Punished || obj != uball) tmp += 2;
games/hack/hack.do.c
433
if(hmon(mon,obj,1) == FALSE)
games/hack/hack.do.c
436
} else if(obj->olet == POTION_SYM && u.ulevel > rn2(15)) {
games/hack/hack.do.c
437
potionhit(mon, obj);
games/hack/hack.do.c
443
if(obj->olet == FOOD_SYM && mon->data->mlet == 'd')
games/hack/hack.do.c
444
if(tamedog(mon,obj)) return(1);
games/hack/hack.do.c
445
if(obj->olet == GEM_SYM && mon->data->mlet == 'u' &&
games/hack/hack.do.c
447
if(obj->dknown && objects[obj->otyp].oc_name_known){
games/hack/hack.do.c
448
if(objects[obj->otyp].g_val > 0){
games/hack/hack.do.c
462
mpickobj(mon, obj);
games/hack/hack.do.c
470
if(obj->otyp == CRYSKNIFE)
games/hack/hack.do.c
471
obj->otyp = WORM_TOOTH;
games/hack/hack.do.c
472
obj->ox = bhitpos.x;
games/hack/hack.do.c
473
obj->oy = bhitpos.y;
games/hack/hack.do.c
474
obj->nobj = fobj;
games/hack/hack.do.c
475
fobj = obj;
games/hack/hack.do.c
478
stackobj(obj);
games/hack/hack.do.c
479
if(Punished && obj == uball &&
games/hack/hack.do.c
511
struct obj *
games/hack/hack.do.c
512
splitobj(struct obj *obj, int num)
games/hack/hack.do.c
514
struct obj *otmp;
games/hack/hack.do.c
517
*otmp = *obj; /* copies whole structure */
games/hack/hack.do.c
520
obj->quan = num;
games/hack/hack.do.c
521
obj->owt = weight(obj);
games/hack/hack.do.c
524
obj->nobj = otmp;
games/hack/hack.do.c
525
if(obj->unpaid) splitbill(obj,otmp);
games/hack/hack.do.c
75
static int drop(struct obj *);
games/hack/hack.do.c
84
drop(struct obj *obj)
games/hack/hack.do.c
86
if(!obj) return(0);
games/hack/hack.do.c
87
if(obj->olet == '$') { /* pseudo object */
games/hack/hack.do.c
88
long amount = OGOLD(obj);
games/hack/hack.do.c
98
free(obj);
games/hack/hack.do_name.c
169
do_oname(struct obj *obj)
games/hack/hack.do_name.c
171
struct obj *otmp, *otmp2;
games/hack/hack.do_name.c
175
pline("What do you want to name %s? ", doname(obj));
games/hack/hack.do_name.c
186
*otmp2 = *obj;
games/hack/hack.do_name.c
190
setworn((struct obj *) 0, obj->owornmask);
games/hack/hack.do_name.c
195
if(obj == invent) invent = otmp2;
games/hack/hack.do_name.c
199
if(otmp->nobj == obj){
games/hack/hack.do_name.c
205
free(obj); /* let us hope nobody else saved a pointer */
games/hack/hack.do_name.c
211
struct obj *obj;
games/hack/hack.do_name.c
218
obj = getobj("#", "name");
games/hack/hack.do_name.c
219
if(obj) do_oname(obj);
games/hack/hack.do_name.c
222
obj = getobj("?!=/", "call");
games/hack/hack.do_name.c
223
if(obj) docall(obj);
games/hack/hack.do_name.c
229
docall(struct obj *obj)
games/hack/hack.do_name.c
232
struct obj otemp;
games/hack/hack.do_name.c
237
otemp = *obj;
games/hack/hack.do_name.c
249
str1 = &(objects[obj->otyp].oc_uname);
games/hack/hack.do_name.c
70
static void do_oname(struct obj *);
games/hack/hack.do_wear.c
138
dorr(struct obj *otmp)
games/hack/hack.do_wear.c
147
cursed(struct obj *otmp)
games/hack/hack.do_wear.c
157
armoroff(struct obj *otmp)
games/hack/hack.do_wear.c
162
setworn((struct obj *) 0, otmp->owornmask & W_ARMOR);
games/hack/hack.do_wear.c
184
struct obj *otmp;
games/hack/hack.do_wear.c
232
setuwep((struct obj *) 0);
games/hack/hack.do_wear.c
245
struct obj *otmp;
games/hack/hack.do_wear.c
288
setuwep((struct obj *) 0);
games/hack/hack.do_wear.c
314
ringoff(struct obj *obj)
games/hack/hack.do_wear.c
318
mask = obj->owornmask & W_RING;
games/hack/hack.do_wear.c
319
setworn((struct obj *) 0, obj->owornmask);
games/hack/hack.do_wear.c
320
if(!(u.uprops[PROP(obj->otyp)].p_flgs & mask))
games/hack/hack.do_wear.c
322
u.uprops[PROP(obj->otyp)].p_flgs &= ~mask;
games/hack/hack.do_wear.c
323
switch(obj->otyp) {
games/hack/hack.do_wear.c
338
u.ustr -= obj->spe;
games/hack/hack.do_wear.c
339
u.ustrmax -= obj->spe;
games/hack/hack.do_wear.c
345
u.udaminc -= obj->spe;
games/hack/hack.do_wear.c
371
struct obj *otmp;
games/hack/hack.do_wear.c
379
if((otmp = uleft) != Null(obj)){
games/hack/hack.do_wear.c
383
if((otmp = uright) != Null(obj)){
games/hack/hack.do_wear.c
388
if((otmp = uwep) != Null(obj)){
games/hack/hack.do_wear.c
390
setuwep((struct obj *) 0);
games/hack/hack.do_wear.c
397
struct obj *
games/hack/hack.do_wear.c
400
struct obj *otmph = uarm;
games/hack/hack.do_wear.c
411
struct obj *otmph = some_armor();
games/hack/hack.do_wear.c
69
static void off_msg(struct obj *);
games/hack/hack.do_wear.c
70
static int dorr(struct obj *);
games/hack/hack.do_wear.c
71
static int cursed(struct obj *);
games/hack/hack.do_wear.c
74
off_msg(struct obj *otmp)
games/hack/hack.do_wear.c
82
struct obj *otmp;
games/hack/hack.dog.c
162
dogfood(struct obj *obj)
games/hack/hack.dog.c
164
switch(obj->olet) {
games/hack/hack.dog.c
167
(obj->otyp == TRIPE_RATION) ? DOGFOOD :
games/hack/hack.dog.c
168
(obj->otyp < CARROT) ? ACCFOOD :
games/hack/hack.dog.c
169
(obj->otyp < CORPSE) ? MANFOOD :
games/hack/hack.dog.c
170
(poisonous(obj) || obj->age + 50 <= moves ||
games/hack/hack.dog.c
171
obj->otyp == DEAD_COCKATRICE)
games/hack/hack.dog.c
175
if(!obj->cursed) return(APPORT);
games/hack/hack.dog.c
193
struct obj *obj;
games/hack/hack.dog.c
242
if ((obj = o_at(omx,omy)))
games/hack/hack.dog.c
243
if(!strchr("0_", obj->olet)){
games/hack/hack.dog.c
244
if((otyp = dogfood(obj)) <= CADAVER){
games/hack/hack.dog.c
249
if (obj->owt < 10*mtmp->data->mlevel)
games/hack/hack.dog.c
252
freeobj(obj);
games/hack/hack.dog.c
253
unpobj(obj);
games/hack/hack.dog.c
256
mpickobj(mtmp,obj);
games/hack/hack.dog.c
264
for(obj = fobj; obj; obj = obj->nobj) {
games/hack/hack.dog.c
265
otyp = dogfood(obj);
games/hack/hack.dog.c
267
if(inroom(obj->ox,obj->oy) != dogroom) continue;
games/hack/hack.dog.c
269
(dogroom >= 0 || DDIST(obj->ox,obj->oy) < 10)) {
games/hack/hack.dog.c
271
DDIST(obj->ox,obj->oy) < DDIST(gx,gy))){
games/hack/hack.dog.c
272
gx = obj->ox;
games/hack/hack.dog.c
273
gy = obj->oy;
games/hack/hack.dog.c
280
gx = obj->ox;
games/hack/hack.dog.c
281
gy = obj->oy;
games/hack/hack.dog.c
322
obj = invent;
games/hack/hack.dog.c
323
while(obj){
games/hack/hack.dog.c
324
if(obj->otyp == TRIPE_RATION){
games/hack/hack.dog.c
328
obj = obj->nobj;
games/hack/hack.dog.c
373
obj = fobj;
games/hack/hack.dog.c
374
while(obj){
games/hack/hack.dog.c
375
if(obj->ox != nx || obj->oy != ny)
games/hack/hack.dog.c
377
if(obj->cursed) goto nxti;
games/hack/hack.dog.c
378
if(obj->olet == FOOD_SYM &&
games/hack/hack.dog.c
379
(otyp = dogfood(obj)) < MANFOOD &&
games/hack/hack.dog.c
388
moves + obj->quan * objects[obj->otyp].oc_delay;
games/hack/hack.dog.c
392
5*obj->quan * objects[obj->otyp].nutrition;
games/hack/hack.dog.c
395
pline("%s ate %s.", Monnam(mtmp), doname(obj));
games/hack/hack.dog.c
399
delobj(obj);
games/hack/hack.dog.c
403
obj = obj->nobj;
games/hack/hack.dog.c
461
tamedog(struct monst *mtmp, struct obj *obj)
games/hack/hack.dog.c
477
if(obj) {
games/hack/hack.dog.c
478
if(dogfood(obj) >= MANFOOD) return(0);
games/hack/hack.dog.c
481
objects[obj->otyp].oc_name);
games/hack/hack.dog.c
483
obfree(obj, (struct obj *) 0);
games/hack/hack.dog.c
71
static int dogfood(struct obj *);
games/hack/hack.eat.c
107
struct obj *tin;
games/hack/hack.eat.c
112
static int eatcorpse(struct obj *);
games/hack/hack.eat.c
166
struct obj *otmp;
games/hack/hack.eat.c
217
struct obj *obj;
games/hack/hack.eat.c
219
obj = splitobj(otmp, 1);
games/hack/hack.eat.c
220
if(otmp == uwep) setuwep(obj);
games/hack/hack.eat.c
441
poisonous(struct obj *otmp)
games/hack/hack.eat.c
448
eatcorpse(struct obj *otmp)
games/hack/hack.end.c
223
struct obj *otmp;
games/hack/hack.engrave.c
207
struct obj *otmp;
games/hack/hack.engrave.c
70
extern struct obj zeroobj;
games/hack/hack.fight.c
204
hmon(struct monst *mon, struct obj *obj, int thrown)
games/hack/hack.fight.c
209
if(!obj){
games/hack/hack.fight.c
216
} else if(obj->olet == WEAPON_SYM || obj->otyp == PICK_AXE) {
games/hack/hack.fight.c
217
if(obj == uwep && (obj->otyp > SPEAR || obj->otyp < BOOMERANG))
games/hack/hack.fight.c
221
tmp = rnd(objects[obj->otyp].wldam);
games/hack/hack.fight.c
222
if(obj->otyp == TWO_HANDED_SWORD) tmp += d(2,6);
games/hack/hack.fight.c
223
else if(obj->otyp == FLAIL) tmp += rnd(4);
games/hack/hack.fight.c
225
tmp = rnd(objects[obj->otyp].wsdam);
games/hack/hack.fight.c
227
tmp += obj->spe;
games/hack/hack.fight.c
228
if(!thrown && obj == uwep && obj->otyp == BOOMERANG
games/hack/hack.fight.c
232
freeinv(obj);
games/hack/hack.fight.c
233
setworn((struct obj *) 0, obj->owornmask);
games/hack/hack.fight.c
234
obfree(obj, (struct obj *) 0);
games/hack/hack.fight.c
238
if(mon->data->mlet == 'O' && obj->otyp == TWO_HANDED_SWORD &&
games/hack/hack.fight.c
239
!strcmp(ONAME(obj), "Orcrist"))
games/hack/hack.fight.c
241
} else switch(obj->otyp) {
games/hack/hack.fight.c
246
freeinv(obj);
games/hack/hack.fight.c
247
if(obj->owornmask)
games/hack/hack.fight.c
248
setworn((struct obj *) 0, obj->owornmask);
games/hack/hack.fight.c
249
obfree(obj, (struct obj *) 0);
games/hack/hack.fight.c
270
tmp = obj->owt/10;
games/hack/hack.fight.c
299
hit( xname(obj) /* or: objects[obj->otyp].oc_name */,
games/hack/hack.fight.c
356
struct obj *obj;
games/hack/hack.fight.c
359
if((obj = o_at(mtmp->mx,mtmp->my)) && !Blind)
games/hack/hack.fight.c
361
mdat->mname, doname(obj));
games/hack/hack.h
104
extern struct obj *invent, *uwep, *uarm, *uarm2, *uarmh, *uarms, *uarmg,
games/hack/hack.h
106
extern struct obj *uchain; /* defined iff PUNISHED */
games/hack/hack.h
107
extern struct obj *uball; /* defined if PUNISHED */
games/hack/hack.h
265
void dropx(struct obj *);
games/hack/hack.h
266
void dropy(struct obj *);
games/hack/hack.h
274
struct obj *splitobj(struct obj *, int);
games/hack/hack.h
283
void docall(struct obj *);
games/hack/hack.h
294
int armoroff(struct obj *);
games/hack/hack.h
297
void ringoff(struct obj *);
games/hack/hack.h
300
struct obj *some_armor(void);
games/hack/hack.h
310
int tamedog(struct monst *, struct obj *);
games/hack/hack.h
321
int poisonous(struct obj *);
games/hack/hack.h
351
boolean hmon(struct monst *, struct obj *, int);
games/hack/hack.h
355
struct obj *addinv(struct obj *);
games/hack/hack.h
356
void useup(struct obj *);
games/hack/hack.h
357
void freeinv(struct obj *);
games/hack/hack.h
358
void delobj(struct obj *);
games/hack/hack.h
359
void freeobj(struct obj *);
games/hack/hack.h
363
struct obj *o_at(int, int);
games/hack/hack.h
364
struct obj *sobj_at(int, int, int);
games/hack/hack.h
365
int carried(struct obj *);
games/hack/hack.h
367
struct obj *o_on(unsigned int, struct obj *);
games/hack/hack.h
370
struct obj *mkgoldobj(long);
games/hack/hack.h
371
struct obj *getobj(char *, char *);
games/hack/hack.h
372
int ckunpaid(struct obj *);
games/hack/hack.h
373
int ggetobj(char *, int (*fn)(struct obj *), int);
games/hack/hack.h
374
int askchain(struct obj *, char *, int, int (*fn)(struct obj *),
games/hack/hack.h
375
int (*ckfn)(struct obj *), int);
games/hack/hack.h
376
void prinv(struct obj *);
games/hack/hack.h
380
void stackobj(struct obj *);
games/hack/hack.h
397
void saveobjchn(int, struct obj *);
games/hack/hack.h
434
struct obj * mkobj_at(int, int, int);
games/hack/hack.h
436
struct obj *mkobj(int);
games/hack/hack.h
437
struct obj *mksobj(int);
games/hack/hack.h
439
int weight(struct obj *);
games/hack/hack.h
484
char *xname(struct obj *);
games/hack/hack.h
485
char *doname(struct obj *);
games/hack/hack.h
487
char *aobjnam(struct obj *, char *);
games/hack/hack.h
488
char *Doname(struct obj *);
games/hack/hack.h
489
struct obj *readobjnam(char *, size_t);
games/hack/hack.h
517
void strange_feeling(struct obj *, char *);
games/hack/hack.h
518
void potionhit(struct monst *, struct obj *);
games/hack/hack.h
519
void potionbreathe(struct obj *);
games/hack/hack.h
547
void unpobj(struct obj *);
games/hack/hack.h
558
int identify(struct obj *);
games/hack/hack.h
571
struct obj *restobjchn(int);
games/hack/hack.h
586
void obfree(struct obj *, struct obj *);
games/hack/hack.h
589
void addtobill(struct obj *);
games/hack/hack.h
590
void splitbill(struct obj *, struct obj *);
games/hack/hack.h
591
void subfrombill(struct obj *);
games/hack/hack.h
593
int shkcatch(struct obj *);
games/hack/hack.h
607
void mpickobj(struct monst *, struct obj *);
games/hack/hack.h
702
void setuwep(struct obj *);
games/hack/hack.h
705
int chwepon(struct obj *, int);
games/hack/hack.h
727
void setworn(struct obj *, long);
games/hack/hack.h
728
void setnotworn(struct obj *);
games/hack/hack.h
731
void bhitm(struct monst *, struct obj *);
games/hack/hack.h
732
boolean bhito(struct obj *, struct obj *);
games/hack/hack.h
738
void (*fhitm)(struct monst *, struct obj *),
games/hack/hack.h
739
boolean (*fhito)(struct obj *, struct obj *), struct obj *);
games/hack/hack.h
742
void fracture_rock(struct obj *);
games/hack/hack.invent.c
115
struct obj *
games/hack/hack.invent.c
116
addinv(struct obj *obj)
games/hack/hack.invent.c
118
struct obj *otmp;
games/hack/hack.invent.c
122
invent = obj;
games/hack/hack.invent.c
126
if(merged(otmp, obj, 0))
games/hack/hack.invent.c
129
otmp->nobj = obj;
games/hack/hack.invent.c
133
obj->nobj = 0;
games/hack/hack.invent.c
136
assigninvlet(obj);
games/hack/hack.invent.c
144
if((invent->invlet ^ 040) > (obj->invlet ^ 040)) {
games/hack/hack.invent.c
145
obj->nobj = invent;
games/hack/hack.invent.c
146
invent = obj;
games/hack/hack.invent.c
150
(otmp->nobj->invlet ^ 040) > (obj->invlet ^ 040)){
games/hack/hack.invent.c
151
obj->nobj = otmp->nobj;
games/hack/hack.invent.c
152
otmp->nobj = obj;
games/hack/hack.invent.c
159
return(obj);
games/hack/hack.invent.c
163
useup(struct obj *obj)
games/hack/hack.invent.c
165
if(obj->quan > 1){
games/hack/hack.invent.c
166
obj->quan--;
games/hack/hack.invent.c
167
obj->owt = weight(obj);
games/hack/hack.invent.c
169
setnotworn(obj);
games/hack/hack.invent.c
170
freeinv(obj);
games/hack/hack.invent.c
171
obfree(obj, (struct obj *) 0);
games/hack/hack.invent.c
176
freeinv(struct obj *obj)
games/hack/hack.invent.c
178
struct obj *otmp;
games/hack/hack.invent.c
180
if(obj == invent)
games/hack/hack.invent.c
183
for(otmp = invent; otmp->nobj != obj; otmp = otmp->nobj)
games/hack/hack.invent.c
185
otmp->nobj = obj->nobj;
games/hack/hack.invent.c
191
delobj(struct obj *obj)
games/hack/hack.invent.c
193
freeobj(obj);
games/hack/hack.invent.c
194
unpobj(obj);
games/hack/hack.invent.c
195
obfree(obj, (struct obj *) 0);
games/hack/hack.invent.c
200
freeobj(struct obj *obj)
games/hack/hack.invent.c
202
struct obj *otmp;
games/hack/hack.invent.c
204
if(obj == fobj) fobj = fobj->nobj;
games/hack/hack.invent.c
206
for(otmp = fobj; otmp->nobj != obj; otmp = otmp->nobj)
games/hack/hack.invent.c
208
otmp->nobj = obj->nobj;
games/hack/hack.invent.c
268
struct obj *
games/hack/hack.invent.c
271
struct obj *otmp;
games/hack/hack.invent.c
278
struct obj *
games/hack/hack.invent.c
281
struct obj *otmp;
games/hack/hack.invent.c
290
carried(struct obj *obj)
games/hack/hack.invent.c
292
struct obj *otmp;
games/hack/hack.invent.c
295
if(otmp == obj) return(1);
games/hack/hack.invent.c
302
struct obj *otmp;
games/hack/hack.invent.c
310
struct obj *
games/hack/hack.invent.c
311
o_on(unsigned int id, struct obj *objchn)
games/hack/hack.invent.c
345
struct obj *
games/hack/hack.invent.c
348
struct obj *otmp;
games/hack/hack.invent.c
365
struct obj *
games/hack/hack.invent.c
368
struct obj *otmp;
games/hack/hack.invent.c
454
return((struct obj *)0);
games/hack/hack.invent.c
456
return(allownone ? &zeroobj : (struct obj *) 0);
games/hack/hack.invent.c
503
struct obj *obj;
games/hack/hack.invent.c
504
obj = splitobj(otmp, (int) cnt);
games/hack/hack.invent.c
505
if(otmp == uwep) setuwep(obj);
games/hack/hack.invent.c
512
ckunpaid(struct obj *otmp)
games/hack/hack.invent.c
520
ggetobj(char *word, int (*fn)(struct obj *), int max)
games/hack/hack.invent.c
528
int (*ckfn)(struct obj *) = NULL;
games/hack/hack.invent.c
535
struct obj *otmp = invent;
games/hack/hack.invent.c
597
askchain(struct obj *objchn, char *olets, int allflag, int (*fn)(struct obj *),
games/hack/hack.invent.c
598
int (*ckfn)(struct obj *), int max)
games/hack/hack.invent.c
600
struct obj *otmp, *otmp2;
games/hack/hack.invent.c
637
obj_to_let(struct obj *obj)
games/hack/hack.invent.c
639
struct obj *otmp;
games/hack/hack.invent.c
643
return(obj->invlet);
games/hack/hack.invent.c
645
for(otmp = invent; otmp && otmp != obj; otmp = otmp->nobj)
games/hack/hack.invent.c
651
prinv(struct obj *obj)
games/hack/hack.invent.c
653
pline("%s", xprname(obj, obj_to_let(obj)));
games/hack/hack.invent.c
657
xprname(struct obj *obj, char let)
games/hack/hack.invent.c
662
flags.invlet_constant ? obj->invlet : let,
games/hack/hack.invent.c
663
doname(obj));
games/hack/hack.invent.c
679
struct obj *otmp;
games/hack/hack.invent.c
70
extern struct obj zeroobj;
games/hack/hack.invent.c
712
struct obj *otmp;
games/hack/hack.invent.c
781
struct obj *otmp, *otmp0;
games/hack/hack.invent.c
82
static void assigninvlet(struct obj *);
games/hack/hack.invent.c
83
static char obj_to_let(struct obj *);
games/hack/hack.invent.c
838
stackobj(struct obj *obj)
games/hack/hack.invent.c
84
static char *xprname(struct obj *, char);
games/hack/hack.invent.c
840
struct obj *otmp = fobj;
games/hack/hack.invent.c
842
for(otmp = fobj; otmp; otmp = otmp->nobj) if(otmp != obj)
games/hack/hack.invent.c
843
if(otmp->ox == obj->ox && otmp->oy == obj->oy &&
games/hack/hack.invent.c
844
merged(obj,otmp,1))
games/hack/hack.invent.c
850
merged(struct obj *otmp, struct obj *obj, int lose)
games/hack/hack.invent.c
852
if(obj->otyp == otmp->otyp &&
games/hack/hack.invent.c
853
obj->unpaid == otmp->unpaid &&
games/hack/hack.invent.c
854
obj->spe == otmp->spe &&
games/hack/hack.invent.c
855
obj->dknown == otmp->dknown &&
games/hack/hack.invent.c
856
obj->cursed == otmp->cursed &&
games/hack/hack.invent.c
857
(strchr("%*?!", obj->olet) ||
games/hack/hack.invent.c
858
(obj->known == otmp->known &&
games/hack/hack.invent.c
859
(obj->olet == WEAPON_SYM && obj->otyp < BOOMERANG)))) {
games/hack/hack.invent.c
86
static int merged(struct obj *, struct obj *, int);
games/hack/hack.invent.c
860
otmp->quan += obj->quan;
games/hack/hack.invent.c
861
otmp->owt += obj->owt;
games/hack/hack.invent.c
862
if(lose) freeobj(obj);
games/hack/hack.invent.c
863
obfree(obj,otmp); /* free(obj), bill->otmp */
games/hack/hack.invent.c
90
assigninvlet(struct obj *otmp)
games/hack/hack.invent.c
94
struct obj *obj;
games/hack/hack.invent.c
97
for(obj = invent; obj; obj = obj->nobj) if(obj != otmp) {
games/hack/hack.invent.c
98
i = obj->invlet;
games/hack/hack.lev.c
139
saveobjchn(int fd, struct obj *otmp)
games/hack/hack.lev.c
141
struct obj *otmp2;
games/hack/hack.lev.c
149
bwrite(fd, otmp, xl + sizeof(struct obj));
games/hack/hack.lev.c
69
extern struct obj *billobjs;
games/hack/hack.mhitu.c
392
struct obj *obj;
games/hack/hack.mhitu.c
393
if ((obj = o_at(mtmp->mx,mtmp->my)))
games/hack/hack.mhitu.c
395
Xmonnam(mtmp), doname(obj));
games/hack/hack.mkobj.c
104
struct obj zeroobj;
games/hack/hack.mkobj.c
106
struct obj *
games/hack/hack.mkobj.c
109
struct obj *otmp;
games/hack/hack.mkobj.c
185
weight(struct obj *obj)
games/hack/hack.mkobj.c
187
int wt = objects[obj->otyp].oc_weight;
games/hack/hack.mkobj.c
189
return(wt ? wt*obj->quan : (obj->quan + 1)/2);
games/hack/hack.mkobj.c
68
struct obj *
games/hack/hack.mkobj.c
71
struct obj *otmp = mkobj(let);
games/hack/hack.mkobj.c
82
struct obj *otmp = mksobj(otyp);
games/hack/hack.mkobj.c
89
struct obj *
games/hack/hack.mkshop.c
71
extern struct obj *mkobj_at(int, int, int);
games/hack/hack.mon.c
418
struct obj *otmp;
games/hack/hack.mon.c
527
struct obj *otmp;
games/hack/hack.mon.c
825
struct obj *obj2 = mkobj_at(tmp,x,y);
games/hack/hack.objnam.c
168
xname(struct obj *obj)
games/hack/hack.objnam.c
171
int nn = objects[obj->otyp].oc_name_known;
games/hack/hack.objnam.c
172
char *an = objects[obj->otyp].oc_name;
games/hack/hack.objnam.c
173
char *dn = objects[obj->otyp].oc_descr;
games/hack/hack.objnam.c
174
char *un = objects[obj->otyp].oc_uname;
games/hack/hack.objnam.c
175
int pl = (obj->quan != 1);
games/hack/hack.objnam.c
178
if(!obj->dknown && !Blind) obj->dknown = 1; /* %% doesn't belong here */
games/hack/hack.objnam.c
179
switch(obj->olet) {
games/hack/hack.objnam.c
181
strlcpy(buf, (obj->spe < 0 && obj->known)
games/hack/hack.objnam.c
193
if(obj->otyp == DEAD_HOMUNCULUS && pl) {
games/hack/hack.objnam.c
201
if(obj->otyp == WORM_TOOTH && pl) {
games/hack/hack.objnam.c
206
if(obj->otyp == CRYSKNIFE && pl) {
games/hack/hack.objnam.c
219
(obj->owt > objects[obj->otyp].oc_weight) ? "very " : "");
games/hack/hack.objnam.c
222
if(nn || un || !obj->dknown) {
games/hack/hack.objnam.c
228
if(!obj->dknown) break;
games/hack/hack.objnam.c
247
if(!obj->dknown) break;
games/hack/hack.objnam.c
260
if(!obj->dknown)
games/hack/hack.objnam.c
270
if(!obj->dknown)
games/hack/hack.objnam.c
280
if(!obj->dknown) {
games/hack/hack.objnam.c
289
if(obj->otyp >= TURQUOISE && obj->otyp <= JADE)
games/hack/hack.objnam.c
294
obj->olet,obj->olet,obj->otyp,obj->spe);
games/hack/hack.objnam.c
321
if(obj->onamelth) {
games/hack/hack.objnam.c
323
strlcat(buf, ONAME(obj), len);
games/hack/hack.objnam.c
329
doname(struct obj *obj)
games/hack/hack.objnam.c
332
char *bp = xname(obj);
games/hack/hack.objnam.c
335
if(obj->quan != 1)
games/hack/hack.objnam.c
336
snprintf(prefix, sizeof prefix, "%u ", obj->quan);
games/hack/hack.objnam.c
339
switch(obj->olet) {
games/hack/hack.objnam.c
345
if(obj->owornmask & W_ARMOR)
games/hack/hack.objnam.c
349
if(obj->known) {
games/hack/hack.objnam.c
350
strlcat(prefix, sitoa(obj->spe), sizeof prefix);
games/hack/hack.objnam.c
355
if(obj->known) {
games/hack/hack.objnam.c
357
snprintf(p, bufr + sizeof bufr - p, " (%d)", obj->spe);
games/hack/hack.objnam.c
361
if(obj->owornmask & W_RINGR)
games/hack/hack.objnam.c
363
if(obj->owornmask & W_RINGL)
games/hack/hack.objnam.c
365
if(obj->known && (objects[obj->otyp].bits & SPEC)) {
games/hack/hack.objnam.c
366
strlcat(prefix, sitoa(obj->spe), sizeof prefix);
games/hack/hack.objnam.c
371
if(obj->owornmask & W_WEP)
games/hack/hack.objnam.c
373
if(obj->unpaid)
games/hack/hack.objnam.c
392
aobjnam(struct obj *otmp, char *verb)
games/hack/hack.objnam.c
418
Doname(struct obj *obj)
games/hack/hack.objnam.c
420
char *s = doname(obj);
games/hack/hack.objnam.c
429
struct obj *
games/hack/hack.objnam.c
611
{ struct obj *otmp;
games/hack/hack.potion.c
272
strange_feeling(struct obj *obj, char *txt)
games/hack/hack.potion.c
278
if(!objects[obj->otyp].oc_name_known && !objects[obj->otyp].oc_uname)
games/hack/hack.potion.c
279
docall(obj);
games/hack/hack.potion.c
280
useup(obj);
games/hack/hack.potion.c
288
potionhit(struct monst *mon, struct obj *obj)
games/hack/hack.potion.c
306
pline("The %s evaporates.", xname(obj));
games/hack/hack.potion.c
308
if(!isyou && !rn2(3)) switch(obj->otyp) {
games/hack/hack.potion.c
353
potionbreathe(obj);
games/hack/hack.potion.c
354
obfree(obj, Null(obj));
games/hack/hack.potion.c
358
potionbreathe(struct obj *obj)
games/hack/hack.potion.c
360
switch(obj->otyp) {
games/hack/hack.potion.c
419
struct obj *potion, *obj;
games/hack/hack.potion.c
421
if(!(obj = getobj("#", "dip")))
games/hack/hack.potion.c
426
if(obj->otyp == ARROW || obj->otyp == DART ||
games/hack/hack.potion.c
427
obj->otyp == CROSSBOW_BOLT) {
games/hack/hack.potion.c
430
if(obj->spe < 7) obj->spe++; /* %% */
games/hack/hack.potion.c
75
struct obj *otmp,*objs;
games/hack/hack.pri.c
378
struct obj *otmp;
games/hack/hack.pri.c
418
struct obj *otmp;
games/hack/hack.pri.c
561
unpobj(struct obj *obj)
games/hack/hack.pri.c
569
if(!vism_at(obj->ox,obj->oy))
games/hack/hack.pri.c
570
newsym(obj->ox,obj->oy);
games/hack/hack.pri.c
576
struct obj *obj, *obj2;
games/hack/hack.pri.c
578
for(obj = fobj; obj; obj = obj2) {
games/hack/hack.pri.c
579
obj2 = obj->nobj;
games/hack/hack.pri.c
580
if(obj->olet == FOOD_SYM && obj->otyp >= CORPSE
games/hack/hack.pri.c
581
&& obj->age + 250 < moves)
games/hack/hack.pri.c
582
delobj(obj);
games/hack/hack.pri.c
584
for(obj = invent; obj; obj = obj2) {
games/hack/hack.pri.c
585
obj2 = obj->nobj;
games/hack/hack.pri.c
586
if(obj->olet == FOOD_SYM && obj->otyp >= CORPSE
games/hack/hack.pri.c
587
&& obj->age + 250 < moves)
games/hack/hack.pri.c
588
useup(obj);
games/hack/hack.read.c
122
struct obj *otmp = some_armor();
games/hack/hack.read.c
184
{ struct obj *obj;
games/hack/hack.read.c
189
for(obj = invent; obj ; obj = obj->nobj)
games/hack/hack.read.c
190
if(obj->owornmask)
games/hack/hack.read.c
191
obj->cursed = confused;
games/hack/hack.read.c
199
uchain = uball = (struct obj *) 0;
games/hack/hack.read.c
344
struct obj *obj;
games/hack/hack.read.c
347
for(obj = fobj; obj; obj = obj->nobj)
games/hack/hack.read.c
348
if(obj->olet == FOOD_SYM) {
games/hack/hack.read.c
349
if(obj->ox == u.ux && obj->oy == u.uy) ctu++;
games/hack/hack.read.c
363
for(obj = fobj; obj; obj = obj->nobj)
games/hack/hack.read.c
364
if(obj->olet == foodsym)
games/hack/hack.read.c
365
at(obj->ox, obj->oy, FOOD_SYM);
games/hack/hack.read.c
501
identify(struct obj *otmp)
games/hack/hack.read.c
73
struct obj *scroll;
games/hack/hack.read.c
97
{ struct obj *otmp = some_armor();
games/hack/hack.save.c
161
struct obj *otmp;
games/hack/hack.save.c
237
struct obj *
games/hack/hack.save.c
240
struct obj *otmp, *otmp2;
games/hack/hack.save.c
241
struct obj *first = 0;
games/hack/hack.save.c
249
mread(fd, (char *) otmp, (unsigned) xl + sizeof(struct obj));
games/hack/hack.shk.c
1104
struct obj *obj, *obj2;
games/hack/hack.shk.c
1107
for(obj = invent; obj; obj = obj2) {
games/hack/hack.shk.c
1108
obj2 = obj->nobj;
games/hack/hack.shk.c
1109
if(obj->owornmask) continue;
games/hack/hack.shk.c
1110
freeinv(obj);
games/hack/hack.shk.c
1111
obj->nobj = shopkeeper->minvent;
games/hack/hack.shk.c
1112
shopkeeper->minvent = obj;
games/hack/hack.shk.c
1113
if(obj->unpaid)
games/hack/hack.shk.c
1114
subfrombill(obj);
games/hack/hack.shk.c
123
shkcatch(struct obj *ign)
games/hack/hack.shk.c
160
struct obj *billobjs; /* objects on bill with bp->useup */
games/hack/hack.shk.c
185
static struct bill_x *onbill(struct obj *);
games/hack/hack.shk.c
188
static struct obj *bp_to_obj(struct bill_x *);
games/hack/hack.shk.c
189
static int getprice(struct obj *);
games/hack/hack.shk.c
227
struct obj *obj;
games/hack/hack.shk.c
230
for(obj = invent; obj; obj = obj->nobj)
games/hack/hack.shk.c
231
obj->unpaid = 0;
games/hack/hack.shk.c
232
for(obj = fobj; obj; obj = obj->nobj)
games/hack/hack.shk.c
233
obj->unpaid = 0;
games/hack/hack.shk.c
234
for(obj = fcobj; obj; obj = obj->nobj)
games/hack/hack.shk.c
235
obj->unpaid = 0;
games/hack/hack.shk.c
237
for(obj = mtmp->minvent; obj; obj = obj->nobj)
games/hack/hack.shk.c
238
obj->unpaid = 0;
games/hack/hack.shk.c
240
for(obj = mtmp->minvent; obj; obj = obj->nobj)
games/hack/hack.shk.c
241
obj->unpaid = 0;
games/hack/hack.shk.c
242
while ((obj = billobjs)) {
games/hack/hack.shk.c
243
billobjs = obj->nobj;
games/hack/hack.shk.c
244
free(obj);
games/hack/hack.shk.c
385
onbill(struct obj *obj)
games/hack/hack.shk.c
391
if(bp->bo_id == obj->o_id) {
games/hack/hack.shk.c
392
if(!obj->unpaid) pline("onbill: paid obj on bill?");
games/hack/hack.shk.c
395
if(obj->unpaid) pline("onbill: unpaid obj not on bill?");
games/hack/hack.shk.c
401
obfree(struct obj *obj, struct obj *merge)
games/hack/hack.shk.c
403
struct bill_x *bp = onbill(obj);
games/hack/hack.shk.c
409
obj->unpaid = 0; /* only for doinvbill */
games/hack/hack.shk.c
410
obj->nobj = billobjs;
games/hack/hack.shk.c
411
billobjs = obj;
games/hack/hack.shk.c
426
free(obj);
games/hack/hack.shk.c
565
struct obj *obj;
games/hack/hack.shk.c
569
obj = bp_to_obj(bp);
games/hack/hack.shk.c
571
if(!obj) {
games/hack/hack.shk.c
577
if(!obj->unpaid && !bp->useup){
games/hack/hack.shk.c
581
obj->unpaid = 0;
games/hack/hack.shk.c
586
doname(obj));
games/hack/hack.shk.c
587
obj->unpaid = 1;
games/hack/hack.shk.c
592
doname(obj), ltmp, plur(ltmp));
games/hack/hack.shk.c
594
struct obj *otmp = billobjs;
games/hack/hack.shk.c
595
if(obj == billobjs)
games/hack/hack.shk.c
596
billobjs = obj->nobj;
games/hack/hack.shk.c
598
while(otmp && otmp->nobj != obj) otmp = otmp->nobj;
games/hack/hack.shk.c
599
if(otmp) otmp->nobj = obj->nobj;
games/hack/hack.shk.c
602
free(obj);
games/hack/hack.shk.c
629
static struct obj *
games/hack/hack.shk.c
632
struct obj *obj;
games/hack/hack.shk.c
637
obj = o_on(id, billobjs);
games/hack/hack.shk.c
638
else if(!(obj = o_on(id, invent)) &&
games/hack/hack.shk.c
639
!(obj = o_on(id, fobj)) &&
games/hack/hack.shk.c
640
!(obj = o_on(id, fcobj))) {
games/hack/hack.shk.c
642
if ((obj = o_on(id, mtmp->minvent)))
games/hack/hack.shk.c
645
if ((obj = o_on(id, mtmp->minvent)))
games/hack/hack.shk.c
648
return(obj);
games/hack/hack.shk.c
653
addtobill(struct obj *obj)
games/hack/hack.shk.c
660
onbill(obj) /* perhaps we threw it away earlier */
games/hack/hack.shk.c
667
bp->bo_id = obj->o_id;
games/hack/hack.shk.c
668
bp->bquan = obj->quan;
games/hack/hack.shk.c
670
bp->price = getprice(obj);
games/hack/hack.shk.c
672
obj->unpaid = 1;
games/hack/hack.shk.c
676
splitbill(struct obj *obj, struct obj *otmp)
games/hack/hack.shk.c
681
bp = onbill(obj);
games/hack/hack.shk.c
708
subfrombill(struct obj *obj)
games/hack/hack.shk.c
712
struct obj *otmp;
games/hack/hack.shk.c
718
if((bp = onbill(obj)) != 0){
games/hack/hack.shk.c
719
obj->unpaid = 0;
games/hack/hack.shk.c
72
struct obj *billobjs = 0;
games/hack/hack.shk.c
720
if(bp->bquan > obj->quan){
games/hack/hack.shk.c
722
*otmp = *obj;
games/hack/hack.shk.c
724
otmp->quan = (bp->bquan -= obj->quan);
games/hack/hack.shk.c
736
if(obj->unpaid){
games/hack/hack.shk.c
738
obj->unpaid = 0;
games/hack/hack.shk.c
746
((tmp = shtypes[rooms[ESHK(shopkeeper)->shoproom].rtype-8]) && tmp != obj->olet)
games/hack/hack.shk.c
747
|| strchr("_0", obj->olet)) {
games/hack/hack.shk.c
75
obfree(struct obj *obj, struct obj *merge)
games/hack/hack.shk.c
751
ltmp = getprice(obj) * obj->quan;
games/hack/hack.shk.c
767
Monnam(shopkeeper), doname(obj));
games/hack/hack.shk.c
769
pline("You sold %s and got %ld gold piece%s.", doname(obj), ltmp,
games/hack/hack.shk.c
77
free(obj);
games/hack/hack.shk.c
778
struct obj *obj;
games/hack/hack.shk.c
788
((obj = bp_to_obj(bp)) && obj->quan < bp->bquan))
games/hack/hack.shk.c
804
obj = bp_to_obj(bp);
games/hack/hack.shk.c
805
if(!obj) {
games/hack/hack.shk.c
809
if(bp->useup || bp->bquan > obj->quan) {
games/hack/hack.shk.c
812
oquan = obj->quan;
games/hack/hack.shk.c
816
obj->quan = uquan; /* cheat doname */
games/hack/hack.shk.c
817
(void) snprintf(buf, sizeof buf, "x - %s", doname(obj));
games/hack/hack.shk.c
818
obj->quan = oquan; /* restore value */
games/hack/hack.shk.c
839
getprice(struct obj *obj)
games/hack/hack.shk.c
843
switch(obj->olet){
games/hack/hack.shk.c
848
tmp = 10*rnd((obj->otyp == EXPENSIVE_CAMERA) ? 150 : 30);
games/hack/hack.shk.c
859
if(obj->otyp == SCR_MAIL)
games/hack/hack.shk.c
873
ac = ARM_BONUS(obj);
games/hack/hack.shk.c
879
if(obj->otyp < BOOMERANG)
games/hack/hack.shk.c
881
else if(obj->otyp == LONG_SWORD ||
games/hack/hack.shk.c
882
obj->otyp == TWO_HANDED_SWORD)
games/hack/hack.shk.c
902
struct obj *otmp = invent;
games/hack/hack.shk.c
91
addtobill(struct obj *ign)
games/hack/hack.shk.c
913
shkcatch(struct obj *obj)
games/hack/hack.shk.c
922
pline("%s nimbly catches the %s.", Monnam(shkp), xname(obj));
games/hack/hack.shk.c
923
obj->nobj = shkp->minvent;
games/hack/hack.shk.c
924
shkp->minvent = obj;
games/hack/hack.shk.c
946
struct obj *ib = 0;
games/hack/hack.shk.c
95
subfrombill(struct obj *ign)
games/hack/hack.shk.c
99
splitbill(struct obj *ign, struct obj *ign2)
games/hack/hack.steal.c
112
struct obj *otmp;
games/hack/hack.steal.c
138
struct obj *otmp;
games/hack/hack.steal.c
170
setworn((struct obj *) 0, otmp->owornmask & W_ARMOR);
games/hack/hack.steal.c
198
setuwep((struct obj *) 0);
games/hack/hack.steal.c
206
uchain = (struct obj *) 0;
games/hack/hack.steal.c
208
uball = (struct obj *) 0; /* superfluous */
games/hack/hack.steal.c
217
mpickobj(struct monst *mtmp, struct obj *otmp)
games/hack/hack.steal.c
226
struct obj *otmp;
games/hack/hack.steal.c
231
if(otmp == uwep) setuwep((struct obj *) 0);
games/hack/hack.steal.c
245
struct obj *otmp, *otmp2;
games/hack/hack.steal.c
257
mtmp->minvent = (struct obj *) 0;
games/hack/hack.trap.c
529
struct obj *obj;
games/hack/hack.trap.c
531
for(obj = invent; obj; obj = obj->nobj)
games/hack/hack.trap.c
532
if(obj->olet == SCROLL_SYM && rn2(12) > u.uluck)
games/hack/hack.trap.c
533
obj->otyp = SCR_BLANK_PAPER;
games/hack/hack.u_init.c
327
struct obj *obj;
games/hack/hack.u_init.c
330
obj = mkobj(trop->trolet);
games/hack/hack.u_init.c
331
obj->known = trop->trknown;
games/hack/hack.u_init.c
333
obj->cursed = 0;
games/hack/hack.u_init.c
334
if(obj->olet == WEAPON_SYM){
games/hack/hack.u_init.c
335
obj->quan = trop->trquan;
games/hack/hack.u_init.c
339
obj->spe = trop->trspe;
games/hack/hack.u_init.c
341
obj->otyp = trop->trotyp;
games/hack/hack.u_init.c
343
if(obj->otyp == WAN_WISHING) /* gitpyr!robert */
games/hack/hack.u_init.c
344
obj->otyp = WAN_DEATH;
games/hack/hack.u_init.c
345
obj->owt = weight(obj); /* defined after setting otyp+quan */
games/hack/hack.u_init.c
346
obj = addinv(obj);
games/hack/hack.u_init.c
347
if(obj->olet == ARMOR_SYM){
games/hack/hack.u_init.c
348
switch(obj->otyp){
games/hack/hack.u_init.c
350
if(!uarms) setworn(obj, W_ARMS);
games/hack/hack.u_init.c
353
if(!uarmh) setworn(obj, W_ARMH);
games/hack/hack.u_init.c
356
if(!uarmg) setworn(obj, W_ARMG);
games/hack/hack.u_init.c
360
setworn(obj, W_ARM);
games/hack/hack.u_init.c
363
if(!uarm) setworn(obj, W_ARM);
games/hack/hack.u_init.c
366
if(obj->olet == WEAPON_SYM)
games/hack/hack.u_init.c
367
if(!uwep) setuwep(obj);
games/hack/hack.unix.c
341
struct obj *obj;
games/hack/hack.unix.c
344
extern struct obj *mksobj();
games/hack/hack.unix.c
348
obj = mksobj(SCR_MAIL);
games/hack/hack.unix.c
363
obj = addinv(obj);
games/hack/hack.unix.c
364
(void) identify(obj); /* set known and do prinv() */
games/hack/hack.wield.c
124
chwepon(struct obj *otmp, int amount)
games/hack/hack.wield.c
66
extern struct obj zeroobj;
games/hack/hack.wield.c
69
setuwep(struct obj *obj)
games/hack/hack.wield.c
71
setworn(obj, W_WEP);
games/hack/hack.wield.c
77
struct obj *wep;
games/hack/hack.wield.c
91
setuwep((struct obj *) 0);
games/hack/hack.wizard.c
115
struct obj *otmp;
games/hack/hack.wizard.c
83
struct obj *otmp;
games/hack/hack.worn.c
100
*(wp->w_obj) = obj;
games/hack/hack.worn.c
101
if(obj) obj->owornmask |= wp->w_mask;
games/hack/hack.worn.c
112
setnotworn(struct obj *obj)
games/hack/hack.worn.c
117
if(obj == *(wp->w_obj)) {
games/hack/hack.worn.c
119
obj->owornmask &= ~wp->w_mask;
games/hack/hack.worn.c
68
struct obj **w_obj;
games/hack/hack.worn.c
84
setworn(struct obj *obj, long mask)
games/hack/hack.worn.c
87
struct obj *oobj;
games/hack/hack.worn.c
94
if(obj && oobj && wp->w_mask == W_ARM){
games/hack/hack.zap.c
138
bhito(struct obj *obj, struct obj *otmp)
games/hack/hack.zap.c
142
if(obj == uball || obj == uchain)
games/hack/hack.zap.c
148
mkobj_at((obj->otyp == ROCK || obj->otyp == ENORMOUS_ROCK)
games/hack/hack.zap.c
149
? GEM_SYM : obj->olet,
games/hack/hack.zap.c
150
obj->ox, obj->oy) -> quan = obj->quan;
games/hack/hack.zap.c
151
delobj(obj);
games/hack/hack.zap.c
154
if(obj->otyp == ENORMOUS_ROCK)
games/hack/hack.zap.c
155
fracture_rock(obj);
games/hack/hack.zap.c
160
if(obj->spe && obj->olet != AMULET_SYM) {
games/hack/hack.zap.c
161
obj->known = 0;
games/hack/hack.zap.c
162
obj->spe = 0;
games/hack/hack.zap.c
166
rloco(obj);
games/hack/hack.zap.c
169
obj->oinvis = 1;
games/hack/hack.zap.c
172
res = revive(obj);
games/hack/hack.zap.c
190
struct obj *obj;
games/hack/hack.zap.c
193
obj = getobj("/", "zap");
games/hack/hack.zap.c
194
if(!obj) return(0);
games/hack/hack.zap.c
195
if(obj->spe < 0 || (obj->spe == 0 && rn2(121))) {
games/hack/hack.zap.c
199
if(obj->spe == 0)
games/hack/hack.zap.c
201
if(!(objects[obj->otyp].bits & NODIR) && !getdir(1))
games/hack/hack.zap.c
203
obj->spe--;
games/hack/hack.zap.c
204
if(objects[obj->otyp].bits & IMMEDIATE) {
games/hack/hack.zap.c
206
bhitm(u.ustuck, obj);
games/hack/hack.zap.c
209
struct obj *otmp = o_at(u.ux, u.uy);
games/hack/hack.zap.c
211
(void) bhito(otmp, obj);
games/hack/hack.zap.c
214
(void) bhit(u.dx,u.dy,rn1(8,6),0,bhitm,bhito,obj);
games/hack/hack.zap.c
216
switch(obj->otyp){
games/hack/hack.zap.c
232
struct obj *otmp;
games/hack/hack.zap.c
315
buzz((int) obj->otyp - WAN_MAGIC_MISSILE,
games/hack/hack.zap.c
319
if(!objects[obj->otyp].oc_name_known) {
games/hack/hack.zap.c
320
objects[obj->otyp].oc_name_known = 1;
games/hack/hack.zap.c
369
void (*fhitm)(struct monst *, struct obj *),
games/hack/hack.zap.c
370
boolean (*fhito)(struct obj *, struct obj *),
games/hack/hack.zap.c
371
struct obj *obj)
games/hack/hack.zap.c
374
struct obj *otmp;
games/hack/hack.zap.c
391
(*fhitm)(mtmp, obj);
games/hack/hack.zap.c
395
if(fhito && (*fhito)(otmp, obj))
games/hack/hack.zap.c
653
revive(struct obj *obj)
games/hack/hack.zap.c
657
if(obj->olet == FOOD_SYM && obj->otyp > CORPSE) {
games/hack/hack.zap.c
661
mtmp = mkmon_at(CORPSE_I_TO_C(obj->otyp),obj->ox,obj->oy);
games/hack/hack.zap.c
662
delobj(obj);
games/hack/hack.zap.c
668
rloco(struct obj *obj)
games/hack/hack.zap.c
672
otx = obj->ox;
games/hack/hack.zap.c
673
oty = obj->oy;
games/hack/hack.zap.c
678
obj->ox = tx;
games/hack/hack.zap.c
679
obj->oy = ty;
games/hack/hack.zap.c
686
fracture_rock(struct obj *obj)
games/hack/hack.zap.c
689
obj->otyp = ROCK;
games/hack/hack.zap.c
690
obj->quan = 7 + rn2(60);
games/hack/hack.zap.c
691
obj->owt = weight(obj);
games/hack/hack.zap.c
692
obj->olet = WEAPON_SYM;
games/hack/hack.zap.c
693
if(cansee(obj->ox,obj->oy))
games/hack/hack.zap.c
694
prl(obj->ox,obj->oy);
games/hack/hack.zap.c
700
struct obj *obj, *obj2;
games/hack/hack.zap.c
703
for(obj = invent; obj; obj = obj2) {
games/hack/hack.zap.c
704
obj2 = obj->nobj;
games/hack/hack.zap.c
705
if(obj->olet == SCROLL_SYM) {
games/hack/hack.zap.c
707
useup(obj);
games/hack/hack.zap.c
80
static boolean revive(struct obj *);
games/hack/hack.zap.c
81
static void rloco(struct obj *);
games/hack/hack.zap.c
87
bhitm(struct monst *mtmp, struct obj *otmp)
games/hunt/huntd/execute.c
595
pickup(PLAYER *pp, int y, int x, int prob, int obj)
games/hunt/huntd/execute.c
600
switch (obj) {
games/hunt/huntd/execute.c
617
add_shot(obj, y, x, LEFTS, req, (PLAYER *) NULL,
lib/libcrypto/asn1/asn1.h
808
ASN1_STRING *ASN1_item_pack(void *obj, const ASN1_ITEM *it,
lib/libcrypto/asn1/asn1_item.c
175
ASN1_item_pack(void *obj, const ASN1_ITEM *it, ASN1_STRING **oct)
lib/libcrypto/asn1/asn1_item.c
190
if (!(octmp->length = ASN1_item_i2d(obj, &octmp->data, it))) {
lib/libcrypto/asn1/p5_pbev2.c
201
ASN1_OBJECT *obj;
lib/libcrypto/asn1/p5_pbev2.c
208
obj = OBJ_nid2obj(alg_nid);
lib/libcrypto/asn1/p5_pbev2.c
216
scheme->algorithm = obj;
lib/libcrypto/asn1/t_req.c
220
ASN1_OBJECT *obj;
lib/libcrypto/asn1/t_req.c
226
obj = X509_EXTENSION_get_object(ex);
lib/libcrypto/asn1/t_req.c
227
i2a_ASN1_OBJECT(bp, obj);
lib/libcrypto/asn1/x_x509a.c
219
X509_add1_trust_object(X509 *x, const ASN1_OBJECT *obj)
lib/libcrypto/asn1/x_x509a.c
225
if (!(objtmp = OBJ_dup(obj)))
lib/libcrypto/asn1/x_x509a.c
242
X509_add1_reject_object(X509 *x, const ASN1_OBJECT *obj)
lib/libcrypto/asn1/x_x509a.c
248
if (!(objtmp = OBJ_dup(obj)))
lib/libcrypto/cms/cms.h
287
int CMS_signed_get_attr_by_OBJ(const CMS_SignerInfo *si, const ASN1_OBJECT *obj,
lib/libcrypto/cms/cms.h
292
int CMS_signed_add1_attr_by_OBJ(CMS_SignerInfo *si, const ASN1_OBJECT *obj,
lib/libcrypto/cms/cms.h
305
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/cms/cms.h
310
const ASN1_OBJECT *obj, int type, const void *bytes, int len);
lib/libcrypto/cms/cms_att.c
110
CMS_signed_add1_attr_by_OBJ(CMS_SignerInfo *si, const ASN1_OBJECT *obj, int type,
lib/libcrypto/cms/cms_att.c
113
if (X509at_add1_attr_by_OBJ(&si->signedAttrs, obj, type, bytes, len))
lib/libcrypto/cms/cms_att.c
162
CMS_unsigned_get_attr_by_OBJ(const CMS_SignerInfo *si, const ASN1_OBJECT *obj,
lib/libcrypto/cms/cms_att.c
165
return X509at_get_attr_by_OBJ(si->unsignedAttrs, obj, lastpos);
lib/libcrypto/cms/cms_att.c
193
CMS_unsigned_add1_attr_by_OBJ(CMS_SignerInfo *si, const ASN1_OBJECT *obj,
lib/libcrypto/cms/cms_att.c
196
if (X509at_add1_attr_by_OBJ(&si->unsignedAttrs, obj, type, bytes, len))
lib/libcrypto/cms/cms_att.c
79
CMS_signed_get_attr_by_OBJ(const CMS_SignerInfo *si, const ASN1_OBJECT *obj,
lib/libcrypto/cms/cms_att.c
82
return X509at_get_attr_by_OBJ(si->signedAttrs, obj, lastpos);
lib/libcrypto/crypto.h
300
int CRYPTO_new_ex_data(int class_index, void *obj, CRYPTO_EX_DATA *ad);
lib/libcrypto/crypto.h
303
void CRYPTO_free_ex_data(int class_index, void *obj, CRYPTO_EX_DATA *ad);
lib/libcrypto/ec/ec_ameth.c
795
const ASN1_OBJECT *obj;
lib/libcrypto/ec/ec_ameth.c
810
X509_ALGOR_get0(&obj, &parameter_type, &parameter, alg);
lib/libcrypto/ec/ec_ameth.c
812
if ((nid = OBJ_obj2nid(obj)) == NID_undef)
lib/libcrypto/ec/ecx_methods.c
580
const ASN1_OBJECT *obj;
lib/libcrypto/ec/ecx_methods.c
586
X509_ALGOR_get0(&obj, &param_type, NULL, signatureAlgorithm);
lib/libcrypto/ec/ecx_methods.c
587
if (OBJ_obj2nid(obj) != NID_ED25519)
lib/libcrypto/objects/obj_dat.c
100
a = ca->obj;
lib/libcrypto/objects/obj_dat.c
135
a = ca->obj;
lib/libcrypto/objects/obj_dat.c
136
b = cb->obj;
lib/libcrypto/objects/obj_dat.c
163
a->obj->nid = 0;
lib/libcrypto/objects/obj_dat.c
164
a->obj->flags |= ASN1_OBJECT_FLAG_DYNAMIC |
lib/libcrypto/objects/obj_dat.c
172
a->obj->nid++;
lib/libcrypto/objects/obj_dat.c
178
if (--a->obj->nid == 0)
lib/libcrypto/objects/obj_dat.c
179
ASN1_OBJECT_free(a->obj);
lib/libcrypto/objects/obj_dat.c
213
OBJ_add_object(const ASN1_OBJECT *obj)
lib/libcrypto/objects/obj_dat.c
223
if (obj == NULL || obj->nid == NID_undef)
lib/libcrypto/objects/obj_dat.c
225
if ((o = OBJ_dup(obj)) == NULL)
lib/libcrypto/objects/obj_dat.c
229
if ((o->length != 0) && (obj->data != NULL))
lib/libcrypto/objects/obj_dat.c
242
ao[i]->obj = o;
lib/libcrypto/objects/obj_dat.c
280
.obj = &aobj,
lib/libcrypto/objects/obj_dat.c
285
return found->obj;
lib/libcrypto/objects/obj_dat.c
344
.obj = (ASN1_OBJECT *)aobj,
lib/libcrypto/objects/obj_dat.c
349
return found->obj->nid;
lib/libcrypto/objects/obj_dat.c
383
.obj = &aobj,
lib/libcrypto/objects/obj_dat.c
388
return found->obj->nid;
lib/libcrypto/objects/obj_dat.c
422
.obj = &aobj,
lib/libcrypto/objects/obj_dat.c
427
return found->obj->nid;
lib/libcrypto/objects/obj_dat.c
470
ASN1_OBJECT *obj;
lib/libcrypto/objects/obj_dat.c
473
obj = OBJ_txt2obj(s, 0);
lib/libcrypto/objects/obj_dat.c
474
nid = OBJ_obj2nid(obj);
lib/libcrypto/objects/obj_dat.c
475
ASN1_OBJECT_free(obj);
lib/libcrypto/objects/obj_dat.c
564
OBJ_length(const ASN1_OBJECT *obj)
lib/libcrypto/objects/obj_dat.c
566
if (obj == NULL)
lib/libcrypto/objects/obj_dat.c
569
if (obj->length < 0)
lib/libcrypto/objects/obj_dat.c
572
return obj->length;
lib/libcrypto/objects/obj_dat.c
577
OBJ_get0_data(const ASN1_OBJECT *obj)
lib/libcrypto/objects/obj_dat.c
579
if (obj == NULL)
lib/libcrypto/objects/obj_dat.c
582
return obj->data;
lib/libcrypto/objects/obj_dat.c
85
ASN1_OBJECT *obj;
lib/libcrypto/objects/objects.h
111
size_t OBJ_length(const ASN1_OBJECT *obj);
lib/libcrypto/objects/objects.h
112
const unsigned char *OBJ_get0_data(const ASN1_OBJECT *obj);
lib/libcrypto/ocsp/ocsp.h
280
int OCSP_REQUEST_get_ext_by_OBJ(OCSP_REQUEST *x, const ASN1_OBJECT *obj,
lib/libcrypto/ocsp/ocsp.h
293
int OCSP_ONEREQ_get_ext_by_OBJ(OCSP_ONEREQ *x, const ASN1_OBJECT *obj,
lib/libcrypto/ocsp/ocsp.h
305
int OCSP_BASICRESP_get_ext_by_OBJ(OCSP_BASICRESP *x, const ASN1_OBJECT *obj,
lib/libcrypto/ocsp/ocsp.h
321
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/ocsp/ocsp_ext.c
167
OCSP_ONEREQ_get_ext_by_OBJ(OCSP_ONEREQ *x, const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/ocsp/ocsp_ext.c
169
return X509v3_get_ext_by_OBJ(x->singleRequestExtensions, obj, lastpos);
lib/libcrypto/ocsp/ocsp_ext.c
236
OCSP_BASICRESP_get_ext_by_OBJ(OCSP_BASICRESP *x, const ASN1_OBJECT *obj,
lib/libcrypto/ocsp/ocsp_ext.c
240
obj, lastpos);
lib/libcrypto/ocsp/ocsp_ext.c
308
OCSP_SINGLERESP_get_ext_by_OBJ(OCSP_SINGLERESP *x, const ASN1_OBJECT *obj,
lib/libcrypto/ocsp/ocsp_ext.c
311
return X509v3_get_ext_by_OBJ(x->singleExtensions, obj, lastpos);
lib/libcrypto/ocsp/ocsp_ext.c
96
OCSP_REQUEST_get_ext_by_OBJ(OCSP_REQUEST *x, const ASN1_OBJECT *obj,
lib/libcrypto/ocsp/ocsp_ext.c
99
return X509v3_get_ext_by_OBJ(x->tbsRequest->requestExtensions, obj,
lib/libcrypto/pkcs12/p12_add.c
70
PKCS12_item_pack_safebag(void *obj, const ASN1_ITEM *it, int nid1, int nid2)
lib/libcrypto/pkcs12/p12_add.c
80
if (!ASN1_item_pack(obj, it, &bag->value.octet)) {
lib/libcrypto/pkcs12/p12_decr.c
160
void *obj, int zbuf)
lib/libcrypto/pkcs12/p12_decr.c
170
inlen = ASN1_item_i2d(obj, &in, it);
lib/libcrypto/pkcs12/pkcs12_local.h
149
const ASN1_ITEM *it, const char *pass, int passlen, void *obj, int zbuf);
lib/libcrypto/pkcs12/pkcs12_local.h
150
PKCS12_SAFEBAG *PKCS12_item_pack_safebag(void *obj, const ASN1_ITEM *it,
lib/libcrypto/pkcs7/pk7_lib.c
173
ASN1_OBJECT *obj;
lib/libcrypto/pkcs7/pk7_lib.c
176
obj=OBJ_nid2obj(type); /* will not fail */
lib/libcrypto/pkcs7/pk7_lib.c
180
p7->type = obj;
lib/libcrypto/pkcs7/pk7_lib.c
190
p7->type = obj;
lib/libcrypto/pkcs7/pk7_lib.c
195
p7->type = obj;
lib/libcrypto/pkcs7/pk7_lib.c
205
p7->type = obj;
lib/libcrypto/pkcs7/pk7_lib.c
214
p7->type = obj;
lib/libcrypto/pkcs7/pk7_lib.c
224
p7->type = obj;
lib/libcrypto/ts/ts.h
246
int TS_REQ_get_ext_by_OBJ(TS_REQ *a, const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/ts/ts.h
313
int TS_TST_INFO_get_ext_by_OBJ(TS_TST_INFO *a, const ASN1_OBJECT *obj,
lib/libcrypto/ts/ts.h
531
int TS_OBJ_print_bio(BIO *bio, const ASN1_OBJECT *obj);
lib/libcrypto/ts/ts_lib.c
100
TS_OBJ_print_bio(BIO *bio, const ASN1_OBJECT *obj)
lib/libcrypto/ts/ts_lib.c
104
int len = OBJ_obj2txt(obj_txt, sizeof(obj_txt), obj, 0);
lib/libcrypto/ts/ts_lib.c
118
ASN1_OBJECT *obj;
lib/libcrypto/ts/ts_lib.c
124
obj = X509_EXTENSION_get_object(ex);
lib/libcrypto/ts/ts_lib.c
125
i2a_ASN1_OBJECT(bio, obj);
lib/libcrypto/ts/ts_req_utils.c
242
TS_REQ_get_ext_by_OBJ(TS_REQ *a, const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/ts/ts_req_utils.c
244
return X509v3_get_ext_by_OBJ(a->extensions, obj, lastpos);
lib/libcrypto/ts/ts_rsp_utils.c
464
TS_TST_INFO_get_ext_by_OBJ(TS_TST_INFO *a, const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/ts/ts_rsp_utils.c
466
return X509v3_get_ext_by_OBJ(a->extensions, obj, lastpos);
lib/libcrypto/x509/x509.h
606
int X509_add1_trust_object(X509 *x, const ASN1_OBJECT *obj);
lib/libcrypto/x509/x509.h
607
int X509_add1_reject_object(X509 *x, const ASN1_OBJECT *obj);
lib/libcrypto/x509/x509.h
712
int X509_REQ_get_attr_by_OBJ(const X509_REQ *req, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509.h
718
const ASN1_OBJECT *obj, int type,
lib/libcrypto/x509/x509.h
788
const ASN1_OBJECT *obj, char *buf, int len);
lib/libcrypto/x509/x509.h
795
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/x509/x509.h
800
int X509_NAME_add_entry_by_OBJ(X509_NAME *name, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509.h
811
const ASN1_OBJECT *obj, int type,
lib/libcrypto/x509/x509.h
814
const ASN1_OBJECT *obj);
lib/libcrypto/x509/x509.h
825
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/x509/x509.h
835
int X509_get_ext_by_OBJ(const X509 *x, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509.h
849
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/x509/x509.h
864
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/x509/x509.h
879
const ASN1_OBJECT *obj, int crit, ASN1_OCTET_STRING *data);
lib/libcrypto/x509/x509.h
881
const ASN1_OBJECT *obj);
lib/libcrypto/x509/x509.h
892
const ASN1_OBJECT *obj, int atrtype, const void *data, int len);
lib/libcrypto/x509/x509.h
895
int X509_ATTRIBUTE_set1_object(X509_ATTRIBUTE *attr, const ASN1_OBJECT *obj);
lib/libcrypto/x509/x509_alt.c
583
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_alt.c
584
if (!(obj = OBJ_txt2obj(value, 0))) {
lib/libcrypto/x509/x509_alt.c
589
gen->d.rid = obj;
lib/libcrypto/x509/x509_att.c
139
X509at_add1_attr_by_OBJ(STACK_OF(X509_ATTRIBUTE) **x, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509_att.c
145
attr = X509_ATTRIBUTE_create_by_OBJ(NULL, obj, type, bytes, len);
lib/libcrypto/x509/x509_att.c
184
X509at_get0_data_by_OBJ(STACK_OF(X509_ATTRIBUTE) *x, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509_att.c
190
i = X509at_get_attr_by_OBJ(x, obj, lastpos);
lib/libcrypto/x509/x509_att.c
193
if ((lastpos <= -2) && (X509at_get_attr_by_OBJ(x, obj, i) != -1))
lib/libcrypto/x509/x509_att.c
205
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_att.c
208
obj = OBJ_nid2obj(nid);
lib/libcrypto/x509/x509_att.c
209
if (obj == NULL) {
lib/libcrypto/x509/x509_att.c
213
ret = X509_ATTRIBUTE_create_by_OBJ(attr, obj, atrtype, data, len);
lib/libcrypto/x509/x509_att.c
215
ASN1_OBJECT_free(obj);
lib/libcrypto/x509/x509_att.c
221
X509_ATTRIBUTE_create_by_OBJ(X509_ATTRIBUTE **attr, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509_att.c
234
if (!X509_ATTRIBUTE_set1_object(ret, obj))
lib/libcrypto/x509/x509_att.c
254
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_att.c
257
obj = OBJ_txt2obj(atrname, 0);
lib/libcrypto/x509/x509_att.c
258
if (obj == NULL) {
lib/libcrypto/x509/x509_att.c
263
nattr = X509_ATTRIBUTE_create_by_OBJ(attr, obj, type, bytes, len);
lib/libcrypto/x509/x509_att.c
264
ASN1_OBJECT_free(obj);
lib/libcrypto/x509/x509_att.c
270
X509_ATTRIBUTE_set1_object(X509_ATTRIBUTE *attr, const ASN1_OBJECT *obj)
lib/libcrypto/x509/x509_att.c
272
if ((attr == NULL) || (obj == NULL))
lib/libcrypto/x509/x509_att.c
275
attr->object = OBJ_dup(obj);
lib/libcrypto/x509/x509_att.c
74
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_att.c
76
obj = OBJ_nid2obj(nid);
lib/libcrypto/x509/x509_att.c
77
if (obj == NULL)
lib/libcrypto/x509/x509_att.c
79
return (X509at_get_attr_by_OBJ(x, obj, lastpos));
lib/libcrypto/x509/x509_att.c
84
const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/x509/x509_att.c
97
if (OBJ_cmp(ex->object, obj) == 0)
lib/libcrypto/x509/x509_conf.c
280
ASN1_OBJECT *obj = NULL;
lib/libcrypto/x509/x509_conf.c
284
if ((obj = OBJ_txt2obj(name, 0)) == NULL) {
lib/libcrypto/x509/x509_conf.c
314
ext = X509_EXTENSION_create_by_OBJ(NULL, obj, crit, oct);
lib/libcrypto/x509/x509_conf.c
317
ASN1_OBJECT_free(obj);
lib/libcrypto/x509/x509_ext.c
145
X509_get_ext_by_OBJ(const X509 *x, const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/x509/x509_ext.c
147
return X509v3_get_ext_by_OBJ(x->cert_info->extensions, obj, lastpos);
lib/libcrypto/x509/x509_ext.c
210
X509_REVOKED_get_ext_by_OBJ(const X509_REVOKED *x, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509_ext.c
213
return X509v3_get_ext_by_OBJ(x->extensions, obj, lastpos);
lib/libcrypto/x509/x509_ext.c
81
X509_CRL_get_ext_by_OBJ(const X509_CRL *x, const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/x509/x509_ext.c
83
return X509v3_get_ext_by_OBJ(x->crl->extensions, obj, lastpos);
lib/libcrypto/x509/x509_extku.c
173
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_extku.c
185
if ((obj = sk_ASN1_OBJECT_value(eku, i)) == NULL)
lib/libcrypto/x509/x509_extku.c
187
if (!i2t_ASN1_OBJECT(obj_tmp, sizeof obj_tmp, obj))
lib/libcrypto/x509/x509_local.h
429
const ASN1_OBJECT *obj, int lastpos);
lib/libcrypto/x509/x509_local.h
433
const ASN1_OBJECT *obj, int type, const unsigned char *bytes, int len);
lib/libcrypto/x509/x509_local.h
439
const ASN1_OBJECT *obj, int lastpos, int type);
lib/libcrypto/x509/x509_lu.c
174
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
176
if ((obj = calloc(1, sizeof(*obj))) == NULL) {
lib/libcrypto/x509/x509_lu.c
181
obj->type = X509_LU_NONE;
lib/libcrypto/x509/x509_lu.c
183
return obj;
lib/libcrypto/x509/x509_lu.c
207
x509_object_dup(const X509_OBJECT *obj)
lib/libcrypto/x509/x509_lu.c
216
copy->type = obj->type;
lib/libcrypto/x509/x509_lu.c
217
copy->data = obj->data;
lib/libcrypto/x509/x509_lu.c
282
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
284
if ((obj = X509_OBJECT_new()) == NULL)
lib/libcrypto/x509/x509_lu.c
286
if (!X509_STORE_CTX_get_by_subject(vs, type, name, obj)) {
lib/libcrypto/x509/x509_lu.c
287
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
291
return obj;
lib/libcrypto/x509/x509_lu.c
336
X509_STORE_add_object(X509_STORE *store, X509_OBJECT *obj)
lib/libcrypto/x509/x509_lu.c
342
if (X509_OBJECT_retrieve_match(store->objs, obj) != NULL) {
lib/libcrypto/x509/x509_lu.c
348
if (sk_X509_OBJECT_push(store->objs, obj) <= 0) {
lib/libcrypto/x509/x509_lu.c
353
obj = NULL;
lib/libcrypto/x509/x509_lu.c
358
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
366
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
371
if ((obj = X509_OBJECT_new()) == NULL)
lib/libcrypto/x509/x509_lu.c
375
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
379
obj->type = X509_LU_X509;
lib/libcrypto/x509/x509_lu.c
380
obj->data.x509 = x;
lib/libcrypto/x509/x509_lu.c
382
return X509_STORE_add_object(store, obj);
lib/libcrypto/x509/x509_lu.c
389
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
394
if ((obj = X509_OBJECT_new()) == NULL)
lib/libcrypto/x509/x509_lu.c
398
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
402
obj->type = X509_LU_CRL;
lib/libcrypto/x509/x509_lu.c
403
obj->data.crl = x;
lib/libcrypto/x509/x509_lu.c
405
return X509_STORE_add_object(store, obj);
lib/libcrypto/x509/x509_lu.c
516
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
529
obj = sk_X509_OBJECT_value(store->objs, idx);
lib/libcrypto/x509/x509_lu.c
531
x = obj->data.x509;
lib/libcrypto/x509/x509_lu.c
557
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
566
obj = X509_STORE_CTX_get_obj_by_subject(ctx, X509_LU_X509, name);
lib/libcrypto/x509/x509_lu.c
567
if (obj == NULL)
lib/libcrypto/x509/x509_lu.c
569
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
581
X509_OBJECT *obj = NULL;
lib/libcrypto/x509/x509_lu.c
588
obj = X509_STORE_CTX_get_obj_by_subject(ctx, X509_LU_CRL, name);
lib/libcrypto/x509/x509_lu.c
589
if (obj == NULL)
lib/libcrypto/x509/x509_lu.c
592
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
593
obj = NULL;
lib/libcrypto/x509/x509_lu.c
604
obj = sk_X509_OBJECT_value(store->objs, idx);
lib/libcrypto/x509/x509_lu.c
606
x = obj->data.crl;
lib/libcrypto/x509/x509_lu.c
630
X509_OBJECT *obj;
lib/libcrypto/x509/x509_lu.c
638
obj = sk_X509_OBJECT_value(h, i);
lib/libcrypto/x509/x509_lu.c
639
if (x509_object_cmp((const X509_OBJECT **)&obj,
lib/libcrypto/x509/x509_lu.c
643
if (!X509_cmp(obj->data.x509, x->data.x509))
lib/libcrypto/x509/x509_lu.c
644
return obj;
lib/libcrypto/x509/x509_lu.c
646
if (!X509_CRL_match(obj->data.crl, x->data.crl))
lib/libcrypto/x509/x509_lu.c
647
return obj;
lib/libcrypto/x509/x509_lu.c
649
return obj;
lib/libcrypto/x509/x509_lu.c
670
X509_OBJECT *obj, *pobj;
lib/libcrypto/x509/x509_lu.c
677
obj = X509_STORE_CTX_get_obj_by_subject(ctx, X509_LU_X509, xn);
lib/libcrypto/x509/x509_lu.c
678
if (obj == NULL)
lib/libcrypto/x509/x509_lu.c
681
if ((issuer = X509_OBJECT_get0_X509(obj)) == NULL) {
lib/libcrypto/x509/x509_lu.c
682
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
686
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
694
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
700
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_lu.c
701
obj = NULL;
lib/libcrypto/x509/x509_lu.c
757
X509_OBJECT *obj = NULL;
lib/libcrypto/x509/x509_lu.c
766
if ((obj = x509_object_dup(sk_X509_OBJECT_value(objs, i))) == NULL)
lib/libcrypto/x509/x509_lu.c
768
if (!sk_X509_OBJECT_push(copy, obj))
lib/libcrypto/x509/x509_lu.c
770
obj = NULL;
lib/libcrypto/x509/x509_lu.c
776
X509_OBJECT_free(obj);
lib/libcrypto/x509/x509_ocsp.c
81
static int i2r_object(const X509V3_EXT_METHOD *method, void *obj, BIO *out,
lib/libcrypto/x509/x509_policy.c
179
is_any_policy(const ASN1_OBJECT *obj)
lib/libcrypto/x509/x509_policy.c
181
return OBJ_obj2nid(obj) == NID_any_policy;
lib/libcrypto/x509/x509_prn.c
174
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_prn.c
179
obj = X509_EXTENSION_get_object(ex);
lib/libcrypto/x509/x509_prn.c
180
i2a_ASN1_OBJECT(bp, obj);
lib/libcrypto/x509/x509_req.c
251
X509_REQ_get_attr_by_OBJ(const X509_REQ *req, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509_req.c
254
return X509at_get_attr_by_OBJ(req->req_info->attributes, obj, lastpos);
lib/libcrypto/x509/x509_req.c
282
X509_REQ_add1_attr_by_OBJ(X509_REQ *req, const ASN1_OBJECT *obj, int type,
lib/libcrypto/x509/x509_req.c
285
if (X509at_add1_attr_by_OBJ(&req->req_info->attributes, obj,
lib/libcrypto/x509/x509_trs.c
104
obj = sk_ASN1_OBJECT_value(aux->trust, i);
lib/libcrypto/x509/x509_trs.c
105
nid = OBJ_obj2nid(obj);
lib/libcrypto/x509/x509_trs.c
90
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_trs.c
97
obj = sk_ASN1_OBJECT_value(aux->reject, i);
lib/libcrypto/x509/x509_trs.c
98
nid = OBJ_obj2nid(obj);
lib/libcrypto/x509/x509_v3.c
102
if (OBJ_cmp(ext->object, obj) == 0)
lib/libcrypto/x509/x509_v3.c
190
const ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_v3.c
192
if ((obj = OBJ_nid2obj(nid)) == NULL) {
lib/libcrypto/x509/x509_v3.c
197
return X509_EXTENSION_create_by_OBJ(out_ext, obj, critical, data);
lib/libcrypto/x509/x509_v3.c
202
X509_EXTENSION_create_by_OBJ(X509_EXTENSION **out_ext, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509_v3.c
214
if (!X509_EXTENSION_set_object(ext, obj))
lib/libcrypto/x509/x509_v3.c
235
X509_EXTENSION_set_object(X509_EXTENSION *ext, const ASN1_OBJECT *obj)
lib/libcrypto/x509/x509_v3.c
237
if (ext == NULL || obj == NULL)
lib/libcrypto/x509/x509_v3.c
241
return (ext->object = OBJ_dup(obj)) != NULL;
lib/libcrypto/x509/x509_v3.c
83
const ASN1_OBJECT *obj;
lib/libcrypto/x509/x509_v3.c
85
if ((obj = OBJ_nid2obj(nid)) == NULL)
lib/libcrypto/x509/x509_v3.c
88
return X509v3_get_ext_by_OBJ(exts, obj, lastpos);
lib/libcrypto/x509/x509_v3.c
94
const ASN1_OBJECT *obj, int lastpos)
lib/libcrypto/x509/x509_vpm.c
481
ASN1_OBJECT *obj = NULL;
lib/libcrypto/x509/x509_vpm.c
488
if ((obj = OBJ_dup(sk_ASN1_OBJECT_value(sk, i))) == NULL)
lib/libcrypto/x509/x509_vpm.c
490
if (sk_ASN1_OBJECT_push(objs, obj) <= 0)
lib/libcrypto/x509/x509_vpm.c
492
obj = NULL;
lib/libcrypto/x509/x509_vpm.c
499
ASN1_OBJECT_free(obj);
lib/libcrypto/x509/x509name.c
135
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509name.c
137
obj = OBJ_nid2obj(nid);
lib/libcrypto/x509/x509name.c
138
if (obj == NULL)
lib/libcrypto/x509/x509name.c
140
return (X509_NAME_get_index_by_OBJ(name, obj, lastpos));
lib/libcrypto/x509/x509name.c
146
X509_NAME_get_index_by_OBJ(const X509_NAME *name, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509name.c
161
if (OBJ_cmp(ne->object, obj) == 0)
lib/libcrypto/x509/x509name.c
219
X509_NAME_add_entry_by_OBJ(X509_NAME *name, const ASN1_OBJECT *obj, int type,
lib/libcrypto/x509/x509name.c
225
ne = X509_NAME_ENTRY_create_by_OBJ(NULL, obj, type, bytes, len);
lib/libcrypto/x509/x509name.c
329
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509name.c
332
obj = OBJ_txt2obj(field, 0);
lib/libcrypto/x509/x509name.c
333
if (obj == NULL) {
lib/libcrypto/x509/x509name.c
338
nentry = X509_NAME_ENTRY_create_by_OBJ(ne, obj, type, bytes, len);
lib/libcrypto/x509/x509name.c
339
ASN1_OBJECT_free(obj);
lib/libcrypto/x509/x509name.c
348
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509name.c
351
obj = OBJ_nid2obj(nid);
lib/libcrypto/x509/x509name.c
352
if (obj == NULL) {
lib/libcrypto/x509/x509name.c
356
nentry = X509_NAME_ENTRY_create_by_OBJ(ne, obj, type, bytes, len);
lib/libcrypto/x509/x509name.c
357
ASN1_OBJECT_free(obj);
lib/libcrypto/x509/x509name.c
363
X509_NAME_ENTRY_create_by_OBJ(X509_NAME_ENTRY **ne, const ASN1_OBJECT *obj,
lib/libcrypto/x509/x509name.c
374
if (!X509_NAME_ENTRY_set_object(ret, obj))
lib/libcrypto/x509/x509name.c
391
X509_NAME_ENTRY_set_object(X509_NAME_ENTRY *ne, const ASN1_OBJECT *obj)
lib/libcrypto/x509/x509name.c
393
if ((ne == NULL) || (obj == NULL)) {
lib/libcrypto/x509/x509name.c
398
ne->object = OBJ_dup(obj);
lib/libcrypto/x509/x509name.c
75
ASN1_OBJECT *obj;
lib/libcrypto/x509/x509name.c
77
obj = OBJ_nid2obj(nid);
lib/libcrypto/x509/x509name.c
78
if (obj == NULL)
lib/libcrypto/x509/x509name.c
80
return (X509_NAME_get_text_by_OBJ(name, obj, buf, len));
lib/libcrypto/x509/x509name.c
85
X509_NAME_get_text_by_OBJ(X509_NAME *name, const ASN1_OBJECT *obj, char *buf,
lib/libcrypto/x509/x509name.c
94
i = X509_NAME_get_index_by_OBJ(name, obj, -1);
lib/libcurses/curses.priv.h
703
#define _nc_mutex_init(obj) /* nothing */
lib/libcurses/tinfo/lib_data.c
354
_nc_mutex_init(pthread_mutex_t * obj)
lib/libcurses/tinfo/lib_data.c
361
pthread_mutex_init(obj, &recattr);
lib/libcurses/tinfo/lib_data.c
366
_nc_mutex_lock(pthread_mutex_t * obj)
lib/libcurses/tinfo/lib_data.c
370
rc = pthread_mutex_lock(obj);
lib/libcurses/tinfo/lib_data.c
375
_nc_mutex_trylock(pthread_mutex_t * obj)
lib/libcurses/tinfo/lib_data.c
379
rc = pthread_mutex_trylock(obj);
lib/libcurses/tinfo/lib_data.c
384
_nc_mutex_unlock(pthread_mutex_t * obj)
lib/libcurses/tinfo/lib_data.c
388
rc = pthread_mutex_unlock(obj);
lib/libkvm/kvm_private.h
93
#define KREAD(kd, addr, obj) \
lib/libkvm/kvm_private.h
94
(kvm_read(kd, addr, (void *)(obj), sizeof(*obj)) != sizeof(*obj))
lib/libtls/tls_ocsp.c
136
X509_OBJECT *obj = NULL;
lib/libtls/tls_ocsp.c
158
if ((obj = X509_STORE_CTX_get_obj_by_subject(storectx, X509_LU_X509,
lib/libtls/tls_ocsp.c
162
cid = OCSP_cert_to_id(NULL, main_cert, X509_OBJECT_get0_X509(obj));
lib/libtls/tls_ocsp.c
166
X509_OBJECT_free(obj);
libexec/ld.so/aarch64/rtld_machine.c
178
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/aarch64/rtld_machine.c
204
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/aarch64/rtld_machine.c
284
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/aarch64/rtld_machine.c
286
if (sr.obj->traced && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/alpha/rtld_machine.c
112
*r_addr += sr.obj->obj_base + sr.sym->st_value +
libexec/ld.so/alpha/rtld_machine.c
138
*r_addr = sr.obj->obj_base + sr.sym->st_value +
libexec/ld.so/alpha/rtld_machine.c
152
prev_value = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/alpha/rtld_machine.c
199
buf.newval = sr.obj->obj_base + sr.sym->st_value + rela->r_addend;
libexec/ld.so/alpha/rtld_machine.c
201
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/amd64/rtld_machine.c
250
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/amd64/rtld_machine.c
273
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/amd64/rtld_machine.c
334
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/amd64/rtld_machine.c
336
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/arm/rtld_machine.c
243
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/arm/rtld_machine.c
269
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/arm/rtld_machine.c
349
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/arm/rtld_machine.c
351
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/dlfcn.c
213
addr = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/dlfcn.c
216
addr = _dl_md_plabel(addr, sr.obj->dyn.pltgot);
libexec/ld.so/dlfcn.c
240
elf_object_t *obj;
libexec/ld.so/dlfcn.c
242
obj = obj_from_addr(data);
libexec/ld.so/dlfcn.c
243
if (obj == NULL) {
libexec/ld.so/dlfcn.c
248
if ((obj->status & STAT_NODELETE) == 0) {
libexec/ld.so/dlfcn.c
249
obj->opencount++;
libexec/ld.so/dlfcn.c
250
obj->status |= STAT_NODELETE;
libexec/ld.so/dlfcn.c
263
elf_object_t *obj;
libexec/ld.so/dlfcn.c
269
obj = n->data;
libexec/ld.so/dlfcn.c
270
_dl_printf("%s\n", obj->load_name);
libexec/ld.so/dlfcn.c
273
for (vec = obj->child_vec, i = 0; i < vec.len; i++)
libexec/ld.so/dlfcn.c
277
TAILQ_FOREACH(m, &obj->grpref_list, next_sib)
libexec/ld.so/hppa/rtld_machine.c
185
sr.obj = object;
libexec/ld.so/hppa/rtld_machine.c
196
*pt = sr.obj->obj_base + sr.sym->st_value +
libexec/ld.so/hppa/rtld_machine.c
201
*pt, sr.obj->load_name));
libexec/ld.so/hppa/rtld_machine.c
222
*pt = _dl_md_plabel(sr.obj->obj_base +
libexec/ld.so/hppa/rtld_machine.c
224
sr.obj->dyn.pltgot);
libexec/ld.so/hppa/rtld_machine.c
228
*pt, sr.obj->load_name));
libexec/ld.so/hppa/rtld_machine.c
241
pt[0] = sr.obj->obj_base + sr.sym->st_value +
libexec/ld.so/hppa/rtld_machine.c
243
pt[1] = (Elf_Addr)sr.obj->dyn.pltgot;
libexec/ld.so/hppa/rtld_machine.c
247
pt[0], pt[1], sr.obj->load_name));
libexec/ld.so/hppa/rtld_machine.c
265
_dl_bcopy((void *)(sr.obj->obj_base +
libexec/ld.so/hppa/rtld_machine.c
269
i, symn, sr.obj->obj_base +
libexec/ld.so/hppa/rtld_machine.c
271
pt, sym->st_size, sr.obj->load_name));
libexec/ld.so/hppa/rtld_machine.c
432
value = sr.obj->obj_base + sr.sym->st_value + rela->r_addend;
libexec/ld.so/hppa/rtld_machine.c
434
buf.newval = ((uint64_t)value << 32) | (Elf_Addr)sr.obj->dyn.pltgot;
libexec/ld.so/hppa/rtld_machine.c
436
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/i386/rtld_machine.c
250
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/i386/rtld_machine.c
273
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/i386/rtld_machine.c
339
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/i386/rtld_machine.c
341
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/loader.c
108
_dl_run_dtors(elf_object_t *obj)
libexec/ld.so/loader.c
110
if (obj->dyn.fini_array) {
libexec/ld.so/loader.c
111
int num = obj->dyn.fini_arraysz / sizeof(Elf_Addr);
libexec/ld.so/loader.c
115
obj, obj->dyn.fini_array, obj->load_name));
libexec/ld.so/loader.c
117
(*obj->dyn.fini_array[i-1])();
libexec/ld.so/loader.c
120
if (obj->dyn.fini) {
libexec/ld.so/loader.c
122
obj, obj->dyn.fini, obj->load_name));
libexec/ld.so/loader.c
123
(*obj->dyn.fini)();
libexec/ld.so/loader.c
979
void *p = (void *)(sr.sym->st_value + sr.obj->obj_base);
libexec/ld.so/loader.c
982
p, sr.obj->load_name, (void *)sr.obj, addr));
libexec/ld.so/m88k/rtld_machine.c
125
_dl_bcopy((void *)(sr.obj->obj_base +
libexec/ld.so/m88k/rtld_machine.c
158
prev_ooff = sr.obj->obj_base;
libexec/ld.so/m88k/rtld_machine.c
315
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/m88k/rtld_machine.c
317
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/mips64/archdep.h
52
#define RELOC_GOT(obj, off) \
libexec/ld.so/mips64/archdep.h
54
struct boot_dyn *__dynld = obj; \
libexec/ld.so/mips64/rtld_machine.c
111
prev_value = sr.obj->obj_base +
libexec/ld.so/mips64/rtld_machine.c
203
sr.obj->obj_base;
libexec/ld.so/mips64/rtld_machine.c
212
*gotp = sr.sym->st_value + sr.obj->obj_base;
libexec/ld.so/mips64/rtld_machine.c
222
*gotp = sr.sym->st_value + sr.obj->obj_base;
libexec/ld.so/mips64/rtld_machine.c
259
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/mips64/rtld_machine.c
261
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/powerpc/rtld_machine.c
123
prev_ooff = sr.obj->obj_base;
libexec/ld.so/powerpc/rtld_machine.c
245
_dl_bcopy((void *)(sr.obj->obj_base + sr.sym->st_value),
libexec/ld.so/powerpc/rtld_machine.c
327
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/powerpc/rtld_machine.c
329
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/powerpc64/rtld_machine.c
119
prev_ooff = sr.obj->obj_base;
libexec/ld.so/powerpc64/rtld_machine.c
243
_dl_bcopy((void *)(sr.obj->obj_base + sr.sym->st_value),
libexec/ld.so/powerpc64/rtld_machine.c
323
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/powerpc64/rtld_machine.c
325
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/resolve.c
534
matched_symbol(elf_object_t *obj, const Elf_Sym *sym, struct symlookup *sl)
libexec/ld.so/resolve.c
565
_dl_strcmp(sl->sl_name, obj->dyn.strtab + sym->st_name))
libexec/ld.so/resolve.c
570
sl->sl_out.obj = obj;
libexec/ld.so/resolve.c
575
sl->sl_weak_out.obj = obj;
libexec/ld.so/resolve.c
584
_dl_find_symbol_obj(elf_object_t *obj, struct symlookup *sl)
libexec/ld.so/resolve.c
586
const Elf_Sym *symt = obj->dyn.symtab;
libexec/ld.so/resolve.c
588
if (obj->status & STAT_GNU_HASH) {
libexec/ld.so/resolve.c
597
bloom_word = obj->bloom_gnu[(hash / ELFSIZE) &
libexec/ld.so/resolve.c
598
obj->mask_bm_gnu];
libexec/ld.so/resolve.c
602
h2 = (hash >> obj->shift2_gnu) & (ELFSIZE - 1);
libexec/ld.so/resolve.c
609
bucket = obj->buckets_gnu[hash % obj->nbuckets];
libexec/ld.so/resolve.c
612
hashval = &obj->chains_gnu[bucket];
libexec/ld.so/resolve.c
616
(hashval - obj->chains_gnu);
libexec/ld.so/resolve.c
618
int r = matched_symbol(obj, sym, sl);
libexec/ld.so/resolve.c
626
for (si = obj->buckets_elf[sl->sl_elf_hash % obj->nbuckets];
libexec/ld.so/resolve.c
627
si != STN_UNDEF; si = obj->chains_elf[si]) {
libexec/ld.so/resolve.c
630
int r = matched_symbol(obj, sym, sl);
libexec/ld.so/resolve.c
739
__progname, req_obj->load_name, sl.sl_out.obj->load_name,
libexec/ld.so/resolve.h
284
const elf_object_t *obj;
libexec/ld.so/resolve.h
407
const elf_object_t *obj;
libexec/ld.so/riscv64/rtld_machine.c
177
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/riscv64/rtld_machine.c
203
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/riscv64/rtld_machine.c
283
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/riscv64/rtld_machine.c
285
if (sr.obj->traced && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/sh/rtld_machine.c
502
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/sh/rtld_machine.c
525
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/sh/rtld_machine.c
612
buf.newval = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/sh/rtld_machine.c
614
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/sparc64/rtld_machine.c
281
prev_value = (Elf_Addr)(sr.obj->obj_base +
libexec/ld.so/sparc64/rtld_machine.c
299
srcaddr = (void *)(sr.obj->obj_base + sr.sym->st_value);
libexec/ld.so/sparc64/rtld_machine.c
627
newvalue = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/sparc64/rtld_machine.c
629
if (__predict_false(sr.obj->traced) && _dl_trace_plt(sr.obj, symn))
libexec/ld.so/sparc64/rtld_machine.c
793
value = sr.obj->obj_base + sr.sym->st_value;
libexec/ld.so/tib.c
100
_dl_bcopy(obj->tls_static_data, addr,
libexec/ld.so/tib.c
101
obj->tls_fsize);
libexec/ld.so/tib.c
103
obj->load_name, obj->tls_offset,
libexec/ld.so/tib.c
104
(void *)addr, obj->tls_msize, obj->tls_fsize));
libexec/ld.so/tib.c
230
struct elf_object *obj;
libexec/ld.so/tib.c
233
for (obj = _dl_objects; obj != NULL; obj = obj->next) {
libexec/ld.so/tib.c
234
if (obj->tls_msize != 0) {
libexec/ld.so/tib.c
235
obj->tls_offset = allocate_tls_offset(obj->tls_msize,
libexec/ld.so/tib.c
236
obj->tls_align, obj->obj_type == OBJTYPE_EXE);
libexec/ld.so/tib.c
63
struct elf_object *obj;
libexec/ld.so/tib.c
93
for (obj = _dl_objects; obj != NULL; obj = obj->next) {
libexec/ld.so/tib.c
94
if (obj->tls_msize != 0) {
libexec/ld.so/tib.c
95
char *addr = TLS_ADDR(tib, obj->tls_offset);
libexec/ld.so/tib.c
97
_dl_memset(addr + obj->tls_fsize, 0,
libexec/ld.so/tib.c
98
obj->tls_msize - obj->tls_fsize);
libexec/ld.so/tib.c
99
if (obj->tls_static_data != NULL)
libexec/snmpd/snmpd_metrics/mib.c
1006
} else if (obj == ifType) {
libexec/snmpd/snmpd_metrics/mib.c
1015
} else if (obj == ifMtu)
libexec/snmpd/snmpd_metrics/mib.c
1017
else if (obj == ifSpeed) {
libexec/snmpd/snmpd_metrics/mib.c
1023
} else if (obj == ifPhysAddress) {
libexec/snmpd/snmpd_metrics/mib.c
1031
} else if (obj == ifAdminStatus) {
libexec/snmpd/snmpd_metrics/mib.c
1035
} else if (obj == ifOperStatus) {
libexec/snmpd/snmpd_metrics/mib.c
1045
} else if (obj == ifLastChange)
libexec/snmpd/snmpd_metrics/mib.c
1047
else if (obj == ifInOctets)
libexec/snmpd/snmpd_metrics/mib.c
1049
else if (obj == ifInUcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1051
else if (obj == ifInNUcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1053
else if (obj == ifInDiscards)
libexec/snmpd/snmpd_metrics/mib.c
1055
else if (obj == ifInErrors)
libexec/snmpd/snmpd_metrics/mib.c
1057
else if (obj == ifInUnknownProtos)
libexec/snmpd/snmpd_metrics/mib.c
1059
else if (obj == ifOutOctets)
libexec/snmpd/snmpd_metrics/mib.c
1061
else if (obj == ifOutUcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1063
else if (obj == ifOutNUcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1065
else if (obj == ifOutDiscards)
libexec/snmpd/snmpd_metrics/mib.c
1067
else if (obj == ifOutErrors)
libexec/snmpd/snmpd_metrics/mib.c
1069
else if (obj == ifOutQLen)
libexec/snmpd/snmpd_metrics/mib.c
1071
else if (obj == ifSpecific)
libexec/snmpd/snmpd_metrics/mib.c
1080
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1086
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1108
if (obj == ifName)
libexec/snmpd/snmpd_metrics/mib.c
1110
else if (obj == ifInMulticastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1112
else if (obj == ifInBroadcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1114
else if (obj == ifOutMulticastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1116
else if (obj == ifOutBroadcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1118
else if (obj == ifHCInOctets)
libexec/snmpd/snmpd_metrics/mib.c
1120
else if (obj == ifHCInUcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1122
else if (obj == ifHCInMulticastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1124
else if (obj == ifHCInBroadcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1126
else if (obj == ifHCOutOctets)
libexec/snmpd/snmpd_metrics/mib.c
1128
else if (obj == ifHCOutUcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1130
else if (obj == ifHCOutMulticastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1132
else if (obj == ifHCOutBroadcastPkts)
libexec/snmpd/snmpd_metrics/mib.c
1134
else if (obj == ifLinkUpDownTrapEnable)
libexec/snmpd/snmpd_metrics/mib.c
1136
else if (obj == ifHighSpeed) {
libexec/snmpd/snmpd_metrics/mib.c
1140
} else if (obj == ifPromiscuousMode) {
libexec/snmpd/snmpd_metrics/mib.c
1144
} else if (obj == ifConnectorPresent) {
libexec/snmpd/snmpd_metrics/mib.c
1148
} else if (obj == ifAlias)
libexec/snmpd/snmpd_metrics/mib.c
1150
else if (obj == ifCounterDiscontinuityTime)
libexec/snmpd/snmpd_metrics/mib.c
1165
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1173
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1215
if (obj == ifRcvAddressStatus) {
libexec/snmpd/snmpd_metrics/mib.c
1219
} else if (obj == ifRcvAddressType) {
libexec/snmpd/snmpd_metrics/mib.c
1276
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1286
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1287
if (obj == pfRunning)
libexec/snmpd/snmpd_metrics/mib.c
1289
else if (obj == pfRuntime) {
libexec/snmpd/snmpd_metrics/mib.c
1294
} else if (obj == pfDebug)
libexec/snmpd/snmpd_metrics/mib.c
1296
else if (obj == pfHostid)
libexec/snmpd/snmpd_metrics/mib.c
1305
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1313
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1314
if (obj == pfCntMatch)
libexec/snmpd/snmpd_metrics/mib.c
1316
else if (obj == pfCntBadOffset)
libexec/snmpd/snmpd_metrics/mib.c
1318
else if (obj == pfCntFragment)
libexec/snmpd/snmpd_metrics/mib.c
1320
else if (obj == pfCntShort)
libexec/snmpd/snmpd_metrics/mib.c
1322
else if (obj == pfCntNormalize)
libexec/snmpd/snmpd_metrics/mib.c
1324
else if (obj == pfCntMemory)
libexec/snmpd/snmpd_metrics/mib.c
1326
else if (obj == pfCntTimestamp)
libexec/snmpd/snmpd_metrics/mib.c
1328
else if (obj == pfCntCongestion)
libexec/snmpd/snmpd_metrics/mib.c
1330
else if (obj == pfCntIpOption)
libexec/snmpd/snmpd_metrics/mib.c
1332
else if (obj == pfCntProtoCksum)
libexec/snmpd/snmpd_metrics/mib.c
1334
else if (obj == pfCntStateMismatch)
libexec/snmpd/snmpd_metrics/mib.c
1336
else if (obj == pfCntStateInsert)
libexec/snmpd/snmpd_metrics/mib.c
1338
else if (obj == pfCntStateLimit)
libexec/snmpd/snmpd_metrics/mib.c
1340
else if (obj == pfCntSrcLimit)
libexec/snmpd/snmpd_metrics/mib.c
1342
else if (obj == pfCntSynproxy)
libexec/snmpd/snmpd_metrics/mib.c
1344
else if (obj == pfCntTranslate)
libexec/snmpd/snmpd_metrics/mib.c
1346
else if (obj == pfCntNoRoute)
libexec/snmpd/snmpd_metrics/mib.c
1355
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1363
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1364
if (obj == pfStateCount)
libexec/snmpd/snmpd_metrics/mib.c
1366
else if (obj == pfStateSearches)
libexec/snmpd/snmpd_metrics/mib.c
1368
else if (obj == pfStateInserts)
libexec/snmpd/snmpd_metrics/mib.c
1370
else if (obj == pfStateRemovals)
libexec/snmpd/snmpd_metrics/mib.c
1379
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1387
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1388
if (obj == pfLogIfName)
libexec/snmpd/snmpd_metrics/mib.c
1390
else if (obj == pfLogIfIpBytesIn)
libexec/snmpd/snmpd_metrics/mib.c
1392
else if (obj == pfLogIfIpBytesOut)
libexec/snmpd/snmpd_metrics/mib.c
1394
else if (obj == pfLogIfIpPktsInPass)
libexec/snmpd/snmpd_metrics/mib.c
1396
else if (obj == pfLogIfIpPktsInDrop)
libexec/snmpd/snmpd_metrics/mib.c
1398
else if (obj == pfLogIfIpPktsOutPass)
libexec/snmpd/snmpd_metrics/mib.c
1400
else if (obj == pfLogIfIpPktsOutDrop)
libexec/snmpd/snmpd_metrics/mib.c
1402
else if (obj == pfLogIfIp6BytesIn)
libexec/snmpd/snmpd_metrics/mib.c
1404
else if (obj == pfLogIfIp6BytesOut)
libexec/snmpd/snmpd_metrics/mib.c
1406
else if (obj == pfLogIfIp6PktsInPass)
libexec/snmpd/snmpd_metrics/mib.c
1408
else if (obj == pfLogIfIp6PktsInDrop)
libexec/snmpd/snmpd_metrics/mib.c
1410
else if (obj == pfLogIfIp6PktsOutPass)
libexec/snmpd/snmpd_metrics/mib.c
1412
else if (obj == pfLogIfIp6PktsOutDrop)
libexec/snmpd/snmpd_metrics/mib.c
1421
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1429
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1430
if (obj == pfSrcTrackCount)
libexec/snmpd/snmpd_metrics/mib.c
1432
else if (obj == pfSrcTrackSearches)
libexec/snmpd/snmpd_metrics/mib.c
1434
else if (obj == pfSrcTrackInserts)
libexec/snmpd/snmpd_metrics/mib.c
1436
else if (obj == pfSrcTrackRemovals)
libexec/snmpd/snmpd_metrics/mib.c
1446
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1450
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1452
if (obj == pfLimitStates)
libexec/snmpd/snmpd_metrics/mib.c
1454
else if (obj == pfLimitSourceNodes)
libexec/snmpd/snmpd_metrics/mib.c
1456
else if (obj == pfLimitFragments)
libexec/snmpd/snmpd_metrics/mib.c
1458
else if (obj == pfLimitMaxTables)
libexec/snmpd/snmpd_metrics/mib.c
1460
else if (obj == pfLimitMaxTableEntries)
libexec/snmpd/snmpd_metrics/mib.c
1477
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1481
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1483
if (obj == pfTimeoutTcpFirst)
libexec/snmpd/snmpd_metrics/mib.c
1485
else if (obj == pfTimeoutTcpOpening)
libexec/snmpd/snmpd_metrics/mib.c
1487
else if (obj == pfTimeoutTcpEstablished)
libexec/snmpd/snmpd_metrics/mib.c
1489
else if (obj == pfTimeoutTcpClosing)
libexec/snmpd/snmpd_metrics/mib.c
1491
else if (obj == pfTimeoutTcpFinWait)
libexec/snmpd/snmpd_metrics/mib.c
1493
else if (obj == pfTimeoutTcpClosed)
libexec/snmpd/snmpd_metrics/mib.c
1495
else if (obj == pfTimeoutUdpFirst)
libexec/snmpd/snmpd_metrics/mib.c
1497
else if (obj == pfTimeoutUdpSingle)
libexec/snmpd/snmpd_metrics/mib.c
1499
else if (obj == pfTimeoutUdpMultiple)
libexec/snmpd/snmpd_metrics/mib.c
1501
else if (obj == pfTimeoutIcmpFirst)
libexec/snmpd/snmpd_metrics/mib.c
1503
else if (obj == pfTimeoutIcmpError)
libexec/snmpd/snmpd_metrics/mib.c
1505
else if (obj == pfTimeoutOtherFirst)
libexec/snmpd/snmpd_metrics/mib.c
1507
else if (obj == pfTimeoutOtherSingle)
libexec/snmpd/snmpd_metrics/mib.c
1509
else if (obj == pfTimeoutOtherMultiple)
libexec/snmpd/snmpd_metrics/mib.c
1511
else if (obj == pfTimeoutFragment)
libexec/snmpd/snmpd_metrics/mib.c
1513
else if (obj == pfTimeoutInterval)
libexec/snmpd/snmpd_metrics/mib.c
1515
else if (obj == pfTimeoutAdaptiveStart)
libexec/snmpd/snmpd_metrics/mib.c
1517
else if (obj == pfTimeoutAdaptiveEnd)
libexec/snmpd/snmpd_metrics/mib.c
1519
else if (obj == pfTimeoutSrcTrack)
libexec/snmpd/snmpd_metrics/mib.c
1547
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1552
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1574
if (obj == pfIfIndex)
libexec/snmpd/snmpd_metrics/mib.c
1576
else if (obj == pfIfDescr)
libexec/snmpd/snmpd_metrics/mib.c
1578
else if (obj == pfIfType) {
libexec/snmpd/snmpd_metrics/mib.c
1582
} else if (obj == pfIfRefs)
libexec/snmpd/snmpd_metrics/mib.c
1584
else if (obj == pfIfRules)
libexec/snmpd/snmpd_metrics/mib.c
1586
else if (obj == pfIfIn4PassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1588
else if (obj == pfIfIn4PassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1590
else if (obj == pfIfIn4BlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1592
else if (obj == pfIfIn4BlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1594
else if (obj == pfIfOut4PassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1596
else if (obj == pfIfOut4PassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1598
else if (obj == pfIfOut4BlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1600
else if (obj == pfIfOut4BlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1602
else if (obj == pfIfIn6PassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1604
else if (obj == pfIfIn6PassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1606
else if (obj == pfIfIn6BlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1608
else if (obj == pfIfIn6BlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1610
else if (obj == pfIfOut6PassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1612
else if (obj == pfIfOut6PassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1614
else if (obj == pfIfOut6BlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1616
else if (obj == pfIfOut6BlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1636
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1642
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1664
if (obj == pfTblIndex)
libexec/snmpd/snmpd_metrics/mib.c
1666
else if (obj == pfTblName)
libexec/snmpd/snmpd_metrics/mib.c
1668
else if (obj == pfTblAddresses)
libexec/snmpd/snmpd_metrics/mib.c
1670
else if (obj == pfTblAnchorRefs)
libexec/snmpd/snmpd_metrics/mib.c
1672
else if (obj == pfTblRuleRefs)
libexec/snmpd/snmpd_metrics/mib.c
1674
else if (obj == pfTblEvalsMatch)
libexec/snmpd/snmpd_metrics/mib.c
1676
else if (obj == pfTblEvalsNoMatch)
libexec/snmpd/snmpd_metrics/mib.c
1678
else if (obj == pfTblInPassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1680
else if (obj == pfTblInPassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1682
else if (obj == pfTblInBlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1684
else if (obj == pfTblInBlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1686
else if (obj == pfTblInXPassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1688
else if (obj == pfTblInXPassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1690
else if (obj == pfTblOutPassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1692
else if (obj == pfTblOutPassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1694
else if (obj == pfTblOutBlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1696
else if (obj == pfTblOutBlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1698
else if (obj == pfTblOutXPassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1700
else if (obj == pfTblOutXPassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1702
else if (obj == pfTblStatsCleared) {
libexec/snmpd/snmpd_metrics/mib.c
1705
} else if (obj == pfTblInMatchPkts)
libexec/snmpd/snmpd_metrics/mib.c
1707
else if (obj == pfTblInMatchBytes)
libexec/snmpd/snmpd_metrics/mib.c
1709
else if (obj == pfTblOutMatchPkts)
libexec/snmpd/snmpd_metrics/mib.c
1711
else if (obj == pfTblOutMatchBytes)
libexec/snmpd/snmpd_metrics/mib.c
1720
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1725
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1767
if (obj == pfTblAddrTblIndex)
libexec/snmpd/snmpd_metrics/mib.c
1769
else if (obj == pfTblAddrNet)
libexec/snmpd/snmpd_metrics/mib.c
1771
else if (obj == pfTblAddrMask)
libexec/snmpd/snmpd_metrics/mib.c
1773
else if (obj == pfTblAddrCleared)
libexec/snmpd/snmpd_metrics/mib.c
1775
else if (obj == pfTblAddrInBlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1777
else if (obj == pfTblAddrInBlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1779
else if (obj == pfTblAddrInPassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1781
else if (obj == pfTblAddrInPassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1783
else if (obj == pfTblAddrOutBlockPkts)
libexec/snmpd/snmpd_metrics/mib.c
1785
else if (obj == pfTblAddrOutBlockBytes)
libexec/snmpd/snmpd_metrics/mib.c
1787
else if (obj == pfTblAddrOutPassPkts)
libexec/snmpd/snmpd_metrics/mib.c
1789
else if (obj == pfTblAddrOutPassBytes)
libexec/snmpd/snmpd_metrics/mib.c
1791
else if (obj == pfTblAddrInMatchPkts)
libexec/snmpd/snmpd_metrics/mib.c
1793
else if (obj == pfTblAddrInMatchBytes)
libexec/snmpd/snmpd_metrics/mib.c
1795
else if (obj == pfTblAddrOutMatchPkts)
libexec/snmpd/snmpd_metrics/mib.c
1797
else if (obj == pfTblAddrOutMatchBytes)
libexec/snmpd/snmpd_metrics/mib.c
1851
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1866
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1909
if (obj == pfLabelIndex)
libexec/snmpd/snmpd_metrics/mib.c
1911
else if (obj == pfLabelName)
libexec/snmpd/snmpd_metrics/mib.c
1913
else if (obj == pfLabelEvals)
libexec/snmpd/snmpd_metrics/mib.c
1915
else if (obj == pfLabelPkts)
libexec/snmpd/snmpd_metrics/mib.c
1917
else if (obj == pfLabelBytes)
libexec/snmpd/snmpd_metrics/mib.c
1919
else if (obj == pfLabelInPkts)
libexec/snmpd/snmpd_metrics/mib.c
1921
else if (obj == pfLabelInBytes)
libexec/snmpd/snmpd_metrics/mib.c
1923
else if (obj == pfLabelOutPkts)
libexec/snmpd/snmpd_metrics/mib.c
1925
else if (obj == pfLabelOutBytes)
libexec/snmpd/snmpd_metrics/mib.c
1927
else if (obj == pfLabelTotalStates)
libexec/snmpd/snmpd_metrics/mib.c
1936
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
1948
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
1949
if (obj == pfsyncIpPktsRecv)
libexec/snmpd/snmpd_metrics/mib.c
1951
else if (obj == pfsyncIp6PktsRecv)
libexec/snmpd/snmpd_metrics/mib.c
1953
else if (obj == pfsyncPktDiscardsForBadInterface)
libexec/snmpd/snmpd_metrics/mib.c
1955
else if (obj == pfsyncPktDiscardsForBadTtl)
libexec/snmpd/snmpd_metrics/mib.c
1957
else if (obj == pfsyncPktShorterThanHeader)
libexec/snmpd/snmpd_metrics/mib.c
1959
else if (obj == pfsyncPktDiscardsForBadVersion)
libexec/snmpd/snmpd_metrics/mib.c
1961
else if (obj == pfsyncPktDiscardsForBadAction)
libexec/snmpd/snmpd_metrics/mib.c
1963
else if (obj == pfsyncPktDiscardsForBadLength)
libexec/snmpd/snmpd_metrics/mib.c
1965
else if (obj == pfsyncPktDiscardsForBadAuth)
libexec/snmpd/snmpd_metrics/mib.c
1967
else if (obj == pfsyncPktDiscardsForStaleState)
libexec/snmpd/snmpd_metrics/mib.c
1969
else if (obj == pfsyncPktDiscardsForBadValues)
libexec/snmpd/snmpd_metrics/mib.c
1971
else if (obj == pfsyncPktDiscardsForBadState)
libexec/snmpd/snmpd_metrics/mib.c
1973
else if (obj == pfsyncIpPktsSent)
libexec/snmpd/snmpd_metrics/mib.c
1975
else if (obj == pfsyncIp6PktsSent)
libexec/snmpd/snmpd_metrics/mib.c
1977
else if (obj == pfsyncNoMemory)
libexec/snmpd/snmpd_metrics/mib.c
1979
else if (obj == pfsyncOutputErrors)
libexec/snmpd/snmpd_metrics/mib.c
2015
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2028
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2082
if (obj == sensorIndex)
libexec/snmpd/snmpd_metrics/mib.c
2084
else if (obj == sensorDescr) {
libexec/snmpd/snmpd_metrics/mib.c
2092
} else if (obj == sensorType)
libexec/snmpd/snmpd_metrics/mib.c
2094
else if (obj == sensorDevice)
libexec/snmpd/snmpd_metrics/mib.c
2096
else if (obj == sensorValue) {
libexec/snmpd/snmpd_metrics/mib.c
2104
} else if (obj == sensorUnits)
libexec/snmpd/snmpd_metrics/mib.c
2106
else if (obj == sensorStatus)
libexec/snmpd/snmpd_metrics/mib.c
2192
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2197
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2198
if (obj == carpAllow)
libexec/snmpd/snmpd_metrics/mib.c
2200
else if (obj == carpPreempt)
libexec/snmpd/snmpd_metrics/mib.c
2202
else if (obj == carpLog)
libexec/snmpd/snmpd_metrics/mib.c
2220
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2234
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2235
if (obj == carpIpPktsRecv)
libexec/snmpd/snmpd_metrics/mib.c
2237
else if (obj == carpIp6PktsRecv)
libexec/snmpd/snmpd_metrics/mib.c
2239
else if (obj == carpPktDiscardsForBadInterface)
libexec/snmpd/snmpd_metrics/mib.c
2241
else if (obj == carpPktDiscardsForWrongTtl)
libexec/snmpd/snmpd_metrics/mib.c
2243
else if (obj == carpPktShorterThanHeader)
libexec/snmpd/snmpd_metrics/mib.c
2245
else if (obj == carpPktDiscardsForBadChecksum)
libexec/snmpd/snmpd_metrics/mib.c
2247
else if (obj == carpPktDiscardsForBadVersion)
libexec/snmpd/snmpd_metrics/mib.c
2249
else if (obj == carpPktDiscardsForTooShort)
libexec/snmpd/snmpd_metrics/mib.c
2251
else if (obj == carpPktDiscardsForBadAuth)
libexec/snmpd/snmpd_metrics/mib.c
2253
else if (obj == carpPktDiscardsForBadVhid)
libexec/snmpd/snmpd_metrics/mib.c
2255
else if (obj == carpPktDiscardsForBadAddressList)
libexec/snmpd/snmpd_metrics/mib.c
2257
else if (obj == carpIpPktsSent)
libexec/snmpd/snmpd_metrics/mib.c
2259
else if (obj == carpIp6PktsSent)
libexec/snmpd/snmpd_metrics/mib.c
2261
else if (obj == carpNoMemory)
libexec/snmpd/snmpd_metrics/mib.c
2263
else if (obj == carpTransitionsToMaster)
libexec/snmpd/snmpd_metrics/mib.c
2346
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2351
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2385
if (obj == carpIfIndex)
libexec/snmpd/snmpd_metrics/mib.c
2387
else if (obj == carpIfDescr)
libexec/snmpd/snmpd_metrics/mib.c
2389
else if (obj == carpIfVhid)
libexec/snmpd/snmpd_metrics/mib.c
2391
else if (obj == carpIfDev)
libexec/snmpd/snmpd_metrics/mib.c
2393
else if (obj == carpIfAdvbase)
libexec/snmpd/snmpd_metrics/mib.c
2395
else if (obj == carpIfAdvskew)
libexec/snmpd/snmpd_metrics/mib.c
2397
else if (obj == carpIfState)
libexec/snmpd/snmpd_metrics/mib.c
2455
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2462
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2490
if (obj == carpGroupName)
libexec/snmpd/snmpd_metrics/mib.c
2492
else if (obj == carpGroupDemote) {
libexec/snmpd/snmpd_metrics/mib.c
2527
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2532
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2554
if (obj == memIfName)
libexec/snmpd/snmpd_metrics/mib.c
2556
else if (obj == memIfLiveLocks)
libexec/snmpd/snmpd_metrics/mib.c
2623
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2632
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2633
if (obj == ipInReceives)
libexec/snmpd/snmpd_metrics/mib.c
2635
else if (obj == ipInUnknownProtos)
libexec/snmpd/snmpd_metrics/mib.c
2637
else if (obj == ipInDelivers)
libexec/snmpd/snmpd_metrics/mib.c
2639
else if (obj == ipOutRequests)
libexec/snmpd/snmpd_metrics/mib.c
2641
else if (obj == ipOutDiscards)
libexec/snmpd/snmpd_metrics/mib.c
2643
else if (obj == ipOutNoRoutes)
libexec/snmpd/snmpd_metrics/mib.c
2645
else if (obj == ipReasmReqds)
libexec/snmpd/snmpd_metrics/mib.c
2647
else if (obj == ipReasmOKs)
libexec/snmpd/snmpd_metrics/mib.c
2649
else if (obj == ipFragOKs)
libexec/snmpd/snmpd_metrics/mib.c
2651
else if (obj == ipFragCreates)
libexec/snmpd/snmpd_metrics/mib.c
2754
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2759
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2792
if (obj == ipAdEntAddr)
libexec/snmpd/snmpd_metrics/mib.c
2795
else if (obj == ipAdEntIfIndex)
libexec/snmpd/snmpd_metrics/mib.c
2797
else if (obj == ipAdEntNetMask)
libexec/snmpd/snmpd_metrics/mib.c
2799
else if (obj == ipAdEntBcastAddr)
libexec/snmpd/snmpd_metrics/mib.c
2801
else if (obj == ipAdEntReasmMaxSize)
libexec/snmpd/snmpd_metrics/mib.c
2810
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2817
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
2886
if (obj == ipNetToMediaIfIndex)
libexec/snmpd/snmpd_metrics/mib.c
2888
else if (obj == ipNetToMediaPhysAddress) {
libexec/snmpd/snmpd_metrics/mib.c
2896
} else if (obj == ipNetToMediaNetAddress)
libexec/snmpd/snmpd_metrics/mib.c
2898
else if (obj == ipNetToMediaType) {
libexec/snmpd/snmpd_metrics/mib.c
2926
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
2937
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
3015
if (obj == inetCidrRouteIfIndex)
libexec/snmpd/snmpd_metrics/mib.c
3017
else if (obj == inetCidrRouteType) {
libexec/snmpd/snmpd_metrics/mib.c
3027
} else if (obj == inetCidrRouteProto) {
libexec/snmpd/snmpd_metrics/mib.c
3055
} else if (obj == inetCidrRouteAge)
libexec/snmpd/snmpd_metrics/mib.c
3057
else if (obj == inetCidrRouteNextHopAS)
libexec/snmpd/snmpd_metrics/mib.c
3059
else if (obj == inetCidrRouteMetric1)
libexec/snmpd/snmpd_metrics/mib.c
3061
else if (obj == inetCidrRouteMetric2)
libexec/snmpd/snmpd_metrics/mib.c
3063
else if (obj == inetCidrRouteMetric3)
libexec/snmpd/snmpd_metrics/mib.c
3065
else if (obj == inetCidrRouteMetric4)
libexec/snmpd/snmpd_metrics/mib.c
3067
else if (obj == inetCidrRouteMetric5)
libexec/snmpd/snmpd_metrics/mib.c
3069
else if (obj == inetCidrRouteStatus)
libexec/snmpd/snmpd_metrics/mib.c
3084
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
3100
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
3139
if (obj == diskIOIndex)
libexec/snmpd/snmpd_metrics/mib.c
3141
else if (obj == diskIODevice)
libexec/snmpd/snmpd_metrics/mib.c
3143
else if (obj == diskIONRead)
libexec/snmpd/snmpd_metrics/mib.c
3146
else if (obj == diskIONWritten)
libexec/snmpd/snmpd_metrics/mib.c
3149
else if (obj == diskIOReads)
libexec/snmpd/snmpd_metrics/mib.c
3152
else if (obj == diskIOWrites)
libexec/snmpd/snmpd_metrics/mib.c
3155
else if (obj == diskIONReadX)
libexec/snmpd/snmpd_metrics/mib.c
3157
else if (obj == diskIONWrittenX)
libexec/snmpd/snmpd_metrics/mib.c
3184
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
3189
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
3212
if (obj == dot1dBasePort)
libexec/snmpd/snmpd_metrics/mib.c
3214
else if (obj == dot1dBasePortIfIndex)
libexec/snmpd/snmpd_metrics/mib.c
3216
else if (obj == dot1dBasePortCircuit)
libexec/snmpd/snmpd_metrics/mib.c
3218
else if (obj == dot1dBasePortDelayExceededDiscards)
libexec/snmpd/snmpd_metrics/mib.c
3220
else if (obj == dot1dBasePortMtuExceededDiscards)
libexec/snmpd/snmpd_metrics/mib.c
330
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
337
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
338
if (obj == hrSystemProcesses) {
libexec/snmpd/snmpd_metrics/mib.c
357
} else if (obj == hrSystemMaxProcesses) {
libexec/snmpd/snmpd_metrics/mib.c
389
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
442
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
512
if (obj == hrStorageIndex)
libexec/snmpd/snmpd_metrics/mib.c
514
else if (obj == hrStorageType)
libexec/snmpd/snmpd_metrics/mib.c
516
else if (obj == hrStorageDescr)
libexec/snmpd/snmpd_metrics/mib.c
518
else if (obj == hrStorageAllocationUnits)
libexec/snmpd/snmpd_metrics/mib.c
520
else if (obj == hrStorageSize)
libexec/snmpd/snmpd_metrics/mib.c
522
else if (obj == hrStorageUsed)
libexec/snmpd/snmpd_metrics/mib.c
524
else if (obj == hrStorageAllocationFailures)
libexec/snmpd/snmpd_metrics/mib.c
533
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
544
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
575
if (obj == hrDeviceIndex)
libexec/snmpd/snmpd_metrics/mib.c
577
else if (obj == hrDeviceType)
libexec/snmpd/snmpd_metrics/mib.c
579
else if (obj == hrDeviceDescr)
libexec/snmpd/snmpd_metrics/mib.c
581
else if (obj == hrDeviceID)
libexec/snmpd/snmpd_metrics/mib.c
583
else if (obj == hrDeviceStatus)
libexec/snmpd/snmpd_metrics/mib.c
585
else if (obj == hrDeviceErrors)
libexec/snmpd/snmpd_metrics/mib.c
594
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
599
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
622
if (obj == hrProcessorFrwID)
libexec/snmpd/snmpd_metrics/mib.c
624
else if (obj == hrProcessorLoad) {
libexec/snmpd/snmpd_metrics/mib.c
645
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
652
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
682
if (obj == hrSWRunIndex)
libexec/snmpd/snmpd_metrics/mib.c
684
else if (obj == hrSWRunName)
libexec/snmpd/snmpd_metrics/mib.c
686
else if (obj == hrSWRunPath) {
libexec/snmpd/snmpd_metrics/mib.c
694
} else if (obj == hrSWRunID)
libexec/snmpd/snmpd_metrics/mib.c
696
else if (obj == hrSWRunParameters) {
libexec/snmpd/snmpd_metrics/mib.c
704
} else if (obj == hrSWRunType) {
libexec/snmpd/snmpd_metrics/mib.c
712
} else if (obj == hrSWRunStatus) {
libexec/snmpd/snmpd_metrics/mib.c
734
} else if (obj == hrSWRunPerfCPU) {
libexec/snmpd/snmpd_metrics/mib.c
738
} else if (obj == hrSWRunPerfMem) {
libexec/snmpd/snmpd_metrics/mib.c
969
struct agentx_object *obj;
libexec/snmpd/snmpd_metrics/mib.c
975
obj = agentx_varbind_get_object(vb);
libexec/snmpd/snmpd_metrics/mib.c
997
if (obj == ifIndex)
libexec/snmpd/snmpd_metrics/mib.c
999
else if (obj == ifDescr) {
regress/lib/libagentx/main.c
685
struct agentx_object *obj;
regress/lib/libagentx/main.c
698
obj = agentx_context_object_find(sac, idx, oidlen, 1, 1);
regress/lib/libagentx/main.c
700
obj = agentx_context_object_nfind(sac, idx, oidlen, 1,
regress/lib/libagentx/main.c
703
if (obj == NULL) {
regress/lib/libagentx/main.c
708
agentx_varbind_set_index_object(vb, regressidx_oid, obj);
regress/lib/libagentx/main.c
709
agentx_varbind_object(vb, obj);
regress/lib/libcrypto/objects/objectstest.c
219
ASN1_OBJECT *obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
226
if ((obj = OBJ_nid2obj(ot->nid)) == NULL) {
regress/lib/libcrypto/objects/objectstest.c
231
if ((nid = OBJ_obj2nid(obj)) != ot->nid) {
regress/lib/libcrypto/objects/objectstest.c
240
ASN1_OBJECT_free(obj);
regress/lib/libcrypto/objects/objectstest.c
260
ASN1_OBJECT *obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
268
if ((obj = OBJ_txt2obj(ot->oid, 0)) == NULL) {
regress/lib/libcrypto/objects/objectstest.c
278
if (!obj_compare_bytes("object data", OBJ_get0_data(obj), OBJ_length(obj),
regress/lib/libcrypto/objects/objectstest.c
282
len = OBJ_obj2txt(buf, sizeof(buf), obj, 1);
regress/lib/libcrypto/objects/objectstest.c
293
if ((OBJ_obj2txt(NULL, 0, obj, 1) != len)) {
regress/lib/libcrypto/objects/objectstest.c
298
if ((OBJ_obj2txt(buf, 3, obj, 1) != len)) {
regress/lib/libcrypto/objects/objectstest.c
307
ASN1_OBJECT_free(obj);
regress/lib/libcrypto/objects/objectstest.c
327
ASN1_OBJECT *obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
337
if ((obj = OBJ_txt2obj(ot->sn, 0)) == NULL) {
regress/lib/libcrypto/objects/objectstest.c
342
if ((nid = OBJ_obj2nid(obj)) != ot->nid) {
regress/lib/libcrypto/objects/objectstest.c
347
ASN1_OBJECT_free(obj);
regress/lib/libcrypto/objects/objectstest.c
348
obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
351
if ((obj = OBJ_txt2obj(ot->ln, 0)) == NULL) {
regress/lib/libcrypto/objects/objectstest.c
356
if ((nid = OBJ_obj2nid(obj)) != ot->nid) {
regress/lib/libcrypto/objects/objectstest.c
361
ASN1_OBJECT_free(obj);
regress/lib/libcrypto/objects/objectstest.c
362
obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
365
if ((obj = OBJ_txt2obj(ot->oid, 0)) == NULL) {
regress/lib/libcrypto/objects/objectstest.c
369
if ((nid = OBJ_obj2nid(obj)) != ot->nid) {
regress/lib/libcrypto/objects/objectstest.c
375
len = OBJ_obj2txt(buf, sizeof(buf), obj, 0);
regress/lib/libcrypto/objects/objectstest.c
394
ASN1_OBJECT_free(obj);
regress/lib/libcrypto/objects/objectstest.c
402
ASN1_OBJECT *obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
418
if ((obj = ASN1_OBJECT_new()) == NULL)
regress/lib/libcrypto/objects/objectstest.c
424
if (OBJ_obj2txt(buf, sizeof(buf), obj, 1) != 0) {
regress/lib/libcrypto/objects/objectstest.c
436
ASN1_OBJECT_free(obj);
regress/lib/libcrypto/objects/objectstest.c
527
ASN1_OBJECT *obj = NULL;
regress/lib/libcrypto/objects/objectstest.c
534
if ((obj = d2i_ASN1_OBJECT(NULL, &p, olt->asn1_der_len)) == NULL) {
regress/lib/libcrypto/objects/objectstest.c
539
len = OBJ_obj2txt(buf, sizeof(buf), obj, 1);
regress/lib/libcrypto/objects/objectstest.c
559
ASN1_OBJECT_free(obj);
sbin/iked/crypto.c
1037
ECDSA_SIG *obj = NULL;
sbin/iked/crypto.c
1060
if ((obj = d2i_ECDSA_SIG(NULL, &p, tmplen)) == NULL)
sbin/iked/crypto.c
1062
ECDSA_SIG_get0(obj, &r, &s);
sbin/iked/crypto.c
1073
ECDSA_SIG_free(obj);
sbin/iked/crypto.c
1121
ECDSA_SIG *obj = NULL;
sbin/iked/crypto.c
1156
if ((obj = ECDSA_SIG_new()) == NULL ||
sbin/iked/crypto.c
1159
ECDSA_SIG_set0(obj, r, s) == 0 ||
sbin/iked/crypto.c
1160
(len = i2d_ECDSA_SIG(obj, &ptr)) <= 0)
sbin/iked/crypto.c
1176
ECDSA_SIG_free(obj);
sbin/isakmpd/x509.c
113
X509_OBJECT *obj = NULL;
sbin/isakmpd/x509.c
164
obj = X509_OBJECT_new();
sbin/isakmpd/x509.c
165
if (obj == NULL) {
sbin/isakmpd/x509.c
173
if (X509_STORE_get_by_subject(csc, X509_LU_X509, issuer, obj) !=
sbin/isakmpd/x509.c
177
if (X509_STORE_get_by_subject(csc, X509_LU_X509, issuer, obj)
sbin/isakmpd/x509.c
189
icert = X509_OBJECT_get0_X509(obj);
sbin/isakmpd/x509.c
200
X509_OBJECT_free(obj);
sbin/isakmpd/x509.c
201
obj = NULL;
sbin/isakmpd/x509.c
473
X509_OBJECT_free(obj);
sys/arch/alpha/include/fpu.h
105
#define FLD_CLEAR(obj, origin, len) \
sys/arch/alpha/include/fpu.h
106
((obj) & ~(FLD_MASK(len) << (origin)))
sys/arch/alpha/include/fpu.h
107
#define FLD_INSERT(obj, origin, len, value) \
sys/arch/alpha/include/fpu.h
108
(FLD_CLEAR(obj, origin, len) | (value) << origin)
sys/arch/amd64/amd64/pmap.c
1151
struct uvm_object *obj;
sys/arch/amd64/amd64/pmap.c
1155
obj = &pmap->pm_obj[lidx];
sys/arch/amd64/amd64/pmap.c
1158
pmap->pm_ptphint[lidx] = RBT_ROOT(uvm_objtree, &obj->memt);
sys/arch/amd64/amd64/pmap.c
1211
struct uvm_object *obj;
sys/arch/amd64/amd64/pmap.c
1236
obj = &pmap->pm_obj[i-2];
sys/arch/amd64/amd64/pmap.c
1237
ptp = uvm_pagealloc(obj, ptp_va2o(va, i - 1), NULL,
sys/arch/amd64/amd64/pmap.c
2544
struct uvm_object *obj;
sys/arch/amd64/amd64/pmap.c
2571
obj = &pmap->pm_obj[2]; /* PML4 UVM object */
sys/arch/amd64/amd64/pmap.c
2572
ptp = uvm_pagealloc(obj, ptp_va2o(gpa, 3), NULL,
sys/arch/amd64/amd64/pmap.c
2615
obj = &pmap->pm_obj[1]; /* PDPT UVM object */
sys/arch/amd64/amd64/pmap.c
2616
ptp = uvm_pagealloc(obj, ptp_va2o(gpa, 2), NULL,
sys/arch/amd64/amd64/pmap.c
2659
obj = &pmap->pm_obj[0]; /* PDE UVM object */
sys/arch/amd64/amd64/pmap.c
2660
ptp = uvm_pagealloc(obj, ptp_va2o(gpa, 1), NULL,
sys/arch/hppa/hppa/pmap.c
115
struct vm_page *pmap_pagealloc(struct uvm_object *obj, voff_t off);
sys/arch/hppa/hppa/pmap.c
141
pmap_pagealloc(struct uvm_object *obj, voff_t off)
sys/arch/hppa/hppa/pmap.c
145
if ((pg = uvm_pagealloc(obj, off, NULL,
sys/dev/fdt/qcspmi.c
101
#define HREAD4(sc, obj, reg) \
sys/dev/fdt/qcspmi.c
102
(bus_space_read_4((sc)->sc_iot, (sc)->sc_ioh[obj], (reg)))
sys/dev/fdt/qcspmi.c
103
#define HWRITE4(sc, obj, reg, val) \
sys/dev/fdt/qcspmi.c
104
bus_space_write_4((sc)->sc_iot, (sc)->sc_ioh[obj], (reg), (val))
sys/dev/fdt/qcspmi.c
105
#define HSET4(sc, obj, reg, bits) \
sys/dev/fdt/qcspmi.c
106
HWRITE4((sc), (obj), (reg), HREAD4((sc), (reg)) | (bits))
sys/dev/fdt/qcspmi.c
107
#define HCLR4(sc, obj, reg, bits) \
sys/dev/fdt/qcspmi.c
108
HWRITE4((sc), (obj), (reg), HREAD4((sc), (reg)) & ~(bits))
sys/dev/fdt/rkdrm.c
124
return drm_gem_handle_create(file, &sfb->obj->base, handle);
sys/dev/fdt/rkdrm.c
133
drm_gem_object_put(&sfb->obj->base);
sys/dev/fdt/rkdrm.c
160
fb->base.obj[0] = gem_obj;
sys/dev/fdt/rkdrm.c
161
fb->obj = to_drm_gem_dma_obj(gem_obj);
sys/dev/fdt/rkdrm.c
289
uint64_t paddr = (uint64_t)sfb->obj->dmamap->dm_segs[0].ds_addr;
sys/dev/fdt/rkdrm.c
290
size_t size = sfb->obj->dmamap->dm_segs[0].ds_len;
sys/dev/fdt/rkdrm.c
432
ri->ri_bits = sfb->obj->vaddr;
sys/dev/fdt/rkdrm.c
500
sfb->obj = drm_gem_dma_create(ddev, size);
sys/dev/fdt/rkdrm.c
501
if (sfb->obj == NULL) {
sys/dev/fdt/rkdrm.c
508
fb->obj[0] = &sfb->obj->base;
sys/dev/fdt/rkdrm.h
84
struct drm_gem_dma_object *obj;
sys/dev/fdt/rkvop.c
331
paddr = (uint64_t)rkfb->obj->dmamap->dm_segs[0].ds_addr;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1001
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1009
tmp->xcp_mode = obj->integer.value & 0xFFFF;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1010
tmp->mem_mode = (obj->integer.value >> 32) & 0xFFFF;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1011
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1014
obj = acpi_evaluate_dsm_typed(xcc_info->handle, &amd_xcc_dsm_guid, 0,
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1018
if (!obj || obj->package.count < 2) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1026
tmp->tmr_base = obj->package.elements[0].integer.value;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1027
tmp->tmr_size = obj->package.elements[1].integer.value;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1028
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1040
if (obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1041
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1054
union acpi_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1058
obj = acpi_evaluate_dsm_typed(xcc_info->handle, &amd_xcc_dsm_guid, 0,
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1062
if (!obj || obj->integer.value != AMD_XCC_DSM_NUM_FUNCS)
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1064
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1067
obj = acpi_evaluate_dsm_typed(xcc_info->handle, &amd_xcc_dsm_guid, 0,
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1071
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1080
xcc_info->phy_id = (obj->integer.value >> 32) & 0xFF;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1082
xcc_info->xcp_node = (obj->integer.value >> 40) & 0xFF;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1084
*sbdf = (obj->integer.value) & 0xFFFF0000;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1086
*sbdf |= (obj->integer.value >> 48) & 0xFFFF;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1087
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1088
obj = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1097
if (obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
1098
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
152
union acpi_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
175
obj = (union acpi_object *)buffer.pointer;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
181
kfree(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
185
if (obj->type != ACPI_TYPE_BUFFER) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
187
obj->type);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
188
kfree(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
192
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
966
union acpi_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
982
obj = acpi_evaluate_dsm_typed(xcc_info->handle, &amd_xcc_dsm_guid, 0,
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
986
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
994
tmp->supp_xcp_mode = obj->integer.value & 0xFFFF;
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
995
ACPI_FREE(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_acpi.c
997
obj = acpi_evaluate_dsm_typed(xcc_info->handle, &amd_xcc_dsm_guid, 0,
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
541
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
554
obj = dma_buf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
555
if (obj->dev->driver != adev_to_drm(adev)->driver)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
559
adev = drm_to_adev(obj->dev);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
560
bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2349
struct drm_gem_object *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2358
bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2368
ret = drm_vma_node_allow(&obj->vma_node, drm_priv);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2414
drm_vma_node_revoke(&obj->vma_node, drm_priv);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2425
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2433
obj = drm_gem_object_lookup(adev->kfd.client.file, handle);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2434
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2439
ret = import_obj_create(adev, obj->dma_buf, obj, va, drm_priv, mem, size,
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2449
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_bios.c
379
union acpi_object atrm_arg_elements[2], *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_bios.c
398
obj = (union acpi_object *)buffer.pointer;
sys/dev/pci/drm/amd/amdgpu/amdgpu_bios.c
399
memcpy(bios+offset, obj->buffer.pointer, obj->buffer.length);
sys/dev/pci/drm/amd/amdgpu/amdgpu_bios.c
400
len = obj->buffer.length;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1208
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1221
drm_exec_for_each_locked_object(&p->exec, index, obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1222
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
862
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
970
drm_exec_for_each_locked_object(&p->exec, index, obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
971
r = amdgpu_cs_bo_validate(p, gem_to_amdgpu_bo(obj));
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1048
if (rfb->base.obj[0]->size < size) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1051
rfb->base.obj[0]->size, size, plane);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1177
rbo = gem_to_amdgpu_bo(amdgpu_fb->base.obj[0]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1201
struct drm_gem_object *obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1205
rfb->base.obj[0] = obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1218
ret = amdgpu_display_framebuffer_init(dev, rfb, mode_cmd, obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1234
rfb->base.obj[0] = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1241
struct drm_gem_object *obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1291
drm_gem_object_get(rfb->base.obj[0]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1292
rfb->base.obj[i] = rfb->base.obj[0];
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1305
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1310
obj = drm_gem_object_lookup(file_priv, mode_cmd->handles[0]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1311
if (obj == NULL) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1320
bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1322
if (drm_gem_is_imported(obj) && !(domains & AMDGPU_GEM_DOMAIN_GTT)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1324
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1330
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1335
info, mode_cmd, obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1338
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1342
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1734
if (!fb || !fb->obj[0])
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1737
robj = gem_to_amdgpu_bo(fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
1839
abo = gem_to_amdgpu_bo(fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
200
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
220
obj = crtc->primary->fb->obj[0];
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
223
work->old_abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
226
obj = fb->obj[0];
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
227
new_abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
683
struct drm_gem_object *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
691
rbo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
89
struct drm_gem_object *obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
901
if (extract_render_dcc_offset(adev, afb->base.obj[0],
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
162
struct drm_gem_object *obj = attach->dmabuf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
163
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
185
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
186
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
209
sgt = drm_prime_pages_to_sg(obj->dev,
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
309
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
310
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
330
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
331
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
457
struct drm_gem_object *obj = attach->importer_priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
458
struct ww_acquire_ctx *ticket = dma_resv_locking_ctx(obj->resv);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
459
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
542
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
545
obj = dma_buf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
546
if (obj->dev == dev) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
551
drm_gem_object_get(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
552
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
556
obj = amdgpu_dma_buf_create_obj(dev, dma_buf);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
557
if (IS_ERR(obj))
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
558
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
563
&amdgpu_dma_buf_attach_ops, obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
565
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
573
obj->import_attach = attach;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
574
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
589
struct drm_gem_object *obj = &bo->tbo.base;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
595
if (drm_gem_is_imported(obj)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
597
struct dma_buf *dma_buf = obj->import_attach->dmabuf;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
61
struct drm_gem_object *obj = attach->importer_priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
62
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
84
struct drm_gem_object *obj = dmabuf->priv;
sys/dev/pci/drm/amd/amdgpu/amdgpu_dma_buf.c
85
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_drv.c
3800
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_drv.c
3840
obj = fb->obj[0];
sys/dev/pci/drm/amd/amdgpu/amdgpu_drv.c
3841
rbo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_eviction_fence.c
51
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_eviction_fence.c
80
drm_exec_for_each_locked_object(exec, index, obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_eviction_fence.c
81
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
247
struct drm_gem_object **obj, int8_t xcp_id_plus1)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
255
*obj = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
273
*obj = &bo->tbo.base;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
313
static int amdgpu_gem_object_open(struct drm_gem_object *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
316
struct amdgpu_bo *abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
370
if (!drm_gem_is_imported(obj) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
371
!dma_buf_is_dynamic(obj->import_attach->dmabuf))
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
392
static void amdgpu_gem_object_close(struct drm_gem_object *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
395
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
447
static int amdgpu_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
449
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
465
return drm_gem_ttm_mmap(obj, vma);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
468
static int amdgpu_gem_object_mmap(struct drm_gem_object *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
471
struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
489
return drm_gem_ttm_mmap(obj, accessprot, off, size);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.h
48
struct drm_gem_object **obj, int8_t xcp_id_plus1);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1068
struct amdgpu_ras_block_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1079
obj = node->ras_obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1080
if (obj->ras_block_match) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1081
if (obj->ras_block_match(obj, block, sub_block_index) == 0)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1082
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1084
if (amdgpu_ras_block_match_default(obj, block) == 0)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1085
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1301
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &query_if->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1305
new_ce = err_data->ce_count - obj->err_data.ce_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1306
new_ue = err_data->ue_count - obj->err_data.ue_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1307
new_de = err_data->de_count - obj->err_data.de_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1331
static void amdgpu_rasmgr_error_data_statistic_update(struct ras_manager *obj, struct ras_err_data *err_data)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1339
amdgpu_ras_error_statistic_de_count(&obj->err_data,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1341
amdgpu_ras_error_statistic_ce_count(&obj->err_data,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1343
amdgpu_ras_error_statistic_ue_count(&obj->err_data,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1348
obj->err_data.ue_count += err_data->ue_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1349
obj->err_data.ce_count += err_data->ce_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1350
obj->err_data.de_count += err_data->de_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1354
static void amdgpu_ras_mgr_virt_error_data_statistics_update(struct ras_manager *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1358
obj->err_data.ue_count = err_data->ue_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1359
obj->err_data.ce_count = err_data->ce_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1360
obj->err_data.de_count = err_data->de_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1376
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1382
obj = get_ras_manager(adev, blk);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1383
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1386
return amdgpu_aca_add_handle(adev, &obj->aca_handle, ras_block_str(blk), aca_info, data);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1391
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1393
obj = get_ras_manager(adev, blk);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1394
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1397
amdgpu_aca_remove_handle(&obj->aca_handle);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1406
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1408
obj = get_ras_manager(adev, blk);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1409
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1412
return amdgpu_aca_get_error_data(adev, &obj->aca_handle, type, err_data, qctx);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1418
struct ras_manager *obj = container_of(handle, struct ras_manager, aca_handle);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1420
.head = obj->head,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1423
if (!amdgpu_ras_get_error_query_ready(obj->adev))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1426
if (amdgpu_ras_query_error_status(obj->adev, &info))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1500
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &info->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1506
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1534
amdgpu_rasmgr_error_data_statistic_update(obj, &err_data);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1542
amdgpu_ras_mgr_virt_error_data_statistics_update(obj, &err_data);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1545
info->ue_count = obj->err_data.ue_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1546
info->ce_count = obj->err_data.ce_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1547
info->de_count = obj->err_data.de_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1613
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &info->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1630
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1725
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1741
list_for_each_entry(obj, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1743
.head = obj->head,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1942
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1947
if (!obj || obj->attr_inuse)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1953
get_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1955
snprintf(obj->fs_data.sysfs_name, sizeof(obj->fs_data.sysfs_name),
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1959
obj->sysfs_attr = (struct device_attribute){
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1961
.name = obj->fs_data.sysfs_name,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1966
sysfs_attr_init(&obj->sysfs_attr.attr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1969
&obj->sysfs_attr.attr,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1971
put_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1976
obj->attr_inuse = 1;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1984
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1989
if (!obj || !obj->attr_inuse)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1995
&obj->sysfs_attr.attr,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1998
obj->attr_inuse = 0;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
1999
put_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2009
struct ras_manager *obj, *tmp;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2011
list_for_each_entry_safe(obj, tmp, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2012
amdgpu_ras_sysfs_remove(adev, &obj->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2090
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &head->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2092
if (!obj || !dir)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2095
get_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2097
memcpy(obj->fs_data.debugfs_name,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2099
sizeof(obj->fs_data.debugfs_name));
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2101
debugfs_create_file(obj->fs_data.debugfs_name, S_IWUGO | S_IRUGO, dir,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2102
obj, &amdgpu_ras_debugfs_ops);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2131
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2143
list_for_each_entry(obj, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2144
if (amdgpu_ras_is_supported(adev, obj->head.block) &&
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2145
(obj->attr_inuse == 1)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2147
get_ras_block_str(&obj->head));
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2148
fs_info.head = obj->head;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2281
static void amdgpu_ras_interrupt_poison_consumption_handler(struct ras_manager *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2285
struct amdgpu_device *adev = obj->adev;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2287
amdgpu_ras_get_ras_block(adev, obj->head.block, 0);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2316
amdgpu_umc_poison_handler(adev, obj->head.block, 0);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2336
static void amdgpu_ras_interrupt_poison_creation_handler(struct ras_manager *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2339
struct amdgpu_device *adev = obj->adev;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2351
if (amdgpu_ip_version(obj->adev, UMC_HWIP, 0) >= IP_VERSION(12, 0, 0)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2352
struct amdgpu_ras *con = amdgpu_ras_get_context(obj->adev);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2361
static void amdgpu_ras_interrupt_umc_handler(struct ras_manager *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2364
struct ras_ih_data *data = &obj->ih_data;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2378
amdgpu_ras_set_fed(obj->adev, true);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2379
ret = data->cb(obj->adev, &err_data, entry);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2389
obj->err_data.ue_count += err_data.ue_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2390
obj->err_data.ce_count += err_data.ce_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2391
obj->err_data.de_count += err_data.de_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2397
static void amdgpu_ras_interrupt_handler(struct ras_manager *obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2399
struct ras_ih_data *data = &obj->ih_data;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2411
if (amdgpu_ras_is_poison_mode_supported(obj->adev)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2412
if (obj->head.block == AMDGPU_RAS_BLOCK__UMC)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2413
amdgpu_ras_interrupt_poison_creation_handler(obj, &entry);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2415
amdgpu_ras_interrupt_poison_consumption_handler(obj, &entry);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2417
if (obj->head.block == AMDGPU_RAS_BLOCK__UMC)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2418
amdgpu_ras_interrupt_umc_handler(obj, &entry);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2420
dev_warn(obj->adev->dev,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2430
struct ras_manager *obj =
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2433
amdgpu_ras_interrupt_handler(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2439
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2442
obj = amdgpu_ras_find_obj(adev, &info->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2443
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2446
data = &obj->ih_data;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2467
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2470
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2473
data = &obj->ih_data;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2481
put_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2489
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2493
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2495
obj = amdgpu_ras_create_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2496
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2499
get_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2503
data = &obj->ih_data;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2520
put_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2533
struct ras_manager *obj, *tmp;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2535
list_for_each_entry_safe(obj, tmp, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2536
amdgpu_ras_interrupt_remove_handler(adev, &obj->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2547
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2552
list_for_each_entry(obj, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2554
.head = obj->head,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2622
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2627
list_for_each_entry(obj, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2629
.head = obj->head,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
276
struct ras_manager *obj = (struct ras_manager *)file_inode(f)->i_private;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
278
.head = obj->head,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
283
if (amdgpu_ras_query_error_status(obj->adev, &info))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
287
if (amdgpu_ip_version(obj->adev, MP0_HWIP, 0) != IP_VERSION(11, 0, 2) &&
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
288
amdgpu_ip_version(obj->adev, MP0_HWIP, 0) != IP_VERSION(11, 0, 4)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
289
if (amdgpu_ras_reset_error_status(obj->adev, info.head.block))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
290
dev_warn(obj->adev->dev, "Failed to reset error counter and error status");
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4397
struct ras_manager *obj, *tmp;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4418
list_for_each_entry_safe(obj, tmp, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4419
if (!amdgpu_ras_is_supported(adev, obj->head.block)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4420
amdgpu_ras_feature_enable(adev, &obj->head, 0);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4422
WARN_ON(alive_obj(obj));
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4444
struct amdgpu_ras_block_object *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4472
obj = node->ras_obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4473
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4478
if (!amdgpu_ras_is_supported(adev, obj->ras_comm.block))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4481
if (obj->ras_late_init) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4482
r = obj->ras_late_init(adev, &obj->ras_comm);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4485
obj->ras_comm.name, r);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4489
amdgpu_ras_block_late_init_default(adev, &obj->ras_comm);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4514
struct amdgpu_ras_block_object *obj = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4525
obj = ras_node->ras_obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4526
if (amdgpu_ras_is_supported(adev, obj->ras_comm.block) &&
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4527
obj->ras_fini)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4528
obj->ras_fini(adev, &obj->ras_comm);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4530
amdgpu_ras_block_late_fini_default(adev, &obj->ras_comm);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
697
struct ras_manager *obj = container_of(attr, struct ras_manager, sysfs_attr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
699
.head = obj->head,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
702
if (!amdgpu_ras_get_error_query_ready(obj->adev))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
705
if (amdgpu_ras_query_error_status(obj->adev, &info))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
708
if (amdgpu_ip_version(obj->adev, MP0_HWIP, 0) != IP_VERSION(11, 0, 2) &&
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
709
amdgpu_ip_version(obj->adev, MP0_HWIP, 0) != IP_VERSION(11, 0, 4)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
710
if (amdgpu_ras_reset_error_status(obj->adev, info.head.block))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
711
dev_warn(obj->adev->dev, "Failed to reset error counter and error status");
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
726
#define get_obj(obj) do { (obj)->use++; } while (0)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
727
#define alive_obj(obj) ((obj)->use)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
729
static inline void put_obj(struct ras_manager *obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
731
if (obj && (--obj->use == 0)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
732
list_del(&obj->node);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
733
amdgpu_ras_error_data_fini(&obj->err_data);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
736
if (obj && (obj->use < 0))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
737
DRM_ERROR("RAS ERROR: Unbalance obj(%s) use\n", get_ras_block_str(&obj->head));
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
745
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
757
obj = &con->objs[AMDGPU_RAS_BLOCK__LAST + head->sub_block_index];
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
759
obj = &con->objs[head->block];
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
762
if (alive_obj(obj))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
765
if (amdgpu_ras_error_data_init(&obj->err_data))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
768
obj->head = *head;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
769
obj->adev = adev;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
770
list_add(&obj->node, &con->head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
771
get_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
773
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
781
struct ras_manager *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
795
obj = &con->objs[AMDGPU_RAS_BLOCK__LAST + head->sub_block_index];
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
797
obj = &con->objs[head->block];
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
799
if (alive_obj(obj))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
800
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
803
obj = &con->objs[i];
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
804
if (alive_obj(obj))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
805
return obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
836
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
848
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
849
obj = amdgpu_ras_create_obj(adev, head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
850
if (!obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
854
get_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
858
if (obj && amdgpu_ras_is_feature_enabled(adev, head)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
860
put_obj(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
980
struct ras_manager *obj, *tmp;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
982
list_for_each_entry_safe(obj, tmp, &con->head, node) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
987
if (__amdgpu_ras_feature_enable(adev, &obj->head, 0))
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
990
if (amdgpu_ras_feature_enable(adev, &obj->head, 0))
sys/dev/pci/drm/amd/amdgpu/amdgpu_umc.c
232
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &head);
sys/dev/pci/drm/amd/amdgpu/amdgpu_umc.c
240
if (ret == AMDGPU_RAS_SUCCESS && obj) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_umc.c
241
obj->err_data.ue_count += err_data.ue_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_umc.c
242
obj->err_data.ce_count += err_data.ce_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_umc.c
243
obj->err_data.de_count += err_data.de_count;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
261
r = amdgpu_bo_create(adev, &bp, &userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
267
r = amdgpu_bo_reserve(userq_obj->obj, true);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
273
r = amdgpu_ttm_alloc_gart(&(userq_obj->obj)->tbo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
279
r = amdgpu_bo_kmap(userq_obj->obj, &userq_obj->cpu_ptr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
285
userq_obj->gpu_addr = amdgpu_bo_gpu_offset(userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
286
amdgpu_bo_unreserve(userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
291
amdgpu_bo_unreserve(userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
294
amdgpu_bo_unref(&userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
301
amdgpu_bo_kunmap(userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
302
amdgpu_bo_unref(&userq_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
321
db_obj->obj = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj));
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
324
r = amdgpu_bo_reserve(db_obj->obj, true);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
331
r = amdgpu_bo_pin(db_obj->obj, AMDGPU_GEM_DOMAIN_DOORBELL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
361
index = amdgpu_doorbell_index_on_bar(uq_mgr->adev, db_obj->obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
365
amdgpu_bo_unreserve(db_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
369
amdgpu_bo_unpin(db_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
371
amdgpu_bo_unreserve(db_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
373
amdgpu_bo_unref(&db_obj->obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
396
r = amdgpu_bo_reserve(queue->db_obj.obj, true);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
398
amdgpu_bo_unpin(queue->db_obj.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
399
amdgpu_bo_unreserve(queue->db_obj.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
401
amdgpu_bo_unref(&queue->db_obj.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
443
if (!queue || !queue->mqd.obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
446
bo = amdgpu_bo_ref(queue->mqd.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
454
seq_printf(m, "mqd_gpu_address: 0x%llx\n", amdgpu_bo_gpu_offset(queue->mqd.obj));
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
47
struct amdgpu_bo *obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_virt.c
252
unsigned int amd_sriov_msg_checksum(void *obj,
sys/dev/pci/drm/amd/amdgpu/amdgpu_virt.c
261
pos = (char *)obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1356
struct drm_gem_object *obj = &bo->tbo.base;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1359
if (drm_gem_is_imported(obj) && bo_va->is_xgmi) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1360
struct dma_buf *dma_buf = obj->import_attach->dmabuf;
sys/dev/pci/drm/amd/amdgpu/amdgv_sriovmsg.h
430
unsigned int amd_sriov_msg_checksum(void *obj, unsigned long obj_size, unsigned int key,
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
1847
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
1871
obj = target_fb->obj[0];
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
1872
abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2072
abo = gem_to_amdgpu_bo(fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2369
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2376
obj = NULL;
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2386
obj = drm_gem_object_lookup(file_priv, handle);
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2387
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2392
aobj = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2395
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2404
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2442
amdgpu_crtc->cursor_bo = obj;
sys/dev/pci/drm/amd/amdgpu/dce_v10_0.c
2558
abo = gem_to_amdgpu_bo(crtc->primary->fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
1885
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
1908
obj = target_fb->obj[0];
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
1909
abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2087
abo = gem_to_amdgpu_bo(fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2342
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2349
obj = NULL;
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2359
obj = drm_gem_object_lookup(file_priv, handle);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2360
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2365
aobj = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2368
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2377
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2415
amdgpu_crtc->cursor_bo = obj;
sys/dev/pci/drm/amd/amdgpu/dce_v6_0.c
2526
abo = gem_to_amdgpu_bo(crtc->primary->fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
1794
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
1818
obj = target_fb->obj[0];
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
1819
abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
1999
abo = gem_to_amdgpu_bo(fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2288
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2295
obj = NULL;
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2305
obj = drm_gem_object_lookup(file_priv, handle);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2306
if (!obj) {
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2311
aobj = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2314
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2323
drm_gem_object_put(obj);
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2361
amdgpu_crtc->cursor_bo = obj;
sys/dev/pci/drm/amd/amdgpu/dce_v8_0.c
2477
abo = gem_to_amdgpu_bo(crtc->primary->fb->obj[0]);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
84
wptr_obj->obj = wptr_mapping->bo_va->base.bo;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
85
if (wptr_obj->obj->tbo.base.size > PAGE_SIZE) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
90
ret = mes_userq_map_gtt_bo_to_gart(wptr_obj->obj);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
96
queue->wptr_obj.gpu_addr = amdgpu_bo_gpu_offset_no_check(wptr_obj->obj);
sys/dev/pci/drm/amd/amdgpu/nbio_v7_4.c
362
struct ras_manager *obj = amdgpu_ras_find_obj(adev, adev->nbio.ras_if);
sys/dev/pci/drm/amd/amdgpu/nbio_v7_4.c
385
if (ras && !ras->disable_ras_err_cnt_harvest && obj) {
sys/dev/pci/drm/amd/amdgpu/nbio_v7_4.c
394
obj->err_data.ue_count += err_data.ue_count;
sys/dev/pci/drm/amd/amdgpu/nbio_v7_4.c
395
obj->err_data.ce_count += err_data.ce_count;
sys/dev/pci/drm/amd/amdgpu/nbio_v7_4.c
400
obj->err_data.ce_count,
sys/dev/pci/drm/amd/amdgpu/nbio_v7_4.c
406
obj->err_data.ue_count,
sys/dev/pci/drm/amd/amdgpu/nbio_v7_9.c
519
struct ras_manager *obj = amdgpu_ras_find_obj(adev, adev->nbio.ras_if);
sys/dev/pci/drm/amd/amdgpu/nbio_v7_9.c
545
obj->err_data.ue_count += err_data.ue_count;
sys/dev/pci/drm/amd/amdgpu/nbio_v7_9.c
546
obj->err_data.ce_count += err_data.ce_count;
sys/dev/pci/drm/amd/amdgpu/nbio_v7_9.c
551
obj->err_data.ce_count,
sys/dev/pci/drm/amd/amdgpu/nbio_v7_9.c
557
obj->err_data.ue_count,
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
12700
struct drm_private_obj *obj = state->private_objs[i].ptr;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
12702
if (obj->funcs == adev->dm.atomic_obj.funcs) {
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
12705
dm_atomic_destroy_state(obj,
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4784
struct drm_private_obj *obj;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4788
for_each_new_private_obj_in_state(state, obj, new_obj_state, i) {
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4789
if (obj->funcs == dm->atomic_obj.funcs)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4797
dm_atomic_duplicate_state(struct drm_private_obj *obj)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4805
__drm_atomic_helper_private_obj_duplicate_state(obj, &new_state->base);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4807
old_state = to_dm_atomic_state(obj->state);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4820
static void dm_atomic_destroy_state(struct drm_private_obj *obj,
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
9566
struct amdgpu_bo *abo = gem_to_amdgpu_bo(fb->obj[0]);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_plane.c
1033
rbo = gem_to_amdgpu_bo(old_state->fb->obj[0]);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_plane.c
928
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_plane.c
941
obj = drm_gem_fb_get_obj(new_state->fb, 0);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_plane.c
942
if (!obj) {
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_plane.c
947
rbo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_wb.c
100
rbo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_wb.c
156
rbo = gem_to_amdgpu_bo(job->fb->obj[0]);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_wb.c
87
struct drm_gem_object *obj;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_wb.c
99
obj = job->fb->obj[0];
sys/dev/pci/drm/apple/iomfb.h
288
char obj[4];
sys/dev/pci/drm/apple/iomfb.h
302
char obj[4];
sys/dev/pci/drm/apple/iomfb.h
310
char obj[4];
sys/dev/pci/drm/apple/iomfb_template.c
1292
struct drm_gem_dma_object *obj;
sys/dev/pci/drm/apple/iomfb_template.c
1353
obj = drm_fb_dma_get_gem_obj(fb, 0);
sys/dev/pci/drm/apple/iomfb_template.c
1354
if (obj)
sys/dev/pci/drm/apple/iomfb_template.c
1355
req->surf_iova[l] = obj->dma_addr + fb->offsets[0];
sys/dev/pci/drm/apple/iomfb_template.c
223
memcmp(req->obj, "SUMP", sizeof(req->obj)) == 0) { /* "PMUS */
sys/dev/pci/drm/apple/iomfb_template.c
240
if (memcmp(req->obj, "FMOI", sizeof(req->obj)) == 0) { /* "IOMF */
sys/dev/pci/drm/apple/iomfb_template.h
148
char obj[4];
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5133
drm_dp_mst_duplicate_state(struct drm_private_obj *obj)
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5136
to_dp_mst_topology_state(obj->state);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5143
__drm_atomic_helper_private_obj_duplicate_state(obj, &state->base);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5175
static void drm_dp_mst_destroy_state(struct drm_private_obj *obj,
sys/dev/pci/drm/display/drm_dp_tunnel.c
1456
tunnel_group_duplicate_state(struct drm_private_obj *obj)
sys/dev/pci/drm/display/drm_dp_tunnel.c
1467
__drm_atomic_helper_private_obj_duplicate_state(obj, &group_state->base);
sys/dev/pci/drm/display/drm_dp_tunnel.c
1469
for_each_tunnel_state(to_group_state(obj->state), tunnel_state) {
sys/dev/pci/drm/display/drm_dp_tunnel.c
1495
static void tunnel_group_destroy_state(struct drm_private_obj *obj, struct drm_private_state *state)
sys/dev/pci/drm/display/drm_dp_tunnel.c
1551
if (to_group(old_group_state->base.obj) == tunnel->group)
sys/dev/pci/drm/display/drm_dp_tunnel.c
1575
if (to_group(new_group_state->base.obj) == tunnel->group)
sys/dev/pci/drm/display/drm_dp_tunnel.c
1614
drm_WARN(to_group(group_state->base.obj)->mgr->dev,
sys/dev/pci/drm/display/drm_dp_tunnel.c
1802
struct drm_dp_tunnel_group *group = to_group(new_group_state->base.obj);
sys/dev/pci/drm/dma-resv.c
138
void dma_resv_init(struct dma_resv *obj)
sys/dev/pci/drm/dma-resv.c
140
ww_mutex_init(&obj->lock, &reservation_ww_class);
sys/dev/pci/drm/dma-resv.c
142
RCU_INIT_POINTER(obj->fences, NULL);
sys/dev/pci/drm/dma-resv.c
150
void dma_resv_fini(struct dma_resv *obj)
sys/dev/pci/drm/dma-resv.c
156
dma_resv_list_free(rcu_dereference_protected(obj->fences, true));
sys/dev/pci/drm/dma-resv.c
157
ww_mutex_destroy(&obj->lock);
sys/dev/pci/drm/dma-resv.c
162
static inline struct dma_resv_list *dma_resv_fences_list(struct dma_resv *obj)
sys/dev/pci/drm/dma-resv.c
164
return rcu_dereference_check(obj->fences, dma_resv_held(obj));
sys/dev/pci/drm/dma-resv.c
182
int dma_resv_reserve_fences(struct dma_resv *obj, unsigned int num_fences)
sys/dev/pci/drm/dma-resv.c
187
dma_resv_assert_held(obj);
sys/dev/pci/drm/dma-resv.c
196
old = dma_resv_fences_list(obj);
sys/dev/pci/drm/dma-resv.c
219
dma_resv_list_entry(old, i, obj, &fence, &usage);
sys/dev/pci/drm/dma-resv.c
235
rcu_assign_pointer(obj->fences, new);
sys/dev/pci/drm/dma-resv.c
245
dma_resv_held(obj));
sys/dev/pci/drm/dma-resv.c
263
void dma_resv_reset_max_fences(struct dma_resv *obj)
sys/dev/pci/drm/dma-resv.c
265
struct dma_resv_list *fences = dma_resv_fences_list(obj);
sys/dev/pci/drm/dma-resv.c
267
dma_resv_assert_held(obj);
sys/dev/pci/drm/dma-resv.c
287
void dma_resv_add_fence(struct dma_resv *obj, struct dma_fence *fence,
sys/dev/pci/drm/dma-resv.c
296
dma_resv_assert_held(obj);
sys/dev/pci/drm/dma-resv.c
303
fobj = dma_resv_fences_list(obj);
sys/dev/pci/drm/dma-resv.c
309
dma_resv_list_entry(fobj, i, obj, &old, &old_usage);
sys/dev/pci/drm/dma-resv.c
343
void dma_resv_replace_fences(struct dma_resv *obj, uint64_t context,
sys/dev/pci/drm/dma-resv.c
350
dma_resv_assert_held(obj);
sys/dev/pci/drm/dma-resv.c
352
list = dma_resv_fences_list(obj);
sys/dev/pci/drm/dma-resv.c
356
dma_resv_list_entry(list, i, obj, &old, NULL);
sys/dev/pci/drm/dma-resv.c
371
cursor->fences = dma_resv_fences_list(cursor->obj);
sys/dev/pci/drm/dma-resv.c
394
cursor->obj, &cursor->fence,
sys/dev/pci/drm/dma-resv.c
426
} while (dma_resv_fences_list(cursor->obj) != cursor->fences);
sys/dev/pci/drm/dma-resv.c
449
restart = dma_resv_fences_list(cursor->obj) != cursor->fences;
sys/dev/pci/drm/dma-resv.c
455
} while (dma_resv_fences_list(cursor->obj) != cursor->fences);
sys/dev/pci/drm/dma-resv.c
475
dma_resv_assert_held(cursor->obj);
sys/dev/pci/drm/dma-resv.c
478
cursor->fences = dma_resv_fences_list(cursor->obj);
sys/dev/pci/drm/dma-resv.c
497
dma_resv_assert_held(cursor->obj);
sys/dev/pci/drm/dma-resv.c
507
cursor->obj, &fence, &cursor->fence_usage);
sys/dev/pci/drm/dma-resv.c
569
int dma_resv_get_fences(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/dma-resv.c
579
dma_resv_iter_begin(&cursor, obj, usage);
sys/dev/pci/drm/dma-resv.c
639
int dma_resv_get_singleton(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/dma-resv.c
647
r = dma_resv_get_fences(obj, usage, &count, &fences);
sys/dev/pci/drm/dma-resv.c
690
long dma_resv_wait_timeout(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/dma-resv.c
697
dma_resv_iter_begin(&cursor, obj, usage);
sys/dev/pci/drm/dma-resv.c
723
void dma_resv_set_deadline(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/dma-resv.c
729
dma_resv_iter_begin(&cursor, obj, usage);
sys/dev/pci/drm/dma-resv.c
750
bool dma_resv_test_signaled(struct dma_resv *obj, enum dma_resv_usage usage)
sys/dev/pci/drm/dma-resv.c
755
dma_resv_iter_begin(&cursor, obj, usage);
sys/dev/pci/drm/dma-resv.c
773
void dma_resv_describe(struct dma_resv *obj, struct seq_file *seq)
sys/dev/pci/drm/dma-resv.c
781
dma_resv_for_each_fence(&cursor, obj, DMA_RESV_USAGE_READ, fence) {
sys/dev/pci/drm/dma-resv.c
795
struct dma_resv obj;
sys/dev/pci/drm/dma-resv.c
802
dma_resv_init(&obj);
sys/dev/pci/drm/dma-resv.c
807
ret = dma_resv_lock(&obj, &ctx);
sys/dev/pci/drm/dma-resv.c
809
dma_resv_lock_slow(&obj, &ctx);
sys/dev/pci/drm/dma-resv.c
822
ww_mutex_unlock(&obj.lock);
sys/dev/pci/drm/drm_atomic.c
1800
struct drm_private_obj *obj = state->obj;
sys/dev/pci/drm/drm_atomic.c
1802
if (obj->funcs->atomic_print_state)
sys/dev/pci/drm/drm_atomic.c
1803
obj->funcs->atomic_print_state(p, state);
sys/dev/pci/drm/drm_atomic.c
1826
struct drm_private_obj *obj;
sys/dev/pci/drm/drm_atomic.c
1846
for_each_new_private_obj_in_state(state, obj, obj_state, i)
sys/dev/pci/drm/drm_atomic.c
1859
struct drm_private_obj *obj;
sys/dev/pci/drm/drm_atomic.c
1889
list_for_each_entry(obj, &config->privobj_list, head) {
sys/dev/pci/drm/drm_atomic.c
1891
drm_modeset_lock(&obj->lock, NULL);
sys/dev/pci/drm/drm_atomic.c
1892
drm_atomic_private_obj_print_state(p, obj->state);
sys/dev/pci/drm/drm_atomic.c
1894
drm_modeset_unlock(&obj->lock);
sys/dev/pci/drm/drm_atomic.c
253
struct drm_private_obj *obj = state->private_objs[i].ptr;
sys/dev/pci/drm/drm_atomic.c
255
obj->funcs->atomic_destroy_state(obj,
sys/dev/pci/drm/drm_atomic.c
787
struct drm_private_obj *obj,
sys/dev/pci/drm/drm_atomic.c
791
memset(obj, 0, sizeof(*obj));
sys/dev/pci/drm/drm_atomic.c
793
drm_modeset_lock_init(&obj->lock);
sys/dev/pci/drm/drm_atomic.c
795
obj->state = state;
sys/dev/pci/drm/drm_atomic.c
796
obj->funcs = funcs;
sys/dev/pci/drm/drm_atomic.c
797
list_add_tail(&obj->head, &dev->mode_config.privobj_list);
sys/dev/pci/drm/drm_atomic.c
799
state->obj = obj;
sys/dev/pci/drm/drm_atomic.c
810
drm_atomic_private_obj_fini(struct drm_private_obj *obj)
sys/dev/pci/drm/drm_atomic.c
812
list_del(&obj->head);
sys/dev/pci/drm/drm_atomic.c
813
obj->funcs->atomic_destroy_state(obj, obj->state);
sys/dev/pci/drm/drm_atomic.c
814
drm_modeset_lock_fini(&obj->lock);
sys/dev/pci/drm/drm_atomic.c
832
struct drm_private_obj *obj)
sys/dev/pci/drm/drm_atomic.c
840
if (obj == state->private_objs[i].ptr)
sys/dev/pci/drm/drm_atomic.c
843
ret = drm_modeset_lock(&obj->lock, state->acquire_ctx);
sys/dev/pci/drm/drm_atomic.c
866
obj_state = obj->funcs->atomic_duplicate_state(obj);
sys/dev/pci/drm/drm_atomic.c
871
state->private_objs[index].old_state = obj->state;
sys/dev/pci/drm/drm_atomic.c
873
state->private_objs[index].ptr = obj;
sys/dev/pci/drm/drm_atomic.c
880
obj, obj_state, state);
sys/dev/pci/drm/drm_atomic.c
896
struct drm_private_obj *obj)
sys/dev/pci/drm/drm_atomic.c
901
if (obj == state->private_objs[i].ptr)
sys/dev/pci/drm/drm_atomic.c
918
struct drm_private_obj *obj)
sys/dev/pci/drm/drm_atomic.c
923
if (obj == state->private_objs[i].ptr)
sys/dev/pci/drm/drm_atomic_helper.c
3268
struct drm_private_obj *obj;
sys/dev/pci/drm/drm_atomic_helper.c
3356
for_each_oldnew_private_obj_in_state(state, obj, old_obj_state, new_obj_state, i) {
sys/dev/pci/drm/drm_atomic_helper.c
3357
WARN_ON(obj->state != old_obj_state);
sys/dev/pci/drm/drm_atomic_helper.c
3363
obj->state = new_obj_state;
sys/dev/pci/drm/drm_atomic_state_helper.c
720
void __drm_atomic_helper_private_obj_duplicate_state(struct drm_private_obj *obj,
sys/dev/pci/drm/drm_atomic_state_helper.c
723
memcpy(state, obj->state, sizeof(*state));
sys/dev/pci/drm/drm_atomic_uapi.c
1013
struct drm_mode_object *obj,
sys/dev/pci/drm/drm_atomic_uapi.c
1025
switch (obj->type) {
sys/dev/pci/drm/drm_atomic_uapi.c
1027
struct drm_connector *connector = obj_to_connector(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1049
struct drm_crtc *crtc = obj_to_crtc(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1070
struct drm_plane *plane = obj_to_plane(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1102
obj->id, plane->name);
sys/dev/pci/drm/drm_atomic_uapi.c
1114
drm_dbg_atomic(prop->dev, "[OBJECT:%d] has no properties\n", obj->id);
sys/dev/pci/drm/drm_atomic_uapi.c
1479
struct drm_mode_object *obj;
sys/dev/pci/drm/drm_atomic_uapi.c
1486
obj = drm_mode_object_find(dev, file_priv, obj_id, DRM_MODE_OBJECT_ANY);
sys/dev/pci/drm/drm_atomic_uapi.c
1487
if (!obj) {
sys/dev/pci/drm/drm_atomic_uapi.c
1493
if (!obj->properties) {
sys/dev/pci/drm/drm_atomic_uapi.c
1495
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1501
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1514
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1519
prop = drm_mode_obj_find_prop_id(obj, prop_id);
sys/dev/pci/drm/drm_atomic_uapi.c
1524
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1532
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1537
ret = drm_atomic_set_property(state, file_priv, obj,
sys/dev/pci/drm/drm_atomic_uapi.c
1540
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
1547
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
883
int drm_atomic_get_property(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_atomic_uapi.c
889
switch (obj->type) {
sys/dev/pci/drm/drm_atomic_uapi.c
891
struct drm_connector *connector = obj_to_connector(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
899
struct drm_crtc *crtc = obj_to_crtc(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
907
struct drm_plane *plane = obj_to_plane(obj);
sys/dev/pci/drm/drm_atomic_uapi.c
915
drm_dbg_atomic(dev, "[OBJECT:%d] has no properties\n", obj->id);
sys/dev/pci/drm/drm_bridge.c
369
drm_bridge_atomic_duplicate_priv_state(struct drm_private_obj *obj)
sys/dev/pci/drm/drm_bridge.c
371
struct drm_bridge *bridge = drm_priv_to_bridge(obj);
sys/dev/pci/drm/drm_bridge.c
379
drm_bridge_atomic_destroy_priv_state(struct drm_private_obj *obj,
sys/dev/pci/drm/drm_bridge.c
383
struct drm_bridge *bridge = drm_priv_to_bridge(obj);
sys/dev/pci/drm/drm_client.c
193
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_client.c
209
obj = drm_gem_object_lookup(client->file, dumb_args.handle);
sys/dev/pci/drm/drm_client.c
210
if (!obj) {
sys/dev/pci/drm/drm_client.c
216
buffer->gem = obj;
sys/dev/pci/drm/drm_connector.c
3245
int drm_connector_set_obj_prop(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_connector.c
3250
struct drm_connector *connector = obj_to_connector(obj);
sys/dev/pci/drm/drm_crtc.c
897
int drm_mode_crtc_set_obj_prop(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc.c
902
struct drm_crtc *crtc = obj_to_crtc(obj);
sys/dev/pci/drm/drm_crtc.c
907
drm_object_property_set_value(obj, property, value);
sys/dev/pci/drm/drm_crtc_internal.h
153
int __drm_mode_object_add(struct drm_device *dev, struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc_internal.h
156
int drm_mode_object_add(struct drm_device *dev, struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc_internal.h
159
struct drm_mode_object *obj);
sys/dev/pci/drm/drm_crtc_internal.h
165
int drm_mode_object_get_properties(struct drm_mode_object *obj, bool atomic,
sys/dev/pci/drm/drm_crtc_internal.h
169
struct drm_property *drm_mode_obj_find_prop_id(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc_internal.h
192
int drm_connector_set_obj_prop(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc_internal.h
261
struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc_internal.h
264
int drm_atomic_get_property(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_crtc_internal.h
72
int drm_mode_crtc_set_obj_prop(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_debugfs.c
127
struct drm_gem_object *obj = ptr;
sys/dev/pci/drm/drm_debugfs.c
131
obj->name, obj->size,
sys/dev/pci/drm/drm_debugfs.c
132
obj->handle_count,
sys/dev/pci/drm/drm_debugfs.c
133
kref_read(&obj->refcount));
sys/dev/pci/drm/drm_debugfs.c
230
(u64)(uintptr_t)va->gem.obj, va->gem.offset);
sys/dev/pci/drm/drm_exec.c
144
struct drm_gem_object *obj)
sys/dev/pci/drm/drm_exec.c
165
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_exec.c
166
exec->objects[exec->num_objects++] = obj;
sys/dev/pci/drm/drm_exec.c
174
struct drm_gem_object *obj = exec->contended;
sys/dev/pci/drm/drm_exec.c
177
if (likely(!obj))
sys/dev/pci/drm/drm_exec.c
183
ret = dma_resv_lock_slow_interruptible(obj->resv,
sys/dev/pci/drm/drm_exec.c
188
dma_resv_lock_slow(obj->resv, &exec->ticket);
sys/dev/pci/drm/drm_exec.c
191
ret = drm_exec_obj_locked(exec, obj);
sys/dev/pci/drm/drm_exec.c
195
exec->prelocked = obj;
sys/dev/pci/drm/drm_exec.c
199
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_exec.c
202
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_exec.c
217
int drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
sys/dev/pci/drm/drm_exec.c
225
if (exec->prelocked == obj) {
sys/dev/pci/drm/drm_exec.c
232
ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket);
sys/dev/pci/drm/drm_exec.c
234
ret = dma_resv_lock(obj->resv, &exec->ticket);
sys/dev/pci/drm/drm_exec.c
237
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_exec.c
238
exec->contended = obj;
sys/dev/pci/drm/drm_exec.c
249
ret = drm_exec_obj_locked(exec, obj);
sys/dev/pci/drm/drm_exec.c
256
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_exec.c
270
void drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
sys/dev/pci/drm/drm_exec.c
275
if (exec->objects[i] == obj) {
sys/dev/pci/drm/drm_exec.c
276
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_exec.c
280
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_exec.c
299
int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
sys/dev/pci/drm/drm_exec.c
304
ret = drm_exec_lock_obj(exec, obj);
sys/dev/pci/drm/drm_exec.c
308
ret = dma_resv_reserve_fences(obj->resv, num_fences);
sys/dev/pci/drm/drm_exec.c
310
drm_exec_unlock_obj(exec, obj);
sys/dev/pci/drm/drm_exec.c
58
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_exec.c
61
drm_exec_for_each_locked_object_reverse(exec, index, obj) {
sys/dev/pci/drm/drm_exec.c
62
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_exec.c
63
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_fb_dma_helper.c
10
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_fb_dma_helper.c
13
obj = fb->obj[plane];
sys/dev/pci/drm/drm_fb_dma_helper.c
14
if (obj)
sys/dev/pci/drm/drm_fb_dma_helper.c
15
return to_drm_gem_dma_obj(obj);
sys/dev/pci/drm/drm_fbdev_shmem.c
46
struct drm_gem_object *obj = drm_gem_fb_get_obj(fb, 0);
sys/dev/pci/drm/drm_fbdev_shmem.c
47
struct drm_gem_shmem_object *shmem = to_drm_gem_shmem_obj(obj);
sys/dev/pci/drm/drm_fbdev_shmem.c
87
struct drm_gem_object *obj = drm_gem_fb_get_obj(fb, 0);
sys/dev/pci/drm/drm_fbdev_shmem.c
88
struct drm_gem_shmem_object *shmem = to_drm_gem_shmem_obj(obj);
sys/dev/pci/drm/drm_fbdev_shmem.c
92
if (fb_WARN_ON_ONCE(info, offset > obj->size))
sys/dev/pci/drm/drm_file.c
956
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_file.c
962
idr_for_each_entry (&file->object_idr, obj, id) {
sys/dev/pci/drm/drm_file.c
964
size_t add_size = (obj->funcs && obj->funcs->rss) ?
sys/dev/pci/drm/drm_file.c
965
obj->funcs->rss(obj) : obj->size;
sys/dev/pci/drm/drm_file.c
967
if (obj->funcs && obj->funcs->status) {
sys/dev/pci/drm/drm_file.c
968
s = obj->funcs->status(obj);
sys/dev/pci/drm/drm_file.c
972
if (drm_gem_object_is_shared_for_memory_stats(obj))
sys/dev/pci/drm/drm_file.c
973
status.shared += obj->size;
sys/dev/pci/drm/drm_file.c
975
status.private += obj->size;
sys/dev/pci/drm/drm_file.c
986
if (!dma_resv_test_signaled(obj->resv, dma_resv_usage_rw(true))) {
sys/dev/pci/drm/drm_framebuffer.c
1209
fb->obj[i] ? "" : "(null)");
sys/dev/pci/drm/drm_framebuffer.c
1210
if (fb->obj[i])
sys/dev/pci/drm/drm_framebuffer.c
1211
drm_gem_print_info(p, indent + 2, fb->obj[i]);
sys/dev/pci/drm/drm_framebuffer.c
602
if (!fb->obj[0] &&
sys/dev/pci/drm/drm_framebuffer.c
647
if (fb->obj[i] == fb->obj[j]) {
sys/dev/pci/drm/drm_framebuffer.c
656
if (fb->obj[i]) {
sys/dev/pci/drm/drm_framebuffer.c
657
ret = drm_gem_handle_create(file_priv, fb->obj[i],
sys/dev/pci/drm/drm_framebuffer.c
876
if (fb->obj[i]) {
sys/dev/pci/drm/drm_framebuffer.c
877
exists = drm_gem_object_handle_get_if_exists_unlocked(fb->obj[i]);
sys/dev/pci/drm/drm_framebuffer.c
909
drm_gem_object_handle_put_unlocked(fb->obj[i]);
sys/dev/pci/drm/drm_framebuffer.c
931
struct drm_mode_object *obj;
sys/dev/pci/drm/drm_framebuffer.c
934
obj = __drm_mode_object_find(dev, file_priv, id, DRM_MODE_OBJECT_FB);
sys/dev/pci/drm/drm_framebuffer.c
935
if (obj)
sys/dev/pci/drm/drm_framebuffer.c
936
fb = obj_to_fb(obj);
sys/dev/pci/drm/drm_framebuffer.c
993
drm_gem_object_handle_put_unlocked(fb->obj[i]);
sys/dev/pci/drm/drm_gem.c
1028
struct drm_gem_object *obj = NULL;
sys/dev/pci/drm/drm_gem.c
1030
objects_lookup(filp, &handle, 1, &obj);
sys/dev/pci/drm/drm_gem.c
1031
return obj;
sys/dev/pci/drm/drm_gem.c
1051
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
1053
obj = drm_gem_object_lookup(filep, handle);
sys/dev/pci/drm/drm_gem.c
1054
if (!obj) {
sys/dev/pci/drm/drm_gem.c
1059
ret = dma_resv_wait_timeout(obj->resv, dma_resv_usage_rw(wait_all),
sys/dev/pci/drm/drm_gem.c
1066
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1092
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
1098
obj = drm_gem_object_lookup(file_priv, args->handle);
sys/dev/pci/drm/drm_gem.c
1099
if (obj == NULL)
sys/dev/pci/drm/drm_gem.c
1104
if (obj->handle_count == 0) {
sys/dev/pci/drm/drm_gem.c
1109
if (!obj->name) {
sys/dev/pci/drm/drm_gem.c
1110
ret = idr_alloc(&dev->object_name_idr, obj, 1, 0, GFP_KERNEL);
sys/dev/pci/drm/drm_gem.c
1114
obj->name = ret;
sys/dev/pci/drm/drm_gem.c
1117
args->name = (uint64_t) obj->name;
sys/dev/pci/drm/drm_gem.c
1122
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1131
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
1139
obj = idr_find(&dev->object_name_idr, (int) args->name);
sys/dev/pci/drm/drm_gem.c
1140
if (obj) {
sys/dev/pci/drm/drm_gem.c
1141
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
1148
ret = drm_gem_handle_create_tail(file_priv, obj, &handle);
sys/dev/pci/drm/drm_gem.c
1153
args->size = obj->size;
sys/dev/pci/drm/drm_gem.c
1156
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1164
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
1175
obj = drm_gem_object_lookup(file_priv, args->handle);
sys/dev/pci/drm/drm_gem.c
1176
if (!obj)
sys/dev/pci/drm/drm_gem.c
1187
ret = idr_alloc(&file_priv->object_idr, obj, handle, handle + 1,
sys/dev/pci/drm/drm_gem.c
1194
if (obj->dma_buf) {
sys/dev/pci/drm/drm_gem.c
1195
ret = drm_prime_add_buf_handle(&file_priv->prime, obj->dma_buf,
sys/dev/pci/drm/drm_gem.c
1216
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1261
drm_gem_object_release(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1264
if (obj->filp)
sys/dev/pci/drm/drm_gem.c
1265
fput(obj->filp);
sys/dev/pci/drm/drm_gem.c
1267
if (obj->uao)
sys/dev/pci/drm/drm_gem.c
1268
uao_detach(obj->uao);
sys/dev/pci/drm/drm_gem.c
1269
if (obj->uobj.pgops)
sys/dev/pci/drm/drm_gem.c
1270
uvm_obj_destroy(&obj->uobj);
sys/dev/pci/drm/drm_gem.c
1273
drm_gem_private_object_fini(obj);
sys/dev/pci/drm/drm_gem.c
1275
drm_gem_free_mmap_offset(obj);
sys/dev/pci/drm/drm_gem.c
1276
drm_gem_lru_remove(obj);
sys/dev/pci/drm/drm_gem.c
1291
struct drm_gem_object *obj =
sys/dev/pci/drm/drm_gem.c
1294
if (WARN_ON(!obj->funcs->free))
sys/dev/pci/drm/drm_gem.c
1297
obj->funcs->free(obj);
sys/dev/pci/drm/drm_gem.c
1311
struct drm_gem_object *obj = vma->vm_private_data;
sys/dev/pci/drm/drm_gem.c
1313
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
1326
struct drm_gem_object *obj = vma->vm_private_data;
sys/dev/pci/drm/drm_gem.c
1328
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
134
ret = dev->driver->gem_fault(obj, ufi, entry->offset + (vaddr -
sys/dev/pci/drm/drm_gem.c
1356
int drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size,
sys/dev/pci/drm/drm_gem.c
1371
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
1373
vma->vm_private_data = obj;
sys/dev/pci/drm/drm_gem.c
1374
vma->vm_ops = obj->funcs->vm_ops;
sys/dev/pci/drm/drm_gem.c
1376
if (obj->funcs->mmap) {
sys/dev/pci/drm/drm_gem.c
1377
ret = obj->funcs->mmap(obj, vma);
sys/dev/pci/drm/drm_gem.c
1395
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1419
struct drm_gem_object *obj = NULL;
sys/dev/pci/drm/drm_gem.c
1431
obj = container_of(node, struct drm_gem_object, vma_node);
sys/dev/pci/drm/drm_gem.c
1442
if (!kref_get_unless_zero(&obj->refcount))
sys/dev/pci/drm/drm_gem.c
1443
obj = NULL;
sys/dev/pci/drm/drm_gem.c
1447
if (!obj)
sys/dev/pci/drm/drm_gem.c
1451
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1455
ret = drm_gem_mmap_obj(obj, drm_vma_node_size(node) << PAGE_SHIFT,
sys/dev/pci/drm/drm_gem.c
1458
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1465
int drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size,
sys/dev/pci/drm/drm_gem.c
1480
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
1483
vma->vm_private_data = obj;
sys/dev/pci/drm/drm_gem.c
1484
vma->vm_ops = obj->funcs->vm_ops;
sys/dev/pci/drm/drm_gem.c
1486
if (obj->uobj.pgops == NULL)
sys/dev/pci/drm/drm_gem.c
1487
uvm_obj_init(&obj->uobj, obj->funcs->vm_ops, 1);
sys/dev/pci/drm/drm_gem.c
1490
if (obj->funcs->mmap) {
sys/dev/pci/drm/drm_gem.c
1491
ret = obj->funcs->mmap(obj, accessprot, off, size);
sys/dev/pci/drm/drm_gem.c
1516
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1526
struct drm_gem_object *obj = NULL;
sys/dev/pci/drm/drm_gem.c
1538
obj = container_of(node, struct drm_gem_object, vma_node);
sys/dev/pci/drm/drm_gem.c
1549
if (!kref_get_unless_zero(&obj->refcount))
sys/dev/pci/drm/drm_gem.c
1550
obj = NULL;
sys/dev/pci/drm/drm_gem.c
1554
if (!obj)
sys/dev/pci/drm/drm_gem.c
1558
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1562
ret = drm_gem_mmap_obj(obj, drm_vma_node_size(node) << PAGE_SHIFT,
sys/dev/pci/drm/drm_gem.c
1565
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1567
return &obj->uobj;
sys/dev/pci/drm/drm_gem.c
157
struct drm_gem_object *obj = NULL;
sys/dev/pci/drm/drm_gem.c
1573
const struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1575
drm_printf_indent(p, indent, "name=%d\n", obj->name);
sys/dev/pci/drm/drm_gem.c
1577
kref_read(&obj->refcount));
sys/dev/pci/drm/drm_gem.c
1579
drm_vma_node_start(&obj->vma_node));
sys/dev/pci/drm/drm_gem.c
1580
drm_printf_indent(p, indent, "size=%zu\n", obj->size);
sys/dev/pci/drm/drm_gem.c
1582
str_yes_no(drm_gem_is_imported(obj)));
sys/dev/pci/drm/drm_gem.c
1584
if (obj->funcs->print_info)
sys/dev/pci/drm/drm_gem.c
1585
obj->funcs->print_info(p, indent, obj);
sys/dev/pci/drm/drm_gem.c
1588
int drm_gem_vmap_locked(struct drm_gem_object *obj, struct iosys_map *map)
sys/dev/pci/drm/drm_gem.c
1592
dma_resv_assert_held(obj->resv);
sys/dev/pci/drm/drm_gem.c
1594
if (!obj->funcs->vmap)
sys/dev/pci/drm/drm_gem.c
1597
ret = obj->funcs->vmap(obj, map);
sys/dev/pci/drm/drm_gem.c
1607
void drm_gem_vunmap_locked(struct drm_gem_object *obj, struct iosys_map *map)
sys/dev/pci/drm/drm_gem.c
1609
dma_resv_assert_held(obj->resv);
sys/dev/pci/drm/drm_gem.c
1614
if (obj->funcs->vunmap)
sys/dev/pci/drm/drm_gem.c
1615
obj->funcs->vunmap(obj, map);
sys/dev/pci/drm/drm_gem.c
1622
void drm_gem_lock(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1624
dma_resv_lock(obj->resv, NULL);
sys/dev/pci/drm/drm_gem.c
1628
void drm_gem_unlock(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1630
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_gem.c
1634
int drm_gem_vmap(struct drm_gem_object *obj, struct iosys_map *map)
sys/dev/pci/drm/drm_gem.c
1638
dma_resv_lock(obj->resv, NULL);
sys/dev/pci/drm/drm_gem.c
1639
ret = drm_gem_vmap_locked(obj, map);
sys/dev/pci/drm/drm_gem.c
1640
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_gem.c
1646
void drm_gem_vunmap(struct drm_gem_object *obj, struct iosys_map *map)
sys/dev/pci/drm/drm_gem.c
1648
dma_resv_lock(obj->resv, NULL);
sys/dev/pci/drm/drm_gem.c
1649
drm_gem_vunmap_locked(obj, map);
sys/dev/pci/drm/drm_gem.c
1650
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_gem.c
1678
struct drm_gem_object *obj = objs[contended];
sys/dev/pci/drm/drm_gem.c
1680
ret = dma_resv_lock_slow_interruptible(obj->resv,
sys/dev/pci/drm/drm_gem.c
1748
drm_gem_lru_remove_locked(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1750
obj->lru->count -= obj->size >> PAGE_SHIFT;
sys/dev/pci/drm/drm_gem.c
1751
WARN_ON(obj->lru->count < 0);
sys/dev/pci/drm/drm_gem.c
1752
list_del(&obj->lru_node);
sys/dev/pci/drm/drm_gem.c
1753
obj->lru = NULL;
sys/dev/pci/drm/drm_gem.c
1764
drm_gem_lru_remove(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1766
struct drm_gem_lru *lru = obj->lru;
sys/dev/pci/drm/drm_gem.c
1772
drm_gem_lru_remove_locked(obj);
sys/dev/pci/drm/drm_gem.c
1786
drm_gem_lru_move_tail_locked(struct drm_gem_lru *lru, struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1790
if (obj->lru)
sys/dev/pci/drm/drm_gem.c
1791
drm_gem_lru_remove_locked(obj);
sys/dev/pci/drm/drm_gem.c
1793
lru->count += obj->size >> PAGE_SHIFT;
sys/dev/pci/drm/drm_gem.c
1794
list_add_tail(&obj->lru_node, &lru->list);
sys/dev/pci/drm/drm_gem.c
1795
obj->lru = lru;
sys/dev/pci/drm/drm_gem.c
1810
drm_gem_lru_move_tail(struct drm_gem_lru *lru, struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1813
drm_gem_lru_move_tail_locked(lru, obj);
sys/dev/pci/drm/drm_gem.c
1838
bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket),
sys/dev/pci/drm/drm_gem.c
1842
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
185
obj = container_of(node, struct drm_gem_object, vma_node);
sys/dev/pci/drm/drm_gem.c
1850
obj = list_first_entry_or_null(&lru->list, typeof(*obj), lru_node);
sys/dev/pci/drm/drm_gem.c
1852
if (!obj)
sys/dev/pci/drm/drm_gem.c
1855
drm_gem_lru_move_tail_locked(&still_in_lru, obj);
sys/dev/pci/drm/drm_gem.c
1862
if (!kref_get_unless_zero(&obj->refcount))
sys/dev/pci/drm/drm_gem.c
1880
if (!ww_mutex_trylock(&obj->resv->lock, ticket)) {
sys/dev/pci/drm/drm_gem.c
1881
*remaining += obj->size >> PAGE_SHIFT;
sys/dev/pci/drm/drm_gem.c
1885
if (shrink(obj, ticket)) {
sys/dev/pci/drm/drm_gem.c
1886
freed += obj->size >> PAGE_SHIFT;
sys/dev/pci/drm/drm_gem.c
1893
WARN_ON(obj->lru == &still_in_lru);
sys/dev/pci/drm/drm_gem.c
1894
WARN_ON(obj->lru == lru);
sys/dev/pci/drm/drm_gem.c
1897
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_gem.c
1903
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
1911
list_for_each_entry (obj, &still_in_lru.list, lru_node)
sys/dev/pci/drm/drm_gem.c
1912
obj->lru = lru;
sys/dev/pci/drm/drm_gem.c
1926
int drm_gem_evict_locked(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
1928
dma_resv_assert_held(obj->resv);
sys/dev/pci/drm/drm_gem.c
1930
if (!dma_resv_test_signaled(obj->resv, DMA_RESV_USAGE_READ))
sys/dev/pci/drm/drm_gem.c
1933
if (obj->funcs->evict)
sys/dev/pci/drm/drm_gem.c
1934
return obj->funcs->evict(obj);
sys/dev/pci/drm/drm_gem.c
196
if (!kref_get_unless_zero(&obj->refcount))
sys/dev/pci/drm/drm_gem.c
197
obj = NULL;
sys/dev/pci/drm/drm_gem.c
201
if (!obj)
sys/dev/pci/drm/drm_gem.c
205
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
209
return &obj->uobj;
sys/dev/pci/drm/drm_gem.c
289
struct drm_gem_object *obj, size_t size,
sys/dev/pci/drm/drm_gem.c
294
drm_gem_private_object_init(dev, obj, size);
sys/dev/pci/drm/drm_gem.c
305
obj->filp = filp;
sys/dev/pci/drm/drm_gem.c
320
int drm_gem_object_init(struct drm_device *dev, struct drm_gem_object *obj,
sys/dev/pci/drm/drm_gem.c
323
return drm_gem_object_init_with_mnt(dev, obj, size, NULL);
sys/dev/pci/drm/drm_gem.c
330
struct drm_gem_object *obj, size_t size)
sys/dev/pci/drm/drm_gem.c
332
drm_gem_private_object_init(dev, obj, size);
sys/dev/pci/drm/drm_gem.c
339
obj->uao = uao_create(size, 0);
sys/dev/pci/drm/drm_gem.c
340
uvm_obj_init(&obj->uobj, &drm_pgops, 1);
sys/dev/pci/drm/drm_gem.c
358
struct drm_gem_object *obj, size_t size)
sys/dev/pci/drm/drm_gem.c
362
obj->dev = dev;
sys/dev/pci/drm/drm_gem.c
364
obj->filp = NULL;
sys/dev/pci/drm/drm_gem.c
366
obj->uao = NULL;
sys/dev/pci/drm/drm_gem.c
367
obj->uobj.pgops = NULL;
sys/dev/pci/drm/drm_gem.c
370
kref_init(&obj->refcount);
sys/dev/pci/drm/drm_gem.c
371
obj->handle_count = 0;
sys/dev/pci/drm/drm_gem.c
372
obj->size = size;
sys/dev/pci/drm/drm_gem.c
373
rw_init(&obj->gpuva.lock, "gpuva");
sys/dev/pci/drm/drm_gem.c
374
dma_resv_init(&obj->_resv);
sys/dev/pci/drm/drm_gem.c
375
if (!obj->resv)
sys/dev/pci/drm/drm_gem.c
376
obj->resv = &obj->_resv;
sys/dev/pci/drm/drm_gem.c
379
drm_gem_gpuva_init(obj);
sys/dev/pci/drm/drm_gem.c
381
drm_vma_node_reset(&obj->vma_node);
sys/dev/pci/drm/drm_gem.c
382
INIT_LIST_HEAD(&obj->lru_node);
sys/dev/pci/drm/drm_gem.c
392
void drm_gem_private_object_fini(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
394
WARN_ON(obj->dma_buf);
sys/dev/pci/drm/drm_gem.c
396
dma_resv_fini(&obj->_resv);
sys/dev/pci/drm/drm_gem.c
397
mutex_destroy(&obj->gpuva.lock);
sys/dev/pci/drm/drm_gem.c
401
static void drm_gem_object_handle_get(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
403
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
407
if (obj->handle_count++ == 0)
sys/dev/pci/drm/drm_gem.c
408
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
422
bool drm_gem_object_handle_get_if_exists_unlocked(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
424
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
433
if (!obj->handle_count)
sys/dev/pci/drm/drm_gem.c
436
drm_gem_object_handle_get(obj);
sys/dev/pci/drm/drm_gem.c
451
static void drm_gem_object_handle_free(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
453
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
456
if (obj->name) {
sys/dev/pci/drm/drm_gem.c
457
idr_remove(&dev->object_name_idr, obj->name);
sys/dev/pci/drm/drm_gem.c
458
obj->name = 0;
sys/dev/pci/drm/drm_gem.c
462
static void drm_gem_object_exported_dma_buf_free(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
465
if (obj->dma_buf) {
sys/dev/pci/drm/drm_gem.c
466
dma_buf_put(obj->dma_buf);
sys/dev/pci/drm/drm_gem.c
467
obj->dma_buf = NULL;
sys/dev/pci/drm/drm_gem.c
478
void drm_gem_object_handle_put_unlocked(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
480
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
483
if (drm_WARN_ON(dev, READ_ONCE(obj->handle_count) == 0))
sys/dev/pci/drm/drm_gem.c
493
if (--obj->handle_count == 0) {
sys/dev/pci/drm/drm_gem.c
494
drm_gem_object_handle_free(obj);
sys/dev/pci/drm/drm_gem.c
495
drm_gem_object_exported_dma_buf_free(obj);
sys/dev/pci/drm/drm_gem.c
501
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
512
struct drm_gem_object *obj = ptr;
sys/dev/pci/drm/drm_gem.c
514
if (drm_WARN_ON(obj->dev, !data))
sys/dev/pci/drm/drm_gem.c
517
if (obj->funcs->close)
sys/dev/pci/drm/drm_gem.c
518
obj->funcs->close(obj, file_priv);
sys/dev/pci/drm/drm_gem.c
526
drm_vma_node_revoke(&obj->vma_node, file_priv);
sys/dev/pci/drm/drm_gem.c
528
drm_gem_object_handle_put_unlocked(obj);
sys/dev/pci/drm/drm_gem.c
545
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
550
obj = idr_replace(&filp->object_idr, NULL, handle);
sys/dev/pci/drm/drm_gem.c
552
if (IS_ERR_OR_NULL(obj))
sys/dev/pci/drm/drm_gem.c
556
drm_gem_object_release_handle(handle, obj, filp);
sys/dev/pci/drm/drm_gem.c
583
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
586
obj = drm_gem_object_lookup(file, handle);
sys/dev/pci/drm/drm_gem.c
587
if (!obj)
sys/dev/pci/drm/drm_gem.c
591
if (drm_gem_is_imported(obj)) {
sys/dev/pci/drm/drm_gem.c
596
ret = drm_gem_create_mmap_offset(obj);
sys/dev/pci/drm/drm_gem.c
600
*offset = drm_vma_node_offset_addr(&obj->vma_node);
sys/dev/pci/drm/drm_gem.c
602
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
624
struct drm_gem_object *obj,
sys/dev/pci/drm/drm_gem.c
627
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
633
drm_gem_object_handle_get(obj);
sys/dev/pci/drm/drm_gem.c
653
ret = drm_vma_node_allow(&obj->vma_node, file_priv);
sys/dev/pci/drm/drm_gem.c
657
if (obj->funcs->open) {
sys/dev/pci/drm/drm_gem.c
658
ret = obj->funcs->open(obj, file_priv);
sys/dev/pci/drm/drm_gem.c
665
obj = idr_replace(&file_priv->object_idr, obj, handle);
sys/dev/pci/drm/drm_gem.c
666
WARN_ON(obj != NULL);
sys/dev/pci/drm/drm_gem.c
672
drm_vma_node_revoke(&obj->vma_node, file_priv);
sys/dev/pci/drm/drm_gem.c
678
drm_gem_object_handle_put_unlocked(obj);
sys/dev/pci/drm/drm_gem.c
696
struct drm_gem_object *obj,
sys/dev/pci/drm/drm_gem.c
699
mutex_lock(&obj->dev->object_name_lock);
sys/dev/pci/drm/drm_gem.c
701
return drm_gem_handle_create_tail(file_priv, obj, handlep);
sys/dev/pci/drm/drm_gem.c
717
drm_gem_free_mmap_offset(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
719
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
721
drm_vma_offset_remove(dev->vma_offset_manager, &obj->vma_node);
sys/dev/pci/drm/drm_gem.c
743
drm_gem_create_mmap_offset_size(struct drm_gem_object *obj, size_t size)
sys/dev/pci/drm/drm_gem.c
745
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem.c
747
return drm_vma_offset_add(dev->vma_offset_manager, &obj->vma_node,
sys/dev/pci/drm/drm_gem.c
75
struct drm_gem_object *obj =
sys/dev/pci/drm/drm_gem.c
766
int drm_gem_create_mmap_offset(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
768
return drm_gem_create_mmap_offset_size(obj, obj->size);
sys/dev/pci/drm/drm_gem.c
78
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
810
struct vm_page **drm_gem_get_pages(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gem.c
821
if (WARN_ON(!obj->filp))
sys/dev/pci/drm/drm_gem.c
825
mapping = obj->filp->f_mapping;
sys/dev/pci/drm/drm_gem.c
831
WARN_ON((obj->size & (PAGE_SIZE - 1)) != 0);
sys/dev/pci/drm/drm_gem.c
833
npages = obj->size >> PAGE_SHIFT;
sys/dev/pci/drm/drm_gem.c
84
struct drm_gem_object *obj =
sys/dev/pci/drm/drm_gem.c
87
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem.c
889
void drm_gem_put_pages(struct drm_gem_object *obj, struct vm_page **pages,
sys/dev/pci/drm/drm_gem.c
898
mapping = file_inode(obj->filp)->i_mapping;
sys/dev/pci/drm/drm_gem.c
905
WARN_ON((obj->size & (PAGE_SIZE - 1)) != 0);
sys/dev/pci/drm/drm_gem.c
907
npages = obj->size >> PAGE_SHIFT;
sys/dev/pci/drm/drm_gem.c
940
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem.c
946
obj = idr_find(&filp->object_idr, handle[i]);
sys/dev/pci/drm/drm_gem.c
947
if (!obj) {
sys/dev/pci/drm/drm_gem.c
951
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gem.c
952
objs[i] = obj;
sys/dev/pci/drm/drm_gem.c
97
struct drm_gem_object *obj =
sys/dev/pci/drm/drm_gem.c
99
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_gem_atomic_helper.c
15
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gem_atomic_helper.c
20
obj = dps->fb->obj[0];
sys/dev/pci/drm/drm_gem_atomic_helper.c
21
if (obj == NULL)
sys/dev/pci/drm/drm_gem_atomic_helper.c
23
r = dma_resv_get_singleton(obj->resv, DMA_RESV_USAGE_WRITE, &f);
sys/dev/pci/drm/drm_gem_dma_helper.c
101
bus_dmamem_unmap(obj->dmat, obj->vaddr, obj->dmasize);
sys/dev/pci/drm/drm_gem_dma_helper.c
104
if (obj->sgt)
sys/dev/pci/drm/drm_gem_dma_helper.c
105
drm_prime_sg_free(obj->sgt);
sys/dev/pci/drm/drm_gem_dma_helper.c
108
bus_dmamem_free(obj->dmat, obj->dmasegs, nsegs);
sys/dev/pci/drm/drm_gem_dma_helper.c
110
free(obj, M_DRM, sizeof(*obj));
sys/dev/pci/drm/drm_gem_dma_helper.c
123
drm_gem_dma_obj_free(struct drm_gem_dma_object *obj)
sys/dev/pci/drm/drm_gem_dma_helper.c
126
bus_dmamap_unload(obj->dmat, obj->dmamap);
sys/dev/pci/drm/drm_gem_dma_helper.c
127
bus_dmamap_destroy(obj->dmat, obj->dmamap);
sys/dev/pci/drm/drm_gem_dma_helper.c
128
bus_dmamem_unmap(obj->dmat, obj->vaddr, obj->dmasize);
sys/dev/pci/drm/drm_gem_dma_helper.c
130
if (obj->sgt)
sys/dev/pci/drm/drm_gem_dma_helper.c
131
drm_prime_sg_free(obj->sgt);
sys/dev/pci/drm/drm_gem_dma_helper.c
134
bus_dmamem_free(obj->dmat, obj->dmasegs, 1);
sys/dev/pci/drm/drm_gem_dma_helper.c
135
free(obj, M_DRM, sizeof(*obj));
sys/dev/pci/drm/drm_gem_dma_helper.c
141
struct drm_gem_dma_object *obj = to_drm_gem_dma_obj(gem_obj);
sys/dev/pci/drm/drm_gem_dma_helper.c
145
drm_gem_dma_obj_free(obj);
sys/dev/pci/drm/drm_gem_dma_helper.c
152
struct drm_gem_dma_object *obj;
sys/dev/pci/drm/drm_gem_dma_helper.c
158
obj = drm_gem_dma_create(ddev, args->size);
sys/dev/pci/drm/drm_gem_dma_helper.c
159
if (obj == NULL)
sys/dev/pci/drm/drm_gem_dma_helper.c
162
error = drm_gem_handle_create(file_priv, &obj->base, &handle);
sys/dev/pci/drm/drm_gem_dma_helper.c
163
drm_gem_object_put(&obj->base);
sys/dev/pci/drm/drm_gem_dma_helper.c
165
drm_gem_dma_obj_free(obj);
sys/dev/pci/drm/drm_gem_dma_helper.c
190
struct drm_gem_dma_object *obj = to_drm_gem_dma_obj(gem_obj);
sys/dev/pci/drm/drm_gem_dma_helper.c
191
struct uvm_object *uobj = &obj->base.uobj;
sys/dev/pci/drm/drm_gem_dma_helper.c
196
offset -= drm_vma_node_offset_addr(&obj->base.vma_node);
sys/dev/pci/drm/drm_gem_dma_helper.c
208
paddr = bus_dmamem_mmap(obj->dmat, obj->dmasegs, 1,
sys/dev/pci/drm/drm_gem_dma_helper.c
236
struct drm_gem_dma_object *obj = to_drm_gem_dma_obj(gem_obj);
sys/dev/pci/drm/drm_gem_dma_helper.c
238
return drm_prime_bus_dmamem_to_sg(obj->dmat, obj->dmasegs, 1);
sys/dev/pci/drm/drm_gem_dma_helper.c
249
struct drm_gem_dma_object *obj;
sys/dev/pci/drm/drm_gem_dma_helper.c
251
obj = drm_gem_dma_create_internal(ddev, size, sgt);
sys/dev/pci/drm/drm_gem_dma_helper.c
252
if (obj == NULL)
sys/dev/pci/drm/drm_gem_dma_helper.c
255
return &obj->base;
sys/dev/pci/drm/drm_gem_dma_helper.c
262
struct drm_gem_dma_object *obj = to_drm_gem_dma_obj(gem_obj);
sys/dev/pci/drm/drm_gem_dma_helper.c
264
iosys_map_set_vaddr(map, obj->vaddr);
sys/dev/pci/drm/drm_gem_dma_helper.c
48
struct drm_gem_dma_object *obj;
sys/dev/pci/drm/drm_gem_dma_helper.c
51
obj = malloc(sizeof(*obj), M_DRM, M_WAITOK | M_ZERO);
sys/dev/pci/drm/drm_gem_dma_helper.c
52
obj->dmat = ddev->dmat;
sys/dev/pci/drm/drm_gem_dma_helper.c
53
obj->dmasize = size;
sys/dev/pci/drm/drm_gem_dma_helper.c
54
obj->base.funcs = &drm_gem_dma_default_funcs;
sys/dev/pci/drm/drm_gem_dma_helper.c
59
error = -drm_prime_sg_to_bus_dmamem(obj->dmat, obj->dmasegs, 1,
sys/dev/pci/drm/drm_gem_dma_helper.c
64
error = bus_dmamem_alloc(obj->dmat, obj->dmasize,
sys/dev/pci/drm/drm_gem_dma_helper.c
65
PAGE_SIZE, 0, obj->dmasegs, 1, &nsegs,
sys/dev/pci/drm/drm_gem_dma_helper.c
70
error = bus_dmamem_map(obj->dmat, obj->dmasegs, nsegs,
sys/dev/pci/drm/drm_gem_dma_helper.c
71
obj->dmasize, &obj->vaddr,
sys/dev/pci/drm/drm_gem_dma_helper.c
75
error = bus_dmamap_create(obj->dmat, obj->dmasize, 1,
sys/dev/pci/drm/drm_gem_dma_helper.c
76
obj->dmasize, 0, BUS_DMA_WAITOK, &obj->dmamap);
sys/dev/pci/drm/drm_gem_dma_helper.c
79
error = bus_dmamap_load(obj->dmat, obj->dmamap, obj->vaddr,
sys/dev/pci/drm/drm_gem_dma_helper.c
80
obj->dmasize, NULL, BUS_DMA_WAITOK);
sys/dev/pci/drm/drm_gem_dma_helper.c
87
memset(obj->vaddr, 0, obj->dmasize);
sys/dev/pci/drm/drm_gem_dma_helper.c
89
error = drm_gem_object_init(ddev, &obj->base, size);
sys/dev/pci/drm/drm_gem_dma_helper.c
93
obj->dma_addr = obj->dmamap->dm_segs[0].ds_addr;
sys/dev/pci/drm/drm_gem_dma_helper.c
94
return obj;
sys/dev/pci/drm/drm_gem_dma_helper.c
97
bus_dmamap_unload(obj->dmat, obj->dmamap);
sys/dev/pci/drm/drm_gem_dma_helper.c
99
bus_dmamap_destroy(obj->dmat, obj->dmamap);
sys/dev/pci/drm/drm_gem_framebuffer_helper.c
13
drm_gem_object_put(fb->obj[i]);
sys/dev/pci/drm/drm_gem_framebuffer_helper.c
22
return drm_gem_handle_create(file, fb->obj[0], handle);
sys/dev/pci/drm/drm_gem_framebuffer_helper.c
47
fb->obj[0] = gem_obj;
sys/dev/pci/drm/drm_gem_framebuffer_helper.c
66
if (plane < nitems(fb->obj))
sys/dev/pci/drm/drm_gem_framebuffer_helper.c
67
return fb->obj[plane];
sys/dev/pci/drm/drm_gem_ttm_helper.c
13
drm_gem_ttm_mmap(struct drm_gem_object *obj,
sys/dev/pci/drm/drm_gem_ttm_helper.c
17
container_of(obj, struct ttm_buffer_object, base);
sys/dev/pci/drm/drm_gem_ttm_helper.c
20
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gem_ttm_helper.c
25
drm_gem_ttm_vmap(struct drm_gem_object *obj, struct iosys_map *ism)
sys/dev/pci/drm/drm_gem_ttm_helper.c
28
container_of(obj, struct ttm_buffer_object, base);
sys/dev/pci/drm/drm_gem_ttm_helper.c
34
drm_gem_ttm_vunmap(struct drm_gem_object *obj, struct iosys_map *ism)
sys/dev/pci/drm/drm_gem_ttm_helper.c
37
container_of(obj, struct ttm_buffer_object, base);
sys/dev/pci/drm/drm_gpuvm.c
1012
drm_gpuvm_gem_object_free(struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gpuvm.c
1014
drm_gem_object_release(obj);
sys/dev/pci/drm/drm_gpuvm.c
1015
kfree(obj);
sys/dev/pci/drm/drm_gpuvm.c
1035
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gpuvm.c
1037
obj = kzalloc(sizeof(*obj), GFP_KERNEL);
sys/dev/pci/drm/drm_gpuvm.c
1038
if (!obj)
sys/dev/pci/drm/drm_gpuvm.c
1041
obj->funcs = &drm_gpuvm_object_funcs;
sys/dev/pci/drm/drm_gpuvm.c
1042
drm_gem_private_object_init(drm, obj, 0);
sys/dev/pci/drm/drm_gpuvm.c
1044
return obj;
sys/dev/pci/drm/drm_gpuvm.c
1159
exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
sys/dev/pci/drm/drm_gpuvm.c
1162
return num_fences ? drm_exec_prepare_obj(exec, obj, num_fences) :
sys/dev/pci/drm/drm_gpuvm.c
1163
drm_exec_lock_obj(exec, obj);
sys/dev/pci/drm/drm_gpuvm.c
1199
ret = exec_prepare_obj(exec, vm_bo->obj, num_fences);
sys/dev/pci/drm/drm_gpuvm.c
1220
ret = exec_prepare_obj(exec, vm_bo->obj, num_fences);
sys/dev/pci/drm/drm_gpuvm.c
1290
struct drm_gem_object *obj = va->gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
1292
ret = exec_prepare_obj(exec, obj, num_fences);
sys/dev/pci/drm/drm_gpuvm.c
1467
dma_resv_assert_held(vm_bo->obj->resv);
sys/dev/pci/drm/drm_gpuvm.c
1516
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_gpuvm.c
1519
drm_exec_for_each_locked_object(exec, index, obj) {
sys/dev/pci/drm/drm_gpuvm.c
1520
dma_resv_assert_held(obj->resv);
sys/dev/pci/drm/drm_gpuvm.c
1521
dma_resv_add_fence(obj->resv, fence,
sys/dev/pci/drm/drm_gpuvm.c
1522
drm_gpuvm_is_extobj(gpuvm, obj) ?
sys/dev/pci/drm/drm_gpuvm.c
1540
struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gpuvm.c
1554
vm_bo->obj = obj;
sys/dev/pci/drm/drm_gpuvm.c
1555
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_gpuvm.c
1575
struct drm_gem_object *obj = vm_bo->obj;
sys/dev/pci/drm/drm_gpuvm.c
1584
drm_gem_gpuva_assert_lock_held(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1593
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_gpuvm.c
1626
struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gpuvm.c
1630
drm_gem_gpuva_assert_lock_held(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1631
drm_gem_for_each_gpuvm_bo(vm_bo, obj)
sys/dev/pci/drm/drm_gpuvm.c
1652
struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gpuvm.c
1654
struct drm_gpuvm_bo *vm_bo = __drm_gpuvm_bo_find(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1677
struct drm_gem_object *obj)
sys/dev/pci/drm/drm_gpuvm.c
1681
vm_bo = drm_gpuvm_bo_find(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1685
vm_bo = drm_gpuvm_bo_create(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1689
drm_gem_gpuva_assert_lock_held(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1690
list_add_tail(&vm_bo->list.entry.gem, &obj->gpuva.list);
sys/dev/pci/drm/drm_gpuvm.c
1716
struct drm_gem_object *obj = __vm_bo->obj;
sys/dev/pci/drm/drm_gpuvm.c
1719
vm_bo = drm_gpuvm_bo_find(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1725
drm_gem_gpuva_assert_lock_held(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1726
list_add_tail(&__vm_bo->list.entry.gem, &obj->gpuva.list);
sys/dev/pci/drm/drm_gpuvm.c
1750
if (drm_gpuvm_is_extobj(gpuvm, vm_bo->obj))
sys/dev/pci/drm/drm_gpuvm.c
1767
struct drm_gem_object *obj = vm_bo->obj;
sys/dev/pci/drm/drm_gpuvm.c
1770
dma_resv_assert_held(obj->resv);
sys/dev/pci/drm/drm_gpuvm.c
1777
if (drm_gpuvm_is_extobj(gpuvm, obj) && !lock)
sys/dev/pci/drm/drm_gpuvm.c
1902
struct drm_gem_object *obj = va->gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
1905
if (unlikely(!obj))
sys/dev/pci/drm/drm_gpuvm.c
1908
drm_WARN_ON(gpuvm->drm, obj != vm_bo->obj);
sys/dev/pci/drm/drm_gpuvm.c
1912
drm_gem_gpuva_assert_lock_held(gpuvm, obj);
sys/dev/pci/drm/drm_gpuvm.c
1937
struct drm_gem_object *obj = va->gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
1940
if (unlikely(!obj))
sys/dev/pci/drm/drm_gpuvm.c
1943
drm_gem_gpuva_assert_lock_held(va->vm, obj);
sys/dev/pci/drm/drm_gpuvm.c
2135
op.map.gem.obj = req->map.gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
2181
struct drm_gem_object *req_obj = req->map.gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
2194
struct drm_gem_object *obj = va->gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
2199
bool merge = !!va->gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
2201
if (madvise && obj)
sys/dev/pci/drm/drm_gpuvm.c
2205
merge &= obj == req_obj &&
sys/dev/pci/drm/drm_gpuvm.c
2226
.gem.obj = obj,
sys/dev/pci/drm/drm_gpuvm.c
2247
.gem.obj = obj,
sys/dev/pci/drm/drm_gpuvm.c
2252
merge &= obj == req_obj &&
sys/dev/pci/drm/drm_gpuvm.c
2289
.gem.obj = obj,
sys/dev/pci/drm/drm_gpuvm.c
2303
merge &= obj == req_obj &&
sys/dev/pci/drm/drm_gpuvm.c
2327
.gem.obj = obj,
sys/dev/pci/drm/drm_gpuvm.c
2369
struct drm_gem_object *obj = va->gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
2378
prev.gem.obj = obj;
sys/dev/pci/drm/drm_gpuvm.c
2387
next.gem.obj = obj;
sys/dev/pci/drm/drm_gpuvm.c
2505
if (op->remap.unmap->va->gem.obj)
sys/dev/pci/drm/drm_gpuvm.c
2506
return drm_exec_lock_obj(exec, op->remap.unmap->va->gem.obj);
sys/dev/pci/drm/drm_gpuvm.c
2509
if (op->unmap.va->gem.obj)
sys/dev/pci/drm/drm_gpuvm.c
2510
return drm_exec_lock_obj(exec, op->unmap.va->gem.obj);
sys/dev/pci/drm/drm_gpuvm.c
2582
struct drm_gem_object *req_obj = req->map.gem.obj;
sys/dev/pci/drm/drm_gpuvm.c
2957
drm_gem_gpuva_assert_lock_held(vm_bo->vm, vm_bo->obj);
sys/dev/pci/drm/drm_internal.h
196
bool drm_gem_object_handle_get_if_exists_unlocked(struct drm_gem_object *obj);
sys/dev/pci/drm/drm_internal.h
197
void drm_gem_object_handle_put_unlocked(struct drm_gem_object *obj);
sys/dev/pci/drm/drm_internal.h
199
struct drm_gem_object *obj,
sys/dev/pci/drm/drm_internal.h
212
const struct drm_gem_object *obj);
sys/dev/pci/drm/drm_internal.h
214
int drm_gem_vmap_locked(struct drm_gem_object *obj, struct iosys_map *map);
sys/dev/pci/drm/drm_internal.h
215
void drm_gem_vunmap_locked(struct drm_gem_object *obj, struct iosys_map *map);
sys/dev/pci/drm/drm_linux.c
1473
union acpi_object *obj;
sys/dev/pci/drm/drm_linux.c
1512
obj = (union acpi_object *)result->pointer;
sys/dev/pci/drm/drm_linux.c
1513
data = (uint8_t *)(obj + 1);
sys/dev/pci/drm/drm_linux.c
1514
obj->type = res.type;
sys/dev/pci/drm/drm_linux.c
1515
obj->buffer.length = res.length;
sys/dev/pci/drm/drm_linux.c
1516
obj->buffer.pointer = data;
sys/dev/pci/drm/drm_mode_object.c
142
struct drm_mode_object *obj = NULL;
sys/dev/pci/drm/drm_mode_object.c
145
obj = idr_find(&dev->mode_config.object_idr, id);
sys/dev/pci/drm/drm_mode_object.c
146
if (obj && type != DRM_MODE_OBJECT_ANY && obj->type != type)
sys/dev/pci/drm/drm_mode_object.c
147
obj = NULL;
sys/dev/pci/drm/drm_mode_object.c
148
if (obj && obj->id != id)
sys/dev/pci/drm/drm_mode_object.c
149
obj = NULL;
sys/dev/pci/drm/drm_mode_object.c
152
if (obj && drm_mode_object_lease_required(obj->type) &&
sys/dev/pci/drm/drm_mode_object.c
153
!_drm_lease_held(file_priv, obj->id)) {
sys/dev/pci/drm/drm_mode_object.c
155
obj = NULL;
sys/dev/pci/drm/drm_mode_object.c
159
if (obj && obj->free_cb) {
sys/dev/pci/drm/drm_mode_object.c
160
if (!kref_get_unless_zero(&obj->refcount))
sys/dev/pci/drm/drm_mode_object.c
161
obj = NULL;
sys/dev/pci/drm/drm_mode_object.c
165
return obj;
sys/dev/pci/drm/drm_mode_object.c
183
struct drm_mode_object *obj = NULL;
sys/dev/pci/drm/drm_mode_object.c
185
obj = __drm_mode_object_find(dev, file_priv, id, type);
sys/dev/pci/drm/drm_mode_object.c
186
return obj;
sys/dev/pci/drm/drm_mode_object.c
198
void drm_mode_object_put(struct drm_mode_object *obj)
sys/dev/pci/drm/drm_mode_object.c
200
if (obj->free_cb) {
sys/dev/pci/drm/drm_mode_object.c
201
DRM_DEBUG("OBJ ID: %d (%d)\n", obj->id, kref_read(&obj->refcount));
sys/dev/pci/drm/drm_mode_object.c
202
kref_put(&obj->refcount, obj->free_cb);
sys/dev/pci/drm/drm_mode_object.c
215
void drm_mode_object_get(struct drm_mode_object *obj)
sys/dev/pci/drm/drm_mode_object.c
217
if (obj->free_cb) {
sys/dev/pci/drm/drm_mode_object.c
218
DRM_DEBUG("OBJ ID: %d (%d)\n", obj->id, kref_read(&obj->refcount));
sys/dev/pci/drm/drm_mode_object.c
219
kref_get(&obj->refcount);
sys/dev/pci/drm/drm_mode_object.c
237
void drm_object_attach_property(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
241
int count = obj->properties->count;
sys/dev/pci/drm/drm_mode_object.c
245
if (obj->type == DRM_MODE_OBJECT_CONNECTOR) {
sys/dev/pci/drm/drm_mode_object.c
246
struct drm_connector *connector = obj_to_connector(obj);
sys/dev/pci/drm/drm_mode_object.c
258
obj->type);
sys/dev/pci/drm/drm_mode_object.c
262
obj->properties->properties[count] = property;
sys/dev/pci/drm/drm_mode_object.c
263
obj->properties->values[count] = init_val;
sys/dev/pci/drm/drm_mode_object.c
264
obj->properties->count++;
sys/dev/pci/drm/drm_mode_object.c
287
int drm_object_property_set_value(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
295
for (i = 0; i < obj->properties->count; i++) {
sys/dev/pci/drm/drm_mode_object.c
296
if (obj->properties->properties[i] == property) {
sys/dev/pci/drm/drm_mode_object.c
297
obj->properties->values[i] = val;
sys/dev/pci/drm/drm_mode_object.c
306
static int __drm_object_property_get_prop_value(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
312
for (i = 0; i < obj->properties->count; i++) {
sys/dev/pci/drm/drm_mode_object.c
313
if (obj->properties->properties[i] == property) {
sys/dev/pci/drm/drm_mode_object.c
314
*val = obj->properties->values[i];
sys/dev/pci/drm/drm_mode_object.c
322
static int __drm_object_property_get_value(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
328
if (obj->type == DRM_MODE_OBJECT_CONNECTOR) {
sys/dev/pci/drm/drm_mode_object.c
329
struct drm_connector *connector = obj_to_connector(obj);
sys/dev/pci/drm/drm_mode_object.c
350
return drm_atomic_get_property(obj, property, val);
sys/dev/pci/drm/drm_mode_object.c
352
return __drm_object_property_get_prop_value(obj, property, val);
sys/dev/pci/drm/drm_mode_object.c
372
int drm_object_property_get_value(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
377
return __drm_object_property_get_value(obj, property, val);
sys/dev/pci/drm/drm_mode_object.c
397
int drm_object_property_get_default_value(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
40
int __drm_mode_object_add(struct drm_device *dev, struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
403
return __drm_object_property_get_prop_value(obj, property, val);
sys/dev/pci/drm/drm_mode_object.c
408
int drm_mode_object_get_properties(struct drm_mode_object *obj, bool atomic,
sys/dev/pci/drm/drm_mode_object.c
415
for (i = 0, count = 0; i < obj->properties->count; i++) {
sys/dev/pci/drm/drm_mode_object.c
416
struct drm_property *prop = obj->properties->properties[i];
sys/dev/pci/drm/drm_mode_object.c
423
ret = __drm_object_property_get_value(obj, prop, &val);
sys/dev/pci/drm/drm_mode_object.c
460
struct drm_mode_object *obj;
sys/dev/pci/drm/drm_mode_object.c
469
obj = drm_mode_object_find(dev, file_priv, arg->obj_id, arg->obj_type);
sys/dev/pci/drm/drm_mode_object.c
470
if (!obj) {
sys/dev/pci/drm/drm_mode_object.c
474
if (!obj->properties) {
sys/dev/pci/drm/drm_mode_object.c
479
ret = drm_mode_object_get_properties(obj, file_priv->atomic,
sys/dev/pci/drm/drm_mode_object.c
485
drm_mode_object_put(obj);
sys/dev/pci/drm/drm_mode_object.c
49
ret = idr_alloc(&dev->mode_config.object_idr, register_obj ? obj : NULL,
sys/dev/pci/drm/drm_mode_object.c
491
struct drm_property *drm_mode_obj_find_prop_id(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
496
for (i = 0; i < obj->properties->count; i++)
sys/dev/pci/drm/drm_mode_object.c
497
if (obj->properties->properties[i]->base.id == prop_id)
sys/dev/pci/drm/drm_mode_object.c
498
return obj->properties->properties[i];
sys/dev/pci/drm/drm_mode_object.c
504
static int set_property_legacy(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
517
switch (obj->type) {
sys/dev/pci/drm/drm_mode_object.c
519
ret = drm_connector_set_obj_prop(obj, prop, prop_value);
sys/dev/pci/drm/drm_mode_object.c
522
ret = drm_mode_crtc_set_obj_prop(obj, prop, prop_value);
sys/dev/pci/drm/drm_mode_object.c
525
ret = drm_mode_plane_set_obj_prop(obj_to_plane(obj),
sys/dev/pci/drm/drm_mode_object.c
535
static int set_property_atomic(struct drm_mode_object *obj,
sys/dev/pci/drm/drm_mode_object.c
554
if (obj->type != DRM_MODE_OBJECT_CONNECTOR) {
sys/dev/pci/drm/drm_mode_object.c
56
obj->id = ret;
sys/dev/pci/drm/drm_mode_object.c
560
obj_to_connector(obj),
sys/dev/pci/drm/drm_mode_object.c
563
} else if (obj->type == DRM_MODE_OBJECT_CONNECTOR &&
sys/dev/pci/drm/drm_mode_object.c
564
prop == (obj_to_connector(obj))->backlight_property) {
sys/dev/pci/drm/drm_mode_object.c
565
struct drm_connector *connector = obj_to_connector(obj);
sys/dev/pci/drm/drm_mode_object.c
57
obj->type = obj_type;
sys/dev/pci/drm/drm_mode_object.c
572
ret = drm_atomic_set_property(state, file_priv, obj, prop, prop_value, false);
sys/dev/pci/drm/drm_mode_object.c
59
obj->free_cb = obj_free_cb;
sys/dev/pci/drm/drm_mode_object.c
60
kref_init(&obj->refcount);
sys/dev/pci/drm/drm_mode_object.c
81
struct drm_mode_object *obj, uint32_t obj_type)
sys/dev/pci/drm/drm_mode_object.c
83
return __drm_mode_object_add(dev, obj, obj_type, true, NULL);
sys/dev/pci/drm/drm_mode_object.c
88
struct drm_mode_object *obj)
sys/dev/pci/drm/drm_mode_object.c
91
idr_replace(&dev->mode_config.object_idr, obj, obj->id);
sys/dev/pci/drm/drm_plane.c
791
struct drm_mode_object *obj = &plane->base;
sys/dev/pci/drm/drm_plane.c
796
drm_object_property_set_value(obj, property, value);
sys/dev/pci/drm/drm_prime.c
1015
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_prime.c
1023
obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
1024
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_prime.c
1025
return obj;
sys/dev/pci/drm/drm_prime.c
1046
obj = dev->driver->gem_prime_import_sg_table(dev, attach, sgt);
sys/dev/pci/drm/drm_prime.c
1047
if (IS_ERR(obj)) {
sys/dev/pci/drm/drm_prime.c
1048
ret = PTR_ERR(obj);
sys/dev/pci/drm/drm_prime.c
1052
obj->import_attach = attach;
sys/dev/pci/drm/drm_prime.c
1053
obj->resv = dma_buf->resv;
sys/dev/pci/drm/drm_prime.c
1055
return obj;
sys/dev/pci/drm/drm_prime.c
1161
void drm_prime_gem_destroy(struct drm_gem_object *obj, struct sg_table *sg)
sys/dev/pci/drm/drm_prime.c
1168
attach = obj->import_attach;
sys/dev/pci/drm/drm_prime.c
241
struct drm_gem_object *obj = exp_info->priv;
sys/dev/pci/drm/drm_prime.c
249
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_prime.c
251
dma_buf->file->f_mapping = obj->dev->anon_inode->i_mapping;
sys/dev/pci/drm/drm_prime.c
269
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
270
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_prime.c
273
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_prime.c
298
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_prime.c
315
obj = dev->driver->gem_prime_import(dev, dma_buf);
sys/dev/pci/drm/drm_prime.c
317
obj = drm_gem_prime_import(dev, dma_buf);
sys/dev/pci/drm/drm_prime.c
318
if (IS_ERR(obj)) {
sys/dev/pci/drm/drm_prime.c
319
ret = PTR_ERR(obj);
sys/dev/pci/drm/drm_prime.c
323
if (obj->dma_buf) {
sys/dev/pci/drm/drm_prime.c
324
WARN_ON(obj->dma_buf != dma_buf);
sys/dev/pci/drm/drm_prime.c
326
obj->dma_buf = dma_buf;
sys/dev/pci/drm/drm_prime.c
331
ret = drm_gem_handle_create_tail(file_priv, obj, handle);
sys/dev/pci/drm/drm_prime.c
332
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_prime.c
377
struct drm_gem_object *obj,
sys/dev/pci/drm/drm_prime.c
383
if (obj->handle_count == 0) {
sys/dev/pci/drm/drm_prime.c
388
if (obj->funcs && obj->funcs->export)
sys/dev/pci/drm/drm_prime.c
389
dmabuf = obj->funcs->export(obj, flags);
sys/dev/pci/drm/drm_prime.c
391
dmabuf = drm_gem_prime_export(obj, flags);
sys/dev/pci/drm/drm_prime.c
404
obj->dma_buf = dmabuf;
sys/dev/pci/drm/drm_prime.c
405
get_dma_buf(obj->dma_buf);
sys/dev/pci/drm/drm_prime.c
436
struct drm_gem_object *obj;
sys/dev/pci/drm/drm_prime.c
441
obj = drm_gem_object_lookup(file_priv, handle);
sys/dev/pci/drm/drm_prime.c
442
if (!obj) {
sys/dev/pci/drm/drm_prime.c
456
if (obj->import_attach) {
sys/dev/pci/drm/drm_prime.c
457
dmabuf = obj->import_attach->dmabuf;
sys/dev/pci/drm/drm_prime.c
463
if (obj->dma_buf) {
sys/dev/pci/drm/drm_prime.c
464
get_dma_buf(obj->dma_buf);
sys/dev/pci/drm/drm_prime.c
465
dmabuf = obj->dma_buf;
sys/dev/pci/drm/drm_prime.c
469
dmabuf = export_and_register_object(dev, obj, flags);
sys/dev/pci/drm/drm_prime.c
493
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_prime.c
607
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
616
!obj->funcs->get_sg_table)
sys/dev/pci/drm/drm_prime.c
618
if (!obj->funcs->get_sg_table)
sys/dev/pci/drm/drm_prime.c
622
if (!obj->funcs->pin)
sys/dev/pci/drm/drm_prime.c
625
ret = dma_resv_lock(obj->resv, NULL);
sys/dev/pci/drm/drm_prime.c
628
ret = obj->funcs->pin(obj);
sys/dev/pci/drm/drm_prime.c
629
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_prime.c
647
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
650
if (!obj->funcs->unpin)
sys/dev/pci/drm/drm_prime.c
653
ret = dma_resv_lock(obj->resv, NULL);
sys/dev/pci/drm/drm_prime.c
654
if (drm_WARN_ON(obj->dev, ret))
sys/dev/pci/drm/drm_prime.c
656
obj->funcs->unpin(obj);
sys/dev/pci/drm/drm_prime.c
657
dma_resv_unlock(obj->resv);
sys/dev/pci/drm/drm_prime.c
678
struct drm_gem_object *obj = attach->dmabuf->priv;
sys/dev/pci/drm/drm_prime.c
685
if (WARN_ON(!obj->funcs->get_sg_table))
sys/dev/pci/drm/drm_prime.c
688
sgt = obj->funcs->get_sg_table(obj);
sys/dev/pci/drm/drm_prime.c
740
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
742
return drm_gem_vmap_locked(obj, map);
sys/dev/pci/drm/drm_prime.c
756
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
758
drm_gem_vunmap_locked(obj, map);
sys/dev/pci/drm/drm_prime.c
773
int drm_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
sys/dev/pci/drm/drm_prime.c
780
vma->vm_pgoff += drm_vma_node_start(&obj->vma_node);
sys/dev/pci/drm/drm_prime.c
782
if (obj->funcs && obj->funcs->mmap) {
sys/dev/pci/drm/drm_prime.c
783
vma->vm_ops = obj->funcs->vm_ops;
sys/dev/pci/drm/drm_prime.c
785
drm_gem_object_get(obj);
sys/dev/pci/drm/drm_prime.c
786
ret = obj->funcs->mmap(obj, vma);
sys/dev/pci/drm/drm_prime.c
788
drm_gem_object_put(obj);
sys/dev/pci/drm/drm_prime.c
791
vma->vm_private_data = obj;
sys/dev/pci/drm/drm_prime.c
803
priv->minor = obj->dev->primary;
sys/dev/pci/drm/drm_prime.c
806
ret = drm_vma_node_allow(&obj->vma_node, priv);
sys/dev/pci/drm/drm_prime.c
810
ret = obj->dev->driver->fops->mmap(fil, vma);
sys/dev/pci/drm/drm_prime.c
812
drm_vma_node_revoke(&obj->vma_node, priv);
sys/dev/pci/drm/drm_prime.c
844
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
846
return drm_gem_prime_mmap(obj, vma);
sys/dev/pci/drm/drm_prime.c
953
struct dma_buf *drm_gem_prime_export(struct drm_gem_object *obj,
sys/dev/pci/drm/drm_prime.c
956
struct drm_device *dev = obj->dev;
sys/dev/pci/drm/drm_prime.c
963
.size = obj->size,
sys/dev/pci/drm/drm_prime.c
965
.priv = obj,
sys/dev/pci/drm/drm_prime.c
966
.resv = obj->resv,
sys/dev/pci/drm/drm_prime.c
987
struct drm_gem_object *obj = dma_buf->priv;
sys/dev/pci/drm/drm_prime.c
989
return (dma_buf->ops == &drm_gem_prime_dmabuf_ops) && (obj->dev == dev);
sys/dev/pci/drm/drm_privacy_screen_x86.c
30
union acpi_object obj = { .type = ACPI_TYPE_INTEGER };
sys/dev/pci/drm/drm_privacy_screen_x86.c
31
struct acpi_object_list args = { .count = 1, .pointer = &obj, };
sys/dev/pci/drm/drm_property.c
654
struct drm_mode_object *obj;
sys/dev/pci/drm/drm_property.c
657
obj = __drm_mode_object_find(dev, NULL, id, DRM_MODE_OBJECT_BLOB);
sys/dev/pci/drm/drm_property.c
658
if (obj)
sys/dev/pci/drm/drm_property.c
659
blob = obj_to_blob(obj);
sys/dev/pci/drm/i915/display/i9xx_wm.c
2298
struct drm_gem_object *obj;
sys/dev/pci/drm/i915/display/i9xx_wm.c
2300
obj = intel_fb_bo(crtc->base.primary->state->fb);
sys/dev/pci/drm/i915/display/i9xx_wm.c
2303
if (!intel_bo_is_tiled(obj))
sys/dev/pci/drm/i915/display/intel_bo.c
12
bool intel_bo_is_tiled(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
14
return i915_gem_object_is_tiled(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.c
17
bool intel_bo_is_userptr(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
19
return i915_gem_object_is_userptr(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.c
22
bool intel_bo_is_shmem(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
24
return i915_gem_object_is_shmem(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.c
27
bool intel_bo_is_protected(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
29
return i915_gem_object_is_protected(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.c
32
void intel_bo_flush_if_display(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
34
i915_gem_object_flush_if_display(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.c
38
int intel_bo_fb_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
sys/dev/pci/drm/i915/display/intel_bo.c
40
return i915_gem_fb_mmap(to_intel_bo(obj), vma);
sys/dev/pci/drm/i915/display/intel_bo.c
44
int intel_bo_read_from_page(struct drm_gem_object *obj, u64 offset, void *dst, int size)
sys/dev/pci/drm/i915/display/intel_bo.c
46
return i915_gem_object_read_from_page(to_intel_bo(obj), offset, dst, size);
sys/dev/pci/drm/i915/display/intel_bo.c
49
struct intel_frontbuffer *intel_bo_get_frontbuffer(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
51
return i915_gem_object_get_frontbuffer(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.c
54
struct intel_frontbuffer *intel_bo_set_frontbuffer(struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_bo.c
57
return i915_gem_object_set_frontbuffer(to_intel_bo(obj), front);
sys/dev/pci/drm/i915/display/intel_bo.c
60
void intel_bo_describe(struct seq_file *m, struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_bo.c
62
i915_debugfs_describe_obj(m, to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_bo.h
15
bool intel_bo_is_tiled(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bo.h
16
bool intel_bo_is_userptr(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bo.h
17
bool intel_bo_is_shmem(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bo.h
18
bool intel_bo_is_protected(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bo.h
19
void intel_bo_flush_if_display(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bo.h
20
int intel_bo_fb_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma);
sys/dev/pci/drm/i915/display/intel_bo.h
21
int intel_bo_read_from_page(struct drm_gem_object *obj, u64 offset, void *dst, int size);
sys/dev/pci/drm/i915/display/intel_bo.h
23
struct intel_frontbuffer *intel_bo_get_frontbuffer(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bo.h
24
struct intel_frontbuffer *intel_bo_set_frontbuffer(struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_bo.h
27
void intel_bo_describe(struct seq_file *m, struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_bw.c
1673
to_intel_bw_state(display->bw.obj.state);
sys/dev/pci/drm/i915/display/intel_bw.c
1704
to_intel_bw_state(display->bw.obj.state);
sys/dev/pci/drm/i915/display/intel_bw.c
1716
intel_bw_duplicate_state(struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/intel_bw.c
1720
state = kmemdup(obj->state, sizeof(*state), GFP_KERNEL);
sys/dev/pci/drm/i915/display/intel_bw.c
1727
static void intel_bw_destroy_state(struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/intel_bw.c
1746
intel_atomic_global_obj_init(display, &display->bw.obj,
sys/dev/pci/drm/i915/display/intel_bw.c
921
bw_state = intel_atomic_get_old_global_obj_state(state, &display->bw.obj);
sys/dev/pci/drm/i915/display/intel_bw.c
932
bw_state = intel_atomic_get_new_global_obj_state(state, &display->bw.obj);
sys/dev/pci/drm/i915/display/intel_bw.c
943
bw_state = intel_atomic_get_global_obj_state(state, &display->bw.obj);
sys/dev/pci/drm/i915/display/intel_cdclk.c
3148
static struct intel_global_state *intel_cdclk_duplicate_state(struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/intel_cdclk.c
3152
cdclk_state = kmemdup(obj->state, sizeof(*cdclk_state), GFP_KERNEL);
sys/dev/pci/drm/i915/display/intel_cdclk.c
3162
static void intel_cdclk_destroy_state(struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/intel_cdclk.c
3179
cdclk_state = intel_atomic_get_global_obj_state(state, &display->cdclk.obj);
sys/dev/pci/drm/i915/display/intel_cdclk.c
3243
intel_atomic_global_obj_init(display, &display->cdclk.obj,
sys/dev/pci/drm/i915/display/intel_cdclk.c
3380
to_intel_bw_state(display->bw.obj.state);
sys/dev/pci/drm/i915/display/intel_cdclk.c
3382
to_intel_cdclk_state(display->cdclk.obj.state);
sys/dev/pci/drm/i915/display/intel_cdclk.c
3929
cdclk_state = to_intel_cdclk_state(display->cdclk.obj.state);
sys/dev/pci/drm/i915/display/intel_cdclk.h
56
to_intel_cdclk_state(intel_atomic_get_old_global_obj_state(state, &to_intel_display(state)->cdclk.obj))
sys/dev/pci/drm/i915/display/intel_cdclk.h
58
to_intel_cdclk_state(intel_atomic_get_new_global_obj_state(state, &to_intel_display(state)->cdclk.obj))
sys/dev/pci/drm/i915/display/intel_display_core.h
332
struct intel_global_obj obj;
sys/dev/pci/drm/i915/display/intel_display_core.h
354
struct intel_global_obj obj;
sys/dev/pci/drm/i915/display/intel_display_core.h
369
struct intel_global_obj obj;
sys/dev/pci/drm/i915/display/intel_display_core.h
494
struct intel_global_obj obj;
sys/dev/pci/drm/i915/display/intel_dpt.c
121
i915_gem_object_put(dpt->obj);
sys/dev/pci/drm/i915/display/intel_dpt.c
137
if (i915_gem_object_is_stolen(dpt->obj))
sys/dev/pci/drm/i915/display/intel_dpt.c
144
err = i915_gem_object_lock(dpt->obj, &ww);
sys/dev/pci/drm/i915/display/intel_dpt.c
148
vma = i915_gem_object_ggtt_pin_ww(dpt->obj, &ww, NULL, 0,
sys/dev/pci/drm/i915/display/intel_dpt.c
169
dpt->obj->mm.dirty = true;
sys/dev/pci/drm/i915/display/intel_dpt.c
21
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_dpt.c
246
struct drm_gem_object *obj = intel_fb_bo(&fb->base);
sys/dev/pci/drm/i915/display/intel_dpt.c
247
struct drm_i915_private *i915 = to_i915(obj->dev);
sys/dev/pci/drm/i915/display/intel_dpt.c
257
size = DIV_ROUND_UP_ULL(obj->size, I915_GTT_PAGE_SIZE);
sys/dev/pci/drm/i915/display/intel_dpt.c
307
dpt->obj = dpt_obj;
sys/dev/pci/drm/i915/display/intel_dpt.c
308
dpt->obj->is_dpt = true;
sys/dev/pci/drm/i915/display/intel_dpt.c
317
dpt->obj->is_dpt = false;
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
38
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
43
obj = i915_gem_object_create_lmem(i915, PAGE_ALIGN(size),
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
45
if (IS_ERR(obj))
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
48
obj = i915_gem_object_create_internal(i915, PAGE_ALIGN(size));
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
49
if (IS_ERR(obj))
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
52
i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE);
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
55
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, 0);
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
57
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
61
buf = i915_gem_object_pin_map_unlocked(vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/display/intel_dsb_buffer.c
81
i915_gem_object_flush_map(dsb_buf->vma->obj);
sys/dev/pci/drm/i915/display/intel_fb.c
1342
struct drm_gem_object *obj = intel_fb_bo(&fb->base);
sys/dev/pci/drm/i915/display/intel_fb.c
1366
if (color_plane == 0 && intel_bo_is_tiled(obj) &&
sys/dev/pci/drm/i915/display/intel_fb.c
1705
struct drm_gem_object *obj = intel_fb_bo(&fb->base);
sys/dev/pci/drm/i915/display/intel_fb.c
1792
if (mul_u32_u32(max_size, tile_size) > obj->size) {
sys/dev/pci/drm/i915/display/intel_fb.c
1795
mul_u32_u32(max_size, tile_size), obj->size);
sys/dev/pci/drm/i915/display/intel_fb.c
2128
struct drm_gem_object *obj = intel_fb_bo(fb);
sys/dev/pci/drm/i915/display/intel_fb.c
2129
struct intel_display *display = to_intel_display(obj->dev);
sys/dev/pci/drm/i915/display/intel_fb.c
2131
if (intel_bo_is_userptr(obj)) {
sys/dev/pci/drm/i915/display/intel_fb.c
2137
return drm_gem_handle_create(file, obj, handle);
sys/dev/pci/drm/i915/display/intel_fb.c
2161
struct drm_gem_object *obj = intel_fb_bo(fb);
sys/dev/pci/drm/i915/display/intel_fb.c
2170
if (dma_resv_test_signaled(obj->resv, dma_resv_usage_rw(false)))
sys/dev/pci/drm/i915/display/intel_fb.c
2173
ret = dma_resv_get_singleton(obj->resv, dma_resv_usage_rw(false),
sys/dev/pci/drm/i915/display/intel_fb.c
2200
intel_bo_flush_if_display(obj);
sys/dev/pci/drm/i915/display/intel_fb.c
2212
struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_fb.c
2216
struct intel_display *display = to_intel_display(obj->dev);
sys/dev/pci/drm/i915/display/intel_fb.c
2230
intel_fb->frontbuffer = intel_frontbuffer_get(obj);
sys/dev/pci/drm/i915/display/intel_fb.c
2236
ret = intel_fb_bo_framebuffer_init(fb, obj, mode_cmd);
sys/dev/pci/drm/i915/display/intel_fb.c
2298
fb->obj[i] = obj;
sys/dev/pci/drm/i915/display/intel_fb.c
2330
intel_fb_bo_framebuffer_fini(obj);
sys/dev/pci/drm/i915/display/intel_fb.c
2346
struct drm_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_fb.c
2349
obj = intel_fb_bo_lookup_valid_bo(dev, filp, &mode_cmd);
sys/dev/pci/drm/i915/display/intel_fb.c
2350
if (IS_ERR(obj))
sys/dev/pci/drm/i915/display/intel_fb.c
2351
return ERR_CAST(obj);
sys/dev/pci/drm/i915/display/intel_fb.c
2353
fb = intel_framebuffer_create(obj, info, &mode_cmd);
sys/dev/pci/drm/i915/display/intel_fb.c
2354
drm_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_fb.c
2371
intel_framebuffer_create(struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_fb.c
2382
ret = intel_framebuffer_init(intel_fb, obj, info, mode_cmd);
sys/dev/pci/drm/i915/display/intel_fb.c
2395
return fb ? fb->obj[0] : NULL;
sys/dev/pci/drm/i915/display/intel_fb.h
104
struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_fb.h
111
intel_framebuffer_create(struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_fb_bo.c
100
return intel_bo_to_drm_bo(obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
16
void intel_fb_bo_framebuffer_fini(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_fb_bo.c
25
struct drm_i915_gem_object *obj = to_intel_bo(_obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
26
struct intel_display *display = to_intel_display(obj->base.dev);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
29
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
30
tiling = i915_gem_object_get_tiling(obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
31
stride = i915_gem_object_get_stride(obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
32
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
86
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_fb_bo.c
88
obj = i915_gem_object_lookup(filp, mode_cmd->handles[0]);
sys/dev/pci/drm/i915/display/intel_fb_bo.c
89
if (!obj)
sys/dev/pci/drm/i915/display/intel_fb_bo.c
93
if (HAS_LMEM(i915) && !i915_gem_object_can_migrate(obj, INTEL_REGION_LMEM_0)) {
sys/dev/pci/drm/i915/display/intel_fb_bo.c
95
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.h
15
void intel_fb_bo_framebuffer_fini(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_fb_bo.h
18
struct drm_gem_object *obj,
sys/dev/pci/drm/i915/display/intel_fb_pin.c
100
i915_gem_object_flush_if_display(obj);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
122
struct drm_i915_gem_object *obj = to_intel_bo(_obj);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
129
if (drm_WARN_ON(dev, !i915_gem_object_is_framebuffer(obj)))
sys/dev/pci/drm/i915/display/intel_fb_pin.c
160
ret = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
162
ret = i915_gem_object_attach_phys(obj, phys_alignment);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
164
ret = i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
166
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
170
vma = i915_gem_object_pin_to_display_plane(obj, &ww, alignment,
sys/dev/pci/drm/i915/display/intel_fb_pin.c
209
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
318
struct drm_i915_gem_object *obj = to_intel_bo(intel_fb_bo(&fb->base));
sys/dev/pci/drm/i915/display/intel_fb_pin.c
320
plane_state->surf = i915_gem_object_get_dma_address(obj, 0) +
sys/dev/pci/drm/i915/display/intel_fb_pin.c
34
struct drm_i915_gem_object *obj = to_intel_bo(_obj);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
46
if (WARN_ON(!i915_gem_object_is_framebuffer(obj)))
sys/dev/pci/drm/i915/display/intel_fb_pin.c
52
ret = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
57
unsigned int flags = obj->flags;
sys/dev/pci/drm/i915/display/intel_fb_pin.c
67
ret = __i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0,
sys/dev/pci/drm/i915/display/intel_fb_pin.c
73
ret = i915_gem_object_set_cache_level(obj, I915_CACHE_NONE);
sys/dev/pci/drm/i915/display/intel_fb_pin.c
77
vma = i915_vma_instance(obj, vm, view);
sys/dev/pci/drm/i915/display/intel_fbdev.c
130
struct drm_gem_object *obj = drm_gem_fb_get_obj(fb_helper->fb, 0);
sys/dev/pci/drm/i915/display/intel_fbdev.c
132
return intel_bo_fb_mmap(obj, vma);
sys/dev/pci/drm/i915/display/intel_fbdev.c
258
struct drm_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_fbdev.c
321
obj = intel_fb_bo(&fb->base);
sys/dev/pci/drm/i915/display/intel_fbdev.c
323
ret = intel_fbdev_fb_fill_info(display, info, obj, vma);
sys/dev/pci/drm/i915/display/intel_fbdev.c
333
if (!intel_bo_is_shmem(obj) && !prealloc)
sys/dev/pci/drm/i915/display/intel_fbdev.c
411
struct drm_gem_object *obj = intel_fb_bo(plane_state->uapi.fb);
sys/dev/pci/drm/i915/display/intel_fbdev.c
420
if (!obj) {
sys/dev/pci/drm/i915/display/intel_fbdev.c
427
if (obj->size > max_size) {
sys/dev/pci/drm/i915/display/intel_fbdev.c
432
max_size = obj->size;
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
115
info->screen_size = intel_bo_to_drm_bo(obj)->size;
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
22
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
28
obj = ERR_PTR(-ENODEV);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
30
obj = i915_gem_object_create_lmem(dev_priv, size,
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
42
obj = i915_gem_object_create_stolen(dev_priv, size);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
43
if (IS_ERR(obj))
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
44
obj = i915_gem_object_create_shmem(dev_priv, size);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
47
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
48
drm_err(drm, "failed to allocate framebuffer (%pe)\n", obj);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
52
fb = intel_framebuffer_create(intel_bo_to_drm_bo(obj),
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
58
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
62
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
73
struct drm_i915_gem_object *obj = to_intel_bo(_obj);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
78
if (i915_gem_object_is_lmem(obj)) {
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
79
struct intel_memory_region *mem = obj->mm.region;
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
84
i915_gem_object_get_dma_address(obj, 0) -
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
86
info->fix.smem_len = obj->base.size;
sys/dev/pci/drm/i915/display/intel_fbdev_fb.c
97
ret = i915_gem_object_lock(vma->obj, &ww);
sys/dev/pci/drm/i915/display/intel_fbdev_fb.h
19
struct drm_gem_object *obj, struct i915_vma *vma);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
176
struct intel_display *display = to_intel_display(front->obj->dev);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
197
struct intel_display *display = to_intel_display(front->obj->dev);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
216
intel_bo_flush_if_display(front->obj);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
257
__releases(&to_intel_display(front->obj->dev)->fb_tracking.lock)
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
261
struct drm_gem_object *obj = front->obj;
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
262
struct intel_display *display = to_intel_display(obj->dev);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
266
i915_ggtt_clear_scanout(to_intel_bo(obj));
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
268
ret = intel_bo_set_frontbuffer(obj, NULL);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
274
drm_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
279
intel_frontbuffer_get(struct drm_gem_object *obj)
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
281
struct intel_display *display = to_intel_display(obj->dev);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
284
front = intel_bo_get_frontbuffer(obj);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
292
drm_gem_object_get(obj);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
294
front->obj = obj;
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
304
cur = intel_bo_set_frontbuffer(obj, front);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
308
drm_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
319
&to_intel_display(front->obj->dev)->fb_tracking.lock);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
348
struct intel_display *display = to_intel_display(old->obj->dev);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
356
struct intel_display *display = to_intel_display(new->obj->dev);
sys/dev/pci/drm/i915/display/intel_frontbuffer.h
48
struct drm_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_frontbuffer.h
81
intel_frontbuffer_get(struct drm_gem_object *obj);
sys/dev/pci/drm/i915/display/intel_global_state.c
112
struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/intel_global_state.c
116
memset(obj, 0, sizeof(*obj));
sys/dev/pci/drm/i915/display/intel_global_state.c
118
state->obj = obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
122
obj->state = state;
sys/dev/pci/drm/i915/display/intel_global_state.c
123
obj->funcs = funcs;
sys/dev/pci/drm/i915/display/intel_global_state.c
124
list_add_tail(&obj->head, &display->global.obj_list);
sys/dev/pci/drm/i915/display/intel_global_state.c
129
struct intel_global_obj *obj, *next;
sys/dev/pci/drm/i915/display/intel_global_state.c
131
list_for_each_entry_safe(obj, next, &display->global.obj_list, head) {
sys/dev/pci/drm/i915/display/intel_global_state.c
132
list_del(&obj->head);
sys/dev/pci/drm/i915/display/intel_global_state.c
134
drm_WARN_ON(display->drm, kref_read(&obj->state->ref) != 1);
sys/dev/pci/drm/i915/display/intel_global_state.c
135
intel_atomic_global_state_put(obj->state);
sys/dev/pci/drm/i915/display/intel_global_state.c
16
#define for_each_new_global_obj_in_state(__state, obj, new_obj_state, __i) \
sys/dev/pci/drm/i915/display/intel_global_state.c
176
struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/intel_global_state.c
185
if (obj == state->global_objs[i].ptr)
sys/dev/pci/drm/i915/display/intel_global_state.c
19
((obj) = (__state)->global_objs[__i].ptr, \
sys/dev/pci/drm/i915/display/intel_global_state.c
209
obj_state = obj->funcs->atomic_duplicate_state(obj);
sys/dev/pci/drm/i915/display/intel_global_state.c
213
obj_state->obj = obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
22
for_each_if(obj)
sys/dev/pci/drm/i915/display/intel_global_state.c
222
intel_atomic_global_state_get(obj->state);
sys/dev/pci/drm/i915/display/intel_global_state.c
224
state->global_objs[index].ptr = obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
230
obj, obj_state, state);
sys/dev/pci/drm/i915/display/intel_global_state.c
237
struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/intel_global_state.c
24
#define for_each_old_global_obj_in_state(__state, obj, old_obj_state, __i) \
sys/dev/pci/drm/i915/display/intel_global_state.c
242
if (obj == state->global_objs[i].ptr)
sys/dev/pci/drm/i915/display/intel_global_state.c
250
struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/intel_global_state.c
255
if (obj == state->global_objs[i].ptr)
sys/dev/pci/drm/i915/display/intel_global_state.c
265
struct intel_global_obj *obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
268
for_each_oldnew_global_obj_in_state(state, obj, old_obj_state,
sys/dev/pci/drm/i915/display/intel_global_state.c
27
((obj) = (__state)->global_objs[__i].ptr, \
sys/dev/pci/drm/i915/display/intel_global_state.c
270
drm_WARN_ON(display->drm, obj->state != old_obj_state);
sys/dev/pci/drm/i915/display/intel_global_state.c
286
intel_atomic_global_state_put(obj->state);
sys/dev/pci/drm/i915/display/intel_global_state.c
287
obj->state = intel_atomic_global_state_get(new_obj_state);
sys/dev/pci/drm/i915/display/intel_global_state.c
30
for_each_if(obj)
sys/dev/pci/drm/i915/display/intel_global_state.c
32
#define for_each_oldnew_global_obj_in_state(__state, obj, old_obj_state, new_obj_state, __i) \
sys/dev/pci/drm/i915/display/intel_global_state.c
35
((obj) = (__state)->global_objs[__i].ptr, \
sys/dev/pci/drm/i915/display/intel_global_state.c
357
struct intel_global_obj *obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
360
for_each_oldnew_global_obj_in_state(state, obj, old_obj_state,
sys/dev/pci/drm/i915/display/intel_global_state.c
39
for_each_if(obj)
sys/dev/pci/drm/i915/display/intel_global_state.c
391
struct intel_global_obj *obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
394
for_each_old_global_obj_in_state(state, obj, old_obj_state, i) {
sys/dev/pci/drm/i915/display/intel_global_state.c
415
struct intel_global_obj *obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
418
for_each_new_global_obj_in_state(state, obj, new_obj_state, i) {
sys/dev/pci/drm/i915/display/intel_global_state.c
91
struct intel_global_obj *obj = obj_state->obj;
sys/dev/pci/drm/i915/display/intel_global_state.c
95
obj->funcs->atomic_destroy_state(obj, obj_state);
sys/dev/pci/drm/i915/display/intel_global_state.h
19
struct intel_global_state *(*atomic_duplicate_state)(struct intel_global_obj *obj);
sys/dev/pci/drm/i915/display/intel_global_state.h
20
void (*atomic_destroy_state)(struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/intel_global_state.h
31
struct intel_global_obj *obj;
sys/dev/pci/drm/i915/display/intel_global_state.h
39
struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/intel_global_state.h
46
struct intel_global_obj *obj);
sys/dev/pci/drm/i915/display/intel_global_state.h
49
struct intel_global_obj *obj);
sys/dev/pci/drm/i915/display/intel_global_state.h
52
struct intel_global_obj *obj);
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
42
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
48
obj = i915_gem_object_create_shmem(i915, 2 * PAGE_SIZE);
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
50
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
52
return PTR_ERR(obj);
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
55
cmd_in = i915_gem_object_pin_map_unlocked(obj, intel_gt_coherent_map_type(gt, obj, true));
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
64
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
74
memset(cmd_in, 0, obj->base.size);
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
84
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/display/intel_hdcp_gsc.c
86
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_modeset_setup.c
125
to_intel_pmdemand_state(display->pmdemand.obj.state);
sys/dev/pci/drm/i915/display/intel_modeset_setup.c
161
to_intel_pmdemand_state(display->pmdemand.obj.state);
sys/dev/pci/drm/i915/display/intel_modeset_setup.c
582
to_intel_pmdemand_state(display->pmdemand.obj.state);
sys/dev/pci/drm/i915/display/intel_modeset_setup.c
693
to_intel_pmdemand_state(display->pmdemand.obj.state);
sys/dev/pci/drm/i915/display/intel_overlay.c
1364
struct drm_i915_gem_object *obj = ERR_PTR(-ENODEV);
sys/dev/pci/drm/i915/display/intel_overlay.c
1369
obj = i915_gem_object_create_stolen(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/display/intel_overlay.c
1370
if (IS_ERR(obj))
sys/dev/pci/drm/i915/display/intel_overlay.c
1371
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/display/intel_overlay.c
1372
if (IS_ERR(obj))
sys/dev/pci/drm/i915/display/intel_overlay.c
1373
return PTR_ERR(obj);
sys/dev/pci/drm/i915/display/intel_overlay.c
1375
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/display/intel_overlay.c
1382
overlay->flip_addr = sg_dma_address(obj->mm.pages->sgl);
sys/dev/pci/drm/i915/display/intel_overlay.c
1393
overlay->reg_bo = obj;
sys/dev/pci/drm/i915/display/intel_overlay.c
1397
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_overlay.c
301
frontbuffer = intel_frontbuffer_get(intel_bo_to_drm_bo(vma->obj));
sys/dev/pci/drm/i915/display/intel_panic.c
19
struct drm_gem_object *obj = intel_fb_bo(&fb->base);
sys/dev/pci/drm/i915/display/intel_panic.c
21
return i915_gem_object_panic_setup(panic, sb, obj, fb->panic_tiling);
sys/dev/pci/drm/i915/display/intel_plane.c
1189
struct drm_gem_object *obj = intel_fb_bo(new_plane_state->hw.fb);
sys/dev/pci/drm/i915/display/intel_plane.c
1217
if (!obj)
sys/dev/pci/drm/i915/display/intel_plane.c
1270
struct drm_gem_object *obj = intel_fb_bo(old_plane_state->hw.fb);
sys/dev/pci/drm/i915/display/intel_plane.c
1272
if (!obj)
sys/dev/pci/drm/i915/display/intel_plane.c
1388
struct drm_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_plane.c
1400
obj = intel_fb_bo(fb);
sys/dev/pci/drm/i915/display/intel_plane.c
1401
if (!obj)
sys/dev/pci/drm/i915/display/intel_plane_initial.c
130
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/display/intel_plane_initial.c
163
obj = i915_gem_object_create_region_at(mem, phys_base, size,
sys/dev/pci/drm/i915/display/intel_plane_initial.c
166
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/display/intel_plane_initial.c
177
i915_gem_object_set_cache_coherency(obj, HAS_WT(i915) ?
sys/dev/pci/drm/i915/display/intel_plane_initial.c
185
obj->tiling_and_stride =
sys/dev/pci/drm/i915/display/intel_plane_initial.c
217
vma = i915_vma_instance(obj, &to_gt(i915)->ggtt->vm, NULL);
sys/dev/pci/drm/i915/display/intel_plane_initial.c
223
if (!i915_gem_object_is_lmem(obj))
sys/dev/pci/drm/i915/display/intel_plane_initial.c
238
if (i915_gem_object_is_tiled(obj) &&
sys/dev/pci/drm/i915/display/intel_plane_initial.c
254
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/display/intel_plane_initial.c
292
intel_bo_to_drm_bo(vma->obj),
sys/dev/pci/drm/i915/display/intel_pmdemand.c
111
&display->pmdemand.obj);
sys/dev/pci/drm/i915/display/intel_pmdemand.c
127
intel_atomic_global_obj_init(display, &display->pmdemand.obj,
sys/dev/pci/drm/i915/display/intel_pmdemand.c
55
intel_pmdemand_duplicate_state(struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/intel_pmdemand.c
59
pmdemand_state = kmemdup(obj->state, sizeof(*pmdemand_state), GFP_KERNEL);
sys/dev/pci/drm/i915/display/intel_pmdemand.c
66
static void intel_pmdemand_destroy_state(struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/intel_pmdemand.c
83
&display->pmdemand.obj);
sys/dev/pci/drm/i915/display/intel_pmdemand.c
97
&display->pmdemand.obj);
sys/dev/pci/drm/i915/display/skl_universal_plane.c
2254
struct drm_gem_object *obj = intel_fb_bo(fb);
sys/dev/pci/drm/i915/display/skl_universal_plane.c
2259
plane_state->decrypt = intel_pxp_key_check(obj, false) == 0;
sys/dev/pci/drm/i915/display/skl_universal_plane.c
2260
plane_state->force_black = intel_bo_is_protected(obj) &&
sys/dev/pci/drm/i915/display/skl_watermark.c
3060
to_intel_dbuf_state(display->dbuf.obj.state);
sys/dev/pci/drm/i915/display/skl_watermark.c
3293
static struct intel_global_state *intel_dbuf_duplicate_state(struct intel_global_obj *obj)
sys/dev/pci/drm/i915/display/skl_watermark.c
3297
dbuf_state = kmemdup(obj->state, sizeof(*dbuf_state), GFP_KERNEL);
sys/dev/pci/drm/i915/display/skl_watermark.c
3304
static void intel_dbuf_destroy_state(struct intel_global_obj *obj,
sys/dev/pci/drm/i915/display/skl_watermark.c
3321
dbuf_state = intel_atomic_get_global_obj_state(state, &display->dbuf.obj);
sys/dev/pci/drm/i915/display/skl_watermark.c
3336
intel_atomic_global_obj_init(display, &display->dbuf.obj,
sys/dev/pci/drm/i915/display/skl_watermark.c
3691
to_intel_dbuf_state(display->dbuf.obj.state);
sys/dev/pci/drm/i915/display/skl_watermark.c
3714
to_intel_dbuf_state(display->dbuf.obj.state);
sys/dev/pci/drm/i915/display/skl_watermark.c
3791
to_intel_dbuf_state(display->dbuf.obj.state);
sys/dev/pci/drm/i915/display/skl_watermark.c
52
to_intel_dbuf_state(intel_atomic_get_old_global_obj_state(state, &to_intel_display(state)->dbuf.obj))
sys/dev/pci/drm/i915/display/skl_watermark.c
54
to_intel_dbuf_state(intel_atomic_get_new_global_obj_state(state, &to_intel_display(state)->dbuf.obj))
sys/dev/pci/drm/i915/gem/i915_gem_busy.c
116
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_busy.c
123
obj = i915_gem_object_lookup_rcu(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_busy.c
124
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_busy.c
145
dma_resv_iter_begin(&cursor, obj->base.resv, DMA_RESV_USAGE_READ);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
104
obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ)
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
107
trace_i915_gem_object_clflush(obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
111
dma_resv_reserve_fences(obj->base.resv, 1) == 0)
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
112
clflush = clflush_work_create(obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
115
obj->base.resv, true,
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
118
dma_resv_add_fence(obj->base.resv, &clflush->base.dma,
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
128
obj->cache_dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
129
} else if (obj->mm.pages) {
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
130
__do_clflush(obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
131
obj->cache_dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
133
GEM_BUG_ON(obj->write_domain != I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
17
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
20
static void __do_clflush(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
22
GEM_BUG_ON(!i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
23
drm_clflush_sg(obj->mm.pages);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
25
i915_gem_object_flush_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
32
__do_clflush(clflush->obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
39
i915_gem_object_unpin_pages(clflush->obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
40
i915_gem_object_put(clflush->obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
49
static struct clflush *clflush_work_create(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
53
GEM_BUG_ON(!obj->cache_dirty);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
59
if (__i915_gem_object_get_pages(obj) < 0) {
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
65
clflush->obj = i915_gem_object_get(obj); /* obj <-> clflush cycle */
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
70
bool i915_gem_clflush_object(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
73
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
76
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
79
WARN_ON_ONCE(obj->cache_dirty);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
90
if (!i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
91
obj->cache_dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_clflush.h
14
bool i915_gem_clflush_object(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_context.c
120
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gem/i915_gem_context.c
123
if (!kref_get_unless_zero(&obj->base.refcount))
sys/dev/pci/drm/i915/gem/i915_gem_context.c
126
spin_lock(&obj->lut_lock);
sys/dev/pci/drm/i915/gem/i915_gem_context.c
127
list_for_each_entry(lut, &obj->lut_list, obj_link) {
sys/dev/pci/drm/i915/gem/i915_gem_context.c
137
spin_unlock(&obj->lut_lock);
sys/dev/pci/drm/i915/gem/i915_gem_context.c
139
if (&lut->obj_link != &obj->lut_list) {
sys/dev/pci/drm/i915/gem/i915_gem_context.c
143
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_context.c
146
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
113
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/i915_gem_create.c
114
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_create.c
117
ret = object_set_placements(obj, placements, n_placements);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
127
ret = mr->ops->init_object(mr, obj, I915_BO_INVALID_OFFSET, size, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
131
GEM_BUG_ON(size != obj->base.size);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
134
obj->flags |= ext_flags;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
136
trace_i915_gem_object_create(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
137
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
140
if (obj->mm.n_placements > 1)
sys/dev/pci/drm/i915/gem/i915_gem_create.c
141
kfree(obj->mm.placements);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
142
i915_gem_object_free(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
172
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
211
obj = __i915_gem_object_create_user(to_i915(dev), args->size, &mr, 1);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
212
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/i915_gem_create.c
213
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
215
return i915_gem_publish(obj, file, &args->size, &args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
230
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
235
obj = __i915_gem_object_create_user(i915, args->size, &mr, 1);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
236
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/i915_gem_create.c
237
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
239
return i915_gem_publish(obj, file, &args->size, &args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
36
static int object_set_placements(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_create.c
448
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
484
obj = __i915_gem_object_create_user_ext(i915, args->size,
sys/dev/pci/drm/i915/gem/i915_gem_create.c
488
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/i915_gem_create.c
489
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
492
i915_gem_object_set_pat_index(obj, ext_data.pat_index);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
494
obj->pat_set_by_user = true;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
497
return i915_gem_publish(obj, file, &args->size, &args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
53
obj->mm.placements = &i915->mm.regions[mr->id];
sys/dev/pci/drm/i915/gem/i915_gem_create.c
54
obj->mm.n_placements = 1;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
65
obj->mm.placements = arr;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
66
obj->mm.n_placements = n_placements;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
72
static int i915_gem_publish(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_create.c
77
u64 size = obj->base.size;
sys/dev/pci/drm/i915/gem/i915_gem_create.c
80
ret = drm_gem_handle_create(file, &obj->base, handle_p);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
82
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_create.c
97
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
101
if (obj->base.size < vma->vm_end - vma->vm_start)
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
105
return drm_gem_prime_mmap(&obj->base, vma);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
107
if (!obj->base.filp)
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
110
ret = vfs_mmap(obj->base.filp, vma);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
114
vma_set_file(vma, obj->base.filp);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
121
struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
128
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
130
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
132
err = i915_gem_object_set_to_cpu_domain(obj, write);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
133
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
146
struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
152
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
154
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
156
err = i915_gem_object_set_to_gtt_domain(obj, false);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
157
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
171
struct drm_i915_gem_object *obj = dma_buf_to_obj(dmabuf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
175
if (!i915_gem_object_can_migrate(obj, INTEL_REGION_SMEM))
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
179
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
183
err = i915_gem_object_migrate(obj, &ww, INTEL_REGION_SMEM);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
187
err = i915_gem_object_wait_migration(obj, 0);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
191
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
200
struct drm_i915_gem_object *obj = dma_buf_to_obj(dmabuf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
202
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
226
struct drm_i915_gem_object *obj = to_intel_bo(gem_obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
233
exp_info.resv = obj->base.resv;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
235
if (obj->ops->dmabuf_export) {
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
236
int ret = obj->ops->dmabuf_export(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
246
static int i915_gem_object_get_pages_dmabuf(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
248
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
251
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
253
sgt = dma_buf_map_attachment(obj->base.import_attach,
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
269
if (i915_gem_object_can_bypass_llc(obj) ||
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
273
__i915_gem_object_set_pages(obj, sgt);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
278
static void i915_gem_object_put_pages_dmabuf(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
281
dma_buf_unmap_attachment(obj->base.import_attach, sgt,
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
298
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
303
obj = dma_buf_to_obj(dma_buf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
305
if (obj->base.dev == dev &&
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
311
return &i915_gem_object_get(obj)->base;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
32
struct drm_i915_gem_object *obj = dma_buf_to_obj(attach->dmabuf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
326
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
327
if (!obj) {
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
332
drm_gem_private_object_init(dev, &obj->base, dma_buf->size);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
333
i915_gem_object_init(obj, &i915_gem_object_dmabuf_ops, &lock_class,
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
335
obj->base.import_attach = attach;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
336
obj->base.resv = dma_buf->resv;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
345
obj->read_domains = I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
346
obj->write_domain = 0;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
348
return &obj->base;
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
47
ret = sg_alloc_table(sgt, obj->mm.pages->orig_nents, GFP_KERNEL);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
52
for_each_sg(obj->mm.pages->sgl, src, obj->mm.pages->orig_nents, i) {
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
74
struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
77
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
89
struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
91
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
92
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
97
struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
sys/dev/pci/drm/i915/gem/i915_gem_dmabuf.c
98
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
100
obj->write_domain = 0;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
103
void i915_gem_object_flush_if_display(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
105
if (!i915_gem_object_is_framebuffer(obj))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
108
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
109
__i915_gem_object_flush_for_display(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
110
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
113
void i915_gem_object_flush_if_display_locked(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
115
if (i915_gem_object_is_framebuffer(obj))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
116
__i915_gem_object_flush_for_display(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
129
i915_gem_object_set_to_wc_domain(struct drm_i915_gem_object *obj, bool write)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
133
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
135
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
142
if (obj->write_domain == I915_GEM_DOMAIN_WC)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
153
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
157
flush_write_domain(obj, ~I915_GEM_DOMAIN_WC);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
163
if ((obj->read_domains & I915_GEM_DOMAIN_WC) == 0)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
169
GEM_BUG_ON((obj->write_domain & ~I915_GEM_DOMAIN_WC) != 0);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
170
obj->read_domains |= I915_GEM_DOMAIN_WC;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
172
obj->read_domains = I915_GEM_DOMAIN_WC;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
173
obj->write_domain = I915_GEM_DOMAIN_WC;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
174
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
177
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
19
static bool gpu_write_needs_clflush(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
191
i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
195
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
197
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
204
if (obj->write_domain == I915_GEM_DOMAIN_GTT)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
21
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
215
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
219
flush_write_domain(obj, ~I915_GEM_DOMAIN_GTT);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
225
if ((obj->read_domains & I915_GEM_DOMAIN_GTT) == 0)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
231
GEM_BUG_ON((obj->write_domain & ~I915_GEM_DOMAIN_GTT) != 0);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
232
obj->read_domains |= I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
236
obj->read_domains = I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
237
obj->write_domain = I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
238
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
240
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
241
for_each_ggtt_vma(vma, obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
244
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
247
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
266
int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
277
if (i915_gem_object_has_cache_level(obj, cache_level))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
280
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
288
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
289
obj->cache_dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
292
return i915_gem_object_unbind(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
301
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
308
obj = i915_gem_object_lookup_rcu(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
309
if (!obj) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
318
if (obj->pat_set_by_user) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
323
if (i915_gem_object_has_cache_level(obj, I915_CACHE_LLC) ||
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
324
i915_gem_object_has_cache_level(obj, I915_CACHE_L3_LLC))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
326
else if (i915_gem_object_has_cache_level(obj, I915_CACHE_WT))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
33
return !(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) ||
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
34
i915_gem_object_has_cache_level(obj, I915_CACHE_WT));
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
340
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
37
bool i915_gem_cpu_write_needs_clflush(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
373
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
374
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
381
if (obj->pat_set_by_user) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
39
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
390
if (i915_gem_object_is_proxy(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
395
if (!i915_gem_object_is_userptr(obj) ||
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
402
ret = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
406
ret = i915_gem_object_set_cache_level(obj, level);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
407
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
41
if (obj->cache_dirty)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
410
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
421
i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
427
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
432
if (HAS_LMEM(i915) && !i915_gem_object_is_lmem(obj))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
445
ret = i915_gem_object_set_cache_level(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
466
vma = i915_gem_object_ggtt_pin_ww(obj, ww, view, 0, alignment,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
47
if (!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
470
vma = i915_gem_object_ggtt_pin_ww(obj, ww, view, 0,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
478
i915_gem_object_flush_if_display_locked(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
493
i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
497
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
499
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
506
flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
509
if ((obj->read_domains & I915_GEM_DOMAIN_CPU) == 0) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
51
return i915_gem_object_is_framebuffer(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
510
i915_gem_clflush_object(obj, I915_CLFLUSH_SYNC);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
511
obj->read_domains |= I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
517
GEM_BUG_ON(obj->write_domain & ~I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
523
__start_cpu_write(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
541
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
55
flush_write_domain(struct drm_i915_gem_object *obj, unsigned int flush_domains)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
563
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
564
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
572
err = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
580
if (i915_gem_object_is_userptr(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
585
err = i915_gem_object_userptr_validate(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
587
err = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
59
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
601
if (i915_gem_object_is_proxy(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
606
err = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
61
if (!(obj->write_domain & flush_domains))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
619
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
633
if (READ_ONCE(obj->write_domain) == read_domains)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
637
err = i915_gem_object_set_to_wc_domain(obj, write_domain);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
639
err = i915_gem_object_set_to_gtt_domain(obj, write_domain);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
64
switch (obj->write_domain) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
641
err = i915_gem_object_set_to_cpu_domain(obj, write_domain);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
644
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
647
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
650
i915_gem_object_invalidate_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
653
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
66
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
662
int i915_gem_object_prepare_read(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
668
if (!i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
67
for_each_ggtt_vma(vma, obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
671
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
673
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
679
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
683
if (obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ ||
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
685
ret = i915_gem_object_set_to_cpu_domain(obj, false);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
69
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
692
flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
699
if (!obj->cache_dirty &&
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
700
!(obj->read_domains & I915_GEM_DOMAIN_CPU))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
708
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
71
i915_gem_object_flush_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
712
int i915_gem_object_prepare_write(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
718
if (!i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
721
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
723
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
730
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
734
if (obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE ||
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
736
ret = i915_gem_object_set_to_cpu_domain(obj, true);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
743
flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
750
if (!obj->cache_dirty) {
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
757
if (!(obj->read_domains & I915_GEM_DOMAIN_CPU))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
762
i915_gem_object_invalidate_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
763
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
768
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
79
i915_gem_clflush_object(obj, I915_CLFLUSH_SYNC);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
83
if (gpu_write_needs_clflush(obj))
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
84
obj->cache_dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
88
obj->write_domain = 0;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
91
static void __i915_gem_object_flush_for_display(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
97
flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
98
if (obj->cache_dirty)
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
99
i915_gem_clflush_object(obj, I915_CLFLUSH_FORCE);
sys/dev/pci/drm/i915/gem/i915_gem_domain.h
12
int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1033
err = dma_resv_reserve_fences(vma->obj->base.resv, eb->num_batches);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1150
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1158
struct vm_page *page = i915_gem_object_get_page(obj, cache->page);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1185
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1191
i915_gem_object_finish_access(obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1214
static void *reloc_kmap(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1227
err = i915_gem_object_prepare_write(obj, &flushes);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1235
cache->node.mm = (void *)obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1240
page = i915_gem_object_get_page(obj, pageno);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1241
if (!obj->mm.dirty)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1255
struct drm_i915_gem_object *obj = batch->obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1268
if (i915_gem_object_is_tiled(obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1271
if (use_cpu_reloc(cache, obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1274
err = i915_gem_object_set_to_gtt_domain(obj, true);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1287
vma = i915_gem_object_ggtt_pin_ww(obj, &eb->ww, NULL, 0, 0,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1316
i915_gem_object_get_dma_address(obj, page),
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1347
vaddr = reloc_kmap(vma->obj, cache, page);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1461
target->vma->obj->pat_index,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1464
reloc_cache_remap(&eb->reloc_cache, ev->vma->obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1753
if (!i915_gem_object_is_userptr(ev->vma->obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1756
ret = i915_gem_object_userptr_submit_init(ev->vma->obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2089
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2116
if (unlikely(obj->cache_dirty & ~obj->cache_coherent)) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2117
if (i915_gem_clflush_object(obj, 0))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2124
(eb_find_first_request_added(eb), obj,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2148
struct drm_i915_gem_object *obj = ev->vma->obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2150
if (!i915_gem_object_is_userptr(obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2153
err = i915_gem_object_userptr_submit_done(obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2232
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2239
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2257
return i915_gem_object_ggtt_pin_ww(vma->obj, &eb->ww, NULL, 0, 0, PIN_VALIDATE);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2306
err = i915_gem_object_lock(pool->obj, &eb->ww);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2310
shadow = shadow_batch_pin(eb, pool->obj, eb->context->vm, PIN_USER);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2315
i915_gem_object_set_readonly(shadow->obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2322
shadow = shadow_batch_pin(eb, pool->obj,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
2337
err = dma_resv_reserve_fences(shadow->obj->base.resv, 1);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
544
i915_gem_object_is_tiled(vma->obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
632
const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
634
if (!i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
650
obj->cache_dirty ||
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
651
!i915_gem_object_has_cache_level(obj, I915_CACHE_NONE));
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
862
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
864
spin_lock(&obj->lut_lock);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
865
if (idr_find(&eb->file->object_idr, handle) == obj) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
866
list_add(&lut->obj_link, &obj->lut_list);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
871
spin_unlock(&obj->lut_lock);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
892
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
904
obj = i915_gem_object_lookup(eb->file, handle);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
905
if (unlikely(!obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
916
i915_gem_object_is_protected(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
917
err = intel_pxp_key_check(intel_bo_to_drm_bo(obj), true);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
919
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
924
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
926
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
934
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
966
if (i915_gem_object_is_userptr(vma->obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
967
err = i915_gem_object_userptr_submit_init(vma->obj);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
988
err = i915_gem_object_lock(vma->obj, &eb->ww);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
108
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
120
static void i915_gem_object_put_pages_internal(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
123
i915_gem_gtt_finish_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
126
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
128
__start_cpu_write(obj);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
144
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
150
if (overflows_type(size, obj->base.size))
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
153
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
154
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
157
drm_gem_private_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
158
i915_gem_object_init(obj, ops, &lock_class, 0);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
159
obj->mem_flags |= I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
168
i915_gem_object_set_volatile(obj);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
170
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
171
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
174
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
176
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
33
static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
35
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
43
if (overflows_type(obj->base.size >> PAGE_SHIFT, npages))
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
46
npages = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_internal.c
98
if (i915_gem_gtt_prepare_pages(obj, st)) {
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
100
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
106
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
107
__i915_gem_object_release_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
109
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
14
i915_gem_object_lmem_io_map(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
20
GEM_BUG_ON(!i915_gem_object_is_contiguous(obj));
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
22
offset = i915_gem_object_get_dma_address(obj, n);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
23
offset -= obj->mm.region->region.start;
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
25
return io_mapping_map_wc(&obj->mm.region->iomap, offset, size);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
42
bool i915_gem_object_is_lmem(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
44
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
47
if (i915_gem_object_migratable(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
48
i915_gem_object_evictable(obj))
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
49
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
89
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
92
obj = i915_gem_object_create_lmem(i915,
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
95
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
96
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
98
map = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/i915_gem_lmem.h
16
i915_gem_object_lmem_io_map(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_lmem.h
20
bool i915_gem_object_is_lmem(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
100
if (!obj->base.filp) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1005
void i915_gem_object_runtime_pm_release_mmap_offset(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1007
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1018
GEM_BUG_ON(!obj->userfault_count);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1019
list_del(&obj->userfault_link);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1020
obj->userfault_count = 0;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1023
void i915_gem_object_release_mmap_offset(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1027
if (obj->ops->unmap_virtual)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1028
obj->ops->unmap_virtual(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1030
spin_lock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1032
&obj->mmo.offsets, offset) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1040
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1043
obj->base.dev->anon_inode->i_mapping);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1045
spin_lock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1047
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
105
if (!obj->base.uao) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1051
lookup_mmo(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1056
spin_lock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1057
rb = obj->mmo.offsets.rb_node;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1063
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1072
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1078
insert_mmo(struct drm_i915_gem_object *obj, struct i915_mmap_offset *mmo)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1082
spin_lock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1084
p = &obj->mmo.offsets.rb_node;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1092
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1093
drm_vma_offset_remove(obj->base.dev->vma_offset_manager,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1105
rb_insert_color(&mmo->offset, &obj->mmo.offsets);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1106
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
111
if (range_overflows(args->offset, args->size, (u64)obj->base.size)) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1112
mmap_offset_attach(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1116
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1120
GEM_BUG_ON(obj->ops->mmap_offset || obj->ops->mmap_ops);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1122
mmo = lookup_mmo(obj, mmap_type);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1130
mmo->obj = obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1134
err = drm_vma_offset_add(obj->base.dev->vma_offset_manager,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1135
&mmo->vma_node, obj->base.size / PAGE_SIZE);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1146
err = drm_vma_offset_add(obj->base.dev->vma_offset_manager,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1147
&mmo->vma_node, obj->base.size / PAGE_SIZE);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1152
mmo = insert_mmo(obj, mmo);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1153
GEM_BUG_ON(lookup_mmo(obj, mmap_type) != mmo);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1165
__assign_mmap_offset(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
117
addr = vm_mmap(obj->base.filp, 0, args->size,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1171
if (i915_gem_object_never_mmap(obj))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1174
if (obj->ops->mmap_offset) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1178
*offset = obj->ops->mmap_offset(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1186
!i915_gem_object_has_struct_page(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1187
!i915_gem_object_has_iomem(obj))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1190
mmo = mmap_offset_attach(obj, mmap_type, file);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1204
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1207
obj = i915_gem_object_lookup(file, handle);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1208
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1211
err = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1214
err = __assign_mmap_offset(obj, mmap_type, offset, file);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1215
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1217
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1319
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
132
if (vma && __vma_matches(vma, obj->base.filp, addr, args->size))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1321
GEM_BUG_ON(!obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1322
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1328
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1330
GEM_BUG_ON(!obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1331
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1385
i915_gem_object_mmap(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1389
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1393
if (i915_gem_object_is_readonly(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1395
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1403
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
141
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1421
if (obj->ops->mmap_ops) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1423
vma->vm_ops = obj->ops->mmap_ops;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1424
vma->vm_private_data = obj->base.vma_node.driver_private;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
144
uao_reference(obj->base.uao);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
146
obj->base.uao, args->offset, 0, UVM_MAPFLAG(PROT_READ | PROT_WRITE,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1473
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1492
obj = i915_gem_object_get_rcu(mmo->obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1494
GEM_BUG_ON(obj && obj->ops->mmap_ops);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1496
obj = i915_gem_object_get_rcu
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
150
uao_detach(obj->base.uao);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1500
GEM_BUG_ON(obj && !obj->ops->mmap_ops);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1505
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1508
return i915_gem_object_mmap(obj, mmo, vma);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
151
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1526
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1545
obj = i915_gem_object_get_rcu(mmo->obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1547
GEM_BUG_ON(obj && obj->ops->mmap_ops);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1549
obj = i915_gem_object_get_rcu
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1553
GEM_BUG_ON(obj && !obj->ops->mmap_ops);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1558
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1561
if (i915_gem_object_is_readonly(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1563
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1568
if (obj->ops->mmap_ops)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1569
uvm_obj_init(&obj->base.uobj, obj->ops->mmap_ops, 1);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1571
return &obj->base.uobj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1577
int i915_gem_fb_mmap(struct drm_i915_gem_object *obj, struct vm_area_struct *vma)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1579
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1589
if (obj->ops->mmap_ops) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1594
vma->vm_pgoff += drm_vma_node_start(&obj->base.vma_node);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1598
mmo = mmap_offset_attach(obj, mmap_type, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
160
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1611
obj = i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
1612
return i915_gem_object_mmap(obj, mmo, vma);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
164
static unsigned int tile_row_pages(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
166
return i915_gem_object_get_tile_row_size(obj) >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
232
compute_partial_view(const struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
238
if (i915_gem_object_is_tiled(obj))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
239
chunk = roundup(chunk, tile_row_pages(obj) ?: 1);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
245
(obj->base.size >> PAGE_SHIFT) - view.partial.offset);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
248
if (chunk >= obj->base.size >> PAGE_SHIFT)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
290
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
296
if (unlikely(i915_gem_object_is_readonly(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
300
if (i915_gem_object_lock_interruptible(obj, NULL))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
303
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
308
if (!i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
309
iomap = obj->mm.region->iomap.base;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
310
iomap -= obj->mm.region->region.start;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
317
obj->mm.pages->sgl, obj_offset, iomap);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
320
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
321
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
324
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
327
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
377
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
378
struct drm_device *dev = obj->base.dev;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
397
trace_i915_gem_object_fault(obj, page_offset, true, write);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
403
ret = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
408
if (i915_gem_object_is_readonly(obj) && write) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
413
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
422
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, NULL, 0, 0,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
429
compute_partial_view(obj, page_offset, MIN_CHUNK_PAGES);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
441
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, &view, 0, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
445
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, &view, 0, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
461
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, &view, 0, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
478
if (!(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) ||
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
505
if (!i915_vma_set_userfault(vma) && !obj->userfault_count++)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
506
list_add(&obj->userfault_link, &to_gt(i915)->ggtt->userfault_list);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
517
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
519
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
529
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
546
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
551
if (i915_gem_object_is_readonly(obj) && write)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
555
if (range_overflows_t(u64, addr, len, obj->base.size))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
560
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
565
vaddr = i915_gem_object_pin_map(obj, I915_MAP_FORCE_WC);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
573
__i915_gem_object_flush_map(obj, addr, len);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
578
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
630
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
641
if (unlikely(i915_gem_object_is_readonly(obj) && write)) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
642
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
646
if (i915_gem_object_lock_interruptible(obj, NULL))
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
649
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
66
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
668
pages = obj->mm.pages;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
681
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
682
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
685
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
688
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
689
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
751
struct drm_i915_gem_object *obj = mmo->obj;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
752
struct drm_device *dev = obj->base.dev;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
771
trace_i915_gem_object_fault(obj, page_offset, true, write);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
777
ret = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
782
if (i915_gem_object_is_readonly(obj) && write) {
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
787
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
796
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, NULL, 0, 0,
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
803
compute_partial_view(obj, page_offset, MIN_CHUNK_PAGES);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
815
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, &view, 0, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
819
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, &view, 0, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
835
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, &view, 0, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
852
if (!(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) ||
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
878
if (!i915_vma_set_userfault(vma) && !obj->userfault_count++)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
879
list_add(&obj->userfault_link, &to_gt(i915)->ggtt->userfault_list);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
890
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
892
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
902
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
911
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
92
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
93
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
938
KASSERT(gem_obj == &mmo->obj->base);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
948
void __i915_gem_object_release_mmap_gtt(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
952
GEM_BUG_ON(!obj->userfault_count);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
954
for_each_ggtt_vma(vma, obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
957
GEM_BUG_ON(obj->userfault_count);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
968
void i915_gem_object_release_mmap_gtt(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
970
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
985
if (!obj->userfault_count)
sys/dev/pci/drm/i915/gem/i915_gem_mman.c
988
__i915_gem_object_release_mmap_gtt(obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.h
32
void __i915_gem_object_release_mmap_gtt(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.h
33
void i915_gem_object_release_mmap_gtt(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.h
35
void i915_gem_object_runtime_pm_release_mmap_offset(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.h
36
void i915_gem_object_release_mmap_offset(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_mman.h
38
int i915_gem_fb_mmap(struct drm_i915_gem_object *obj, struct vm_area_struct *vma);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
1005
bool i915_gem_object_has_unknown_state(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
1013
return obj->mm.unknown_state;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
102
void i915_gem_object_init(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
110
BUILD_BUG_ON(offsetof(typeof(*obj), base) !=
sys/dev/pci/drm/i915/gem/i915_gem_object.c
111
offsetof(typeof(*obj), __do_not_access.base));
sys/dev/pci/drm/i915/gem/i915_gem_object.c
113
mtx_init(&obj->vma.lock, IPL_NONE);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
114
INIT_LIST_HEAD(&obj->vma.list);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
116
INIT_LIST_HEAD(&obj->mm.link);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
119
INIT_LIST_HEAD(&obj->client_link);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
122
INIT_LIST_HEAD(&obj->lut_list);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
123
mtx_init(&obj->lut_lock, IPL_NONE);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
125
mtx_init(&obj->mmo.lock, IPL_NONE);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
126
obj->mmo.offsets = RB_ROOT;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
128
init_rcu_head(&obj->rcu);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
130
obj->ops = ops;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
132
obj->flags = flags;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
134
obj->mm.madv = I915_MADV_WILLNEED;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
135
INIT_RADIX_TREE(&obj->mm.get_page.radix, GFP_KERNEL | __GFP_NOWARN);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
136
rw_init(&obj->mm.get_page.lock, "mmget");
sys/dev/pci/drm/i915/gem/i915_gem_object.c
137
INIT_RADIX_TREE(&obj->mm.get_dma_page.radix, GFP_KERNEL | __GFP_NOWARN);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
138
rw_init(&obj->mm.get_dma_page.lock, "mmgetd");
sys/dev/pci/drm/i915/gem/i915_gem_object.c
150
void __i915_gem_object_fini(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
152
mutex_destroy(&obj->mm.get_page.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
153
mutex_destroy(&obj->mm.get_dma_page.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
154
dma_resv_fini(&obj->base._resv);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
163
void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
166
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
168
obj->pat_index = i915_gem_get_pat_index(i915, cache_level);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
171
obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ |
sys/dev/pci/drm/i915/gem/i915_gem_object.c
174
obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
176
obj->cache_coherent = 0;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
178
obj->cache_dirty =
sys/dev/pci/drm/i915/gem/i915_gem_object.c
179
!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) &&
sys/dev/pci/drm/i915/gem/i915_gem_object.c
191
void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
194
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
196
if (obj->pat_index == pat_index)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
199
obj->pat_index = pat_index;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
202
obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ |
sys/dev/pci/drm/i915/gem/i915_gem_object.c
205
obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
207
obj->cache_coherent = 0;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
209
obj->cache_dirty =
sys/dev/pci/drm/i915/gem/i915_gem_object.c
210
!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) &&
sys/dev/pci/drm/i915/gem/i915_gem_object.c
214
bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
216
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
222
if (!(obj->flags & I915_BO_ALLOC_USER))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
228
if (obj->pat_set_by_user)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
248
struct drm_i915_gem_object *obj = to_intel_bo(gem);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
255
spin_lock(&obj->lut_lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
256
list_for_each_entry_safe(lut, ln, &obj->lut_list, obj_link) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
265
if (&ln->obj_link != &obj->lut_list) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
267
if (cond_resched_lock(&obj->lut_lock))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
272
spin_unlock(&obj->lut_lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
274
spin_lock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
275
rbtree_postorder_for_each_entry_safe(mmo, mn, &obj->mmo.offsets, offset)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
277
spin_unlock(&obj->mmo.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
291
GEM_BUG_ON(vma->obj != obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
299
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
305
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_object.c
306
container_of(head, typeof(*obj), rcu);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
307
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
310
if (obj->mm.n_placements > 1)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
311
kfree(obj->mm.placements);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
314
if (obj->base.uao)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
315
uao_detach(obj->base.uao);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
318
i915_gem_object_free(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
324
static void __i915_gem_object_free_mmaps(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
328
if (obj->userfault_count && !IS_DGFX(to_i915(obj->base.dev)))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
329
i915_gem_object_release_mmap_gtt(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
331
if (!RB_EMPTY_ROOT(&obj->mmo.offsets)) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
334
i915_gem_object_release_mmap_offset(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
337
&obj->mmo.offsets,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
339
drm_vma_offset_remove(obj->base.dev->vma_offset_manager,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
343
obj->mmo.offsets = RB_ROOT;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
356
void __i915_gem_object_pages_fini(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
358
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
360
if (!list_empty(&obj->vma.list)) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
363
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
364
while ((vma = list_first_entry_or_null(&obj->vma.list,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
367
GEM_BUG_ON(vma->obj != obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
368
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
372
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
374
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
377
__i915_gem_object_free_mmaps(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
379
atomic_set(&obj->mm.pages_pin_count, 0);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
387
if (obj->base.import_attach)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
388
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
390
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
392
if (obj->base.import_attach)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
393
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
395
GEM_BUG_ON(i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.c
398
void __i915_gem_free_object(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
400
trace_i915_gem_object_destroy(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
402
GEM_BUG_ON(!list_empty(&obj->lut_list));
sys/dev/pci/drm/i915/gem/i915_gem_object.c
404
bitmap_free(obj->bit_17);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
406
if (obj->base.import_attach)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
407
drm_prime_gem_destroy(&obj->base, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
409
drm_gem_free_mmap_offset(&obj->base);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
411
if (obj->ops->release)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
412
obj->ops->release(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
414
if (obj->shares_resv_from)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
415
i915_vm_resv_put(obj->shares_resv_from);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
417
__i915_gem_object_fini(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
423
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
425
llist_for_each_entry_safe(obj, on, freed, freed) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
427
if (obj->ops->delayed_free) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
428
obj->ops->delayed_free(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
432
__i915_gem_object_pages_fini(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
433
__i915_gem_free_object(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
436
call_rcu(&obj->rcu, __i915_gem_free_object_rcu);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
459
struct drm_i915_gem_object *obj = to_intel_bo(gem_obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
460
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
462
GEM_BUG_ON(i915_gem_object_is_framebuffer(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.c
464
i915_drm_client_remove_object(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
485
if (llist_add(&obj->freed, &i915->mm.free_list))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
489
void __i915_gem_object_flush_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
494
front = i915_gem_object_get_frontbuffer(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
501
void __i915_gem_object_invalidate_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
506
front = i915_gem_object_get_frontbuffer(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
514
i915_gem_object_read_from_page_kmap(struct drm_i915_gem_object *obj, u64 offset, void *dst, int size)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
519
src_ptr = kmap_local_page(i915_gem_object_get_page(obj, idx))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
521
if (!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
529
i915_gem_object_read_from_page_iomap(struct drm_i915_gem_object *obj, u64 offset, void *dst, int size)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
532
dma_addr_t dma = i915_gem_object_get_dma_address(obj, idx);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
536
src_map = io_mapping_map_wc(&obj->mm.region->iomap,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
537
dma - obj->mm.region->region.start,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
547
static bool object_has_mappable_iomem(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
549
GEM_BUG_ON(!i915_gem_object_has_iomem(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.c
551
if (IS_DGFX(to_i915(obj->base.dev)))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
552
return i915_ttm_resource_mappable(i915_gem_to_ttm(obj)->resource);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
571
int i915_gem_object_read_from_page(struct drm_i915_gem_object *obj, u64 offset, void *dst, int size)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
574
GEM_BUG_ON(offset >= obj->base.size);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
576
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.c
578
if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
579
i915_gem_object_read_from_page_kmap(obj, offset, dst, size);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
580
else if (i915_gem_object_has_iomem(obj) && object_has_mappable_iomem(obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
581
i915_gem_object_read_from_page_iomap(obj, offset, dst, size);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
59
bool i915_gem_object_has_cache_level(const struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
601
bool i915_gem_object_evictable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
604
int pin_count = atomic_read(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
609
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
610
list_for_each_entry(vma, &obj->vma.list, obj_link) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
612
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
618
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
632
bool i915_gem_object_migratable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
634
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
639
return obj->mm.n_placements > 1;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
651
bool i915_gem_object_has_struct_page(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
654
if (IS_DGFX(to_i915(obj->base.dev)) &&
sys/dev/pci/drm/i915/gem/i915_gem_object.c
655
i915_gem_object_evictable((void __force *)obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
656
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
658
return obj->mem_flags & I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
67
if (obj->pat_set_by_user)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
670
bool i915_gem_object_has_iomem(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
673
if (IS_DGFX(to_i915(obj->base.dev)) &&
sys/dev/pci/drm/i915/gem/i915_gem_object.c
674
i915_gem_object_evictable((void __force *)obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
675
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
677
return obj->mem_flags & I915_BO_FLAG_IOMEM;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
697
bool i915_gem_object_can_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
700
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
701
unsigned int num_allowed = obj->mm.n_placements;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
706
GEM_BUG_ON(obj->mm.madv != I915_MADV_WILLNEED);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
712
if (!IS_ALIGNED(obj->base.size, mr->min_page_size))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
715
if (obj->mm.region == mr)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
718
if (!i915_gem_object_evictable(obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
721
if (!obj->ops->migrate)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
724
if (!(obj->flags & I915_BO_ALLOC_USER))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
731
if (mr == obj->mm.placements[i])
sys/dev/pci/drm/i915/gem/i915_gem_object.c
74
return obj->pat_index == i915_gem_get_pat_index(obj_to_i915(obj), lvl);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
763
int i915_gem_object_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
767
return __i915_gem_object_migrate(obj, ww, id, obj->flags);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
79
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
797
int __i915_gem_object_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
802
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
806
GEM_BUG_ON(obj->mm.madv != I915_MADV_WILLNEED);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
807
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
812
if (!i915_gem_object_can_migrate(obj, id))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
815
if (!obj->ops->migrate) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
816
if (GEM_WARN_ON(obj->mm.region != mr))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
82
obj = kmem_cache_zalloc(slab_objects, GFP_KERNEL);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
821
return obj->ops->migrate(obj, mr, flags);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
832
bool i915_gem_object_placement_possible(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
837
if (!obj->mm.n_placements) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
84
obj = pool_get(&slab_objects, PR_WAITOK | PR_ZERO);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
840
return i915_gem_object_has_iomem(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
842
return i915_gem_object_has_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
850
for (i = 0; i < obj->mm.n_placements; i++) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
851
if (obj->mm.placements[i]->type == type)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
86
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
867
bool i915_gem_object_needs_ccs_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
872
if (!HAS_FLAT_CCS(to_i915(obj->base.dev)))
sys/dev/pci/drm/i915/gem/i915_gem_object.c
875
if (obj->flags & I915_BO_ALLOC_CCS_AUX)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
878
for (i = 0; i < obj->mm.n_placements; i++) {
sys/dev/pci/drm/i915/gem/i915_gem_object.c
88
obj->base.funcs = &i915_gem_object_funcs;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
880
if (obj->mm.placements[i]->type == INTEL_MEMORY_SYSTEM)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
883
obj->mm.placements[i]->type == INTEL_MEMORY_LOCAL)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
893
struct drm_i915_gem_object *obj = to_intel_bo(gem_obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
896
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
90
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.c
908
struct drm_i915_gem_object *obj = to_intel_bo(gem_obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
910
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
911
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
93
void i915_gem_object_free(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
96
return kmem_cache_free(slab_objects, obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
961
int i915_gem_object_get_moving_fence(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
964
return dma_resv_get_singleton(obj->base.resv, DMA_RESV_USAGE_KERNEL,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
98
pool_put(&slab_objects, obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
980
int i915_gem_object_wait_moving_fence(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
985
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
987
ret = dma_resv_wait_timeout(obj->base. resv, DMA_RESV_USAGE_KERNEL,
sys/dev/pci/drm/i915/gem/i915_gem_object.c
991
else if (ret > 0 && i915_gem_object_has_unknown_state(obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.h
107
i915_gem_object_get_rcu(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
109
if (obj && !kref_get_unless_zero(&obj->base.refcount))
sys/dev/pci/drm/i915/gem/i915_gem_object.h
110
obj = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
112
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
118
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
121
obj = i915_gem_object_lookup_rcu(file, handle);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
122
obj = i915_gem_object_get_rcu(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
125
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
134
i915_gem_object_get(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
136
drm_gem_object_get(&obj->base);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
137
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
142
i915_gem_object_put(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
144
__drm_gem_object_put(&obj->base);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
147
#define assert_object_held(obj) dma_resv_assert_held((obj)->base.resv)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
152
static inline void assert_object_held_shared(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
159
kref_read(&obj->base.refcount) > 0)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
160
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
163
static inline int __i915_gem_object_lock(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
170
ret = dma_resv_lock_interruptible(obj->base.resv, ww ? &ww->ctx : NULL);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
172
ret = dma_resv_lock(obj->base.resv, ww ? &ww->ctx : NULL);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
175
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
176
list_add_tail(&obj->obj_link, &ww->obj_list);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
182
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
183
ww->contended = obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
189
static inline int i915_gem_object_lock(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
192
return __i915_gem_object_lock(obj, ww, ww && ww->intr);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
195
static inline int i915_gem_object_lock_interruptible(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
199
return __i915_gem_object_lock(obj, ww, true);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
202
static inline bool i915_gem_object_trylock(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
206
return dma_resv_trylock(obj->base.resv);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
208
return ww_mutex_trylock(&obj->base.resv->lock, &ww->ctx);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
211
static inline void i915_gem_object_unlock(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
213
if (obj->ops->adjust_lru)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
214
obj->ops->adjust_lru(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
216
dma_resv_unlock(obj->base.resv);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
220
i915_gem_object_set_readonly(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
222
obj->flags |= I915_BO_READONLY;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
226
i915_gem_object_is_readonly(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
228
return obj->flags & I915_BO_READONLY;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
232
i915_gem_object_is_contiguous(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
234
return obj->flags & I915_BO_ALLOC_CONTIGUOUS;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
238
i915_gem_object_is_volatile(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
240
return obj->flags & I915_BO_ALLOC_VOLATILE;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
244
i915_gem_object_set_volatile(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
246
obj->flags |= I915_BO_ALLOC_VOLATILE;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
250
i915_gem_object_has_tiling_quirk(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
252
return test_bit(I915_TILING_QUIRK_BIT, &obj->flags);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
256
i915_gem_object_set_tiling_quirk(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
258
set_bit(I915_TILING_QUIRK_BIT, &obj->flags);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
262
i915_gem_object_clear_tiling_quirk(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
264
clear_bit(I915_TILING_QUIRK_BIT, &obj->flags);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
268
i915_gem_object_is_protected(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
27
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
270
return obj->flags & I915_BO_PROTECTED;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
274
i915_gem_object_type_has(const struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
277
return obj->ops->flags & flags;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
280
bool i915_gem_object_has_struct_page(const struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
282
bool i915_gem_object_has_iomem(const struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
285
i915_gem_object_is_shrinkable(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
287
return i915_gem_object_type_has(obj, I915_GEM_OBJECT_IS_SHRINKABLE);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
29
if (overflows_type(size, obj->base.size))
sys/dev/pci/drm/i915/gem/i915_gem_object.h
291
i915_gem_object_has_self_managed_shrink_list(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
293
return i915_gem_object_type_has(obj, I915_GEM_OBJECT_SELF_MANAGED_SHRINK_LIST);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
297
i915_gem_object_is_proxy(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
299
return i915_gem_object_type_has(obj, I915_GEM_OBJECT_IS_PROXY);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
303
i915_gem_object_never_mmap(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
305
return i915_gem_object_type_has(obj, I915_GEM_OBJECT_NO_MMAP);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
309
i915_gem_object_is_framebuffer(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
311
return READ_ONCE(obj->frontbuffer) || obj->is_dpt;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
315
i915_gem_object_get_tiling(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
317
return obj->tiling_and_stride & TILING_MASK;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
321
i915_gem_object_is_tiled(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
323
return i915_gem_object_get_tiling(obj) != I915_TILING_NONE;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
327
i915_gem_object_get_stride(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
329
return obj->tiling_and_stride & STRIDE_MASK;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
340
i915_gem_object_get_tile_height(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
342
return i915_gem_tile_height(i915_gem_object_get_tiling(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.h
346
i915_gem_object_get_tile_row_size(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
348
return (i915_gem_object_get_stride(obj) *
sys/dev/pci/drm/i915/gem/i915_gem_object.h
349
i915_gem_object_get_tile_height(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.h
352
int i915_gem_object_set_tiling(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
37
bool i915_gem_object_has_cache_level(const struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
375
__i915_gem_object_page_iter_get_sg(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
400
#define i915_gem_object_page_iter_get_sg(obj, it, n, offset) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
402
__i915_gem_object_page_iter_get_sg(obj, it, n, offset); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
424
__i915_gem_object_get_sg(struct drm_i915_gem_object *obj, pgoff_t n,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
427
return __i915_gem_object_page_iter_get_sg(obj, &obj->mm.get_page, n, offset);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
445
#define i915_gem_object_get_sg(obj, n, offset) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
447
__i915_gem_object_get_sg(obj, n, offset); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
45
void i915_gem_object_free(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
469
__i915_gem_object_get_sg_dma(struct drm_i915_gem_object *obj, pgoff_t n,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
47
void i915_gem_object_init(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
472
return __i915_gem_object_page_iter_get_sg(obj, &obj->mm.get_dma_page, n, offset);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
490
#define i915_gem_object_get_sg_dma(obj, n, offset) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
492
__i915_gem_object_get_sg_dma(obj, n, offset); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
511
__i915_gem_object_get_page(struct drm_i915_gem_object *obj, pgoff_t n);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
52
void __i915_gem_object_fini(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
526
#define i915_gem_object_get_page(obj, n) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
528
__i915_gem_object_get_page(obj, n); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
546
__i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj, pgoff_t n);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
561
#define i915_gem_object_get_dirty_page(obj, n) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
563
__i915_gem_object_get_dirty_page(obj, n); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
580
__i915_gem_object_get_dma_address_len(struct drm_i915_gem_object *obj, pgoff_t n,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
599
#define i915_gem_object_get_dma_address_len(obj, n, len) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
601
__i915_gem_object_get_dma_address_len(obj, n, len); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
617
__i915_gem_object_get_dma_address(struct drm_i915_gem_object *obj, pgoff_t n);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
634
#define i915_gem_object_get_dma_address(obj, n) ({ \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
636
__i915_gem_object_get_dma_address(obj, n); \
sys/dev/pci/drm/i915/gem/i915_gem_object.h
639
void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
642
int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
643
int __i915_gem_object_get_pages(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
646
i915_gem_object_pin_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
648
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
650
if (atomic_inc_not_zero(&obj->mm.pages_pin_count))
sys/dev/pci/drm/i915/gem/i915_gem_object.h
653
return __i915_gem_object_get_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
656
int i915_gem_object_pin_pages_unlocked(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
659
i915_gem_object_has_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
661
return !IS_ERR_OR_NULL(READ_ONCE(obj->mm.pages));
sys/dev/pci/drm/i915/gem/i915_gem_object.h
665
__i915_gem_object_pin_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
667
GEM_BUG_ON(!i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.h
669
atomic_inc(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
67
void __i915_gem_object_release_shmem(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
673
i915_gem_object_has_pinned_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
675
return atomic_read(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
679
__i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
681
GEM_BUG_ON(!i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.h
682
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_object.h
684
atomic_dec(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
688
i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
690
__i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
693
int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
694
int i915_gem_object_truncate(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
71
int i915_gem_object_pwrite_phys(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
717
void *__must_check i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
720
void *__must_check i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
723
void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
726
static inline void i915_gem_object_flush_map(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
728
__i915_gem_object_flush_map(obj, 0, obj->base.size);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
73
int i915_gem_object_pread_phys(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
740
static inline void i915_gem_object_unpin_map(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
742
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
745
void __i915_gem_object_release_map(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
747
int i915_gem_object_prepare_read(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
749
int i915_gem_object_prepare_write(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
756
i915_gem_object_finish_access(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
758
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
76
int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
761
int i915_gem_object_get_moving_fence(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
763
int i915_gem_object_wait_moving_fence(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
765
bool i915_gem_object_has_unknown_state(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
767
void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
769
void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
77
void i915_gem_object_put_pages_shmem(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
771
bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
772
void i915_gem_object_flush_if_display(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
773
void i915_gem_object_flush_if_display_locked(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
774
bool i915_gem_cpu_write_needs_clflush(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
777
i915_gem_object_set_to_wc_domain(struct drm_i915_gem_object *obj, bool write);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
779
i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
781
i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
783
i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
789
void i915_gem_object_make_unshrinkable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
79
void i915_gem_object_put_pages_phys(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
790
void i915_gem_object_make_shrinkable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
791
void __i915_gem_object_make_shrinkable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
792
void __i915_gem_object_make_purgeable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
793
void i915_gem_object_make_purgeable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
795
static inline void __start_cpu_write(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
797
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
798
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
799
if (i915_gem_cpu_write_needs_clflush(obj))
sys/dev/pci/drm/i915/gem/i915_gem_object.h
800
obj->cache_dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
806
int i915_gem_object_wait(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
809
int i915_gem_object_wait_priority(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
813
int i915_gem_object_read_from_page(struct drm_i915_gem_object *obj, u64 offset, void *dst, int size);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
815
bool i915_gem_object_is_shmem(const struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
819
void __i915_gem_object_pages_fini(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
821
void __i915_gem_free_object(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
823
bool i915_gem_object_evictable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
825
bool i915_gem_object_migratable(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
827
int i915_gem_object_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
830
int __i915_gem_object_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
835
bool i915_gem_object_can_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
838
int i915_gem_object_wait_migration(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
841
bool i915_gem_object_placement_possible(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
844
bool i915_gem_object_needs_ccs_pages(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
85
__i915_gem_object_unset_pages(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
858
struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
861
struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
867
i915_gem_object_is_userptr(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object.h
869
return obj->userptr.notifier.mm;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
872
int i915_gem_object_userptr_submit_init(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
873
int i915_gem_object_userptr_submit_done(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
874
int i915_gem_object_userptr_validate(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
876
static inline bool i915_gem_object_is_userptr(struct drm_i915_gem_object *obj) { return false; }
sys/dev/pci/drm/i915/gem/i915_gem_object.h
878
static inline int i915_gem_object_userptr_submit_init(struct drm_i915_gem_object *obj) { GEM_BUG_ON(1); return -ENODEV; }
sys/dev/pci/drm/i915/gem/i915_gem_object.h
879
static inline int i915_gem_object_userptr_submit_done(struct drm_i915_gem_object *obj) { GEM_BUG_ON(1); return -ENODEV; }
sys/dev/pci/drm/i915/gem/i915_gem_object.h
880
static inline int i915_gem_object_userptr_validate(struct drm_i915_gem_object *obj) { GEM_BUG_ON(1); return -ENODEV; }
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
15
void __i915_gem_object_flush_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
17
void __i915_gem_object_invalidate_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
21
i915_gem_object_flush_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
24
if (unlikely(rcu_access_pointer(obj->frontbuffer)))
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
25
__i915_gem_object_flush_frontbuffer(obj, origin);
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
29
i915_gem_object_invalidate_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
32
if (unlikely(rcu_access_pointer(obj->frontbuffer)))
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
33
__i915_gem_object_invalidate_frontbuffer(obj, origin);
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
46
i915_gem_object_get_frontbuffer(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
50
if (likely(!rcu_access_pointer(obj->frontbuffer)))
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
55
front = rcu_dereference(obj->frontbuffer);
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
62
if (likely(front == rcu_access_pointer(obj->frontbuffer)))
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
85
i915_gem_object_set_frontbuffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
91
RCU_INIT_POINTER(obj->frontbuffer, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
92
} else if (rcu_access_pointer(obj->frontbuffer)) {
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
93
cur = rcu_dereference_protected(obj->frontbuffer, true);
sys/dev/pci/drm/i915/gem/i915_gem_object_frontbuffer.h
96
rcu_assign_pointer(obj->frontbuffer, front);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
104
void (*delayed_free)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
110
int (*migrate)(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
114
void (*release)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
229
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
58
int (*get_pages)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
59
void (*put_pages)(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
61
int (*truncate)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
81
int (*shrink)(struct drm_i915_gem_object *obj, unsigned int flags);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
83
int (*pread)(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
85
int (*pwrite)(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
87
u64 (*mmap_offset)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
88
void (*unmap_virtual)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
90
int (*dmabuf_export)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
99
void (*adjust_lru)(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
100
int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
102
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
105
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
107
if (unlikely(obj->mm.madv != I915_MADV_WILLNEED)) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
113
err = obj->ops->get_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
114
GEM_BUG_ON(!err && !i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
126
int __i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
130
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
132
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
134
if (unlikely(!i915_gem_object_has_pages(obj))) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
135
GEM_BUG_ON(i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
137
err = ____i915_gem_object_get_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
143
atomic_inc(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
148
int i915_gem_object_pin_pages_unlocked(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
155
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
157
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
169
int i915_gem_object_truncate(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
171
if (obj->ops->truncate)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
172
return obj->ops->truncate(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
177
static void __i915_gem_object_reset_page_iter(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
183
radix_tree_for_each_slot(slot, &obj->mm.get_page.radix, &iter, 0)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
184
radix_tree_delete(&obj->mm.get_page.radix, iter.index);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
185
radix_tree_for_each_slot(slot, &obj->mm.get_dma_page.radix, &iter, 0)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
186
radix_tree_delete(&obj->mm.get_dma_page.radix, iter.index);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
190
static void unmap_object(struct drm_i915_gem_object *obj, void *ptr)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
193
vunmap(ptr, obj->base.size);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
196
static void flush_tlb_invalidate(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
198
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
203
if (!obj->mm.tlb[id])
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
206
intel_gt_invalidate_tlb_full(gt, obj->mm.tlb[id]);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
207
obj->mm.tlb[id] = 0;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
21
void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
212
__i915_gem_object_unset_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
216
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
218
pages = fetch_and_zero(&obj->mm.pages);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
222
if (i915_gem_object_is_volatile(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
223
obj->mm.madv = I915_MADV_WILLNEED;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
225
if (!i915_gem_object_has_self_managed_shrink_list(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
226
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
228
if (obj->mm.mapping) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
229
unmap_object(obj, page_mask_bits(obj->mm.mapping));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
230
obj->mm.mapping = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
233
__i915_gem_object_reset_page_iter(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
234
obj->mm.page_sizes.phys = obj->mm.page_sizes.sg = 0;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
236
flush_tlb_invalidate(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
24
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
241
int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
245
if (i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
249
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
251
i915_gem_object_release_mmap_offset(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
258
pages = __i915_gem_object_unset_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
267
obj->ops->put_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
273
static void *i915_gem_object_map_page(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
276
unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
29
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
305
if (n_pages == 1 && !PageHighMem(sg_page(obj->mm.pages->sgl)))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
306
return page_address(sg_page(obj->mm.pages->sgl));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
31
if (i915_gem_object_is_volatile(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
32
obj->mm.madv = I915_MADV_DONTNEED;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
323
for_each_sgt_page(page, iter, obj->mm.pages)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
332
static void *i915_gem_object_map_pfn(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
335
resource_size_t iomap = obj->mm.region->iomap.base -
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
336
obj->mm.region->region.start;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
337
unsigned long n_pfn = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
35
if (obj->cache_dirty) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
353
for_each_sgt_daddr(addr, iter, obj->mm.pages)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
37
obj->write_domain = 0;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
377
static struct vm_page **i915_gem_object_panic_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
379
unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
38
if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
390
for_each_sgt_page(page, iter, obj->mm.pages)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
40
obj->cache_dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
43
obj->mm.get_page.sg_pos = pages->sgl;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
44
obj->mm.get_page.sg_idx = 0;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
45
obj->mm.get_dma_page.sg_pos = pages->sgl;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
454
struct drm_i915_gem_object *obj = to_intel_bo(_obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
457
ptr = page_unpack_bits(obj->mm.mapping, &has_type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
459
if (i915_gem_object_has_iomem(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
46
obj->mm.get_dma_page.sg_idx = 0;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
468
if (i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
469
panic->pages = i915_gem_object_panic_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
48
obj->mm.pages = pages;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
488
void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
496
if (!i915_gem_object_has_struct_page(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
497
!i915_gem_object_has_iomem(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
50
obj->mm.page_sizes.phys = i915_sg_dma_sizes(pages->sgl);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
500
if (WARN_ON_ONCE(obj->flags & I915_BO_ALLOC_GPU_ONLY))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
503
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
508
if (!atomic_inc_not_zero(&obj->mm.pages_pin_count)) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
509
if (unlikely(!i915_gem_object_has_pages(obj))) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
51
GEM_BUG_ON(!obj->mm.page_sizes.phys);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
510
GEM_BUG_ON(i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
512
err = ____i915_gem_object_get_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
518
atomic_inc(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
521
GEM_BUG_ON(!i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
540
if (i915_gem_object_placement_possible(obj, INTEL_MEMORY_LOCAL)) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
541
if (type != I915_MAP_WC && !obj->mm.n_placements) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
547
} else if (IS_DGFX(to_i915(obj->base.dev))) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
551
ptr = page_unpack_bits(obj->mm.mapping, &has_type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
558
unmap_object(obj, ptr);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
560
ptr = obj->mm.mapping = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
564
err = i915_gem_object_wait_moving_fence(obj, true);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
572
else if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
573
ptr = i915_gem_object_map_page(obj, type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
575
ptr = i915_gem_object_map_pfn(obj, type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
579
obj->mm.mapping = page_pack_bits(ptr, type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
585
atomic_dec(&obj->mm.pages_pin_count);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
589
void *i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
594
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
595
ret = i915_gem_object_pin_map(obj, type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
596
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
601
void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
608
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
609
GEM_BUG_ON(range_overflows_t(typeof(obj->base.size),
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
61
obj->mm.page_sizes.sg = 0;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
610
offset, size, obj->base.size));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
613
obj->mm.dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
615
if (obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
618
ptr = page_unpack_bits(obj->mm.mapping, &has_type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
623
if (size == obj->base.size) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
624
obj->write_domain &= ~I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
625
obj->cache_dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
629
void __i915_gem_object_release_map(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
63
if (obj->mm.page_sizes.phys & ~0u << i)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
631
GEM_BUG_ON(!obj->mm.mapping);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
639
unmap_object(obj, page_mask_bits(fetch_and_zero(&obj->mm.mapping)));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
64
obj->mm.page_sizes.sg |= BIT(i);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
641
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
645
__i915_gem_object_page_iter_get_sg(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
651
const bool dma = iter == &obj->mm.get_dma_page ||
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
652
iter == &obj->ttm.get_io_page;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
657
GEM_BUG_ON(n >= obj->base.size >> PAGE_SHIFT);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
658
if (!i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
659
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
66
GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
68
shrinkable = i915_gem_object_is_shrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
70
if (i915_gem_object_is_tiled(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
72
GEM_BUG_ON(i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
73
i915_gem_object_set_tiling_quirk(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
74
GEM_BUG_ON(!list_empty(&obj->mm.link));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
75
atomic_inc(&obj->mm.shrink_pin);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
762
__i915_gem_object_get_page(struct drm_i915_gem_object *obj, pgoff_t n)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
767
GEM_BUG_ON(!i915_gem_object_has_struct_page(obj));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
769
sg = i915_gem_object_get_sg(obj, n, &offset);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
775
__i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj, pgoff_t n)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
779
page = i915_gem_object_get_page(obj, n);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
780
if (!obj->mm.dirty)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
787
__i915_gem_object_get_dma_address_len(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
79
if (shrinkable && !i915_gem_object_has_self_managed_shrink_list(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
793
sg = i915_gem_object_get_sg_dma(obj, n, &offset);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
802
__i915_gem_object_get_dma_address(struct drm_i915_gem_object *obj, pgoff_t n)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
804
return i915_gem_object_get_dma_address_len(obj, n, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
83
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
87
i915->mm.shrink_memory += obj->base.size;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
89
if (obj->mm.madv != I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
93
list_add_tail(&obj->mm.link, list);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
95
atomic_set(&obj->mm.shrink_pin, 0);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
108
uvm_obj_unwire(obj->base.uao, i * PAGE_SIZE,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
117
obj->mem_flags &= ~I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
118
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
126
dma_free_coherent(obj->base.dev->dev,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
127
roundup_pow_of_two(obj->base.size),
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
136
i915_gem_object_put_pages_phys(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
145
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
148
__i915_gem_object_release_shmem(obj, pages, false);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
150
if (obj->mm.dirty) {
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
152
struct address_space *mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
157
for (i = 0; i < obj->base.size / PAGE_SIZE; i++) {
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
167
if (uvm_obj_wire(obj->base.uao, i * PAGE_SIZE,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
178
if (obj->mm.madv == I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
182
uvm_obj_unwire(obj->base.uao, i * PAGE_SIZE,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
188
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
195
dma_free_coherent(obj->base.dev->dev,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
196
roundup_pow_of_two(obj->base.size),
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
203
int i915_gem_object_pwrite_phys(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
207
void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
209
struct drm_dmamem *dmah = (void *)sg_page(obj->mm.pages->sgl);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
21
static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
213
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
216
err = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
227
i915_gem_object_invalidate_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
235
i915_gem_object_flush_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
239
int i915_gem_object_pread_phys(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
24
struct address_space *mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
243
void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
245
struct drm_dmamem *dmah = (void *)sg_page(obj->mm.pages->sgl);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
251
err = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
264
static int i915_gem_object_shmem_to_phys(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
269
pages = __i915_gem_object_unset_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
271
err = i915_gem_object_get_pages_phys(obj);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
276
__i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
279
i915_gem_object_put_pages_shmem(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
281
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
286
__i915_gem_object_set_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
29
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
290
int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align)
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
294
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
296
if (align > obj->base.size)
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
299
if (!i915_gem_object_is_shmem(obj))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
302
if (!i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
305
err = i915_gem_object_unbind(obj, I915_GEM_OBJECT_UNBIND_ACTIVE);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
309
if (obj->mm.madv != I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
312
if (i915_gem_object_has_tiling_quirk(obj))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
315
if (obj->mm.mapping || i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
318
if (unlikely(obj->mm.madv != I915_MADV_WILLNEED)) {
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
319
drm_dbg(obj->base.dev,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
324
return i915_gem_object_shmem_to_phys(obj);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
38
if (overflows_type(obj->base.size, sg->length))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
41
if (GEM_WARN_ON(i915_gem_object_needs_bit17_swizzle(obj)))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
50
vaddr = dma_alloc_coherent(obj->base.dev->dev,
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
51
roundup_pow_of_two(obj->base.size),
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
57
roundup_pow_of_two(obj->base.size),
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
59
roundup_pow_of_two(obj->base.size), flags, 0);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
75
sg->length = obj->base.size;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
83
sg_dma_len(sg) = obj->base.size;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
86
for (i = 0; i < obj->base.size / PAGE_SIZE; i++) {
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
96
if (uvm_obj_wire(obj->base.uao, i * PAGE_SIZE,
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
138
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
177
list_for_each_entry(obj, *phase, mm.link) {
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
178
if (!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ))
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
179
flush |= (obj->read_domains & I915_GEM_DOMAIN_CPU) == 0;
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
180
__start_cpu_write(obj); /* presume auto-hibernate */
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
200
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
223
list_for_each_entry(obj, &i915->mm.shrink_list, mm.link)
sys/dev/pci/drm/i915/gem/i915_gem_pm.c
224
__start_cpu_write(obj);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
102
i915_gem_object_free(obj);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
13
void i915_gem_object_init_memory_region(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_region.c
158
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_region.c
16
obj->mm.region = mem;
sys/dev/pci/drm/i915/gem/i915_gem_region.c
173
obj = list_first_entry_or_null(&mr->objects.list, typeof(*obj),
sys/dev/pci/drm/i915/gem/i915_gem_region.c
175
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_region.c
178
list_move_tail(&obj->mm.region_link, &still_in_list);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
179
if (!kref_get_unless_zero(&obj->base.refcount))
sys/dev/pci/drm/i915/gem/i915_gem_region.c
19
list_add(&obj->mm.region_link, &mem->objects.list);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
190
ret = i915_gem_object_lock(obj, apply->ww);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
194
if (obj->mm.region == mr)
sys/dev/pci/drm/i915/gem/i915_gem_region.c
195
ret = ops->process_obj(apply, obj);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
199
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
23
void i915_gem_object_release_memory_region(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_region.c
25
struct intel_memory_region *mem = obj->mm.region;
sys/dev/pci/drm/i915/gem/i915_gem_region.c
28
list_del(&obj->mm.region_link);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
39
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_region.c
79
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/i915_gem_region.c
80
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_region.c
94
err = mem->ops->init_object(mem, obj, offset, size, page_size, flags);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
98
trace_i915_gem_object_create(obj);
sys/dev/pci/drm/i915/gem/i915_gem_region.c
99
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_region.h
32
struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_region.h
50
void i915_gem_object_init_memory_region(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_region.h
52
void i915_gem_object_release_memory_region(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
204
if (uvm_obj_wire(obj->base.uao, 0, obj->base.size, &plist)) {
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
228
shmem_sg_free_table(st, mapping, false, false, obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
249
static int shmem_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
251
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
252
struct intel_memory_region *mem = obj->mm.region;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
254
struct address_space *mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
265
GEM_BUG_ON(obj->read_domains & I915_GEM_GPU_DOMAINS);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
266
GEM_BUG_ON(obj->write_domain & I915_GEM_GPU_DOMAINS);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
274
ret = shmem_sg_alloc_table(i915, st, obj->base.size, mem, mapping,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
277
ret = shmem_sg_alloc_table(i915, st, obj->base.size, mem, NULL,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
278
max_segment, obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
283
ret = i915_gem_gtt_prepare_pages(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
294
shmem_sg_free_table(st, NULL, false, false, obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
304
obj->base.size >> PAGE_SHIFT);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
309
if (i915_gem_object_needs_bit17_swizzle(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
310
i915_gem_object_do_bit_17_swizzle(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
312
if (i915_gem_object_can_bypass_llc(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
313
obj->cache_dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
315
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
323
shmem_sg_free_table(st, NULL, false, false, obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
344
shmem_truncate(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
353
shmem_truncate_range(file_inode(obj->base.filp), 0, (loff_t)-1);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
355
rw_enter(obj->base.uao->vmobjlock, RW_WRITE);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
356
obj->base.uao->pgops->pgo_flush(obj->base.uao, 0, obj->base.size,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
358
rw_exit(obj->base.uao->vmobjlock);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
360
obj->mm.madv = __I915_MADV_PURGED;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
361
obj->mm.pages = ERR_PTR(-EFAULT);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
37
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
395
shmem_writeback(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
399
__shmem_writeback(obj->base.size, obj->base.filp->f_mapping);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
403
static int shmem_shrink(struct drm_i915_gem_object *obj, unsigned int flags)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
405
switch (obj->mm.madv) {
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
407
return i915_gem_object_truncate(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
413
shmem_writeback(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
419
__i915_gem_object_release_shmem(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
423
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
425
GEM_BUG_ON(obj->mm.madv == __I915_MADV_PURGED);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
427
if (obj->mm.madv == I915_MADV_DONTNEED)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
428
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
431
(obj->read_domains & I915_GEM_DOMAIN_CPU) == 0 &&
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
432
!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
435
__start_cpu_write(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
445
obj->cache_dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
448
void i915_gem_object_put_pages_shmem(struct drm_i915_gem_object *obj, struct sg_table *pages)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
450
__i915_gem_object_release_shmem(obj, pages, true);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
452
i915_gem_gtt_finish_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
454
if (i915_gem_object_needs_bit17_swizzle(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
455
i915_gem_object_save_bit_17_swizzle(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
458
shmem_sg_free_table(pages, file_inode(obj->base.filp)->i_mapping,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
459
obj->mm.dirty, obj->mm.madv == I915_MADV_WILLNEED);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
462
obj->mm.dirty, obj->mm.madv == I915_MADV_WILLNEED, obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
465
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
469
shmem_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
471
if (likely(i915_gem_object_has_struct_page(obj)))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
472
i915_gem_object_put_pages_shmem(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
474
i915_gem_object_put_pages_phys(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
478
shmem_pwrite(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
483
struct file *file = obj->base.filp;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
497
if (!i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
498
return i915_gem_object_pwrite_phys(obj, arg);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
509
if (i915_gem_object_has_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
512
if (obj->mm.madv != I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
561
if (uvm_obj_wire(obj->base.uao, trunc_page(pos),
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
571
uvm_obj_unwire(obj->base.uao, trunc_page(pos),
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
589
shmem_pread(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
592
if (!i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
593
return i915_gem_object_pread_phys(obj, arg);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
598
static void shmem_release(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
600
if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
601
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
604
fput(obj->base.filp);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
625
struct drm_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
631
drm_gem_private_object_init(&i915->drm, obj, size);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
662
obj->filp = filp;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
668
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
682
ret = __create_shmem(i915, &obj->base, size);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
684
ret = drm_gem_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
697
mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
702
i915_gem_object_init(obj, &i915_gem_shmem_ops, &lock_class, flags);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
703
obj->mem_flags |= I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
704
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
705
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
72
uvm_obj_unwire(obj->base.uao, 0, obj->base.size);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
731
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
733
i915_gem_object_init_memory_region(obj, mem);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
752
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
758
obj = i915_gem_object_create_shmem(i915, round_up(size, PAGE_SIZE));
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
759
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
760
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
762
GEM_BUG_ON(obj->write_domain != I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
764
file = obj->base.filp;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
775
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
778
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
786
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
792
obj = i915_gem_object_create_shmem(dev_priv, round_up(size, PAGE_SIZE));
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
793
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
794
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
796
GEM_BUG_ON(obj->write_domain != I915_GEM_DOMAIN_CPU);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
798
uao = obj->base.uao;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
82
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
826
return obj;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
829
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
864
bool i915_gem_object_is_shmem(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
866
return obj->ops == &i915_gem_shmem_ops;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
178
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
195
(obj = list_first_entry_or_null(phase->list,
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
196
typeof(*obj),
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
198
list_move_tail(&obj->mm.link, &still_in_list);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
201
!is_vmalloc_addr(obj->mm.mapping))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
205
i915_gem_object_is_framebuffer(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
208
if (!can_release_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
211
if (!kref_get_unless_zero(&obj->base.refcount))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
218
if (!i915_gem_object_trylock(obj, NULL))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
221
err = i915_gem_object_lock(obj, ww);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
226
if (drop_pages(obj, shrink, trylock_vm) &&
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
227
!__i915_gem_object_put_pages(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
228
!try_to_writeback(obj, shrink))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
229
count += obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
232
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
234
scanned += obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
236
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
25
static bool can_release_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
28
if (!i915_gem_object_is_shrinkable(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
350
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
36
return swap_available() || obj->mm.madv == I915_MADV_DONTNEED;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
368
list_for_each_entry(obj, &i915->mm.shrink_list, mm.link) {
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
369
if (!can_release_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
370
unevictable += obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
372
available += obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
39
static bool drop_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
408
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
413
if (!i915_gem_object_trylock(obj, NULL))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
419
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
489
void i915_gem_object_make_unshrinkable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
491
struct drm_i915_private *i915 = obj_to_i915(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
500
if (atomic_add_unless(&obj->mm.shrink_pin, 1, 0))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
504
if (!atomic_fetch_inc(&obj->mm.shrink_pin) &&
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
505
!list_empty(&obj->mm.link)) {
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
506
list_del_init(&obj->mm.link);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
508
i915->mm.shrink_memory -= obj->base.size;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
513
static void ___i915_gem_object_make_shrinkable(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
516
struct drm_i915_private *i915 = obj_to_i915(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
519
if (!i915_gem_object_is_shrinkable(obj))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
52
if (i915_gem_object_unbind(obj, flags) == 0)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
522
if (atomic_add_unless(&obj->mm.shrink_pin, -1, 1))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
526
GEM_BUG_ON(!kref_read(&obj->base.refcount));
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
527
if (atomic_dec_and_test(&obj->mm.shrink_pin)) {
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
528
GEM_BUG_ON(!list_empty(&obj->mm.link));
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
530
list_add_tail(&obj->mm.link, head);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
532
i915->mm.shrink_memory += obj->base.size;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
548
void __i915_gem_object_make_shrinkable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
550
___i915_gem_object_make_shrinkable(obj,
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
551
&obj_to_i915(obj)->mm.shrink_list);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
564
void __i915_gem_object_make_purgeable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
566
___i915_gem_object_make_shrinkable(obj,
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
567
&obj_to_i915(obj)->mm.purge_list);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
58
static int try_to_writeback(struct drm_i915_gem_object *obj, unsigned int flags)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
580
void i915_gem_object_make_shrinkable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
582
GEM_BUG_ON(!i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
583
__i915_gem_object_make_shrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
597
void i915_gem_object_make_purgeable(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
599
GEM_BUG_ON(!i915_gem_object_has_pages(obj));
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
60
if (obj->ops->shrink) {
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
600
__i915_gem_object_make_purgeable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
69
return obj->ops->shrink(obj, shrink_flags);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
1062
bool i915_gem_object_is_stolen(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
1064
return obj->ops == &i915_gem_object_stolen_ops;
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
648
static int i915_gem_object_get_pages_stolen(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
650
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
652
i915_pages_create_for_stolen(obj->base.dev,
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
653
obj->stolen->start,
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
654
obj->stolen->size);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
663
__i915_gem_object_set_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
668
static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
671
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
684
i915_gem_object_release_stolen(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
686
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
687
struct drm_mm_node *stolen = fetch_and_zero(&obj->stolen);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
693
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
704
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
718
drm_gem_private_object_init(&mem->i915->drm, &obj->base, stolen->size);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
719
i915_gem_object_init(obj, &i915_gem_object_stolen_ops, &lock_class, flags);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
721
obj->stolen = stolen;
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
722
obj->read_domains = I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
724
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
726
if (WARN_ON(!i915_gem_object_trylock(obj, NULL)))
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
729
i915_gem_object_init_memory_region(obj, mem);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
731
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
733
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
734
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
740
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_stolen.c
785
ret = __i915_gem_object_create_stolen(mem, obj, stolen);
sys/dev/pci/drm/i915/gem/i915_gem_stolen.h
37
bool i915_gem_object_is_stolen(const struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
118
i915_tiling_ok(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
121
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
182
i915_gem_object_fence_prepare(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
185
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
196
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
197
for_each_ggtt_vma(vma, obj) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
205
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
221
bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
223
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
226
i915_gem_object_is_tiled(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
230
i915_gem_object_set_tiling(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
233
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
240
GEM_BUG_ON(!i915_tiling_ok(obj, tiling, stride));
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
243
if ((tiling | stride) == obj->tiling_and_stride)
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
246
if (i915_gem_object_is_framebuffer(obj))
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
262
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
263
if (i915_gem_object_is_framebuffer(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
264
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
268
err = i915_gem_object_fence_prepare(obj, tiling, stride);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
270
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
278
if (i915_gem_object_has_pages(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
279
obj->mm.madv == I915_MADV_WILLNEED &&
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
282
GEM_BUG_ON(!i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
283
i915_gem_object_clear_tiling_quirk(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
284
i915_gem_object_make_shrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
286
if (!i915_gem_object_is_tiled(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
287
GEM_BUG_ON(i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
288
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
289
i915_gem_object_set_tiling_quirk(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
293
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
294
for_each_ggtt_vma(vma, obj) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
304
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
306
obj->tiling_and_stride = tiling | stride;
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
309
if (i915_gem_object_needs_bit17_swizzle(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
310
if (!obj->bit_17) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
311
obj->bit_17 = bitmap_zalloc(obj->base.size >> PAGE_SHIFT,
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
315
bitmap_free(obj->bit_17);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
316
obj->bit_17 = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
319
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
322
i915_gem_object_release_mmap_gtt(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
347
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
353
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
354
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
361
if (i915_gem_object_is_proxy(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
366
if (!i915_tiling_ok(obj, args->tiling_mode, args->stride)) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
400
err = i915_gem_object_set_tiling(obj, args->tiling_mode, args->stride);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
403
args->stride = i915_gem_object_get_stride(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
404
args->tiling_mode = i915_gem_object_get_tiling(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
407
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
430
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
437
obj = i915_gem_object_lookup_rcu(file, args->handle);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
438
if (obj) {
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
440
READ_ONCE(obj->tiling_and_stride) & TILING_MASK;
sys/dev/pci/drm/i915/gem/i915_gem_tiling.h
14
bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1002
obj->mm.ttm_shrinkable = shrinkable;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1004
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1014
} else if (obj->mm.madv != I915_MADV_WILLNEED) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1016
} else if (!i915_gem_object_has_pages(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1032
!(obj->flags & I915_BO_ALLOC_GPU_ONLY))
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1057
static void i915_ttm_delayed_free(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1059
GEM_BUG_ON(!obj->ttm.created);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1061
ttm_bo_put(i915_gem_to_ttm(obj));
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1071
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1077
if (unlikely(i915_gem_object_is_readonly(obj) &&
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1085
if (obj->mm.madv != I915_MADV_WILLNEED) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1114
for (i = 0; i < obj->mm.n_placements; i++) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1115
struct intel_memory_region *mr = obj->mm.placements[i];
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
112
i915_ttm_select_tt_caching(const struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1121
flags = obj->flags;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1123
err = __i915_ttm_migrate(obj, mr, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1139
wakeref = intel_runtime_pm_get(&to_i915(obj->base.dev)->runtime_pm);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1156
if (ret == VM_FAULT_NOPAGE && wakeref && !obj->userfault_count) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1157
obj->userfault_count = 1;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1158
spin_lock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1159
list_add(&obj->userfault_link, &to_i915(obj->base.dev)->runtime_pm.lmem_userfault_list);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1160
spin_unlock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1166
intel_wakeref_auto(&to_i915(obj->base.dev)->runtime_pm.userfault_wakeref,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1169
i915_ttm_adjust_lru(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1175
intel_runtime_pm_put(&to_i915(obj->base.dev)->runtime_pm, wakeref);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1184
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1187
if (i915_gem_object_is_readonly(obj) && write)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
119
if (obj->mm.n_placements <= 1)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1195
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1199
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1204
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1208
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1229
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1236
if (unlikely(i915_gem_object_is_readonly(obj) && write)) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1237
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1254
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1258
if (obj->mm.madv != I915_MADV_WILLNEED) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1260
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1282
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1289
for (i = 0; i < obj->mm.n_placements; i++) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1290
struct intel_memory_region *mr = obj->mm.placements[i];
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1296
flags = obj->flags;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1298
err = __i915_ttm_migrate(obj, mr, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1313
wakeref = intel_runtime_pm_get(&to_i915(obj->base.dev)->runtime_pm);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1337
if (ret == VM_FAULT_NOPAGE && wakeref && !obj->userfault_count) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1338
obj->userfault_count = 1;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1339
spin_lock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1340
list_add(&obj->userfault_link, &to_i915(obj->base.dev)->runtime_pm.lmem_userfault_list);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1341
spin_unlock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1347
intel_wakeref_auto(&to_i915(obj->base.dev)->runtime_pm.userfault_wakeref,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1350
i915_ttm_adjust_lru(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1368
intel_runtime_pm_put(&to_i915(obj->base.dev)->runtime_pm, wakeref);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1370
uvmfault_unlockall(ufi, NULL, &obj->base.uobj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1378
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1381
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1387
struct drm_i915_gem_object *obj =
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1390
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1401
static u64 i915_ttm_mmap_offset(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1404
GEM_BUG_ON(!drm_mm_node_allocated(&obj->base.vma_node.vm_node));
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1406
return drm_vma_node_offset_addr(&obj->base.vma_node);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1409
static void i915_ttm_unmap_virtual(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1411
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1414
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1417
wakeref = intel_runtime_pm_get(&to_i915(obj->base.dev)->runtime_pm);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1420
if (obj->userfault_count) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1421
spin_lock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1422
list_del(&obj->userfault_link);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1423
spin_unlock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1424
obj->userfault_count = 0;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1428
GEM_WARN_ON(obj->userfault_count);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1430
ttm_bo_unmap_virtual(i915_gem_to_ttm(obj));
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1433
intel_runtime_pm_put(&to_i915(obj->base.dev)->runtime_pm, wakeref);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1457
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1459
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1460
mutex_destroy(&obj->ttm.get_io_page.lock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1462
if (obj->ttm.created) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1472
if (obj->mm.ttm_shrinkable)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1473
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1475
i915_ttm_backup_free(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1478
__i915_gem_free_object(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1480
call_rcu(&obj->rcu, __i915_gem_free_object_rcu);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1482
__i915_gem_object_fini(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1496
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1511
drm_gem_private_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1512
i915_gem_object_init(obj, &i915_gem_ttm_obj_ops, &lock_class, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1514
obj->bo_offset = offset;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1517
obj->mm.region = mem;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1518
INIT_LIST_HEAD(&obj->mm.region_link);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1520
INIT_RADIX_TREE(&obj->ttm.get_io_page.radix, GFP_KERNEL | __GFP_NOWARN);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1521
rw_init(&obj->ttm.get_io_page.lock, "i915ttm");
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1522
bo_type = (obj->flags & I915_BO_ALLOC_USER) ? ttm_bo_type_device :
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1525
obj->base.vma_node.driver_private = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1528
GEM_BUG_ON(page_size && obj->mm.n_placements);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1536
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1545
ret = ttm_bo_init_reserved(&i915->bdev, i915_gem_to_ttm(obj), bo_type,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1562
obj->ttm.created = true;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1563
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1564
i915_gem_object_init_memory_region(obj, mem);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1565
i915_ttm_adjust_domains_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1566
i915_ttm_adjust_gem_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
1567
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
157
i915_ttm_placement_from_obj(const struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
161
unsigned int num_allowed = obj->mm.n_placements;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
162
unsigned int flags = obj->flags;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
165
i915_ttm_place_from_region(num_allowed ? obj->mm.placements[0] :
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
166
obj->mm.region, &places[0], obj->bo_offset,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
167
obj->base.size, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
171
i915_ttm_place_from_region(obj->mm.placements[i],
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
172
&places[i + 1], obj->bo_offset,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
173
obj->base.size, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
277
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
290
if (obj->flags & I915_BO_ALLOC_CPU_CLEAR && (!bo->resource ||
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
294
caching = i915_ttm_select_tt_caching(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
295
if (i915_gem_object_is_shrinkable(obj) && caching == ttm_cached) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
301
if (i915_gem_object_needs_ccs_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
313
i915_tt->dev = obj->base.dev->dev;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
367
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
381
if (!i915_gem_object_evictable(obj))
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
400
void i915_ttm_free_cached_io_rsgt(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
405
if (!obj->ttm.cached_io_rsgt)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
409
radix_tree_for_each_slot(slot, &obj->ttm.get_io_page.radix, &iter, 0)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
410
radix_tree_delete(&obj->ttm.get_io_page.radix, iter.index);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
413
i915_refct_sgt_put(obj->ttm.cached_io_rsgt);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
414
obj->ttm.cached_io_rsgt = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
426
int i915_ttm_purge(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
428
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
438
if (obj->mm.madv == __I915_MADV_PURGED)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
456
rw_enter(obj->base.uao->vmobjlock, RW_WRITE);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
457
obj->base.uao->pgops->pgo_flush(obj->base.uao, 0, obj->base.size,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
459
rw_exit(obj->base.uao->vmobjlock);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
464
obj->write_domain = 0;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
465
obj->read_domains = 0;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
466
i915_ttm_adjust_gem_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
467
i915_ttm_free_cached_io_rsgt(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
468
obj->mm.madv = __I915_MADV_PURGED;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
473
static int i915_ttm_shrink(struct drm_i915_gem_object *obj, unsigned int flags)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
475
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
497
switch (obj->mm.madv) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
499
return i915_ttm_purge(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
516
__shmem_writeback(obj->base.size, i915_tt->filp->f_mapping);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
526
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
535
__i915_gem_object_pages_fini(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
536
i915_ttm_free_cached_io_rsgt(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
587
i915_ttm_resource_get_st(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
590
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
598
page_alignment = obj->mm.region->min_page_size;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
606
if (!obj->ttm.cached_io_rsgt) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
609
rsgt = intel_region_ttm_resource_to_rsgt(obj->mm.region,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
615
obj->ttm.cached_io_rsgt = rsgt;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
617
return i915_refct_sgt_get(obj->ttm.cached_io_rsgt);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
620
return intel_region_ttm_resource_to_rsgt(obj->mm.region, res,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
624
static int i915_ttm_truncate(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
626
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
629
WARN_ON_ONCE(obj->mm.madv == I915_MADV_WILLNEED);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
642
return i915_ttm_purge(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
647
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
655
GEM_WARN_ON(obj->ttm.cached_io_rsgt);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
656
if (!ret && obj->mm.madv != I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
657
i915_ttm_purge(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
680
struct drm_i915_gem_object *obj = i915_ttm_to_gem(mem->bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
686
if (!kref_get_unless_zero(&obj->base.refcount))
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
689
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
691
unknown_state = i915_gem_object_has_unknown_state(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
692
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
711
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
719
base = obj->mm.region->iomap.base - obj->mm.region->region.start;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
720
sg = i915_gem_object_page_iter_get_sg(obj, &obj->ttm.get_io_page, page_offset, &ofs);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
732
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
733
resource_size_t iomap = obj->mm.region->iomap.base -
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
734
obj->mm.region->region.start;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
752
daddr = i915_gem_object_get_dma_address(obj, page);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
802
static int __i915_ttm_get_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
805
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
844
i915_ttm_adjust_domains_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
845
i915_ttm_adjust_gem_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
848
if (!i915_gem_object_has_pages(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
850
i915_ttm_resource_get_st(obj, bo->resource);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
855
GEM_BUG_ON(obj->mm.rsgt);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
856
obj->mm.rsgt = rsgt;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
857
__i915_gem_object_set_pages(obj, &rsgt->table);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
860
GEM_BUG_ON(bo->ttm && ((obj->base.size >> PAGE_SHIFT) < bo->ttm->num_pages));
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
861
i915_ttm_adjust_lru(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
865
static int i915_ttm_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
871
if (overflows_type(obj->base.size >> PAGE_SHIFT, unsigned int))
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
874
GEM_BUG_ON(obj->mm.n_placements > I915_TTM_MAX_PLACEMENTS);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
877
i915_ttm_placement_from_obj(obj, places, &placement);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
879
return __i915_ttm_get_pages(obj, &placement);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
896
static int __i915_ttm_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
904
i915_ttm_place_from_region(mr, &requested, obj->bo_offset,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
905
obj->base.size, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
909
ret = __i915_ttm_get_pages(obj, &placement);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
918
if (obj->mm.region != mr) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
919
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
920
i915_gem_object_init_memory_region(obj, mr);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
926
static int i915_ttm_migrate(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
930
return __i915_ttm_migrate(obj, mr, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
933
static void i915_ttm_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
944
if (obj->mm.rsgt)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
945
i915_refct_sgt_put(fetch_and_zero(&obj->mm.rsgt));
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
952
void i915_ttm_adjust_lru(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
954
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
991
if (kref_get_unless_zero(&obj->base.refcount)) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
992
if (shrinkable != obj->mm.ttm_shrinkable) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
994
if (obj->mm.madv == I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
995
__i915_gem_object_make_shrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
997
__i915_gem_object_make_purgeable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
999
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
20
i915_gem_to_ttm(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
22
return &obj->__do_not_access;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
56
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
71
void i915_ttm_free_cached_io_rsgt(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
74
i915_ttm_resource_get_st(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
77
void i915_ttm_adjust_lru(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.h
79
int i915_ttm_purge(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
102
void i915_ttm_adjust_gem_after_move(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
104
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
134
if (intel_region_to_ttm_type(obj->mm.region) != mem_type) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
135
for (i = 0; i < obj->mm.n_placements; ++i) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
136
struct intel_memory_region *mr = obj->mm.placements[i];
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
139
mr != obj->mm.region) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
140
i915_gem_object_release_memory_region(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
141
i915_gem_object_init_memory_region(obj, mr);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
147
obj->mem_flags &= ~(I915_BO_FLAG_STRUCT_PAGE | I915_BO_FLAG_IOMEM);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
148
obj->mem_flags |= mem_flags;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
150
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
164
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
175
ret = i915_gem_object_unbind(obj, I915_GEM_OBJECT_UNBIND_ACTIVE |
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
180
ret = __i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
196
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
223
i915_ttm_resource_get_st(obj, bo->resource);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
302
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
318
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
333
&obj->ttm.cached_io_rsgt->table,
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
340
i915_ttm_resource_get_st(obj, bo->resource);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
383
copy_work->obj->mm.unknown_state = true;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
391
i915_gem_object_put(copy_work->obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
403
i915_gem_object_put(copy_work->obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
478
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
509
copy_work->obj = i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
548
i915_gem_object_put(copy_work->obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
555
i915_gem_object_put(copy_work->obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
578
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
622
if (obj->mm.madv != I915_MADV_WILLNEED) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
623
i915_ttm_purge(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
635
dst_rsgt = i915_ttm_resource_get_st(obj, dst_mem);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
640
prealloc_bo = obj->flags & I915_BO_PREALLOC;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
677
i915_ttm_adjust_domains_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
678
i915_ttm_free_cached_io_rsgt(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
681
obj->ttm.cached_io_rsgt = dst_rsgt;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
682
obj->ttm.get_io_page.sg_pos = dst_rsgt->table.sgl;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
683
obj->ttm.get_io_page.sg_idx = 0;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
688
i915_ttm_adjust_lru(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
689
i915_ttm_adjust_gem_after_move(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
82
void i915_ttm_adjust_domains_after_move(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
84
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
87
obj->write_domain = I915_GEM_DOMAIN_WC;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
88
obj->read_domains = I915_GEM_DOMAIN_WC;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
90
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
91
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.h
38
void i915_ttm_adjust_domains_after_move(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.h
40
void i915_ttm_adjust_gem_after_move(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
106
obj->ttm.backup = backup;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
118
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
120
i915_ttm_backup_free(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
169
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
173
struct drm_i915_gem_object *backup = obj->ttm.backup;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
181
if (!pm_apply->allow_gpu && !(obj->flags & I915_BO_ALLOC_PM_EARLY))
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
194
err = i915_gem_obj_copy_ttm(obj, backup, pm_apply->allow_gpu,
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
199
obj->ttm.backup = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
22
void i915_ttm_backup_free(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
24
if (obj->ttm.backup) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
25
i915_gem_object_put(obj->ttm.backup);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
26
obj->ttm.backup = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
43
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
47
struct ttm_buffer_object *bo = i915_gem_to_ttm(obj);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
56
if (!i915_ttm_cpu_maps_iomem(bo->resource) || obj->ttm.backup)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
59
if (pm_apply->allow_gpu && i915_gem_object_evictable(obj))
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
63
(pm_apply->allow_gpu && (obj->flags & I915_BO_ALLOC_PM_EARLY)))
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
66
if (obj->flags & I915_BO_ALLOC_PM_VOLATILE)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
77
if (i915_gem_object_needs_ccs_pages(obj)) {
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
78
WARN_ON_ONCE(!i915_gem_object_is_framebuffer(obj));
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
84
obj->base.size, 0, flags);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.c
97
err = i915_gem_obj_copy_ttm(backup, obj, pm_apply->allow_gpu, false);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_pm.h
24
void i915_ttm_backup_free(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
100
unsigned int max_segment = i915_sg_segment_size(obj->base.dev->dev);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
106
if (overflows_type(obj->base.size >> PAGE_SHIFT, num_pages))
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
109
num_pages = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
114
if (!obj->userptr.page_ref) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
119
obj->userptr.page_ref++;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
120
pvec = obj->userptr.pvec;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
129
ret = i915_gem_gtt_prepare_pages(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
141
WARN_ON_ONCE(!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE));
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
142
if (i915_gem_object_can_bypass_llc(obj))
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
143
obj->cache_dirty = true;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
145
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
150
i915_gem_object_userptr_drop_ref(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
157
i915_gem_userptr_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
166
__i915_gem_object_release_shmem(obj, pages, true);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
167
i915_gem_gtt_finish_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
174
if (i915_gem_object_is_readonly(obj))
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
175
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
178
if (obj->mm.dirty && trylock_page(page)) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
203
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
208
i915_gem_object_userptr_drop_ref(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
211
static int i915_gem_object_userptr_unbind(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
216
err = i915_gem_object_unbind(obj, I915_GEM_OBJECT_UNBIND_ACTIVE);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
220
if (GEM_WARN_ON(i915_gem_object_has_pinned_pages(obj)))
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
223
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
225
pages = __i915_gem_object_unset_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
227
i915_gem_userptr_put_pages(obj, pages);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
232
int i915_gem_object_userptr_submit_init(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
234
const unsigned long num_pages = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
240
if (obj->userptr.notifier.mm != current->mm)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
243
notifier_seq = mmu_interval_read_begin(&obj->userptr.notifier);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
245
ret = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
249
if (notifier_seq == obj->userptr.notifier_seq && obj->userptr.pvec) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
250
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
254
ret = i915_gem_object_userptr_unbind(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
255
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
263
if (!i915_gem_object_is_readonly(obj))
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
268
ret = pin_user_pages_fast(obj->userptr.ptr + pinned * PAGE_SIZE,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
277
ret = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
281
if (mmu_interval_read_retry(&obj->userptr.notifier,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
282
!obj->userptr.page_ref ? notifier_seq :
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
283
obj->userptr.notifier_seq)) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
288
if (!obj->userptr.page_ref++) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
289
obj->userptr.pvec = pvec;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
290
obj->userptr.notifier_seq = notifier_seq;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
292
ret = ____i915_gem_object_get_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
295
obj->userptr.page_ref--;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
298
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
309
int i915_gem_object_userptr_submit_done(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
311
if (mmu_interval_read_retry(&obj->userptr.notifier,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
312
obj->userptr.notifier_seq)) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
321
int i915_gem_object_userptr_validate(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
325
err = i915_gem_object_userptr_submit_init(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
329
err = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
336
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
338
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
340
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
347
i915_gem_userptr_release(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
349
GEM_WARN_ON(obj->userptr.page_ref);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
351
if (!obj->userptr.notifier.mm)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
354
mmu_interval_notifier_remove(&obj->userptr.notifier);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
355
obj->userptr.notifier.mm = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
359
i915_gem_userptr_dmabuf_export(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
361
drm_dbg(obj->base.dev, "Exporting userptr no longer allowed\n");
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
367
i915_gem_userptr_pwrite(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
370
drm_dbg(obj->base.dev, "pwrite to userptr no longer allowed\n");
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
376
i915_gem_userptr_pread(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
379
drm_dbg(obj->base.dev, "pread from userptr no longer allowed\n");
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
470
struct drm_i915_gem_object __maybe_unused *obj;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
526
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
527
if (obj == NULL)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
530
drm_gem_private_object_init(dev, &obj->base, args->user_size);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
531
i915_gem_object_init(obj, &i915_gem_userptr_ops, &lock_class,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
533
obj->mem_flags = I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
534
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
535
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
536
i915_gem_object_set_cache_coherency(obj, I915_CACHE_LLC);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
538
obj->userptr.ptr = args->user_ptr;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
539
obj->userptr.notifier_seq = ULONG_MAX;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
541
i915_gem_object_set_readonly(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
547
ret = i915_gem_userptr_init__mmu_notifier(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
549
ret = drm_gem_handle_create(file, &obj->base, &handle);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
552
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
71
i915_gem_userptr_init__mmu_notifier(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
73
return mmu_interval_notifier_insert(&obj->userptr.notifier, current->mm,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
74
obj->userptr.ptr, obj->base.size,
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
78
static void i915_gem_object_userptr_drop_ref(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
82
assert_object_held_shared(obj);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
84
if (!--obj->userptr.page_ref) {
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
85
pvec = obj->userptr.pvec;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
86
obj->userptr.pvec = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
88
GEM_BUG_ON(obj->userptr.page_ref < 0);
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
91
const unsigned long num_pages = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_userptr.c
98
static int i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
142
i915_gem_object_wait_priority(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
149
dma_resv_iter_begin(&cursor, obj->base.resv,
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
164
i915_gem_object_wait(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
171
timeout = i915_gem_object_wait_reservation(obj->base.resv,
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
229
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
236
obj = i915_gem_object_lookup(file, args->bo_handle);
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
237
if (!obj)
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
242
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
268
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
283
int i915_gem_object_wait_migration(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/i915_gem_wait.c
288
return i915_gem_object_wait_moving_fence(obj, !!(flags & I915_WAIT_INTERRUPTIBLE));
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
106
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
11
static void huge_free_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
113
if (overflows_type(dma_size, obj->base.size))
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
116
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
117
if (!obj)
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
120
drm_gem_private_object_init(&i915->drm, &obj->base, dma_size);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
121
i915_gem_object_init(obj, &huge_ops, &lock_class, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
122
obj->mem_flags |= I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
124
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
125
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
127
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
128
obj->scratch = phys_size;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
130
return obj;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
14
unsigned long nreal = obj->scratch / PAGE_SIZE;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
28
static int huge_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
31
const unsigned long nreal = obj->scratch / PAGE_SIZE;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
37
if (overflows_type(obj->base.size / PAGE_SIZE, npages))
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
40
npages = obj->base.size / PAGE_SIZE;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
72
if (i915_gem_gtt_prepare_pages(obj, pages))
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
75
__i915_gem_object_set_pages(obj, pages);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
80
huge_free_pages(obj, pages);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
85
static void huge_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
88
i915_gem_gtt_finish_pages(obj, pages);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
89
huge_free_pages(obj, pages);
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.c
91
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.h
22
huge_gem_object_phys_size(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.h
24
return obj->scratch;
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.h
28
huge_gem_object_dma_size(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_gem_object.h
30
return obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
100
rem = obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1028
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1029
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1030
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1031
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1032
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1043
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1044
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1045
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1047
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1062
i915_gem_object_lock(vma->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1063
err = i915_gem_object_set_to_gtt_domain(vma->obj, true);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1064
i915_gem_object_unlock(vma->obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1073
__cpu_check_shmem(struct drm_i915_gem_object *obj, u32 dword, u32 val)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1079
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1080
err = i915_gem_object_prepare_read(obj, &needs_flush);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1084
for (n = 0; n < obj->base.size >> PAGE_SHIFT; ++n) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1085
u32 *ptr = kmap_local_page(i915_gem_object_get_page(obj, n));
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1101
i915_gem_object_finish_access(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1103
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1108
static int __cpu_check_vmap(struct drm_i915_gem_object *obj, u32 dword, u32 val)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1110
unsigned long n = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1114
err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1118
ptr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1134
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1138
static int cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1140
if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1141
return __cpu_check_shmem(obj, dword, val);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1143
return __cpu_check_vmap(obj, dword, val);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1147
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1155
vma = i915_vma_instance(obj, ce->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1181
err = cpu_check(obj, dword, val);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1193
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1221
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1223
size = obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1224
if (obj->mm.page_sizes.sg & I915_GTT_PAGE_SIZE_64K &&
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1254
max_page_size = rounddown_pow_of_two(obj->mm.page_sizes.sg);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1285
if (obj->mm.page_sizes.sg & I915_GTT_PAGE_SIZE_64K)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1289
err = __igt_write_huge(ce, obj, size, offset_low,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1294
err = __igt_write_huge(ce, obj, size, offset_high,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1374
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
139
if (i915_gem_gtt_prepare_pages(obj, st))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1401
obj = backends[i].fn(i915, size, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1402
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1403
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1415
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1418
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
142
GEM_BUG_ON(sg_page_sizes != obj->mm.page_mask);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1425
if (obj->mm.page_sizes.phys < min) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
143
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1432
err = igt_write_huge(i915, obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1438
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1439
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1440
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1441
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1443
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1499
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1503
obj = backends[i].fn(i915, size, backends[i].flags);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1504
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1505
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1515
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1517
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1521
GEM_BUG_ON(pages > obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1525
obj->mm.page_sizes.sg = pages;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1527
err = igt_write_huge(i915, obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1529
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1530
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1531
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1532
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1533
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
155
static void put_huge_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1555
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1577
obj = i915_gem_object_create_lmem(i915, SZ_4M, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1578
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1579
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
158
i915_gem_gtt_finish_pages(obj, pages);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1581
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1585
if (obj->mm.page_sizes.phys < I915_GTT_PAGE_SIZE_64K) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1594
obj->mm.page_sizes.sg = I915_GTT_PAGE_SIZE_64K;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1596
err = igt_write_huge(i915, obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1601
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1603
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
161
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1615
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
163
__start_cpu_write(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1664
obj = i915_gem_object_create_region(mr, sz, 0, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1665
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1666
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1670
list_add_tail(&obj->st_link, &objects);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1672
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1689
addr += obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1715
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1718
addr = round_up(addr, obj->mm.region->min_page_size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1725
err = __igt_write_huge(ce, obj, obj->base.size, addr, 0, rnd);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1729
err = __igt_write_huge(ce, obj, obj->base.size, addr,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1734
err = __igt_write_huge(ce, obj, obj->base.size, addr,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1740
addr += obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1748
list_for_each_entry_safe(obj, on, &objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1749
list_del(&obj->st_link);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1750
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1765
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
179
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1790
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1791
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1792
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1796
vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1803
__i915_gem_object_flush_map(obj, 0, 64);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1804
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1806
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1820
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1835
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1867
obj = i915_gem_object_create_shmem(i915, SZ_2M);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1868
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1869
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1873
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
188
if (overflows_type(size, obj->base.size))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1885
if (obj->mm.page_sizes.phys < I915_GTT_PAGE_SIZE_2M) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
191
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
192
if (!obj)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1928
if (should_swap == i915_gem_object_has_pages(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1935
if (should_swap == (obj->mm.page_sizes.sg || obj->mm.page_sizes.phys)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1947
err = cpu_check(obj, n, 0xdeadbeaf);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
195
drm_gem_private_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
1957
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
196
i915_gem_object_init(obj, &huge_page_ops, &lock_class, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
197
obj->mem_flags |= I915_BO_FLAG_STRUCT_PAGE;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
198
i915_gem_object_set_volatile(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
200
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
201
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
204
i915_gem_object_set_cache_coherency(obj, cache_level);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
206
obj->mm.page_mask = page_mask;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
208
return obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
211
static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
213
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
220
if (overflows_type(obj->base.size >> PAGE_SHIFT, unsigned int))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
227
if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
233
rem = obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
261
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
266
static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
268
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
285
page_size = get_largest_page_size(i915, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
289
sg->length = obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
290
sg_dma_len(sg) = obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
293
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
299
static void fake_free_huge_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
306
static void fake_put_huge_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
309
fake_free_huge_pages(obj, pages);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
310
obj->mm.dirty = false;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
331
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
339
if (overflows_type(size, obj->base.size))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
342
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
343
if (!obj)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
346
drm_gem_private_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
349
i915_gem_object_init(obj, &fake_ops_single, &lock_class, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
351
i915_gem_object_init(obj, &fake_ops, &lock_class, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
353
i915_gem_object_set_volatile(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
355
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
356
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
357
obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
359
return obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
366
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
386
if (vma->page_sizes.phys != obj->mm.page_sizes.phys) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
388
vma->page_sizes.phys, obj->mm.page_sizes.phys);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
392
if (vma->page_sizes.sg != obj->mm.page_sizes.sg) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
394
vma->page_sizes.sg, obj->mm.page_sizes.sg);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
410
if (i915_gem_object_is_lmem(obj) &&
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
427
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
448
obj = fake_huge_pages_object(i915, combination, !!single);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
449
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
450
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
454
if (obj->base.size != combination) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
456
obj->base.size, combination);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
461
vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
480
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
490
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
504
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
521
obj = i915_gem_object_create_region(mem,
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
524
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
525
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
529
vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
543
phys = i915_gem_object_get_dma_address(obj, 0);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
560
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
561
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
570
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
581
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
603
obj = fake_huge_pages_object(i915, size, true);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
604
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
605
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
607
if (obj->base.size != size) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
609
obj->base.size, size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
614
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
619
obj->mm.page_sizes.sg = page_size;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
621
vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
679
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
680
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
681
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
682
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
683
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
689
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
690
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
691
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
693
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
700
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
702
list_for_each_entry_safe(obj, on, objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
703
list_del(&obj->st_link);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
704
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
705
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
706
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
707
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
708
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
743
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
749
obj = fake_huge_pages_object(i915, size, single);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
750
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
751
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
755
if (obj->base.size != size) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
757
obj->base.size, size);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
758
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
763
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
765
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
769
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
771
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
78
static int get_huge_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
806
if (!has_pte64 && (obj->base.size < I915_GTT_PAGE_SIZE_2M ||
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
81
unsigned int page_mask = obj->mm.page_mask;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
833
obj->base.size, str_yes_no(!!single));
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
840
__func__, obj->base.size))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
861
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
88
if (overflows_type(obj->base.size >> PAGE_SHIFT, unsigned int))
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
95
if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
966
obj = fake_huge_pages_object(i915, size, !!single);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
967
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
968
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
972
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
980
obj->mm.page_sizes.sg &= ~I915_GTT_PAGE_SIZE_2M;
sys/dev/pci/drm/i915/gem/selftests/huge_pages.c
982
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
274
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
278
obj = i915_gem_object_create_lmem(i915, size, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
280
obj = i915_gem_object_create_shmem(i915, size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
281
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
282
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
284
vma = i915_vma_instance(obj, t->ce->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
286
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
341
i915_gem_object_flush_map(t->scratch.vma->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
428
vaddr = i915_gem_object_pin_map_unlocked(buf->vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
449
i915_gem_object_unpin_map(buf->vma->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
497
err = prepare_blit(t, dst, src, t->batch->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
606
map = i915_gem_object_pin_map_unlocked(t->scratch.vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_client_blt.c
626
i915_gem_object_unpin_map(t->scratch.vma->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
123
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
124
err = i915_gem_object_set_to_gtt_domain(ctx->obj, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
125
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
129
vma = i915_gem_object_ggtt_pin(ctx->obj, NULL, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
155
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
156
err = i915_gem_object_set_to_wc_domain(ctx->obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
157
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
161
map = i915_gem_object_pin_map_unlocked(ctx->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
167
__i915_gem_object_flush_map(ctx->obj, offset, sizeof(*map));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
168
i915_gem_object_unpin_map(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
178
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
179
err = i915_gem_object_set_to_wc_domain(ctx->obj, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
180
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
184
map = i915_gem_object_pin_map_unlocked(ctx->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
189
i915_gem_object_unpin_map(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
19
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
201
vma = i915_gem_object_ggtt_pin(ctx->obj, NULL, 0, 0, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
205
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
206
err = i915_gem_object_set_to_gtt_domain(ctx->obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
247
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
30
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
31
err = i915_gem_object_prepare_write(ctx->obj, &needs_clflush);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
35
page = i915_gem_object_get_page(ctx->obj, offset >> PAGE_SHIFT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
362
ctx.obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
363
if (IS_ERR(ctx.obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
364
err = PTR_ERR(ctx.obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
411
i915_gem_object_put(ctx.obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
423
i915_gem_object_put(ctx.obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
47
i915_gem_object_finish_access(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
50
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
61
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
62
err = i915_gem_object_prepare_read(ctx->obj, &needs_clflush);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
66
page = i915_gem_object_get_page(ctx->obj, offset >> PAGE_SHIFT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
75
i915_gem_object_finish_access(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
78
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
89
i915_gem_object_lock(ctx->obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
90
err = i915_gem_object_set_to_gtt_domain(ctx->obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
91
i915_gem_object_unlock(ctx->obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_coherency.c
95
vma = i915_gem_object_ggtt_pin(ctx->obj, NULL, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1091
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1101
ret = emit_rpcs_query(obj, ce, &rq);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1113
buf = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1131
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1165
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1179
ret = __read_slice_count(ce, obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1185
ret = __read_slice_count(ce->engine->kernel_context, obj, NULL, &rpcs);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1197
ret = __read_slice_count(ce, obj, NULL, &rpcs);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1209
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1225
ret = __sseu_finish(name, flags, ce, obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1244
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1254
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1255
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1256
ret = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1302
ret = __sseu_test(name, flags, ce, obj, engine->sseu);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1307
ret = __sseu_test(name, flags, ce, obj, pg_sseu);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1312
ret = __sseu_test(name, flags, ce, obj, engine->sseu);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1317
ret = __sseu_test(name, flags, ce, obj, pg_sseu);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1331
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1370
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1423
if (!obj) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1424
obj = create_test_object(ce->vm, file, &objects);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1425
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1426
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1432
i915_gem_object_set_readonly(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1435
err = gpu_fill(ce, obj, dw);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1438
ndwords, dw, max_dwords(obj),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1452
if (++dw == max_dwords(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1453
obj = NULL;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1465
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1467
min_t(unsigned int, ndwords - dw, max_dwords(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1471
if (i915_gem_object_is_readonly(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1474
err = cpu_check(obj, idx++, num_writes);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1510
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1526
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1540
__i915_gem_object_flush_map(obj, 0, 64);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1541
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1546
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1591
err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1598
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1620
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1630
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1647
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1648
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1656
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1666
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1681
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1682
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1714
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1715
err = i915_gem_object_set_to_cpu_domain(obj, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1716
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1720
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1727
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
1739
err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
431
static unsigned long real_page_count(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
433
return huge_gem_object_phys_size(obj) >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
436
static unsigned long fake_page_count(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
438
return huge_gem_object_dma_size(obj) >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
442
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
448
GEM_BUG_ON(obj->base.size > ce->vm->total);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
451
vma = i915_vma_instance(obj, ce->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
468
(dw * real_page_count(obj)) << PAGE_SHIFT |
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
470
real_page_count(obj),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
477
static int cpu_fill(struct drm_i915_gem_object *obj, u32 value)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
479
const bool has_llc = HAS_LLC(to_i915(obj->base.dev));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
484
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
485
err = i915_gem_object_prepare_write(obj, &need_flush);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
489
for (n = 0; n < real_page_count(obj); n++) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
492
map = kmap_local_page(i915_gem_object_get_page(obj, n));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
500
i915_gem_object_finish_access(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
501
obj->read_domains = I915_GEM_DOMAIN_GTT | I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
502
obj->write_domain = 0;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
504
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
508
static noinline int cpu_check(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
515
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
516
err = i915_gem_object_prepare_read(obj, &needs_flush);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
520
for (n = 0; n < real_page_count(obj); n++) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
523
map = kmap_local_page(i915_gem_object_get_page(obj, n));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
531
n, real_page_count(obj), m, max,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
554
i915_gem_object_finish_access(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
556
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
560
static int file_add_object(struct file *file, struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
564
GEM_BUG_ON(obj->base.handle_count);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
568
&obj->base, 1, 0, GFP_KERNEL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
572
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
573
obj->base.handle_count++;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
582
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
592
obj = huge_gem_object(vm->i915, DW_PER_PAGE * PAGE_SIZE, size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
593
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
594
return obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
596
err = file_add_object(file, obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
597
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
601
err = cpu_fill(obj, STACK_MAGIC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
608
list_add_tail(&obj->st_link, objects);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
609
return obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
612
static unsigned long max_dwords(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
614
unsigned long npages = fake_page_count(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
675
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
713
if (!obj) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
714
obj = create_test_object(ce->vm, file, &objects);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
715
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
716
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
723
err = gpu_fill(ce, obj, dw);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
726
ndwords, dw, max_dwords(obj),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
742
if (++dw == max_dwords(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
743
obj = NULL;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
758
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
760
min_t(unsigned int, ndwords - dw, max_dwords(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
762
err = cpu_check(obj, ncontexts++, rem);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
823
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
846
if (!obj) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
847
obj = create_test_object(parent->vm,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
849
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
850
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
857
err = gpu_fill(ce, obj, dw);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
860
ndwords, dw, max_dwords(obj),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
876
if (++dw == max_dwords(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
877
obj = NULL;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
891
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
893
min_t(unsigned int, ndwords - dw, max_dwords(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
895
err = cpu_check(obj, ncontexts++, rem);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
940
emit_rpcs_query(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
944
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
957
vma = i915_vma_instance(obj, ce->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
973
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
977
err = i915_gem_object_set_to_gtt_domain(obj, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
107
obj = __i915_gem_object_create_user(i915, PAGE_SIZE, &lmem, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
108
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
110
PTR_ERR(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
111
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
115
dmabuf = i915_gem_prime_export(&obj->base, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
142
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
21
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
223
struct drm_i915_gem_object *obj, *import_obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
233
obj = __i915_gem_object_create_user(i915, SZ_8M,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
235
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
237
PTR_ERR(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
238
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
24
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
242
dmabuf = i915_gem_prime_export(&obj->base, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
25
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
259
if (import == &obj->base) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
26
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
278
if (obj->mm.region != i915->mm.regions[INTEL_REGION_SMEM]) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
28
dmabuf = i915_gem_prime_export(&obj->base, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
285
err = verify_access(i915, obj, import_obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
29
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
290
import_attach = dma_buf_attach(dmabuf, obj->base.dev->dev);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
317
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
347
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
358
obj = to_intel_bo(i915_gem_prime_import(&i915->drm, dmabuf));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
359
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
361
(int)PTR_ERR(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
362
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
366
if (obj->base.dev != &i915->drm) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
372
if (obj->base.size != PAGE_SIZE) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
374
(long long)obj->base.size, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
388
obj_map = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
400
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
410
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
415
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
422
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
43
struct drm_i915_gem_object *obj, *import_obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
431
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
452
obj = to_intel_bo(i915_gem_prime_import(&i915->drm, dmabuf));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
453
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
455
(int)PTR_ERR(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
456
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
462
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
469
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
471
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
48
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
482
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
488
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
489
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
49
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
490
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
492
dmabuf = i915_gem_prime_export(&obj->base, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
499
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
50
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
52
dmabuf = i915_gem_prime_export(&obj->base, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
524
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
69
if (import != &obj->base) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
89
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_dmabuf.c
97
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
112
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
118
err = i915_gem_object_lock(obj, ww);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
123
err = i915_vma_pin_ww(vma, ww, obj->base.size, 0,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
140
if (i915_gem_object_is_lmem(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
141
err = i915_gem_object_migrate(obj, ww, INTEL_REGION_SMEM);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
149
if (i915_gem_object_is_lmem(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
15
static int igt_fill_check_buffer(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
154
if (!i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
160
err = i915_gem_object_migrate(obj, ww, INTEL_REGION_LMEM_0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
168
if (i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
173
if (!i915_gem_object_is_lmem(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
19
unsigned int i, count = obj->base.size / sizeof(u32);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
190
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
199
obj = i915_gem_object_create_lmem(i915, SZ_2M, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
200
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
201
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
204
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
21
intel_gt_coherent_map_type(gt, obj, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
213
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
217
err = ____i915_gem_object_get_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
222
obj->mm.pages->sgl, obj->pat_index,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
223
i915_gem_object_is_lmem(obj),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
226
err = dma_resv_reserve_fences(obj->base.resv, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
228
dma_resv_add_fence(obj->base.resv, &rq->fence,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
236
err = igt_fill_check_buffer(obj, gt, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
25
assert_object_held(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
250
err = lmem_pages_migrate_one(&ww, obj, vma,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
256
err = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
26
cur = i915_gem_object_pin_map(obj, map_type);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
270
err = i915_gem_object_wait_migration(obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
279
err = igt_fill_check_buffer(obj, gt, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
283
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
285
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
41
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
52
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
60
obj = i915_gem_object_create_region(src_mr, dst_mr->min_page_size, 0, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
61
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
62
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
65
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
69
err = igt_fill_check_buffer(obj, gt, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
73
err = i915_gem_object_migrate(obj, &ww, dst);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
77
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
81
if (i915_gem_object_can_migrate(obj, src))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
84
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
85
err = i915_gem_object_wait_migration(obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
89
err = igt_fill_check_buffer(obj, gt, false);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_migrate.c
91
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1000
err = __igt_mmap(i915, obj, I915_MMAP_TYPE_GTT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1002
err = __igt_mmap(i915, obj, I915_MMAP_TYPE_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1004
err = __igt_mmap(i915, obj, I915_MMAP_TYPE_FIXED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1006
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1018
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1020
list_for_each_entry_safe(obj, on, objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1021
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1022
if (i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1023
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1025
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1026
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1027
list_del(&obj->st_link);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1028
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1038
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1040
list_for_each_entry(obj, objects, st_link) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1041
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1042
if (i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1043
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1044
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1059
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1061
obj = i915_gem_object_create_region(mr, size, 0, 0);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1062
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1063
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1067
list_add(&obj->st_link, objects);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1069
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
108
err = i915_gem_object_set_tiling(obj, tile->tiling, tile->stride);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1083
total += obj->base.size;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1095
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1102
obj->mm.region->name, I915_MMAP_TYPE_FIXED, addr);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1104
for (i = 0; i < obj->base.size / sizeof(u32); i++) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1111
obj->mm.region->name, i * sizeof(x));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1120
obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1126
obj->mm.region->name,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1134
obj->mm.region->name, i * sizeof(x));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1142
obj->flags &= ~I915_BO_ALLOC_GPU_ONLY;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1143
return wc_check(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
115
GEM_BUG_ON(i915_gem_object_get_tiling(obj) != tile->tiling);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1157
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
116
GEM_BUG_ON(i915_gem_object_get_stride(obj) != tile->stride);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1165
obj = __i915_gem_object_create_user(i915, PAGE_SIZE,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1168
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1169
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1172
obj->flags |= I915_BO_ALLOC_GPU_ONLY;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1174
err = __assign_mmap_offset(obj, I915_MMAP_TYPE_FIXED, &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
118
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1184
addr = igt_mmap_offset(i915, offset, obj->base.size,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
119
err = i915_gem_object_set_to_gtt_domain(obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1196
obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
120
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1207
err = i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1211
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1213
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1218
obj->mm.pages->sgl, obj->pat_index,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1219
i915_gem_object_is_lmem(obj),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1221
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1223
err = dma_resv_reserve_fences(obj->base.resv, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1225
dma_resv_add_fence(obj->base.resv, &rq->fence,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1229
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1237
err = i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1245
err = i915_gem_object_wait_moving_fence(obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1246
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1252
err = ___igt_mmap_migrate(i915, obj, addr,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1255
if (!err && obj->mm.region != expected_mr) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
127
view = compute_partial_view(obj, page, MIN_CHUNK_PAGES);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1286
if (!i915_gem_object_has_unknown_state(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
129
vma = i915_gem_object_ggtt_pin(obj, &view, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1293
vm_munmap(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1296
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1415
static bool can_access(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1419
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1420
access = i915_gem_object_has_struct_page(obj) ||
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1421
i915_gem_object_has_iomem(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1422
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1428
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1441
if (!can_mmap(obj, type) || !can_access(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1444
err = __assign_mmap_offset(obj, type, &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1448
addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1456
obj->mm.region->name, repr_mmap_type(type));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1465
obj->mm.region->name, repr_mmap_type(type));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1472
obj->mm.region->name, repr_mmap_type(type));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1481
obj->mm.region->name, repr_mmap_type(type));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1487
obj->mm.region->name, repr_mmap_type(type),
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1494
vm_munmap(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1505
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1511
obj = __i915_gem_object_create_user(i915, PAGE_SIZE, &mr, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1512
if (obj == ERR_PTR(-ENODEV))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1515
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1516
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1518
err = __igt_mmap_access(i915, obj, I915_MMAP_TYPE_GTT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
152
if (offset >= obj->base.size)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1520
err = __igt_mmap_access(i915, obj, I915_MMAP_TYPE_WB);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1522
err = __igt_mmap_access(i915, obj, I915_MMAP_TYPE_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1524
err = __igt_mmap_access(i915, obj, I915_MMAP_TYPE_UC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1526
err = __igt_mmap_access(i915, obj, I915_MMAP_TYPE_FIXED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1528
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1537
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1553
if (!can_mmap(obj, type))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1556
err = wc_set(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1558
err = gtt_set(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1562
err = __assign_mmap_offset(obj, type, &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1566
addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
157
p = i915_gem_object_get_page(obj, offset >> PAGE_SHIFT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1573
pr_err("%s: Unable to write to mmap\n", obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1586
vma = i915_vma_instance(obj, engine->kernel_context->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1594
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1617
__func__, engine->name, obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1640
vm_munmap(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1651
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1657
obj = __i915_gem_object_create_user(i915, PAGE_SIZE, &mr, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1658
if (obj == ERR_PTR(-ENODEV))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
166
tile->tiling ? tile_row_pages(obj) : 0,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1661
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1662
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1664
err = __igt_mmap_gpu(i915, obj, I915_MMAP_TYPE_GTT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1666
err = __igt_mmap_gpu(i915, obj, I915_MMAP_TYPE_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1668
err = __igt_mmap_gpu(i915, obj, I915_MMAP_TYPE_FIXED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1670
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1735
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1742
if (!can_mmap(obj, type))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1745
err = __assign_mmap_offset(obj, type, &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1749
addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1753
err = prefault_range(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1757
err = check_present(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1759
pr_err("%s: was not present\n", obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1768
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1769
err = i915_gem_object_unbind(obj, I915_GEM_OBJECT_UNBIND_ACTIVE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1770
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1777
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1778
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1779
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1780
if (i915_gem_object_has_pages(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1787
err = check_absent(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1789
pr_err("%s: was not absent\n", obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
179
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1794
vm_munmap(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1805
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
181
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1811
obj = __i915_gem_object_create_user(i915, PAGE_SIZE, &mr, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1812
if (obj == ERR_PTR(-ENODEV))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1815
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1816
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1818
err = __igt_mmap_revoke(i915, obj, I915_MMAP_TYPE_GTT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1820
err = __igt_mmap_revoke(i915, obj, I915_MMAP_TYPE_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1822
err = __igt_mmap_revoke(i915, obj, I915_MMAP_TYPE_FIXED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1824
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
185
static int check_partial_mappings(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
189
const unsigned int nreal = obj->scratch / PAGE_SIZE;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
190
const unsigned long npages = obj->base.size / PAGE_SIZE;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
191
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
196
err = i915_gem_object_set_tiling(obj, tile->tiling, tile->stride);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
203
GEM_BUG_ON(i915_gem_object_get_tiling(obj) != tile->tiling);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
204
GEM_BUG_ON(i915_gem_object_get_stride(obj) != tile->stride);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
206
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
207
err = i915_gem_object_set_to_gtt_domain(obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
208
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
216
compute_partial_view(obj, page, MIN_CHUNK_PAGES);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
226
vma = i915_gem_object_ggtt_pin(obj, &view, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
248
if (offset >= obj->base.size)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
253
p = i915_gem_object_get_page(obj, offset >> PAGE_SHIFT);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
262
tile->tiling ? tile_row_pages(obj) : 0,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
276
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
278
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
319
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
335
obj = huge_gem_object(i915,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
338
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
339
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
341
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
344
nreal, obj->base.size / PAGE_SIZE, err);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
361
err = check_partial_mappings(obj, &tile, end);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
399
err = check_partial_mappings(obj, &tile, end);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
407
err = check_partial_mappings(obj, &tile, end);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
416
err = check_partial_mappings(obj, &tile, end);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
427
err = check_partial_mappings(obj, &tile, end);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
440
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
442
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
450
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
472
obj = huge_gem_object(i915,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
475
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
476
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
478
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
481
nreal, obj->base.size / PAGE_SIZE, err);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
524
err = check_partial_mapping(obj, &tile, &prng);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
534
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
536
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
540
static int make_obj_busy(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
542
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
551
vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
557
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
586
i915_gem_object_put(obj); /* leave it only alive via its active ref */
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
616
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
620
obj = create_sys_or_internal(i915, size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
621
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
622
return expected && expected == PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
624
ret = __assign_mmap_offset(obj, default_mapping(i915), &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
625
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
660
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
722
obj = create_sys_or_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
723
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
724
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
729
err = __assign_mmap_offset(obj, default_mapping(i915), &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
741
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
748
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
749
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
750
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
754
err = make_obj_busy(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
775
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
779
static int gtt_set(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
786
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
798
memset_io(map, POISON_INUSE, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
806
static int gtt_check(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
813
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
825
if (memchr_inv((void __force *)map, POISON_FREE, obj->base.size)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
827
obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
837
static int wc_set(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
841
vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
845
memset(vaddr, POISON_INUSE, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
846
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
847
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
852
static int wc_check(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
857
vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
861
if (memchr_inv(vaddr, POISON_FREE, obj->base.size)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
863
obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
866
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
871
static bool can_mmap(struct drm_i915_gem_object *obj, enum i915_mmap_type type)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
873
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
876
if (obj->ops->mmap_offset)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
885
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
887
!i915_gem_object_has_struct_page(obj) &&
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
888
!i915_gem_object_has_iomem(obj));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
889
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
896
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
904
if (!can_mmap(obj, type))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
907
err = wc_set(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
909
err = gtt_set(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
913
err = __assign_mmap_offset(obj, type, &offset, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
917
addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
92
static int check_partial_mapping(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
921
pr_debug("igt_mmap(%s, %d) @ %lx\n", obj->mm.region->name, type, addr);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
928
obj->mm.region->name);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
933
for (i = 0; i < obj->base.size / sizeof(u32); i++) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
939
obj->mm.region->name, i * sizeof(x));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
946
obj->mm.region->name,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
955
obj->mm.region->name, i * sizeof(x));
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
96
const unsigned long npages = obj->base.size / PAGE_SIZE;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
964
err = wc_check(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
966
err = gtt_check(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
968
vm_munmap(addr, obj->base.size);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
97
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
990
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
993
obj = __i915_gem_object_create_user(i915, sizes[i], &mr, 1);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
994
if (obj == ERR_PTR(-ENODEV))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
997
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
998
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
16
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
21
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
22
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
23
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
29
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
38
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
44
obj = huge_gem_object(i915,
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
47
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
48
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
50
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
53
nreal, obj->base.size / PAGE_SIZE, err);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
57
for (n = 0; n < obj->base.size / PAGE_SIZE; n++) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
58
if (i915_gem_object_get_page(obj, n) !=
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
59
i915_gem_object_get_page(obj, n % nreal)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
68
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_object.c
70
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
14
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
21
obj = i915_gem_object_create_shmem(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
22
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
23
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
28
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
29
if (!i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
30
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
36
err = i915_gem_object_attach_phys(obj, PAGE_SIZE);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
37
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
43
if (i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
49
if (!atomic_read(&obj->mm.pages_pin_count)) {
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
56
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
57
err = i915_gem_object_set_to_gtt_domain(obj, true);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
58
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_phys.c
66
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
107
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
47
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
55
obj = i915_gem_object_create_internal(vma->vm->i915, size);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
56
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
57
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
59
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
89
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
90
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gem/selftests/igt_gem_utils.c
94
vma = i915_vma_instance(obj, vma->vm, NULL);
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
362
static int pd_dummy_obj_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
364
obj->mm.pages = ZERO_SIZE_PTR;
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
368
static void pd_dummy_obj_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/gen7_renderclear.c
442
GEM_BUG_ON(vma->obj->base.size < bv.size);
sys/dev/pci/drm/i915/gt/gen7_renderclear.c
444
batch = i915_gem_object_pin_map(vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/gen7_renderclear.c
450
i915_gem_object_flush_map(vma->obj);
sys/dev/pci/drm/i915/gt/gen7_renderclear.c
451
__i915_gem_object_release_map(vma->obj);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
226
if (vm->rsvd.obj)
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
227
i915_gem_object_put(vm->rsvd.obj);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
863
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
865
obj = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
866
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
867
ret = PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
871
ret = map_pt_dma(vm, obj);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
873
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
877
fill_px(obj, vm->scratch[i - 1]->encode);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
878
obj->encode = gen8_pde_encode(px_dma(obj), I915_CACHE_NONE);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
880
vm->scratch[i] = obj;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
960
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
968
obj = i915_gem_object_create_lmem(i915, PAGE_SIZE,
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
971
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
972
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
973
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
974
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
976
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
987
vm->rsvd.obj = obj;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
991
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_context.c
144
vma->obj->mm.dirty = true;
sys/dev/pci/drm/i915/gt/intel_context.c
246
err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, ww);
sys/dev/pci/drm/i915/gt/intel_context.c
248
err = i915_gem_object_lock(ce->ring->vma->obj, ww);
sys/dev/pci/drm/i915/gt/intel_context.c
250
err = i915_gem_object_lock(ce->state->obj, ww);
sys/dev/pci/drm/i915/gt/intel_context.c
315
i915_gem_ww_unlock_single(ce->timeline->hwsp_ggtt->obj);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1042
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1043
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1073
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1088
obj = i915_gem_object_create_internal(engine->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1089
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1091
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1094
i915_gem_object_set_cache_coherency(obj, I915_CACHE_LLC);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1096
vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1104
ret = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1110
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1131
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
41
struct drm_i915_gem_object *obj = ce->state->obj;
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
42
int type = intel_gt_coherent_map_type(ce->engine->gt, obj, true);
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
45
if (!i915_gem_object_trylock(obj, NULL))
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
48
map = i915_gem_object_pin_map(obj, type);
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
50
memset(map, CONTEXT_REDZONE, obj->base.size);
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
51
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
52
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
54
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2619
__i915_gem_object_flush_map(ce->state->obj, 0, engine->context_size);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1119
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1122
trylock = i915_gem_object_trylock(obj, NULL);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1127
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
165
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
173
if (WARN_ON(!i915_gem_object_trylock(obj, NULL))) {
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1774
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
178
i915_gem_object_get(obj);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1786
obj ? obj->pat_index :
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1791
if (obj) { /* only used during resume => exclusive access */
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1792
write_domain_objs |= fetch_and_zero(&obj->write_domain);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1793
obj->read_domains |= I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
182
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
184
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
185
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
198
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
211
GEM_BUG_ON(!i915_gem_object_get_stride(vma->obj) ||
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
212
!i915_gem_object_get_tiling(vma->obj));
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
227
fence->stride = i915_gem_object_get_stride(vma->obj);
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
228
fence->tiling = i915_gem_object_get_tiling(vma->obj);
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
367
struct i915_vma *set = i915_gem_object_is_tiled(vma->obj) ? vma : NULL;
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
428
if (!vma->fence && !i915_gem_object_is_tiled(vma->obj))
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
778
i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
785
if (obj->bit_17 == NULL)
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
792
if ((new_bit_17 & 0x1) != (test_bit(i, obj->bit_17) != 0)) {
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
811
i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
814
const unsigned int page_count = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
819
if (obj->bit_17 == NULL) {
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
820
obj->bit_17 = bitmap_zalloc(page_count, GFP_KERNEL);
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
821
if (obj->bit_17 == NULL) {
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
822
drm_err(obj->base.dev,
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
832
__set_bit(i, obj->bit_17);
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.c
834
__clear_bit(i, obj->bit_17);
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.h
49
void i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/intel_ggtt_fencing.h
51
void i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/intel_gsc.c
52
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gsc.c
55
obj = i915_gem_object_create_lmem(gt->i915, size,
sys/dev/pci/drm/i915/gt/intel_gsc.c
58
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_gsc.c
60
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/intel_gsc.c
63
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gt/intel_gsc.c
69
intf->gem_obj = obj;
sys/dev/pci/drm/i915/gt/intel_gsc.c
74
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_gsc.c
80
struct drm_i915_gem_object *obj = fetch_and_zero(&intf->gem_obj);
sys/dev/pci/drm/i915/gt/intel_gsc.c
82
if (!obj)
sys/dev/pci/drm/i915/gt/intel_gsc.c
85
if (i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gt/intel_gsc.c
86
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/gt/intel_gsc.c
88
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_gt.c
1100
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/intel_gt.c
1107
if (i915_gem_object_is_lmem(obj) || intel_gt_needs_wa_22016122933(gt))
sys/dev/pci/drm/i915/gt/intel_gt.c
466
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gt.c
470
obj = i915_gem_object_create_lmem(i915, size,
sys/dev/pci/drm/i915/gt/intel_gt.c
473
if (IS_ERR(obj) && !IS_METEORLAKE(i915)) /* Wa_22018444074 */
sys/dev/pci/drm/i915/gt/intel_gt.c
474
obj = i915_gem_object_create_stolen(i915, size);
sys/dev/pci/drm/i915/gt/intel_gt.c
475
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_gt.c
476
obj = i915_gem_object_create_internal(i915, size);
sys/dev/pci/drm/i915/gt/intel_gt.c
477
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_gt.c
479
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/intel_gt.c
482
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_gt.c
497
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_gt.c
599
state = shmem_create_from_object(rq->context->state->obj);
sys/dev/pci/drm/i915/gt/intel_gt.c
601
state = uao_create_from_object(rq->context->state->obj);
sys/dev/pci/drm/i915/gt/intel_gt.h
204
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
105
struct list_head *list = bucket_for_size(pool, node->obj->base.size);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
109
i915_gem_object_unpin_pages(node->obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
112
i915_gem_object_make_purgeable(node->obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
128
assert_object_held(node->obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
133
__i915_gem_object_pin_pages(node->obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
135
i915_gem_object_make_unshrinkable(node->obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
145
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
157
obj = i915_gem_object_create_internal(gt->i915, sz);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
158
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
161
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
164
i915_gem_object_set_readonly(obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
167
node->obj = obj;
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
187
if (node->obj->base.size < size)
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
33
i915_gem_object_put(node->obj);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool_types.h
24
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
104
type = intel_gt_coherent_map_type(vm->gt, obj, true);
sys/dev/pci/drm/i915/gt/intel_gtt.c
115
vaddr = i915_gem_object_pin_map_unlocked(obj, type);
sys/dev/pci/drm/i915/gt/intel_gtt.c
119
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
123
int map_pt_dma_locked(struct i915_address_space *vm, struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/intel_gtt.c
128
type = intel_gt_coherent_map_type(vm->gt, obj, true);
sys/dev/pci/drm/i915/gt/intel_gtt.c
139
vaddr = i915_gem_object_pin_map(obj, type);
sys/dev/pci/drm/i915/gt/intel_gtt.c
143
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
152
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
154
if (!i915_gem_object_get_rcu(obj)) {
sys/dev/pci/drm/i915/gt/intel_gtt.c
177
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
381
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
383
obj = vm->alloc_scratch_dma(vm, size);
sys/dev/pci/drm/i915/gt/intel_gtt.c
384
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_gtt.c
387
if (map_pt_dma(vm, obj))
sys/dev/pci/drm/i915/gt/intel_gtt.c
391
if (obj->mm.page_sizes.sg < size)
sys/dev/pci/drm/i915/gt/intel_gtt.c
395
if (__px_dma(obj) & (size - 1))
sys/dev/pci/drm/i915/gt/intel_gtt.c
407
poison_scratch_page(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
409
vm->scratch[0] = obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
414
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
43
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
57
obj = __i915_gem_object_create_lmem_with_ps(vm->i915, sz, sz,
sys/dev/pci/drm/i915/gt/intel_gtt.c
64
if (!IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_gtt.c
65
obj->base.resv = i915_vm_resv_get(vm);
sys/dev/pci/drm/i915/gt/intel_gtt.c
66
obj->shares_resv_from = vm;
sys/dev/pci/drm/i915/gt/intel_gtt.c
69
i915_drm_client_add_object(vm->fpriv->client, obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
697
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
700
obj = i915_gem_object_create_internal(vm->i915, PAGE_ALIGN(size));
sys/dev/pci/drm/i915/gt/intel_gtt.c
701
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_gtt.c
702
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
704
i915_gem_object_set_cache_coherency(obj, I915_CACHE_LLC);
sys/dev/pci/drm/i915/gt/intel_gtt.c
706
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/intel_gtt.c
708
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
72
return obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
77
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
82
obj = i915_gem_object_create_internal(vm->i915, sz);
sys/dev/pci/drm/i915/gt/intel_gtt.c
88
if (!IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_gtt.c
89
obj->base.resv = i915_vm_resv_get(vm);
sys/dev/pci/drm/i915/gt/intel_gtt.c
90
obj->shares_resv_from = vm;
sys/dev/pci/drm/i915/gt/intel_gtt.c
93
i915_drm_client_add_object(vm->fpriv->client, obj);
sys/dev/pci/drm/i915/gt/intel_gtt.c
96
return obj;
sys/dev/pci/drm/i915/gt/intel_gtt.c
99
int map_pt_dma(struct i915_address_space *vm, struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/intel_gtt.h
253
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_gtt.h
452
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/intel_gtt.h
454
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
sys/dev/pci/drm/i915/gt/intel_gtt.h
642
int map_pt_dma(struct i915_address_space *vm, struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gt/intel_gtt.h
643
int map_pt_dma_locked(struct i915_address_space *vm, struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1078
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_lrc.c
1098
obj = i915_gem_object_create_lmem(engine->i915, context_size,
sys/dev/pci/drm/i915/gt/intel_lrc.c
1100
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/intel_lrc.c
1101
obj = i915_gem_object_create_shmem(engine->i915, context_size);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1102
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_lrc.c
1103
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1111
i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1114
vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1116
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1203
*vaddr = i915_gem_object_pin_map(ce->state->obj,
sys/dev/pci/drm/i915/gt/intel_lrc.c
1205
ce->state->obj,
sys/dev/pci/drm/i915/gt/intel_lrc.c
1238
i915_gem_object_unpin_map(ce->state->obj);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1811
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_lrc.c
1815
obj = i915_gem_object_create_shmem(engine->i915, CTX_WA_BB_SIZE);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1816
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_lrc.c
1817
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1819
vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1829
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1882
err = i915_gem_object_lock(wa_ctx->vma->obj, &ww);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1888
batch = i915_gem_object_pin_map(wa_ctx->vma->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1913
__i915_gem_object_flush_map(wa_ctx->vma->obj, 0, batch_ptr - batch);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1914
__i915_gem_object_release_map(wa_ctx->vma->obj);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
135
__i915_gem_object_flush_map(so->vma->obj, 0, i * sizeof(u32));
sys/dev/pci/drm/i915/gt/intel_renderstate.c
136
__i915_gem_object_release_map(so->vma->obj);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
146
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gt/intel_renderstate.c
156
obj = i915_gem_object_create_internal(engine->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
157
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_renderstate.c
158
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
160
so->vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
177
err = i915_gem_object_lock(so->vma->obj, &so->ww);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
203
if (obj)
sys/dev/pci/drm/i915/gt/intel_renderstate.c
204
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
251
i915_gem_object_put(so->vma->obj);
sys/dev/pci/drm/i915/gt/intel_renderstate.c
58
d = i915_gem_object_pin_map(so->vma->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/intel_reset.c
873
struct drm_i915_private *dev_priv = vma->obj->base.dev->dev_private;
sys/dev/pci/drm/i915/gt/intel_ring.c
105
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/intel_ring.c
115
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_ring.c
118
obj = i915_gem_object_create_lmem(i915, size, I915_BO_ALLOC_VOLATILE |
sys/dev/pci/drm/i915/gt/intel_ring.c
120
if (IS_ERR(obj) && i915_ggtt_has_aperture(ggtt) && !HAS_LLC(i915))
sys/dev/pci/drm/i915/gt/intel_ring.c
121
obj = i915_gem_object_create_stolen(i915, size);
sys/dev/pci/drm/i915/gt/intel_ring.c
122
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_ring.c
123
obj = i915_gem_object_create_internal(i915, size);
sys/dev/pci/drm/i915/gt/intel_ring.c
124
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_ring.c
125
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_ring.c
132
i915_gem_object_set_readonly(obj);
sys/dev/pci/drm/i915/gt/intel_ring.c
134
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/intel_ring.c
141
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_ring.c
48
if (i915_gem_object_is_stolen(vma->obj))
sys/dev/pci/drm/i915/gt/intel_ring.c
60
int type = intel_gt_coherent_map_type(vma->vm->gt, vma->obj, false);
sys/dev/pci/drm/i915/gt/intel_ring.c
62
addr = i915_gem_object_pin_map(vma->obj, type);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1317
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1332
obj = i915_gem_object_create_internal(engine->i915, size);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1333
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1334
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1336
vma = i915_vma_instance(obj, engine->gt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1338
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1346
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1408
err = i915_gem_object_lock(timeline->hwsp_ggtt->obj, &ww);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1410
err = i915_gem_object_lock(gen7_wa_vma->obj, &ww);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1412
err = i915_gem_object_lock(engine->legacy.ring->vma->obj, &ww);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
1451
i915_gem_object_put(gen7_wa_vma->obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
506
struct drm_i915_gem_object *obj = ce->state->obj;
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
509
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
519
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
520
__i915_gem_object_release_map(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
569
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
573
obj = i915_gem_object_create_shmem(i915, engine->context_size);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
574
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
575
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
593
i915_gem_object_set_cache_coherency(obj, I915_CACHE_L3_LLC);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
595
vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
604
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
66
struct drm_i915_gem_object *obj = engine->status_page.vma->obj;
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
68
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gt/intel_ring_submission.c
69
return sg_page(obj->mm.pages->sgl);
sys/dev/pci/drm/i915/gt/intel_timeline.c
129
i915_gem_object_unpin_map(timeline->hwsp_ggtt->obj);
sys/dev/pci/drm/i915/gt/intel_timeline.c
22
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/intel_timeline.c
25
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/intel_timeline.c
26
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/intel_timeline.c
27
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/intel_timeline.c
29
i915_gem_object_set_cache_coherency(obj, I915_CACHE_LLC);
sys/dev/pci/drm/i915/gt/intel_timeline.c
31
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/intel_timeline.c
33
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/intel_timeline.c
60
struct drm_i915_gem_object *obj = timeline->hwsp_ggtt->obj;
sys/dev/pci/drm/i915/gt/intel_timeline.c
64
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/intel_workarounds.c
3050
err = i915_gem_object_lock(vma->obj, &ww);
sys/dev/pci/drm/i915/gt/intel_workarounds.c
3084
results = i915_gem_object_pin_map(vma->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/intel_workarounds.c
3099
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/mock_engine.c
20
if (WARN_ON(!i915_gem_object_trylock(tl->hwsp_ggtt->obj, NULL)))
sys/dev/pci/drm/i915/gt/mock_engine.c
24
i915_gem_object_unlock(tl->hwsp_ggtt->obj);
sys/dev/pci/drm/i915/gt/mock_engine.c
42
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/mock_engine.c
45
obj = i915_gem_object_create_internal(i915, size);
sys/dev/pci/drm/i915/gt/mock_engine.c
46
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/mock_engine.c
47
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/mock_engine.c
49
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/mock_engine.c
56
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1001
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1007
vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1042
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1044
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1287
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1304
obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1305
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1306
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1308
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1314
vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1404
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1406
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1530
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1555
obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1556
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1557
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1561
map = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1567
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1704
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1706
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2704
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2715
obj = i915_gem_object_create_internal(engine->i915, 4096);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2716
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2717
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2721
vma = i915_vma_instance(obj, ce->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2731
cs = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2758
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2759
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2779
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
2790
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3008
cs = i915_gem_object_pin_map_unlocked(rq->batch->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3011
i915_gem_object_unpin_map(rq->batch->obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3051
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3057
obj = i915_gem_object_create_internal(engine->i915, 4096);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3058
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3059
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3061
vma = i915_vma_instance(obj, result->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3063
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3073
cs = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3115
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3116
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3123
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3127
obj = i915_gem_object_create_internal(gt->i915, sz);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3128
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3129
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3131
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3133
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3160
vma = i915_vma_instance(global->obj, ce->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3274
result = i915_gem_object_pin_map_unlocked(global->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
4266
cs = i915_gem_object_pin_map_unlocked(scratch->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
4281
i915_gem_object_unpin_map(scratch->obj);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
979
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
997
obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
998
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
999
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
108
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
116
obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
117
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
119
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
122
vaddr = i915_gem_object_pin_map_unlocked(obj, intel_gt_coherent_map_type(gt, obj, false));
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
124
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
129
i915_gem_object_unpin_map(h->obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
130
i915_gem_object_put(h->obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
132
h->obj = obj;
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
135
vma = i915_vma_instance(h->obj, vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1402
err = i915_gem_object_set_tiling(arg->vma->obj, I915_TILING_Y, 512);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1432
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1456
obj = i915_gem_object_create_internal(gt->i915, SZ_1M);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1457
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1458
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1464
err = i915_gem_object_set_tiling(obj, I915_TILING_X, 512);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1471
arg.vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1572
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
263
i915_gem_object_unpin_map(h->obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
264
i915_gem_object_put(h->obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
35
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
61
h->obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
62
if (IS_ERR(h->obj)) {
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
63
err = PTR_ERR(h->obj);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
75
vaddr = i915_gem_object_pin_map_unlocked(h->obj,
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
76
intel_gt_coherent_map_type(gt, h->obj, false));
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
88
i915_gem_object_put(h->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1053
i915_gem_object_flush_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1054
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1154
cs = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1162
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1211
i915_gem_object_flush_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1212
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1281
A[0] = i915_gem_object_pin_map_unlocked(ref[0]->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1285
A[1] = i915_gem_object_pin_map_unlocked(ref[1]->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1291
B[0] = i915_gem_object_pin_map_unlocked(result[0]->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1297
B[1] = i915_gem_object_pin_map_unlocked(result[1]->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1303
lrc = i915_gem_object_pin_map_unlocked(ce->state->obj,
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1305
ce->state->obj,
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1374
i915_gem_object_unpin_map(ce->state->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1376
i915_gem_object_unpin_map(result[1]->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1378
i915_gem_object_unpin_map(result[0]->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1380
i915_gem_object_unpin_map(ref[1]->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1382
i915_gem_object_unpin_map(ref[0]->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1397
ptr = i915_gem_object_pin_map_unlocked(vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1404
i915_gem_object_flush_map(vma->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
1405
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
425
err = i915_gem_object_lock(scratch->obj, &ww);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
471
cs = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
486
i915_gem_object_unpin_map(scratch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
669
cs = i915_gem_object_pin_map_unlocked(scratch->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
686
i915_gem_object_unpin_map(scratch->obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
936
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_lrc.c
940
obj = i915_gem_object_create_internal(vm->i915, size);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
941
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_lrc.c
942
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
944
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
946
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
952
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
983
cs = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
991
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
258
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
264
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
27
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
272
obj = create_lmem_or_internal(i915, sz);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
273
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_migrate.c
277
sz = obj->base.size;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
279
if (HAS_FLAT_CCS(i915) && i915_gem_object_is_lmem(obj))
sys/dev/pci/drm/i915/gt/selftest_migrate.c
285
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
289
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
29
obj = i915_gem_object_create_lmem(i915, size, 0);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
297
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
30
if (!IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_migrate.c
302
obj->mm.pages->sgl,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
303
obj->pat_index,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
31
return obj;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
318
err = fn(migrate, &ww, obj, val, &rq);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
330
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
350
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
353
obj->mm.pages->sgl,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
354
obj->pat_index,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
370
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
392
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
403
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
406
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
452
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
457
obj->mm.pages->sgl,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
458
obj->pat_index,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
459
i915_gem_object_is_lmem(obj),
sys/dev/pci/drm/i915/gt/selftest_migrate.c
465
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
470
obj->mm.pages->sgl,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
471
obj->pat_index,
sys/dev/pci/drm/i915/gt/selftest_migrate.c
472
i915_gem_object_is_lmem(obj),
sys/dev/pci/drm/i915/gt/selftest_migrate.c
551
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
568
obj = i915_gem_object_create_internal(i915, 2 * PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
569
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/selftest_migrate.c
570
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
574
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
650
it = sg_sgt(obj->mm.pages->sgl);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
651
len = emit_pte(rq, &it, obj->pat_index, false, 0, CHUNK_SZ);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
670
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
813
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
817
obj = i915_gem_object_create_lmem(gt->i915, sz, 0);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
819
if (IS_ERR_OR_NULL(obj)) {
sys/dev/pci/drm/i915/gt/selftest_migrate.c
820
obj = i915_gem_object_create_internal(gt->i915, sz);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
821
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_migrate.c
822
return obj;
sys/dev/pci/drm/i915/gt/selftest_migrate.c
825
i915_gem_object_trylock(obj, NULL);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
826
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
828
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
829
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_migrate.c
833
return obj;
sys/dev/pci/drm/i915/gt/selftest_mocs.c
84
arg->vaddr = i915_gem_object_pin_map_unlocked(arg->scratch->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
11
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
16
obj = i915_gem_object_create_internal(engine->i915, 4096);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
17
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
18
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
20
vma = i915_vma_instance(obj, engine->gt->vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
215
result = i915_gem_object_pin_map_unlocked(bb->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
22
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
28
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
34
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
38
cs = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
40
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
58
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
59
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
64
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
136
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
147
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
69
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_rps.c
721
i915_gem_object_flush_map(vma->obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
722
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
76
obj = i915_gem_object_create_internal(vm->i915, 64 << 10);
sys/dev/pci/drm/i915/gt/selftest_rps.c
77
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_rps.c
78
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
80
end = obj->base.size / sizeof(u32) - 1;
sys/dev/pci/drm/i915/gt/selftest_rps.c
82
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_rps.c
859
i915_gem_object_flush_map(vma->obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
860
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/selftest_rps.c
94
base = i915_gem_object_pin_map(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
29
struct drm_i915_gem_object *obj = tl->hwsp_ggtt->obj;
sys/dev/pci/drm/i915/gt/selftest_timeline.c
31
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gt/selftest_timeline.c
32
return sg_page(obj->mm.pages->sgl);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
49
err = i915_gem_object_lock(tl->hwsp_ggtt->obj, &ww);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
831
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_timeline.c
834
obj = i915_gem_object_create_internal(gt->i915, SZ_2M);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
835
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_timeline.c
836
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
839
i915_gem_object_set_pat_index(obj, tl->hwsp_ggtt->obj->pat_index);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
840
w->map = i915_gem_object_pin_map_unlocked(obj,
sys/dev/pci/drm/i915/gt/selftest_timeline.c
841
page_unmask_bits(tl->hwsp_ggtt->obj->mm.mapping));
sys/dev/pci/drm/i915/gt/selftest_timeline.c
843
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
847
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, 0);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
849
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_tlb.c
151
ce->engine->name, va->obj->mm.region->name ?: "smem",
sys/dev/pci/drm/i915/gt/selftest_tlb.c
158
.bi.pages = vb->obj->mm.pages,
sys/dev/pci/drm/i915/gt/selftest_tlb.c
159
.bi.page_sizes = vb->obj->mm.page_sizes,
sys/dev/pci/drm/i915/gt/selftest_tlb.c
166
if (i915_gem_object_is_lmem(vb->obj))
sys/dev/pci/drm/i915/gt/selftest_tlb.c
27
memset64(page_mask_bits(vma->obj->mm.mapping) +
sys/dev/pci/drm/i915/gt/selftest_tlb.c
320
if (BIT_ULL(bit) < i915_vm_obj_min_alignment(va->vm, va->obj))
sys/dev/pci/drm/i915/gt/selftest_tlb.c
95
ce->engine->name, va->obj->mm.region->name ?: "smem",
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
1029
a = i915_gem_object_pin_map_unlocked(A->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
1033
b = i915_gem_object_pin_map_unlocked(B->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
1051
i915_gem_object_unpin_map(B->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
1053
i915_gem_object_unpin_map(A->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
380
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
384
obj = i915_gem_object_create_internal(vm->i915, 16 * PAGE_SIZE);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
385
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
386
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
388
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
401
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
540
err = i915_gem_object_lock(scratch->obj, &ww);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
542
err = i915_gem_object_lock(batch->obj, &ww);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
548
cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
554
results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
614
i915_gem_object_flush_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
615
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
730
i915_gem_object_unpin_map(scratch->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
733
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
898
cs = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
919
i915_gem_object_flush_map(batch->obj);
sys/dev/pci/drm/i915/gt/selftest_workarounds.c
945
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/gt/shmem_utils.c
196
uao_create_from_object(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/shmem_utils.c
202
if (i915_gem_object_is_shmem(obj)) {
sys/dev/pci/drm/i915/gt/shmem_utils.c
203
uao_reference(obj->base.uao);
sys/dev/pci/drm/i915/gt/shmem_utils.c
204
return obj->base.uao;
sys/dev/pci/drm/i915/gt/shmem_utils.c
207
map_type = i915_gem_object_is_lmem(obj) ? I915_MAP_WC : I915_MAP_WB;
sys/dev/pci/drm/i915/gt/shmem_utils.c
208
ptr = i915_gem_object_pin_map_unlocked(obj, map_type);
sys/dev/pci/drm/i915/gt/shmem_utils.c
212
uao = uao_create_from_data("", ptr, obj->base.size);
sys/dev/pci/drm/i915/gt/shmem_utils.c
213
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/shmem_utils.c
37
struct file *shmem_create_from_object(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/shmem_utils.c
43
if (i915_gem_object_is_shmem(obj)) {
sys/dev/pci/drm/i915/gt/shmem_utils.c
44
file = obj->base.filp;
sys/dev/pci/drm/i915/gt/shmem_utils.c
49
map_type = i915_gem_object_is_lmem(obj) ? I915_MAP_WC : I915_MAP_WB;
sys/dev/pci/drm/i915/gt/shmem_utils.c
50
ptr = i915_gem_object_pin_map_unlocked(obj, map_type);
sys/dev/pci/drm/i915/gt/shmem_utils.c
54
file = shmem_create_from_data("", ptr, obj->base.size);
sys/dev/pci/drm/i915/gt/shmem_utils.c
55
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/shmem_utils.h
17
struct file *shmem_create_from_object(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_fw.c
338
src = i915_gem_object_pin_map_unlocked(gsc->fw.obj,
sys/dev/pci/drm/i915/gt/uc/intel_gsc_fw.c
339
intel_gt_coherent_map_type(gt, gsc->fw.obj, true));
sys/dev/pci/drm/i915/gt/uc/intel_gsc_fw.c
348
i915_gem_object_unpin_map(gsc->fw.obj);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
149
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
169
obj = i915_gem_object_create_stolen(gt->i915, size);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
170
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
171
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
173
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, 0);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
194
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
207
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc_heci_cmd_submit.c
150
err = i915_gem_object_lock(pkt->bb_vma->obj, &ww);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc_heci_cmd_submit.c
153
err = i915_gem_object_lock(pkt->heci_pkt_vma->obj, &ww);
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
780
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
786
obj = i915_gem_object_create_lmem(gt->i915, size,
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
791
obj = i915_gem_object_create_shmem(gt->i915, size);
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
793
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
794
return ERR_CAST(obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
802
i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE);
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
804
vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
818
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
844
vaddr = i915_gem_object_pin_map_unlocked(vma->obj,
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
846
vma->obj, true));
sys/dev/pci/drm/i915/gt/uc/intel_guc.c
955
if (i915_gem_object_is_lmem(guc->ct.vma->obj)) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_ads.c
1004
if (i915_gem_object_is_lmem(guc->ads_vma->obj))
sys/dev/pci/drm/i915/gt/uc/intel_guc_ads.c
951
i915_gem_object_flush_map(guc->ads_vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
342
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(ct->vma->obj));
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
346
blob = __px_vaddr(ct->vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
304
relay_reserve(log->relay.channel, log->vma->obj->base.size -
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
511
i915_gem_object_get(log->vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
521
i915_gem_object_put(log->vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
660
vaddr = i915_gem_object_pin_map_unlocked(log->vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
909
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
918
obj = uc->load_err_log;
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
920
obj = guc->log.vma->obj;
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
922
if (!obj)
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
931
map = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
939
for (i = 0; i < obj->base.size; i += PAGE_SIZE) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
951
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc_log_debugfs.c
15
static u32 obj_to_guc_log_dump_size(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gt/uc/intel_guc_log_debugfs.c
19
if (!obj)
sys/dev/pci/drm/i915/gt/uc/intel_guc_log_debugfs.c
23
size = ((obj->base.size * 11) + 3) / 4;
sys/dev/pci/drm/i915/gt/uc/intel_guc_log_debugfs.c
41
return obj_to_guc_log_dump_size(log->vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2809
GEM_BUG_ON(i915_gem_object_is_lmem(guc->ct.vma->obj) !=
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2810
i915_gem_object_is_lmem(ce->ring->vma->obj));
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2876
GEM_BUG_ON(i915_gem_object_is_lmem(guc->ct.vma->obj) !=
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2877
i915_gem_object_is_lmem(ce->ring->vma->obj));
sys/dev/pci/drm/i915/gt/uc/intel_huc_fw.c
105
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gt/uc/intel_huc_fw.c
30
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/uc/intel_huc_fw.c
41
obj = huc->heci_pkt->obj;
sys/dev/pci/drm/i915/gt/uc/intel_huc_fw.c
44
pkt_vaddr = i915_gem_object_pin_map_unlocked(obj,
sys/dev/pci/drm/i915/gt/uc/intel_huc_fw.c
45
intel_gt_coherent_map_type(gt, obj, true));
sys/dev/pci/drm/i915/gt/uc/intel_huc_fw.c
62
msg_in->huc_in.huc_size = huc->fw.obj->base.size;
sys/dev/pci/drm/i915/gt/uc/intel_uc.c
159
uc->load_err_log = i915_gem_object_get(guc->log.vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1036
GEM_BUG_ON(offset + uc_fw->obj->base.size > node->size);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1037
GEM_BUG_ON(uc_fw->obj->base.size > INTEL_UC_RSVD_GGTT_PER_FW);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1044
struct drm_i915_gem_object *obj = uc_fw->obj;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1053
vma_res->node_size = obj->base.size;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1054
vma_res->bi.pages = obj->mm.pages;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1056
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1059
if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1062
if (i915_gem_object_is_lmem(obj))
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1226
vaddr = i915_gem_object_pin_map_unlocked(vma->obj,
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1227
intel_gt_coherent_map_type(gt, vma->obj, true));
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1235
i915_gem_object_unpin_map(vma->obj);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1266
err = i915_gem_object_pin_pages_unlocked(uc_fw->obj);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1285
i915_gem_object_unpin_pages(uc_fw->obj);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1295
if (i915_gem_object_has_pinned_pages(uc_fw->obj))
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1296
i915_gem_object_unpin_pages(uc_fw->obj);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1306
if (!i915_gem_object_has_pinned_pages(uc_fw->obj))
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1323
i915_gem_object_put(fetch_and_zero(&uc_fw->obj));
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1339
struct intel_memory_region *mr = uc_fw->obj->mm.region;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1351
if (i915_gem_object_has_struct_page(uc_fw->obj)) {
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1354
for_each_sgt_page(page, iter, uc_fw->obj->mm.pages) {
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
1374
for_each_sgt_daddr(addr, iter, uc_fw->obj->mm.pages) {
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
900
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
979
obj = i915_gem_object_create_lmem_from_data(i915, fw->data, fw->size);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
980
if (!IS_ERR(obj))
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
981
obj->flags |= I915_BO_ALLOC_PM_EARLY;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
983
obj = i915_gem_object_create_shmem_from_data(i915, fw->data, fw->size);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
986
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
987
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.c
991
uc_fw->obj = obj;
sys/dev/pci/drm/i915/gt/uc/intel_uc_fw.h
100
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1941
bb->obj = i915_gem_object_create_shmem(s->engine->i915,
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1944
if (IS_ERR(bb->obj)) {
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1945
ret = PTR_ERR(bb->obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1949
bb->va = i915_gem_object_pin_map(bb->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1968
i915_gem_object_unlock(bb->obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1991
i915_gem_object_unpin_map(bb->obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
1993
i915_gem_object_put(bb->obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3009
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3013
obj = i915_gem_object_create_shmem(workload->engine->i915,
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3016
if (IS_ERR(obj))
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3017
return PTR_ERR(obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3020
map = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3027
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3028
ret = i915_gem_object_set_to_cpu_domain(obj, false);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3029
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3044
wa_ctx->indirect_ctx.obj = obj;
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3049
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/gvt/cmd_parser.c
3051
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
112
static void vgpu_gem_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/gvt/dmabuf.c
117
if (obj->base.dma_buf) {
sys/dev/pci/drm/i915/gvt/dmabuf.c
118
struct intel_vgpu_fb_info *fb_info = obj->gvt_info;
sys/dev/pci/drm/i915/gvt/dmabuf.c
119
struct intel_vgpu_dmabuf_obj *obj = fb_info->obj;
sys/dev/pci/drm/i915/gvt/dmabuf.c
120
struct intel_vgpu *vgpu = obj->vgpu;
sys/dev/pci/drm/i915/gvt/dmabuf.c
134
struct intel_vgpu_dmabuf_obj *obj =
sys/dev/pci/drm/i915/gvt/dmabuf.c
136
struct intel_vgpu *vgpu = obj->vgpu;
sys/dev/pci/drm/i915/gvt/dmabuf.c
144
if (dmabuf_obj == obj) {
sys/dev/pci/drm/i915/gvt/dmabuf.c
155
kfree(obj->info);
sys/dev/pci/drm/i915/gvt/dmabuf.c
156
kfree(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
161
static inline void dmabuf_obj_get(struct intel_vgpu_dmabuf_obj *obj)
sys/dev/pci/drm/i915/gvt/dmabuf.c
163
kref_get(&obj->kref);
sys/dev/pci/drm/i915/gvt/dmabuf.c
166
static inline void dmabuf_obj_put(struct intel_vgpu_dmabuf_obj *obj)
sys/dev/pci/drm/i915/gvt/dmabuf.c
168
kref_put(&obj->kref, dmabuf_gem_object_free);
sys/dev/pci/drm/i915/gvt/dmabuf.c
175
struct intel_vgpu_dmabuf_obj *obj = fb_info->obj;
sys/dev/pci/drm/i915/gvt/dmabuf.c
176
struct intel_vgpu *vgpu = obj->vgpu;
sys/dev/pci/drm/i915/gvt/dmabuf.c
181
dmabuf_obj_put(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
186
dmabuf_obj_put(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
203
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gvt/dmabuf.c
205
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/gvt/dmabuf.c
206
if (obj == NULL)
sys/dev/pci/drm/i915/gvt/dmabuf.c
209
drm_gem_private_object_init(dev, &obj->base,
sys/dev/pci/drm/i915/gvt/dmabuf.c
211
i915_gem_object_init(obj, &intel_vgpu_gem_ops, &lock_class, 0);
sys/dev/pci/drm/i915/gvt/dmabuf.c
212
i915_gem_object_set_readonly(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
214
obj->read_domains = I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/gvt/dmabuf.c
215
obj->write_domain = 0;
sys/dev/pci/drm/i915/gvt/dmabuf.c
237
obj->tiling_and_stride = tiling_mode | stride;
sys/dev/pci/drm/i915/gvt/dmabuf.c
239
obj->tiling_and_stride = info->drm_format_mod ?
sys/dev/pci/drm/i915/gvt/dmabuf.c
243
return obj;
sys/dev/pci/drm/i915/gvt/dmabuf.c
46
static int vgpu_gem_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/gvt/dmabuf.c
465
((struct intel_vgpu_fb_info *)dmabuf_obj->info)->obj = dmabuf_obj;
sys/dev/pci/drm/i915/gvt/dmabuf.c
48
struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/gvt/dmabuf.c
505
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gvt/dmabuf.c
519
obj = vgpu_create_gem(dev, dmabuf_obj->info);
sys/dev/pci/drm/i915/gvt/dmabuf.c
520
if (obj == NULL) {
sys/dev/pci/drm/i915/gvt/dmabuf.c
526
obj->gvt_info = dmabuf_obj->info;
sys/dev/pci/drm/i915/gvt/dmabuf.c
528
dmabuf = i915_gem_prime_export(&obj->base, DRM_CLOEXEC | DRM_RDWR);
sys/dev/pci/drm/i915/gvt/dmabuf.c
557
kref_read(&obj->base.refcount));
sys/dev/pci/drm/i915/gvt/dmabuf.c
559
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
566
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/gvt/dmabuf.c
57
if (overflows_type(obj->base.size >> PAGE_SHIFT, page_num))
sys/dev/pci/drm/i915/gvt/dmabuf.c
60
page_num = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gvt/dmabuf.c
61
fb_info = (struct intel_vgpu_fb_info *)obj->gvt_info;
sys/dev/pci/drm/i915/gvt/dmabuf.c
65
vgpu = fb_info->obj->vgpu;
sys/dev/pci/drm/i915/gvt/dmabuf.c
94
__i915_gem_object_set_pages(obj, st);
sys/dev/pci/drm/i915/gvt/dmabuf.h
48
struct intel_vgpu_dmabuf_obj *obj;
sys/dev/pci/drm/i915/gvt/scheduler.c
414
if (!wa_ctx->indirect_ctx.obj)
sys/dev/pci/drm/i915/gvt/scheduler.c
417
i915_gem_object_lock(wa_ctx->indirect_ctx.obj, NULL);
sys/dev/pci/drm/i915/gvt/scheduler.c
418
i915_gem_object_unpin_map(wa_ctx->indirect_ctx.obj);
sys/dev/pci/drm/i915/gvt/scheduler.c
419
i915_gem_object_unlock(wa_ctx->indirect_ctx.obj);
sys/dev/pci/drm/i915/gvt/scheduler.c
420
i915_gem_object_put(wa_ctx->indirect_ctx.obj);
sys/dev/pci/drm/i915/gvt/scheduler.c
422
wa_ctx->indirect_ctx.obj = NULL;
sys/dev/pci/drm/i915/gvt/scheduler.c
555
i915_gem_object_lock(bb->obj, &ww);
sys/dev/pci/drm/i915/gvt/scheduler.c
557
bb->vma = i915_gem_object_ggtt_pin_ww(bb->obj, &ww,
sys/dev/pci/drm/i915/gvt/scheduler.c
580
i915_gem_object_flush_map(bb->obj);
sys/dev/pci/drm/i915/gvt/scheduler.c
621
i915_gem_object_lock(wa_ctx->indirect_ctx.obj, &ww);
sys/dev/pci/drm/i915/gvt/scheduler.c
623
vma = i915_gem_object_ggtt_pin_ww(wa_ctx->indirect_ctx.obj, &ww, NULL,
sys/dev/pci/drm/i915/gvt/scheduler.c
668
if (bb->obj) {
sys/dev/pci/drm/i915/gvt/scheduler.c
669
i915_gem_object_lock(bb->obj, NULL);
sys/dev/pci/drm/i915/gvt/scheduler.c
671
i915_gem_object_unpin_map(bb->obj);
sys/dev/pci/drm/i915/gvt/scheduler.c
676
i915_gem_object_unlock(bb->obj);
sys/dev/pci/drm/i915/gvt/scheduler.c
677
i915_gem_object_put(bb->obj);
sys/dev/pci/drm/i915/gvt/scheduler.h
126
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/gvt/scheduler.h
65
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_cmd_parser.c
1466
cmd = copy_batch(shadow->obj, batch->obj,
sys/dev/pci/drm/i915/i915_cmd_parser.c
1556
cmd = page_mask_bits(shadow->obj->mm.mapping);
sys/dev/pci/drm/i915/i915_cmd_parser.c
1577
i915_gem_object_flush_map(shadow->obj);
sys/dev/pci/drm/i915/i915_cmd_parser.c
1581
i915_gem_object_unpin_map(shadow->obj);
sys/dev/pci/drm/i915/i915_debugfs.c
141
static const char *i915_cache_level_str(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_debugfs.c
143
struct drm_i915_private *i915 = obj_to_i915(obj);
sys/dev/pci/drm/i915/i915_debugfs.c
146
switch (obj->pat_index) {
sys/dev/pci/drm/i915/i915_debugfs.c
155
switch (obj->pat_index) {
sys/dev/pci/drm/i915/i915_debugfs.c
163
switch (obj->pat_index) {
sys/dev/pci/drm/i915/i915_debugfs.c
175
i915_debugfs_describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_debugfs.c
181
&obj->base,
sys/dev/pci/drm/i915/i915_debugfs.c
182
get_tiling_flag(obj),
sys/dev/pci/drm/i915/i915_debugfs.c
183
get_global_flag(obj),
sys/dev/pci/drm/i915/i915_debugfs.c
184
get_pin_mapped_flag(obj),
sys/dev/pci/drm/i915/i915_debugfs.c
185
obj->base.size / 1024,
sys/dev/pci/drm/i915/i915_debugfs.c
186
obj->read_domains,
sys/dev/pci/drm/i915/i915_debugfs.c
187
obj->write_domain,
sys/dev/pci/drm/i915/i915_debugfs.c
188
i915_cache_level_str(obj),
sys/dev/pci/drm/i915/i915_debugfs.c
189
obj->mm.dirty ? " dirty" : "",
sys/dev/pci/drm/i915/i915_debugfs.c
190
obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : "");
sys/dev/pci/drm/i915/i915_debugfs.c
191
if (obj->base.name)
sys/dev/pci/drm/i915/i915_debugfs.c
192
seq_printf(m, " (name: %d)", obj->base.name);
sys/dev/pci/drm/i915/i915_debugfs.c
194
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_debugfs.c
195
list_for_each_entry(vma, &obj->vma.list, obj_link) {
sys/dev/pci/drm/i915/i915_debugfs.c
199
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_debugfs.c
258
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_debugfs.c
260
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_debugfs.c
263
if (i915_gem_object_is_stolen(obj))
sys/dev/pci/drm/i915/i915_debugfs.c
264
seq_printf(m, " (stolen: %08llx)", obj->stolen->start);
sys/dev/pci/drm/i915/i915_debugfs.c
265
if (i915_gem_object_is_framebuffer(obj))
sys/dev/pci/drm/i915/i915_debugfs.c
80
static char get_tiling_flag(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_debugfs.c
82
switch (i915_gem_object_get_tiling(obj)) {
sys/dev/pci/drm/i915/i915_debugfs.c
90
static char get_global_flag(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_debugfs.c
92
return READ_ONCE(obj->userfault_count) ? 'g' : ' ';
sys/dev/pci/drm/i915/i915_debugfs.c
95
static char get_pin_mapped_flag(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_debugfs.c
97
return obj->mm.mapping ? 'M' : ' ';
sys/dev/pci/drm/i915/i915_debugfs.h
16
void i915_debugfs_describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/i915_debugfs.h
19
static inline void i915_debugfs_describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) {}
sys/dev/pci/drm/i915/i915_drm_client.c
182
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_drm_client.c
186
GEM_WARN_ON(obj->client);
sys/dev/pci/drm/i915/i915_drm_client.c
187
GEM_WARN_ON(!list_empty(&obj->client_link));
sys/dev/pci/drm/i915/i915_drm_client.c
190
obj->client = i915_drm_client_get(client);
sys/dev/pci/drm/i915/i915_drm_client.c
191
list_add_tail_rcu(&obj->client_link, &client->objects_list);
sys/dev/pci/drm/i915/i915_drm_client.c
195
void i915_drm_client_remove_object(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_drm_client.c
197
struct i915_drm_client *client = fetch_and_zero(&obj->client);
sys/dev/pci/drm/i915/i915_drm_client.c
205
list_del_rcu(&obj->client_link);
sys/dev/pci/drm/i915/i915_drm_client.c
215
i915_drm_client_add_object(client, ce->state->obj);
sys/dev/pci/drm/i915/i915_drm_client.c
218
i915_drm_client_add_object(client, ce->ring->vma->obj);
sys/dev/pci/drm/i915/i915_drm_client.c
49
obj_meminfo(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_drm_client.c
52
const enum intel_region_id id = obj->mm.region ?
sys/dev/pci/drm/i915/i915_drm_client.c
53
obj->mm.region->id : INTEL_REGION_SMEM;
sys/dev/pci/drm/i915/i915_drm_client.c
54
const u64 sz = obj->base.size;
sys/dev/pci/drm/i915/i915_drm_client.c
56
if (drm_gem_object_is_shared_for_memory_stats(&obj->base))
sys/dev/pci/drm/i915/i915_drm_client.c
61
if (i915_gem_object_has_pages(obj)) {
sys/dev/pci/drm/i915/i915_drm_client.c
64
if (!dma_resv_test_signaled(obj->base.resv,
sys/dev/pci/drm/i915/i915_drm_client.c
67
else if (i915_gem_object_is_shrinkable(obj) &&
sys/dev/pci/drm/i915/i915_drm_client.c
68
obj->mm.madv == I915_MADV_DONTNEED)
sys/dev/pci/drm/i915/i915_drm_client.c
79
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_drm_client.c
86
idr_for_each_entry(&file->object_idr, obj, id)
sys/dev/pci/drm/i915/i915_drm_client.c
87
obj_meminfo(obj, stats);
sys/dev/pci/drm/i915/i915_drm_client.c
93
obj = i915_gem_object_get_rcu(list_entry(pos, typeof(*obj),
sys/dev/pci/drm/i915/i915_drm_client.c
95
if (!obj)
sys/dev/pci/drm/i915/i915_drm_client.c
97
obj_meminfo(obj, stats);
sys/dev/pci/drm/i915/i915_drm_client.c
98
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_drm_client.h
70
struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/i915_drm_client.h
71
void i915_drm_client_remove_object(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/i915_drm_client.h
76
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_drm_client.h
81
i915_drm_client_remove_object(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_gem.c
1005
i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
1014
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/i915_gem.c
1018
ret = i915_gem_object_ggtt_pin_ww(obj, &ww, view, size,
sys/dev/pci/drm/i915/i915_gem.c
1033
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_gem.c
1044
obj = i915_gem_object_lookup(file_priv, args->handle);
sys/dev/pci/drm/i915/i915_gem.c
1045
if (!obj)
sys/dev/pci/drm/i915/i915_gem.c
1048
err = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/i915_gem.c
1052
if (i915_gem_object_has_pages(obj) &&
sys/dev/pci/drm/i915/i915_gem.c
1053
i915_gem_object_is_tiled(obj) &&
sys/dev/pci/drm/i915/i915_gem.c
1055
if (obj->mm.madv == I915_MADV_WILLNEED) {
sys/dev/pci/drm/i915/i915_gem.c
1056
GEM_BUG_ON(!i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/i915_gem.c
1057
i915_gem_object_clear_tiling_quirk(obj);
sys/dev/pci/drm/i915/i915_gem.c
1058
i915_gem_object_make_shrinkable(obj);
sys/dev/pci/drm/i915/i915_gem.c
1061
GEM_BUG_ON(i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/i915_gem.c
1062
i915_gem_object_make_unshrinkable(obj);
sys/dev/pci/drm/i915/i915_gem.c
1063
i915_gem_object_set_tiling_quirk(obj);
sys/dev/pci/drm/i915/i915_gem.c
1067
if (obj->mm.madv != __I915_MADV_PURGED) {
sys/dev/pci/drm/i915/i915_gem.c
1068
obj->mm.madv = args->madv;
sys/dev/pci/drm/i915/i915_gem.c
1069
if (obj->ops->adjust_lru)
sys/dev/pci/drm/i915/i915_gem.c
1070
obj->ops->adjust_lru(obj);
sys/dev/pci/drm/i915/i915_gem.c
1073
if (i915_gem_object_has_pages(obj) ||
sys/dev/pci/drm/i915/i915_gem.c
1074
i915_gem_object_has_self_managed_shrink_list(obj)) {
sys/dev/pci/drm/i915/i915_gem.c
1078
if (!list_empty(&obj->mm.link)) {
sys/dev/pci/drm/i915/i915_gem.c
1081
if (obj->mm.madv != I915_MADV_WILLNEED)
sys/dev/pci/drm/i915/i915_gem.c
1085
list_move_tail(&obj->mm.link, list);
sys/dev/pci/drm/i915/i915_gem.c
1092
if (obj->mm.madv == I915_MADV_DONTNEED &&
sys/dev/pci/drm/i915/i915_gem.c
1093
!i915_gem_object_has_pages(obj))
sys/dev/pci/drm/i915/i915_gem.c
1094
i915_gem_object_truncate(obj);
sys/dev/pci/drm/i915/i915_gem.c
1096
args->retained = obj->mm.madv != __I915_MADV_PURGED;
sys/dev/pci/drm/i915/i915_gem.c
1098
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem.c
1100
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_gem.c
114
int i915_gem_object_unbind(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
117
struct intel_runtime_pm *rpm = &to_i915(obj->base.dev)->runtime_pm;
sys/dev/pci/drm/i915/i915_gem.c
124
assert_object_held(obj);
sys/dev/pci/drm/i915/i915_gem.c
126
if (list_empty(&obj->vma.list))
sys/dev/pci/drm/i915/i915_gem.c
139
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_gem.c
140
while (!ret && (vma = list_first_entry_or_null(&obj->vma.list,
sys/dev/pci/drm/i915/i915_gem.c
163
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_gem.c
173
assert_object_held(vma->obj);
sys/dev/pci/drm/i915/i915_gem.c
190
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_gem.c
192
list_splice_init(&still_in_list, &obj->vma.list);
sys/dev/pci/drm/i915/i915_gem.c
193
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_gem.c
225
i915_gem_shmem_pread(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
235
ret = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/i915_gem.c
239
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
243
ret = i915_gem_object_prepare_read(obj, &needs_clflush);
sys/dev/pci/drm/i915/i915_gem.c
247
i915_gem_object_finish_access(obj);
sys/dev/pci/drm/i915/i915_gem.c
248
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem.c
254
struct vm_page *page = i915_gem_object_get_page(obj, idx);
sys/dev/pci/drm/i915/i915_gem.c
267
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
271
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
273
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem.c
301
static struct i915_vma *i915_gem_gtt_prepare(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
305
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_gem.c
314
ret = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/i915_gem.c
318
ret = i915_gem_object_set_to_gtt_domain(obj, write);
sys/dev/pci/drm/i915/i915_gem.c
322
if (!i915_gem_object_is_tiled(obj))
sys/dev/pci/drm/i915/i915_gem.c
323
vma = i915_gem_object_ggtt_pin_ww(obj, &ww, NULL, 0, 0,
sys/dev/pci/drm/i915/i915_gem.c
341
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
362
static void i915_gem_gtt_cleanup(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
366
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_gem.c
369
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
379
i915_gem_gtt_pread(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
382
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_gem.c
397
vma = i915_gem_gtt_prepare(obj, &node, false);
sys/dev/pci/drm/i915/i915_gem.c
420
i915_gem_object_get_dma_address(obj,
sys/dev/pci/drm/i915/i915_gem.c
440
i915_gem_gtt_cleanup(obj, &node, vma);
sys/dev/pci/drm/i915/i915_gem.c
460
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_gem.c
476
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/i915_gem.c
477
if (!obj)
sys/dev/pci/drm/i915/i915_gem.c
481
if (range_overflows_t(u64, args->offset, args->size, obj->base.size)) {
sys/dev/pci/drm/i915/i915_gem.c
486
trace_i915_gem_object_pread(obj, args->offset, args->size);
sys/dev/pci/drm/i915/i915_gem.c
488
if (obj->ops->pread)
sys/dev/pci/drm/i915/i915_gem.c
489
ret = obj->ops->pread(obj, args);
sys/dev/pci/drm/i915/i915_gem.c
493
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/i915_gem.c
499
ret = i915_gem_shmem_pread(obj, args);
sys/dev/pci/drm/i915/i915_gem.c
501
ret = i915_gem_gtt_pread(obj, args);
sys/dev/pci/drm/i915/i915_gem.c
504
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_gem.c
542
i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
545
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_gem.c
559
if (i915_gem_object_has_struct_page(obj)) {
sys/dev/pci/drm/i915/i915_gem.c
575
vma = i915_gem_gtt_prepare(obj, &node, true);
sys/dev/pci/drm/i915/i915_gem.c
581
i915_gem_object_invalidate_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/i915_gem.c
601
i915_gem_object_get_dma_address(obj,
sys/dev/pci/drm/i915/i915_gem.c
628
i915_gem_object_flush_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/i915_gem.c
630
i915_gem_gtt_cleanup(obj, &node, vma);
sys/dev/pci/drm/i915/i915_gem.c
664
i915_gem_shmem_pwrite(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
675
ret = i915_gem_object_lock_interruptible(obj, NULL);
sys/dev/pci/drm/i915/i915_gem.c
679
ret = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
683
ret = i915_gem_object_prepare_write(obj, &needs_clflush);
sys/dev/pci/drm/i915/i915_gem.c
687
i915_gem_object_finish_access(obj);
sys/dev/pci/drm/i915/i915_gem.c
688
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem.c
702
struct vm_page *page = i915_gem_object_get_page(obj, idx);
sys/dev/pci/drm/i915/i915_gem.c
716
i915_gem_object_flush_frontbuffer(obj, ORIGIN_CPU);
sys/dev/pci/drm/i915/i915_gem.c
718
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
722
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/i915_gem.c
724
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem.c
742
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_gem.c
757
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/i915_gem.c
758
if (!obj)
sys/dev/pci/drm/i915/i915_gem.c
762
if (range_overflows_t(u64, args->offset, args->size, obj->base.size)) {
sys/dev/pci/drm/i915/i915_gem.c
768
if (i915_gem_object_is_readonly(obj)) {
sys/dev/pci/drm/i915/i915_gem.c
773
trace_i915_gem_object_pwrite(obj, args->offset, args->size);
sys/dev/pci/drm/i915/i915_gem.c
776
if (obj->ops->pwrite)
sys/dev/pci/drm/i915/i915_gem.c
777
ret = obj->ops->pwrite(obj, args);
sys/dev/pci/drm/i915/i915_gem.c
781
ret = i915_gem_object_wait(obj,
sys/dev/pci/drm/i915/i915_gem.c
795
if (!i915_gem_object_has_struct_page(obj) ||
sys/dev/pci/drm/i915/i915_gem.c
796
i915_gem_cpu_write_needs_clflush(obj))
sys/dev/pci/drm/i915/i915_gem.c
801
ret = i915_gem_gtt_pwrite_fast(obj, args);
sys/dev/pci/drm/i915/i915_gem.c
804
if (i915_gem_object_has_struct_page(obj))
sys/dev/pci/drm/i915/i915_gem.c
805
ret = i915_gem_shmem_pwrite(obj, args);
sys/dev/pci/drm/i915/i915_gem.c
809
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_gem.c
824
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_gem.c
826
obj = i915_gem_object_lookup(file, args->handle);
sys/dev/pci/drm/i915/i915_gem.c
827
if (!obj)
sys/dev/pci/drm/i915/i915_gem.c
836
i915_gem_object_flush_if_display(obj);
sys/dev/pci/drm/i915/i915_gem.c
837
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_gem.c
844
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/i915_gem.c
853
list_for_each_entry_safe(obj, on,
sys/dev/pci/drm/i915/i915_gem.c
855
__i915_gem_object_release_mmap_gtt(obj);
sys/dev/pci/drm/i915/i915_gem.c
857
list_for_each_entry_safe(obj, on,
sys/dev/pci/drm/i915/i915_gem.c
859
i915_gem_object_runtime_pm_release_mmap_offset(obj);
sys/dev/pci/drm/i915/i915_gem.c
891
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_gem.c
893
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_gem.c
895
rb_erase(&vma->obj_node, &obj->vma.tree);
sys/dev/pci/drm/i915/i915_gem.c
898
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_gem.c
902
i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.c
907
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_gem.c
924
if (obj->base.size > ggtt->mappable_end)
sys/dev/pci/drm/i915/i915_gem.c
944
obj->base.size > ggtt->mappable_end / 2)
sys/dev/pci/drm/i915/i915_gem.c
949
vma = i915_vma_instance(obj, &ggtt->vm, view);
sys/dev/pci/drm/i915/i915_gem.c
989
if (vma->fence && !i915_gem_object_is_tiled(obj)) {
sys/dev/pci/drm/i915/i915_gem.h
56
i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.h
62
i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem.h
66
int i915_gem_object_unbind(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem_evict.c
43
return !kref_read(&vma->obj->base.refcount);
sys/dev/pci/drm/i915/i915_gem_evict.c
491
if (!i915_gem_object_get_rcu(vma->obj) ||
sys/dev/pci/drm/i915/i915_gem_evict.c
492
(ww && (dma_resv_locking_ctx(vma->obj->base.resv) == &ww->ctx))) {
sys/dev/pci/drm/i915/i915_gem_evict.c
498
if (!i915_gem_object_trylock(vma->obj, ww)) {
sys/dev/pci/drm/i915/i915_gem_evict.c
500
*busy_bo = vma->obj; /* holds ref */
sys/dev/pci/drm/i915/i915_gem_evict.c
504
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/i915_gem_evict.c
524
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/i915_gem_evict.c
535
i915_gem_object_unlock(vma->obj);
sys/dev/pci/drm/i915/i915_gem_evict.c
536
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/i915_gem_evict.c
73
if (i915_gem_object_get_rcu(vma->obj)) {
sys/dev/pci/drm/i915/i915_gem_evict.c
74
if (!i915_gem_object_trylock(vma->obj, ww)) {
sys/dev/pci/drm/i915/i915_gem_evict.c
75
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/i915_gem_evict.c
91
i915_gem_object_unlock(vma->obj);
sys/dev/pci/drm/i915/i915_gem_evict.c
92
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/i915_gem_gtt.c
27
int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem_gtt.c
32
if (dma_map_sg_attrs(obj->base.dev->dev,
sys/dev/pci/drm/i915/i915_gem_gtt.c
47
GEM_BUG_ON(obj->mm.pages == pages);
sys/dev/pci/drm/i915/i915_gem_gtt.c
48
} while (i915_gem_shrink(NULL, to_i915(obj->base.dev),
sys/dev/pci/drm/i915/i915_gem_gtt.c
49
obj->base.size >> PAGE_SHIFT, NULL,
sys/dev/pci/drm/i915/i915_gem_gtt.c
59
void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem_gtt.c
62
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_gem_gtt.h
23
int __must_check i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem_gtt.h
25
void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_gem_ww.c
19
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_gem_ww.c
21
while ((obj = list_first_entry_or_null(&ww->obj_list, struct drm_i915_gem_object, obj_link))) {
sys/dev/pci/drm/i915/i915_gem_ww.c
22
list_del(&obj->obj_link);
sys/dev/pci/drm/i915/i915_gem_ww.c
23
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem_ww.c
24
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_gem_ww.c
28
void i915_gem_ww_unlock_single(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_gem_ww.c
30
list_del(&obj->obj_link);
sys/dev/pci/drm/i915/i915_gem_ww.c
31
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_gem_ww.c
32
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_gem_ww.h
20
void i915_gem_ww_unlock_single(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/i915_perf.c
1400
val = i915_gem_object_pin_map_unlocked(scratch->obj, I915_MAP_WB);
sys/dev/pci/drm/i915/i915_perf.c
1407
i915_gem_object_unpin_map(scratch->obj);
sys/dev/pci/drm/i915/i915_perf.c
2217
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_perf.c
2234
obj = i915_gem_object_create_shmem(stream->perf->i915, config_length);
sys/dev/pci/drm/i915/i915_perf.c
2235
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/i915_perf.c
2236
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/i915_perf.c
2242
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/i915_perf.c
2246
cs = i915_gem_object_pin_map(obj, I915_MAP_WB);
sys/dev/pci/drm/i915/i915_perf.c
2269
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/i915_perf.c
2270
__i915_gem_object_release_map(obj);
sys/dev/pci/drm/i915/i915_perf.c
2272
oa_bo->vma = i915_vma_instance(obj,
sys/dev/pci/drm/i915/i915_perf.c
2292
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_perf.c
2343
err = i915_gem_object_lock(vma->obj, &ww);
sys/dev/pci/drm/i915/i915_request.c
1687
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_request.c
1694
dma_resv_for_each_fence(&cursor, obj->base.resv,
sys/dev/pci/drm/i915/i915_request.h
423
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
1028
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_vma.c
1031
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_vma.c
1050
sg = rotate_pages(obj, rot_info->plane[i].offset,
sys/dev/pci/drm/i915/i915_vma.c
1063
obj->base.size, rot_info->plane[0].width,
sys/dev/pci/drm/i915/i915_vma.c
1089
remap_tiled_color_plane_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
1117
addr = i915_gem_object_get_dma_address_len(obj, offset, &length);
sys/dev/pci/drm/i915/i915_vma.c
1148
remap_contiguous_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
1156
iter = i915_gem_object_get_sg_dma(obj, obj_offset, &offset);
sys/dev/pci/drm/i915/i915_vma.c
1181
remap_linear_color_plane_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
1193
sg = remap_contiguous_pages(obj, obj_offset, size, st, sg);
sys/dev/pci/drm/i915/i915_vma.c
1203
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
1214
sg = remap_linear_color_plane_pages(obj,
sys/dev/pci/drm/i915/i915_vma.c
1222
sg = remap_tiled_color_plane_pages(obj,
sys/dev/pci/drm/i915/i915_vma.c
1237
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_vma.c
1240
struct drm_i915_private *i915 = to_i915(obj->base.dev);
sys/dev/pci/drm/i915/i915_vma.c
1260
sg = remap_color_plane_pages(rem_info, obj, i, st, sg, &gtt_offset);
sys/dev/pci/drm/i915/i915_vma.c
1271
obj->base.size, rem_info->plane[0].width,
sys/dev/pci/drm/i915/i915_vma.c
1279
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_vma.c
1296
sg = remap_contiguous_pages(obj, view->partial.offset, count, st, st->sgl);
sys/dev/pci/drm/i915/i915_vma.c
1320
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(vma->obj));
sys/dev/pci/drm/i915/i915_vma.c
1327
pages = vma->obj->mm.pages;
sys/dev/pci/drm/i915/i915_vma.c
1332
intel_rotate_pages(&vma->gtt_view.rotated, vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1337
intel_remap_pages(&vma->gtt_view.remapped, vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1341
pages = intel_partial_pages(&vma->gtt_view, vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1364
err = i915_gem_object_pin_pages(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1372
vma->page_sizes = vma->obj->mm.page_sizes;
sys/dev/pci/drm/i915/i915_vma.c
1378
__i915_gem_object_unpin_pages(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1410
if (vma->pages != vma->obj->mm.pages) {
sys/dev/pci/drm/i915/i915_vma.c
1416
i915_gem_object_unpin_pages(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1491
err = i915_gem_object_get_moving_fence(vma->obj, &moving);
sys/dev/pci/drm/i915/i915_vma.c
154
vma_create(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
1578
vma->obj->pat_index,
sys/dev/pci/drm/i915/i915_vma.c
1639
err = i915_gem_object_lock(vma->obj, &ww);
sys/dev/pci/drm/i915/i915_vma.c
1709
lockdep_assert_not_held(&vma->obj->base.resv->lock.base);
sys/dev/pci/drm/i915/i915_vma.c
171
vma->obj = obj;
sys/dev/pci/drm/i915/i915_vma.c
1712
err = i915_gem_object_lock(vma->obj, &_ww);
sys/dev/pci/drm/i915/i915_vma.c
172
vma->size = obj->base.size;
sys/dev/pci/drm/i915/i915_vma.c
1728
void i915_ggtt_clear_scanout(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/i915_vma.c
1732
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
1733
for_each_ggtt_vma(vma, obj) {
sys/dev/pci/drm/i915/i915_vma.c
1737
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
1803
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
1807
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
1810
rb_erase(&vma->obj_node, &obj->vma.tree);
sys/dev/pci/drm/i915/i915_vma.c
1812
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
1885
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
1890
if (!kref_get_unless_zero(&obj->base.refcount))
sys/dev/pci/drm/i915/i915_vma.c
1894
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_vma.c
1904
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
1907
if (i915_gem_object_trylock(obj, NULL)) {
sys/dev/pci/drm/i915/i915_vma.c
1910
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/i915_vma.c
1918
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_vma.c
1931
__i915_gem_object_release_map(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
1946
GEM_BUG_ON(!vma->obj->userfault_count);
sys/dev/pci/drm/i915/i915_vma.c
1956
struct drm_i915_private *dev_priv = vma->obj->base.dev->dev_private;
sys/dev/pci/drm/i915/i915_vma.c
196
obj->base.size >> PAGE_SHIFT));
sys/dev/pci/drm/i915/i915_vma.c
1966
if (!--vma->obj->userfault_count)
sys/dev/pci/drm/i915/i915_vma.c
1967
list_del(&vma->obj->userfault_link);
sys/dev/pci/drm/i915/i915_vma.c
199
GEM_BUG_ON(vma->size > obj->base.size);
sys/dev/pci/drm/i915/i915_vma.c
1993
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
1996
assert_object_held(obj);
sys/dev/pci/drm/i915/i915_vma.c
2001
err = i915_request_await_object(rq, vma->obj, flags & EXEC_OBJECT_WRITE);
sys/dev/pci/drm/i915/i915_vma.c
2019
err = dma_resv_reserve_fences(vma->obj->base.resv, idx);
sys/dev/pci/drm/i915/i915_vma.c
2027
front = i915_gem_object_get_frontbuffer(obj);
sys/dev/pci/drm/i915/i915_vma.c
2042
obj->write_domain = I915_GEM_DOMAIN_RENDER;
sys/dev/pci/drm/i915/i915_vma.c
2043
obj->read_domains = 0;
sys/dev/pci/drm/i915/i915_vma.c
2046
obj->write_domain = 0;
sys/dev/pci/drm/i915/i915_vma.c
2050
dma_resv_add_fence(vma->obj->base.resv, curr, usage);
sys/dev/pci/drm/i915/i915_vma.c
2056
obj->read_domains |= I915_GEM_GPU_DOMAINS;
sys/dev/pci/drm/i915/i915_vma.c
2057
obj->mm.dirty = true;
sys/dev/pci/drm/i915/i915_vma.c
2113
vma->obj->mm.tlb);
sys/dev/pci/drm/i915/i915_vma.c
2130
vma_invalidate_tlb(vma->vm, vma->obj->mm.tlb);
sys/dev/pci/drm/i915/i915_vma.c
2184
&vma->obj->mm.rsgt->table != vma->resource->bi.pages)
sys/dev/pci/drm/i915/i915_vma.c
2215
assert_object_held_shared(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
223
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
2249
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
2259
assert_object_held(obj);
sys/dev/pci/drm/i915/i915_vma.c
2269
if (!obj->mm.rsgt)
sys/dev/pci/drm/i915/i915_vma.c
2272
err = dma_resv_reserve_fences(obj->base.resv, 2);
sys/dev/pci/drm/i915/i915_vma.c
229
i915_gem_object_get_tiling(obj),
sys/dev/pci/drm/i915/i915_vma.c
230
i915_gem_object_get_stride(obj));
sys/dev/pci/drm/i915/i915_vma.c
2300
dma_resv_add_fence(obj->base.resv, fence, DMA_RESV_USAGE_READ);
sys/dev/pci/drm/i915/i915_vma.c
2313
i915_gem_object_lock(vma->obj, NULL);
sys/dev/pci/drm/i915/i915_vma.c
2315
i915_gem_object_unlock(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
2322
i915_gem_object_make_unshrinkable(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
2328
i915_gem_object_make_shrinkable(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
2333
i915_gem_object_make_purgeable(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
238
i915_gem_object_get_tiling(obj),
sys/dev/pci/drm/i915/i915_vma.c
239
i915_gem_object_get_stride(obj));
sys/dev/pci/drm/i915/i915_vma.c
246
p = &obj->vma.tree.rb_node;
sys/dev/pci/drm/i915/i915_vma.c
267
rb_insert_color(&vma->obj_node, &obj->vma.tree);
sys/dev/pci/drm/i915/i915_vma.c
276
list_add(&vma->obj_link, &obj->vma.list);
sys/dev/pci/drm/i915/i915_vma.c
278
list_add_tail(&vma->obj_link, &obj->vma.list);
sys/dev/pci/drm/i915/i915_vma.c
280
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
286
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
295
i915_vma_lookup(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
301
rb = obj->vma.tree.rb_node;
sys/dev/pci/drm/i915/i915_vma.c
333
i915_vma_instance(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma.c
342
spin_lock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
343
vma = i915_vma_lookup(obj, vm, view);
sys/dev/pci/drm/i915/i915_vma.c
344
spin_unlock(&obj->vma.lock);
sys/dev/pci/drm/i915/i915_vma.c
348
vma = vma_create(obj, vm, view);
sys/dev/pci/drm/i915/i915_vma.c
359
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_vma.c
376
if (i915_gem_object_has_unknown_state(vw->obj))
sys/dev/pci/drm/i915/i915_vma.c
387
if (vw->obj)
sys/dev/pci/drm/i915/i915_vma.c
388
i915_gem_object_put(vw->obj);
sys/dev/pci/drm/i915/i915_vma.c
460
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
463
obj->mm.rsgt, i915_gem_object_is_readonly(obj),
sys/dev/pci/drm/i915/i915_vma.c
464
i915_gem_object_is_lmem(obj), obj->mm.region,
sys/dev/pci/drm/i915/i915_vma.c
56
assert_object_held_shared(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
574
work->obj = i915_gem_object_get(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
576
ret = i915_gem_object_wait_moving_fence(vma->obj, true);
sys/dev/pci/drm/i915/i915_vma.c
596
if (WARN_ON_ONCE(vma->obj->flags & I915_BO_ALLOC_GPU_ONLY))
sys/dev/pci/drm/i915/i915_vma.c
611
if (i915_gem_object_is_lmem(vma->obj)) {
sys/dev/pci/drm/i915/i915_vma.c
612
ptr = i915_gem_object_lmem_io_map(vma->obj, 0,
sys/dev/pci/drm/i915/i915_vma.c
613
vma->obj->base.size);
sys/dev/pci/drm/i915/i915_vma.c
620
i915_gem_object_pin_map(vma->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/i915_vma.c
635
__i915_gem_object_release_map(vma->obj);
sys/dev/pci/drm/i915/i915_vma.c
680
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/i915_vma.c
686
obj = vma->obj;
sys/dev/pci/drm/i915/i915_vma.c
687
GEM_BUG_ON(!obj);
sys/dev/pci/drm/i915/i915_vma.c
692
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/i915_vma.c
694
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/i915_vma.c
843
alignment = max(alignment, i915_vm_obj_min_alignment(vma->vm, vma->obj));
sys/dev/pci/drm/i915/i915_vma.c
851
drm_dbg(vma->obj->base.dev,
sys/dev/pci/drm/i915/i915_vma.c
860
color = vma->obj->pat_index;
sys/dev/pci/drm/i915/i915_vma.c
88
drm_dbg(vma->obj->base.dev,
sys/dev/pci/drm/i915/i915_vma.c
95
drm_dbg(vma->obj->base.dev,
sys/dev/pci/drm/i915/i915_vma.c
978
rotate_pages(struct drm_i915_gem_object *obj, unsigned int offset,
sys/dev/pci/drm/i915/i915_vma.c
999
i915_gem_object_get_dma_address(obj, src_idx);
sys/dev/pci/drm/i915/i915_vma.h
194
i915_gem_object_get(vma->obj);
sys/dev/pci/drm/i915/i915_vma.h
200
if (likely(kref_get_unless_zero(&vma->obj->base.refcount)))
sys/dev/pci/drm/i915/i915_vma.h
208
i915_gem_object_put(vma->obj);
sys/dev/pci/drm/i915/i915_vma.h
280
#define assert_vma_held(vma) dma_resv_assert_held((vma)->obj->base.resv)
sys/dev/pci/drm/i915/i915_vma.h
284
dma_resv_lock(vma->obj->base.resv, NULL);
sys/dev/pci/drm/i915/i915_vma.h
289
dma_resv_unlock(vma->obj->base.resv);
sys/dev/pci/drm/i915/i915_vma.h
433
void i915_ggtt_clear_scanout(struct drm_i915_gem_object *obj);
sys/dev/pci/drm/i915/i915_vma.h
44
i915_vma_instance(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/i915_vma_types.h
143
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/intel_memory_region.h
52
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/pxp/intel_pxp.c
465
struct drm_i915_gem_object *obj = to_intel_bo(_obj);
sys/dev/pci/drm/i915/pxp/intel_pxp.c
472
if (!i915_gem_object_is_protected(obj))
sys/dev/pci/drm/i915/pxp/intel_pxp.c
483
if (!obj->pxp_key_instance && assign)
sys/dev/pci/drm/i915/pxp/intel_pxp.c
484
obj->pxp_key_instance = pxp->key_instance;
sys/dev/pci/drm/i915/pxp/intel_pxp.c
486
if (obj->pxp_key_instance != pxp->key_instance)
sys/dev/pci/drm/i915/pxp/intel_pxp.h
34
int intel_pxp_key_check(struct drm_gem_object *obj, bool assign);
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
323
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
326
obj = i915_gem_object_create_internal(i915, size);
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
327
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
329
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
333
*vma = i915_vma_instance(obj, gt->vm, NULL);
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
341
*map = i915_gem_object_pin_map_unlocked(obj, intel_gt_coherent_map_type(gt, obj, true));
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
358
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_gsccs.c
360
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_huc.c
34
huc_phys_addr = i915_gem_object_get_dma_address(huc->fw.obj, 0);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
131
GEM_BUG_ON(!pxp->stream_cmd.obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
133
sg = i915_gem_object_get_sg_dma(pxp->stream_cmd.obj, 0, &offset);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
239
struct drm_i915_gem_object *obj = NULL;
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
243
pxp->stream_cmd.obj = NULL;
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
250
obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, I915_BO_ALLOC_CONTIGUOUS);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
251
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
253
return PTR_ERR(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
256
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
263
cmd = i915_gem_object_pin_map_unlocked(obj,
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
265
obj, true));
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
272
memset(cmd, 0, obj->base.size);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
274
pxp->stream_cmd.obj = obj;
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
280
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
282
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
288
struct drm_i915_gem_object *obj = fetch_and_zero(&pxp->stream_cmd.obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
290
if (!obj)
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
293
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
294
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_tee.c
295
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/pxp/intel_pxp_types.h
100
struct drm_i915_gem_object *obj; /* contains PXP command memory */
sys/dev/pci/drm/i915/selftests/i915_gem.c
213
struct drm_i915_gem_object *obj, *obj2;
sys/dev/pci/drm/i915/selftests/i915_gem.c
217
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_gem.c
218
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_gem.c
219
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem.c
230
err = i915_gem_object_lock(obj, &ww);
sys/dev/pci/drm/i915/selftests/i915_gem.c
232
err = i915_gem_object_lock_interruptible(obj, &ww);
sys/dev/pci/drm/i915/selftests/i915_gem.c
246
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
100
i915_gem_object_set_tiling_quirk(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
101
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
157
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
170
obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
171
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
172
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
176
quirk_add(obj, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
178
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, 0);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
251
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
265
obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
266
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
267
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
270
i915_gem_object_set_cache_coherency(obj, I915_CACHE_LLC);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
271
quirk_add(obj, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
273
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0,
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
281
obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
282
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
283
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
286
i915_gem_object_set_cache_coherency(obj, I915_CACHE_LLC);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
287
quirk_add(obj, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
290
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0,
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
39
static void quirk_add(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
43
GEM_BUG_ON(i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
44
i915_gem_object_set_tiling_quirk(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
45
list_add(&obj->st_link, objects);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
50
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
57
obj = i915_gem_object_create_internal(ggtt->vm.i915,
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
59
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
60
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
62
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, 0);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
64
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
71
quirk_add(obj, objects);
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
90
if (i915_gem_object_has_tiling_quirk(vma->obj))
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
96
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
98
list_for_each_entry_safe(obj, on, list, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_evict.c
99
GEM_BUG_ON(!i915_gem_object_has_tiling_quirk(obj));
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
101
static void fake_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
104
fake_free_pages(obj, pages);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
105
obj->mm.dirty = false;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1085
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1092
obj = i915_gem_object_create_region(mr, size, 0, I915_BO_ALLOC_GPU_ONLY);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1093
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1095
if (PTR_ERR(obj) == -ENODEV && is_stolen)
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1097
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1100
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1124
if (HAS_64K_PAGES(vm->i915) && i915_gem_object_is_lmem(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1142
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
119
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
124
if (overflows_type(size, obj->base.size))
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
127
obj = i915_gem_object_alloc();
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
128
if (!obj)
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
131
drm_gem_private_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
132
i915_gem_object_init(obj, &fake_ops, &lock_class, 0);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
134
i915_gem_object_set_volatile(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1350
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1359
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
136
obj->write_domain = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1360
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1361
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1363
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
137
obj->read_domains = I915_GEM_DOMAIN_CPU;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
138
obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1384
i915_gem_object_get_dma_address(obj, 0),
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
141
if (i915_gem_object_pin_pages_unlocked(obj))
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1433
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1435
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
144
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1441
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1443
__i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1447
__i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1448
vma->pages = obj->mm.pages;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
145
return obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
148
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1511
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1519
err = i915_gem_gtt_reserve(vm, NULL, &vma->node, obj->base.size,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1521
obj->pat_index,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1537
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1555
obj = i915_gem_object_create_internal(ggtt->vm.i915,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1557
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1558
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1562
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1564
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1568
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1569
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1600
obj = i915_gem_object_create_internal(ggtt->vm.i915,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1602
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1603
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1607
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1609
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1613
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1615
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1641
list_for_each_entry_safe(obj, on, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1645
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1682
list_for_each_entry_safe(obj, on, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1683
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1684
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1693
struct drm_i915_gem_object *obj = vma->obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1701
err = i915_gem_gtt_insert(vm, NULL, &vma->node, obj->base.size, 0,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1702
obj->pat_index, 0, vm->total, 0);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1717
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1778
obj = i915_gem_object_create_internal(ggtt->vm.i915,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1780
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1781
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1785
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1787
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1791
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1793
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1802
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1816
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1819
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1835
list_for_each_entry_safe(obj, on, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1839
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1877
obj = i915_gem_object_create_internal(ggtt->vm.i915,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1879
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1880
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1884
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1886
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1890
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1892
vma = i915_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1910
list_for_each_entry_safe(obj, on, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1911
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
1912
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
261
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
297
obj = fake_dma_object(vm->i915, BIT_ULL(size));
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
298
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
303
GEM_BUG_ON(obj->base.size != BIT_ULL(size));
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
305
if (i915_gem_object_pin_pages_unlocked(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
306
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
357
mock_vma_res->bi.pages = obj->mm.pages;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
379
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
380
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
394
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
397
list_for_each_entry_safe(obj, on, objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
400
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
404
list_del(&obj->st_link);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
405
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
414
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
444
obj = fake_dma_object(vm->i915, full_size);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
445
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
448
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
459
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
460
u64 aligned_size = round_up(obj->base.size,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
463
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
499
list_for_each_entry(obj, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
500
u64 aligned_size = round_up(obj->base.size,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
503
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
51
static void fake_free_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
538
list_for_each_entry_reverse(obj, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
539
u64 aligned_size = round_up(obj->base.size,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
542
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
578
list_for_each_entry_reverse(obj, &objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
579
u64 aligned_size = round_up(obj->base.size,
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
58
static int fake_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
582
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
64
typeof(obj->base.size) rem;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
655
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
660
obj = fake_dma_object(vm->i915, size << PAGE_SHIFT);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
661
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
664
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
671
addr + obj->base.size < hole_end;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
672
addr += round_up(obj->base.size, min_alignment)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
70
rem = round_up(obj->base.size, BIT(31)) >> 31;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
708
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
722
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
735
obj = i915_gem_object_create_internal(vm->i915, 2 * I915_GTT_PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
736
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
737
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
739
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
789
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
810
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
82
rem = obj->base.size;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
843
obj = fake_dma_object(vm->i915, BIT_ULL(size));
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
844
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
849
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
893
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
908
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
924
obj = fake_dma_object(vm->i915, size);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
925
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
926
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
930
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
932
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_gem_gtt.c
95
__i915_gem_object_set_pages(obj, pages);
sys/dev/pci/drm/i915/selftests/i915_request.c
1003
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
1119
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_request.c
1125
obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_request.c
1126
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_request.c
1127
return ERR_CAST(obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
1129
vma = i915_vma_instance(obj, gt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_request.c
1139
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/selftests/i915_request.c
1158
__i915_gem_object_flush_map(obj, 0, 64);
sys/dev/pci/drm/i915/selftests/i915_request.c
1159
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
1166
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
1174
cmd = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
sys/dev/pci/drm/i915/selftests/i915_request.c
1180
__i915_gem_object_flush_map(batch->obj, 0, sizeof(*cmd));
sys/dev/pci/drm/i915/selftests/i915_request.c
1181
i915_gem_object_unpin_map(batch->obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
1435
cmd = i915_gem_object_pin_map_unlocked(request[idx]->batch->obj,
sys/dev/pci/drm/i915/selftests/i915_request.c
1440
__i915_gem_object_flush_map(request[idx]->batch->obj,
sys/dev/pci/drm/i915/selftests/i915_request.c
1442
i915_gem_object_unpin_map(request[idx]->batch->obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
961
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_request.c
966
obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_request.c
967
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_request.c
968
return ERR_CAST(obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
970
cmd = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/selftests/i915_request.c
978
__i915_gem_object_flush_map(obj, 0, 64);
sys/dev/pci/drm/i915/selftests/i915_request.c
979
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/selftests/i915_request.c
983
vma = i915_vma_instance(obj, gt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_vma.c
1013
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/i915_vma.c
1014
err = i915_gem_object_set_to_gtt_domain(obj, true);
sys/dev/pci/drm/i915/selftests/i915_vma.c
1015
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
1023
vma = i915_gem_object_ggtt_pin(obj, &view, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
1054
vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0, PIN_MAPPABLE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
1100
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
111
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_vma.c
115
list_for_each_entry(obj, objects, st_link) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
123
vma = checked_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_vma.c
128
if (!assert_vma(vma, obj, ctx)) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
153
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/selftests/i915_vma.c
170
obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
171
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_vma.c
174
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/i915_vma.c
214
list_for_each_entry_safe(obj, on, &objects, st_link)
sys/dev/pci/drm/i915/selftests/i915_vma.c
215
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
313
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_vma.c
324
obj = i915_gem_object_create_internal(ggtt->vm.i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
325
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_vma.c
326
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
328
vma = checked_vma_instance(obj, &ggtt->vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_vma.c
359
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
373
assert_rotated(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_vma.c
38
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_vma.c
393
src = i915_gem_object_get_dma_address(obj, src_idx);
sys/dev/pci/drm/i915/selftests/i915_vma.c
450
assert_remapped(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_vma.c
474
src = i915_gem_object_get_dma_address(obj, src_idx);
sys/dev/pci/drm/i915/selftests/i915_vma.c
48
if (vma->size != obj->base.size) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
50
vma->size, obj->base.size);
sys/dev/pci/drm/i915/selftests/i915_vma.c
551
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_vma.c
588
obj = i915_gem_object_create_internal(vm->i915, max_pages * PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
589
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_vma.c
623
vma = checked_vma_instance(obj, vm, &view);
sys/dev/pci/drm/i915/selftests/i915_vma.c
64
checked_vma_instance(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_vma.c
667
if (vma->pages == obj->mm.pages) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
676
sg = assert_rotated(obj, &view.rotated, n, sg);
sys/dev/pci/drm/i915/selftests/i915_vma.c
678
sg = assert_remapped(obj, &view.remapped, n, sg);
sys/dev/pci/drm/i915/selftests/i915_vma.c
71
vma = i915_vma_instance(obj, vm, view);
sys/dev/pci/drm/i915/selftests/i915_vma.c
712
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
717
static bool assert_partial(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/i915_vma.c
733
src = i915_gem_object_get_dma_address(obj, offset);
sys/dev/pci/drm/i915/selftests/i915_vma.c
773
if (vma->pages == vma->obj->mm.pages) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
785
if (vma->pages != vma->obj->mm.pages) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
799
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_vma.c
816
obj = i915_gem_object_create_internal(vm->i915, npages * PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
817
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_vma.c
835
vma = checked_vma_instance(obj, vm, &view);
sys/dev/pci/drm/i915/selftests/i915_vma.c
852
if (!assert_partial(obj, vma, offset, sz)) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
872
list_for_each_entry(vma, &obj->vma.list, obj_link)
sys/dev/pci/drm/i915/selftests/i915_vma.c
882
vma = checked_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/i915_vma.c
892
if (!assert_pin(vma, NULL, obj->base.size, p->name)) {
sys/dev/pci/drm/i915/selftests/i915_vma.c
907
list_for_each_entry(vma, &obj->vma.list, obj_link)
sys/dev/pci/drm/i915/selftests/i915_vma.c
917
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/i915_vma.c
989
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/i915_vma.c
996
obj = i915_gem_object_create_internal(i915, 10 * 10 * PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/i915_vma.c
997
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/i915_vma.c
998
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
101
mode = intel_gt_coherent_map_type(spin->gt, spin->obj, false);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
102
vaddr = igt_spinner_pin_obj(ce, ww, spin->obj, mode, &spin->batch_vma);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
243
i915_gem_object_unpin_map(spin->obj);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
245
i915_gem_object_put(spin->obj);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
29
spin->obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
30
if (IS_ERR(spin->obj)) {
sys/dev/pci/drm/i915/selftests/igt_spinner.c
31
err = PTR_ERR(spin->obj);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
45
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/igt_spinner.c
51
*vma = i915_vma_instance(obj, ce->vm, NULL);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
55
ret = i915_gem_object_lock(obj, ww);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
59
vaddr = i915_gem_object_pin_map(obj, mode);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
62
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/igt_spinner.c
73
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/selftests/igt_spinner.h
22
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1013
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1044
obj = i915_gem_object_create_lmem(i915, sz, I915_BO_ALLOC_CONTIGUOUS);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1045
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1046
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1048
vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1054
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1056
err = dma_resv_reserve_fences(obj->base.resv, 1);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1058
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1065
obj->mm.pages->sgl,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1070
dma_resv_add_fence(obj->base.resv, &rq->fence,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1077
err = i915_gem_object_set_to_wc_domain(obj, true);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1078
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1109
offset = igt_random_offset(&prng, 0, obj->base.size,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
112
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1134
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1136
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
115
obj = i915_gem_object_create_region(mem, size, 0, flags);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1157
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
116
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1160
obj = i915_gem_object_create_region(mr, size, 0, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1161
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1162
if (PTR_ERR(obj) == -ENOSPC) /* Stolen memory */
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1164
return obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1167
addr = i915_gem_object_pin_map_unlocked(obj, type);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1169
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
117
return obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
1176
return obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
119
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
123
list_add(&obj->st_link, objects);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
124
return obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
127
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
131
static void igt_object_release(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
133
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
134
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
135
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
136
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
137
list_del(&obj->st_link);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
138
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
141
static bool is_contiguous(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
146
for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
161
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
211
obj = igt_object_create(mem, &objects, size, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
212
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
213
if (PTR_ERR(obj) == -ENXIO)
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
216
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
239
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
252
obj = igt_object_create(mem, &objects, PAGE_SIZE,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
254
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
255
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
257
if (!is_contiguous(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
263
igt_object_release(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
266
obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
267
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
268
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
270
if (!is_contiguous(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
276
igt_object_release(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
284
obj = igt_object_create(mem, &objects, target,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
286
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
287
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
289
if (obj->base.size != target) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
291
obj->base.size, target);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
296
if (!is_contiguous(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
302
igt_object_release(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
320
obj = igt_object_create(mem, list, target,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
322
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
323
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
334
obj = igt_object_create(mem, &objects, target, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
335
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
336
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
340
igt_object_release(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
350
obj = igt_object_create(mem, &objects, target,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
352
if (should_fail != IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
36
struct drm_i915_gem_object *obj, *on;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
373
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
38
list_for_each_entry_safe(obj, on, objects, st_link) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
39
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
391
obj = igt_object_create(mem, &objects, size, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
392
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
393
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
397
res = to_ttm_buddy_resource(obj->mm.res);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
40
if (i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
41
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
416
obj = igt_object_create(mem, &objects, roundup_pow_of_two(size),
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
418
if (!IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
425
obj = igt_object_create(mem, &objects, size, I915_BO_ALLOC_CONTIGUOUS);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
426
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
429
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
43
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
44
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
449
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
45
list_del(&obj->st_link);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
46
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
479
obj = igt_object_create(mem, &objects, size, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
480
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
481
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
485
res = to_ttm_buddy_resource(obj->mm.res);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
500
for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
525
static u64 igt_object_mappable_total(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
527
struct intel_memory_region *mr = obj->mm.region;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
529
to_ttm_buddy_resource(obj->mm.res);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
550
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
589
obj = igt_object_create(mr, &objects, size,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
591
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
594
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
598
mappable_theft_total += igt_object_mappable_total(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
612
obj = igt_object_create(mr, &objects, io_size,
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
614
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
616
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
628
obj = igt_object_create(mr, &objects, size, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
629
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
632
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
636
if (igt_object_mappable_total(obj) != size) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
649
obj = igt_object_create(mr, &objects, ps, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
650
if (!IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
675
static int igt_cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
677
unsigned long n = obj->base.size >> PAGE_SHIFT;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
681
err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
685
ptr = i915_gem_object_pin_map(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
701
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
706
struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
71
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
720
GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
73
obj = i915_gem_object_create_region(mem, size, 0, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
74
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
740
vma = i915_vma_instance(obj, vm, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
75
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
765
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
766
err = igt_cpu_check(obj, dword, rng);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
767
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
785
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
788
obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
789
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
79
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
790
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
792
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
796
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
798
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
81
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
810
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
813
obj = __i915_gem_object_create_lmem_with_ps(i915, ps, ps, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
814
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
815
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
825
if (obj->base.size != ps) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
827
__func__, obj->base.size, ps);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
832
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
833
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
843
daddr = i915_gem_object_get_dma_address(obj, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
85
list_add(&obj->st_link, &objects);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
852
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
853
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
855
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
856
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
880
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
897
obj = i915_gem_object_create_lmem(i915, size, flags);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
898
if (IS_ERR(obj))
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
899
return PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
901
i915_gem_object_lock(obj, NULL);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
902
err = i915_gem_object_pin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
910
err = igt_cpu_check(obj, dword, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
918
vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
926
memset32(vaddr, val, obj->base.size / sizeof(u32));
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
928
i915_gem_object_flush_map(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
929
i915_gem_object_unpin_map(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
931
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
932
__i915_gem_object_put_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
934
i915_gem_object_unlock(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
935
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
950
struct drm_i915_gem_object *obj;
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
969
obj = i915_gem_object_create_lmem(i915, sz, 0);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
970
if (IS_ERR(obj)) {
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
971
err = PTR_ERR(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
975
err = i915_gem_object_pin_pages_unlocked(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
979
err = igt_gpu_write(ctx, obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
983
i915_gem_object_unpin_pages(obj);
sys/dev/pci/drm/i915/selftests/intel_memory_region.c
985
i915_gem_object_put(obj);
sys/dev/pci/drm/i915/selftests/mock_region.c
15
static void mock_region_put_pages(struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/mock_region.c
18
i915_refct_sgt_put(obj->mm.rsgt);
sys/dev/pci/drm/i915/selftests/mock_region.c
19
obj->mm.rsgt = NULL;
sys/dev/pci/drm/i915/selftests/mock_region.c
20
intel_region_ttm_resource_free(obj->mm.region, obj->mm.res);
sys/dev/pci/drm/i915/selftests/mock_region.c
23
static int mock_region_get_pages(struct drm_i915_gem_object *obj)
sys/dev/pci/drm/i915/selftests/mock_region.c
28
obj->mm.res = intel_region_ttm_resource_alloc(obj->mm.region,
sys/dev/pci/drm/i915/selftests/mock_region.c
29
obj->bo_offset,
sys/dev/pci/drm/i915/selftests/mock_region.c
30
obj->base.size,
sys/dev/pci/drm/i915/selftests/mock_region.c
31
obj->flags);
sys/dev/pci/drm/i915/selftests/mock_region.c
32
if (IS_ERR(obj->mm.res))
sys/dev/pci/drm/i915/selftests/mock_region.c
33
return PTR_ERR(obj->mm.res);
sys/dev/pci/drm/i915/selftests/mock_region.c
35
obj->mm.rsgt = intel_region_ttm_resource_to_rsgt(obj->mm.region,
sys/dev/pci/drm/i915/selftests/mock_region.c
36
obj->mm.res,
sys/dev/pci/drm/i915/selftests/mock_region.c
37
obj->mm.region->min_page_size);
sys/dev/pci/drm/i915/selftests/mock_region.c
38
if (IS_ERR(obj->mm.rsgt)) {
sys/dev/pci/drm/i915/selftests/mock_region.c
39
err = PTR_ERR(obj->mm.rsgt);
sys/dev/pci/drm/i915/selftests/mock_region.c
43
pages = &obj->mm.rsgt->table;
sys/dev/pci/drm/i915/selftests/mock_region.c
44
__i915_gem_object_set_pages(obj, pages);
sys/dev/pci/drm/i915/selftests/mock_region.c
49
intel_region_ttm_resource_free(obj->mm.region, obj->mm.res);
sys/dev/pci/drm/i915/selftests/mock_region.c
61
struct drm_i915_gem_object *obj,
sys/dev/pci/drm/i915/selftests/mock_region.c
73
drm_gem_private_object_init(&i915->drm, &obj->base, size);
sys/dev/pci/drm/i915/selftests/mock_region.c
74
i915_gem_object_init(obj, &mock_region_obj_ops, &lock_class, flags);
sys/dev/pci/drm/i915/selftests/mock_region.c
76
obj->bo_offset = offset;
sys/dev/pci/drm/i915/selftests/mock_region.c
78
obj->read_domains = I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT;
sys/dev/pci/drm/i915/selftests/mock_region.c
80
i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE);
sys/dev/pci/drm/i915/selftests/mock_region.c
82
i915_gem_object_init_memory_region(obj, mem);
sys/dev/pci/drm/include/drm/drm_atomic.h
1092
#define for_each_oldnew_private_obj_in_state(__state, obj, old_obj_state, new_obj_state, __i) \
sys/dev/pci/drm/include/drm/drm_atomic.h
1095
((obj) = (__state)->private_objs[__i].ptr, \
sys/dev/pci/drm/include/drm/drm_atomic.h
1111
#define for_each_old_private_obj_in_state(__state, obj, old_obj_state, __i) \
sys/dev/pci/drm/include/drm/drm_atomic.h
1114
((obj) = (__state)->private_objs[__i].ptr, \
sys/dev/pci/drm/include/drm/drm_atomic.h
1129
#define for_each_new_private_obj_in_state(__state, obj, new_obj_state, __i) \
sys/dev/pci/drm/include/drm/drm_atomic.h
1132
((obj) = (__state)->private_objs[__i].ptr, \
sys/dev/pci/drm/include/drm/drm_atomic.h
1133
(void)(obj) /* Only to avoid unused-but-set-variable warning */, \
sys/dev/pci/drm/include/drm/drm_atomic.h
221
struct drm_private_state *(*atomic_duplicate_state)(struct drm_private_obj *obj);
sys/dev/pci/drm/include/drm/drm_atomic.h
228
void (*atomic_destroy_state)(struct drm_private_obj *obj,
sys/dev/pci/drm/include/drm/drm_atomic.h
340
struct drm_private_obj *obj;
sys/dev/pci/drm/include/drm/drm_atomic.h
607
struct drm_private_obj *obj,
sys/dev/pci/drm/include/drm/drm_atomic.h
610
void drm_atomic_private_obj_fini(struct drm_private_obj *obj);
sys/dev/pci/drm/include/drm/drm_atomic.h
614
struct drm_private_obj *obj);
sys/dev/pci/drm/include/drm/drm_atomic.h
617
struct drm_private_obj *obj);
sys/dev/pci/drm/include/drm/drm_atomic.h
620
struct drm_private_obj *obj);
sys/dev/pci/drm/include/drm/drm_atomic_state_helper.h
87
void __drm_atomic_helper_private_obj_duplicate_state(struct drm_private_obj *obj,
sys/dev/pci/drm/include/drm/drm_exec.h
141
int drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_exec.h
142
void drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_exec.h
143
int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
sys/dev/pci/drm/include/drm/drm_exec.h
76
#define drm_exec_for_each_locked_object(exec, index, obj) \
sys/dev/pci/drm/include/drm/drm_exec.h
77
for ((index) = 0; ((obj) = drm_exec_obj(exec, index)); ++(index))
sys/dev/pci/drm/include/drm/drm_exec.h
90
#define drm_exec_for_each_locked_object_reverse(exec, index, obj) \
sys/dev/pci/drm/include/drm/drm_exec.h
92
((obj) = drm_exec_obj(exec, index)); --(index))
sys/dev/pci/drm/include/drm/drm_framebuffer.h
208
struct drm_gem_object *obj[DRM_FORMAT_MAX_PLANES];
sys/dev/pci/drm/include/drm/drm_gem.h
114
const struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
124
struct dma_buf *(*export)(struct drm_gem_object *obj, int flags);
sys/dev/pci/drm/include/drm/drm_gem.h
134
int (*pin)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
143
void (*unpin)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
157
struct sg_table *(*get_sg_table)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
168
int (*vmap)(struct drm_gem_object *obj, struct iosys_map *map);
sys/dev/pci/drm/include/drm/drm_gem.h
179
void (*vunmap)(struct drm_gem_object *obj, struct iosys_map *map);
sys/dev/pci/drm/include/drm/drm_gem.h
193
int (*mmap)(struct drm_gem_object *obj, struct vm_area_struct *vma);
sys/dev/pci/drm/include/drm/drm_gem.h
207
int (*evict)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
220
enum drm_gem_object_status (*status)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
229
size_t (*rss)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
513
void drm_gem_object_release(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
516
struct drm_gem_object *obj, size_t size);
sys/dev/pci/drm/include/drm/drm_gem.h
519
struct drm_gem_object *obj, size_t size,
sys/dev/pci/drm/include/drm/drm_gem.h
523
struct drm_gem_object *obj, size_t size);
sys/dev/pci/drm/include/drm/drm_gem.h
524
void drm_gem_private_object_fini(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
528
int drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size,
sys/dev/pci/drm/include/drm/drm_gem.h
542
static inline void drm_gem_object_get(struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gem.h
544
kref_get(&obj->refcount);
sys/dev/pci/drm/include/drm/drm_gem.h
549
__drm_gem_object_put(struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gem.h
551
kref_put(&obj->refcount, drm_gem_object_free);
sys/dev/pci/drm/include/drm/drm_gem.h
561
drm_gem_object_put(struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gem.h
563
if (obj)
sys/dev/pci/drm/include/drm/drm_gem.h
564
__drm_gem_object_put(obj);
sys/dev/pci/drm/include/drm/drm_gem.h
568
struct drm_gem_object *obj,
sys/dev/pci/drm/include/drm/drm_gem.h
573
void drm_gem_free_mmap_offset(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
574
int drm_gem_create_mmap_offset(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
575
int drm_gem_create_mmap_offset_size(struct drm_gem_object *obj, size_t size);
sys/dev/pci/drm/include/drm/drm_gem.h
577
struct vm_page **drm_gem_get_pages(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
578
void drm_gem_put_pages(struct drm_gem_object *obj, struct vm_page **pages,
sys/dev/pci/drm/include/drm/drm_gem.h
581
void drm_gem_lock(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
582
void drm_gem_unlock(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
584
int drm_gem_vmap(struct drm_gem_object *obj, struct iosys_map *map);
sys/dev/pci/drm/include/drm/drm_gem.h
585
void drm_gem_vunmap(struct drm_gem_object *obj, struct iosys_map *map);
sys/dev/pci/drm/include/drm/drm_gem.h
600
void drm_gem_lru_remove(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
601
void drm_gem_lru_move_tail_locked(struct drm_gem_lru *lru, struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
602
void drm_gem_lru_move_tail(struct drm_gem_lru *lru, struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
607
bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket),
sys/dev/pci/drm/include/drm/drm_gem.h
610
int drm_gem_evict_locked(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
620
static inline bool drm_gem_object_is_shared_for_memory_stats(struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gem.h
622
return (obj->handle_count > 1) || obj->dma_buf;
sys/dev/pci/drm/include/drm/drm_gem.h
632
static inline bool drm_gem_is_imported(const struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gem.h
634
return !!obj->import_attach;
sys/dev/pci/drm/include/drm/drm_gem.h
638
#define drm_gem_gpuva_assert_lock_held(gpuvm, obj) \
sys/dev/pci/drm/include/drm/drm_gem.h
640
lockdep_is_held(&(obj)->gpuva.lock) : \
sys/dev/pci/drm/include/drm/drm_gem.h
641
dma_resv_held((obj)->resv))
sys/dev/pci/drm/include/drm/drm_gem.h
643
#define drm_gem_gpuva_assert_lock_held(gpuvm, obj) do {} while (0)
sys/dev/pci/drm/include/drm/drm_gem.h
657
static inline void drm_gem_gpuva_init(struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gem.h
659
INIT_LIST_HEAD(&obj->gpuva.list);
sys/dev/pci/drm/include/drm/drm_gem.h
80
void (*free)(struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gem.h
89
int (*open)(struct drm_gem_object *obj, struct drm_file *file);
sys/dev/pci/drm/include/drm/drm_gem.h
98
void (*close)(struct drm_gem_object *obj, struct drm_file *file);
sys/dev/pci/drm/include/drm/drm_gpuvm.h
1112
va->gem.obj = op->gem.obj;
sys/dev/pci/drm/include/drm/drm_gpuvm.h
117
struct drm_gem_object *obj;
sys/dev/pci/drm/include/drm/drm_gpuvm.h
436
struct drm_gem_object *obj)
sys/dev/pci/drm/include/drm/drm_gpuvm.h
438
return obj && obj->resv != drm_gpuvm_resv(gpuvm);
sys/dev/pci/drm/include/drm/drm_gpuvm.h
670
struct drm_gem_object *obj;
sys/dev/pci/drm/include/drm/drm_gpuvm.h
723
struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gpuvm.h
727
struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gpuvm.h
751
struct drm_gem_object *obj);
sys/dev/pci/drm/include/drm/drm_gpuvm.h
764
drm_gpuvm_bo_gem_evict(struct drm_gem_object *obj, bool evict)
sys/dev/pci/drm/include/drm/drm_gpuvm.h
768
drm_gem_for_each_gpuvm_bo(vm_bo, obj) {
sys/dev/pci/drm/include/drm/drm_gpuvm.h
769
drm_gem_gpuva_assert_lock_held(vm_bo->vm, obj);
sys/dev/pci/drm/include/drm/drm_gpuvm.h
872
struct drm_gem_object *obj;
sys/dev/pci/drm/include/drm/drm_mode_object.h
124
void drm_mode_object_get(struct drm_mode_object *obj);
sys/dev/pci/drm/include/drm/drm_mode_object.h
125
void drm_mode_object_put(struct drm_mode_object *obj);
sys/dev/pci/drm/include/drm/drm_mode_object.h
127
int drm_object_property_set_value(struct drm_mode_object *obj,
sys/dev/pci/drm/include/drm/drm_mode_object.h
130
int drm_object_property_get_value(struct drm_mode_object *obj,
sys/dev/pci/drm/include/drm/drm_mode_object.h
133
int drm_object_property_get_default_value(struct drm_mode_object *obj,
sys/dev/pci/drm/include/drm/drm_mode_object.h
137
void drm_object_attach_property(struct drm_mode_object *obj,
sys/dev/pci/drm/include/drm/drm_prime.h
101
struct dma_buf *drm_gem_prime_export(struct drm_gem_object *obj,
sys/dev/pci/drm/include/drm/drm_prime.h
116
void drm_prime_gem_destroy(struct drm_gem_object *obj, struct sg_table *sg);
sys/dev/pci/drm/include/drm/drm_prime.h
93
int drm_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma);
sys/dev/pci/drm/include/drm/drm_syncobj.h
80
drm_syncobj_get(struct drm_syncobj *obj)
sys/dev/pci/drm/include/drm/drm_syncobj.h
82
kref_get(&obj->refcount);
sys/dev/pci/drm/include/drm/drm_syncobj.h
90
drm_syncobj_put(struct drm_syncobj *obj)
sys/dev/pci/drm/include/drm/drm_syncobj.h
92
kref_put(&obj->refcount, drm_syncobj_free);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
675
struct drm_gem_object *obj,
sys/dev/pci/drm/include/linux/dma-resv.h
196
struct dma_resv *obj;
sys/dev/pci/drm/include/linux/dma-resv.h
232
struct dma_resv *obj,
sys/dev/pci/drm/include/linux/dma-resv.h
235
cursor->obj = obj;
sys/dev/pci/drm/include/linux/dma-resv.h
307
#define dma_resv_for_each_fence(cursor, obj, usage, fence) \
sys/dev/pci/drm/include/linux/dma-resv.h
308
for (dma_resv_iter_begin(cursor, obj, usage), \
sys/dev/pci/drm/include/linux/dma-resv.h
312
#define dma_resv_held(obj) lockdep_is_held(&(obj)->lock.base)
sys/dev/pci/drm/include/linux/dma-resv.h
313
#define dma_resv_assert_held(obj) lockdep_assert_held(&(obj)->lock.base)
sys/dev/pci/drm/include/linux/dma-resv.h
316
void dma_resv_reset_max_fences(struct dma_resv *obj);
sys/dev/pci/drm/include/linux/dma-resv.h
318
static inline void dma_resv_reset_max_fences(struct dma_resv *obj) {}
sys/dev/pci/drm/include/linux/dma-resv.h
343
static inline int dma_resv_lock(struct dma_resv *obj,
sys/dev/pci/drm/include/linux/dma-resv.h
346
return ww_mutex_lock(&obj->lock, ctx);
sys/dev/pci/drm/include/linux/dma-resv.h
370
static inline int dma_resv_lock_interruptible(struct dma_resv *obj,
sys/dev/pci/drm/include/linux/dma-resv.h
373
return ww_mutex_lock_interruptible(&obj->lock, ctx);
sys/dev/pci/drm/include/linux/dma-resv.h
387
static inline void dma_resv_lock_slow(struct dma_resv *obj,
sys/dev/pci/drm/include/linux/dma-resv.h
390
ww_mutex_lock_slow(&obj->lock, ctx);
sys/dev/pci/drm/include/linux/dma-resv.h
403
static inline int dma_resv_lock_slow_interruptible(struct dma_resv *obj,
sys/dev/pci/drm/include/linux/dma-resv.h
406
return ww_mutex_lock_slow_interruptible(&obj->lock, ctx);
sys/dev/pci/drm/include/linux/dma-resv.h
423
static inline bool __must_check dma_resv_trylock(struct dma_resv *obj)
sys/dev/pci/drm/include/linux/dma-resv.h
425
return ww_mutex_trylock(&obj->lock, NULL);
sys/dev/pci/drm/include/linux/dma-resv.h
434
static inline bool dma_resv_is_locked(struct dma_resv *obj)
sys/dev/pci/drm/include/linux/dma-resv.h
436
return ww_mutex_is_locked(&obj->lock);
sys/dev/pci/drm/include/linux/dma-resv.h
451
static inline struct ww_acquire_ctx *dma_resv_locking_ctx(struct dma_resv *obj)
sys/dev/pci/drm/include/linux/dma-resv.h
453
return READ_ONCE(obj->lock.ctx);
sys/dev/pci/drm/include/linux/dma-resv.h
462
static inline void dma_resv_unlock(struct dma_resv *obj)
sys/dev/pci/drm/include/linux/dma-resv.h
464
dma_resv_reset_max_fences(obj);
sys/dev/pci/drm/include/linux/dma-resv.h
465
ww_mutex_unlock(&obj->lock);
sys/dev/pci/drm/include/linux/dma-resv.h
468
void dma_resv_init(struct dma_resv *obj);
sys/dev/pci/drm/include/linux/dma-resv.h
469
void dma_resv_fini(struct dma_resv *obj);
sys/dev/pci/drm/include/linux/dma-resv.h
470
int dma_resv_reserve_fences(struct dma_resv *obj, unsigned int num_fences);
sys/dev/pci/drm/include/linux/dma-resv.h
471
void dma_resv_add_fence(struct dma_resv *obj, struct dma_fence *fence,
sys/dev/pci/drm/include/linux/dma-resv.h
473
void dma_resv_replace_fences(struct dma_resv *obj, uint64_t context,
sys/dev/pci/drm/include/linux/dma-resv.h
476
int dma_resv_get_fences(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/include/linux/dma-resv.h
478
int dma_resv_get_singleton(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/include/linux/dma-resv.h
481
long dma_resv_wait_timeout(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/include/linux/dma-resv.h
483
void dma_resv_set_deadline(struct dma_resv *obj, enum dma_resv_usage usage,
sys/dev/pci/drm/include/linux/dma-resv.h
485
bool dma_resv_test_signaled(struct dma_resv *obj, enum dma_resv_usage usage);
sys/dev/pci/drm/include/linux/dma-resv.h
486
void dma_resv_describe(struct dma_resv *obj, struct seq_file *seq);
sys/dev/pci/drm/include/linux/hashtable.h
56
#define hash_for_each_possible(table, obj, member, key) \
sys/dev/pci/drm/include/linux/hashtable.h
57
hlist_for_each_entry(obj, __hash(table, key), member)
sys/dev/pci/drm/include/linux/hashtable.h
58
#define hash_for_each_safe(table, i, tmp, obj, member) \
sys/dev/pci/drm/include/linux/hashtable.h
60
hlist_for_each_entry_safe(obj, tmp, &table[i], member)
sys/dev/pci/drm/include/linux/kobject.h
23
kobject_init(struct kobject *obj, struct kobj_type *type)
sys/dev/pci/drm/include/linux/kobject.h
25
kref_init(&obj->kref);
sys/dev/pci/drm/include/linux/kobject.h
26
obj->type = type;
sys/dev/pci/drm/include/linux/kobject.h
30
kobject_init_and_add(struct kobject *obj, struct kobj_type *type,
sys/dev/pci/drm/include/linux/kobject.h
33
kobject_init(obj, type);
sys/dev/pci/drm/include/linux/kobject.h
38
kobject_get(struct kobject *obj)
sys/dev/pci/drm/include/linux/kobject.h
40
if (obj != NULL)
sys/dev/pci/drm/include/linux/kobject.h
41
kref_get(&obj->kref);
sys/dev/pci/drm/include/linux/kobject.h
42
return (obj);
sys/dev/pci/drm/include/linux/kobject.h
48
struct kobject *obj = container_of(ref, struct kobject, kref);
sys/dev/pci/drm/include/linux/kobject.h
49
if (obj->type && obj->type->release)
sys/dev/pci/drm/include/linux/kobject.h
50
obj->type->release(obj);
sys/dev/pci/drm/include/linux/kobject.h
54
kobject_put(struct kobject *obj)
sys/dev/pci/drm/include/linux/kobject.h
56
if (obj != NULL)
sys/dev/pci/drm/include/linux/kobject.h
57
kref_put(&obj->kref, kobject_release);
sys/dev/pci/drm/include/linux/kobject.h
61
kobject_del(struct kobject *obj)
sys/dev/pci/drm/include/linux/kobject.h
65
#define kobject_uevent_env(obj, act, envp) 0
sys/dev/pci/drm/radeon/atombios_crtc.c
1142
struct drm_gem_object *obj;
sys/dev/pci/drm/radeon/atombios_crtc.c
1166
obj = target_fb->obj[0];
sys/dev/pci/drm/radeon/atombios_crtc.c
1167
rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/atombios_crtc.c
1441
rbo = gem_to_radeon_bo(fb->obj[0]);
sys/dev/pci/drm/radeon/atombios_crtc.c
1462
struct drm_gem_object *obj;
sys/dev/pci/drm/radeon/atombios_crtc.c
1483
obj = target_fb->obj[0];
sys/dev/pci/drm/radeon/atombios_crtc.c
1484
rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/atombios_crtc.c
1649
rbo = gem_to_radeon_bo(fb->obj[0]);
sys/dev/pci/drm/radeon/atombios_crtc.c
2157
rbo = gem_to_radeon_bo(crtc->primary->fb->obj[0]);
sys/dev/pci/drm/radeon/radeon.h
548
struct drm_gem_object **obj);
sys/dev/pci/drm/radeon/radeon_bios.c
288
union acpi_object atrm_arg_elements[2], *obj;
sys/dev/pci/drm/radeon/radeon_bios.c
307
obj = (union acpi_object *)buffer.pointer;
sys/dev/pci/drm/radeon/radeon_bios.c
308
memcpy(bios+offset, obj->buffer.pointer, obj->buffer.length);
sys/dev/pci/drm/radeon/radeon_bios.c
309
len = obj->buffer.length;
sys/dev/pci/drm/radeon/radeon_cursor.c
286
struct drm_gem_object *obj;
sys/dev/pci/drm/radeon/radeon_cursor.c
293
obj = NULL;
sys/dev/pci/drm/radeon/radeon_cursor.c
303
obj = drm_gem_object_lookup(file_priv, handle);
sys/dev/pci/drm/radeon/radeon_cursor.c
304
if (!obj) {
sys/dev/pci/drm/radeon/radeon_cursor.c
309
robj = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_cursor.c
312
drm_gem_object_put(obj);
sys/dev/pci/drm/radeon/radeon_cursor.c
322
drm_gem_object_put(obj);
sys/dev/pci/drm/radeon/radeon_cursor.c
360
radeon_crtc->cursor_bo = obj;
sys/dev/pci/drm/radeon/radeon_device.c
1628
if (fb == NULL || fb->obj[0] == NULL) {
sys/dev/pci/drm/radeon/radeon_device.c
1631
robj = gem_to_radeon_bo(fb->obj[0]);
sys/dev/pci/drm/radeon/radeon_display.c
1302
struct drm_gem_object *obj)
sys/dev/pci/drm/radeon/radeon_display.c
1305
fb->obj[0] = obj;
sys/dev/pci/drm/radeon/radeon_display.c
1309
fb->obj[0] = NULL;
sys/dev/pci/drm/radeon/radeon_display.c
1321
struct drm_gem_object *obj;
sys/dev/pci/drm/radeon/radeon_display.c
1325
obj = drm_gem_object_lookup(file_priv, mode_cmd->handles[0]);
sys/dev/pci/drm/radeon/radeon_display.c
1326
if (obj == NULL) {
sys/dev/pci/drm/radeon/radeon_display.c
1333
if (obj->import_attach) {
sys/dev/pci/drm/radeon/radeon_display.c
1335
drm_gem_object_put(obj);
sys/dev/pci/drm/radeon/radeon_display.c
1341
drm_gem_object_put(obj);
sys/dev/pci/drm/radeon/radeon_display.c
1345
ret = radeon_framebuffer_init(dev, fb, info, mode_cmd, obj);
sys/dev/pci/drm/radeon/radeon_display.c
1348
drm_gem_object_put(obj);
sys/dev/pci/drm/radeon/radeon_display.c
488
struct drm_gem_object *obj;
sys/dev/pci/drm/radeon/radeon_display.c
509
obj = crtc->primary->fb->obj[0];
sys/dev/pci/drm/radeon/radeon_display.c
512
drm_gem_object_get(obj);
sys/dev/pci/drm/radeon/radeon_display.c
513
work->old_rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_display.c
515
obj = fb->obj[0];
sys/dev/pci/drm/radeon/radeon_display.c
516
new_rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_gem.c
167
struct drm_gem_object **obj)
sys/dev/pci/drm/radeon/radeon_gem.c
173
*obj = NULL;
sys/dev/pci/drm/radeon/radeon_gem.c
203
*obj = &robj->tbo.base;
sys/dev/pci/drm/radeon/radeon_gem.c
271
static int radeon_gem_object_open(struct drm_gem_object *obj, struct drm_file *file_priv)
sys/dev/pci/drm/radeon/radeon_gem.c
273
struct radeon_bo *rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_gem.c
301
static void radeon_gem_object_close(struct drm_gem_object *obj,
sys/dev/pci/drm/radeon/radeon_gem.c
304
struct radeon_bo *rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_gem.c
342
static int radeon_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
sys/dev/pci/drm/radeon/radeon_gem.c
344
struct radeon_bo *bo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_gem.c
350
return drm_gem_ttm_mmap(obj, vma);
sys/dev/pci/drm/radeon/radeon_gem.c
354
radeon_gem_object_mmap(struct drm_gem_object *obj,
sys/dev/pci/drm/radeon/radeon_gem.c
357
struct radeon_bo *bo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_gem.c
363
return drm_gem_ttm_mmap(obj, accessprot, off, size);
sys/dev/pci/drm/radeon/radeon_gem.c
43
struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_gem.c
44
int radeon_gem_prime_pin(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_gem.c
45
void radeon_gem_prime_unpin(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_legacy_crtc.c
1095
rbo = gem_to_radeon_bo(crtc->primary->fb->obj[0]);
sys/dev/pci/drm/radeon/radeon_legacy_crtc.c
381
struct drm_gem_object *obj;
sys/dev/pci/drm/radeon/radeon_legacy_crtc.c
424
obj = target_fb->obj[0];
sys/dev/pci/drm/radeon/radeon_legacy_crtc.c
425
rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_legacy_crtc.c
452
old_rbo = gem_to_radeon_bo(fb->obj[0]);
sys/dev/pci/drm/radeon/radeon_legacy_crtc.c
559
rbo = gem_to_radeon_bo(fb->obj[0]);
sys/dev/pci/drm/radeon/radeon_mode.h
896
struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_prime.c
37
struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj)
sys/dev/pci/drm/radeon/radeon_prime.c
39
struct radeon_bo *bo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_prime.c
41
return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages,
sys/dev/pci/drm/radeon/radeon_prime.c
73
int radeon_gem_prime_pin(struct drm_gem_object *obj)
sys/dev/pci/drm/radeon/radeon_prime.c
75
struct radeon_bo *bo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_prime.c
86
void radeon_gem_prime_unpin(struct drm_gem_object *obj)
sys/dev/pci/drm/radeon/radeon_prime.c
88
struct radeon_bo *bo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_prime.h
31
struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_prime.h
32
int radeon_gem_prime_pin(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_prime.h
33
void radeon_gem_prime_unpin(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_prime.h
34
void *radeon_gem_prime_vmap(struct drm_gem_object *obj);
sys/dev/pci/drm/radeon/radeon_prime.h
35
void radeon_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr);
sys/dev/pci/drm/scheduler/sched_main.c
1004
struct drm_gem_object *obj,
sys/dev/pci/drm/scheduler/sched_main.c
1007
return drm_sched_job_add_resv_dependencies(job, obj->resv,
sys/dev/pci/drm/ttm/tests/ttm_kunit_helpers.c
176
struct dma_resv *obj)
sys/dev/pci/drm/ttm/tests/ttm_kunit_helpers.c
187
if (obj)
sys/dev/pci/drm/ttm/tests/ttm_kunit_helpers.c
188
bo->base.resv = obj;
sys/dev/pci/drm/ttm/tests/ttm_kunit_helpers.h
40
struct dma_resv *obj);
sys/dev/pv/viogpu.c
691
sfb->obj = drm_gem_cma_create(ddev, size);
sys/dev/pv/viogpu.c
692
if (sfb->obj == NULL) {
sys/dev/pv/viogpu.c
699
fb->obj[0] = &sfb->obj->base;
sys/dev/pv/viogpu.c
717
error = viogpu_attach_backing(sc, 1, sfb->obj->dmamap);
sys/kern/sys_futex.c
177
struct uvm_object *obj = NULL;
sys/kern/sys_futex.c
191
obj = entry->object.uvm_obj;
sys/kern/sys_futex.c
204
f->ft_obj = obj;
sys/uvm/uvm_device.c
116
obj = udv_attach_drm(device, accessprot, off, size);
sys/uvm/uvm_device.c
117
if (obj)
sys/uvm/uvm_device.c
118
return(obj);
sys/uvm/uvm_device.c
97
struct uvm_object *obj;
sys/uvm/uvm_page.c
1208
uvm_pagelookup(struct uvm_object *obj, voff_t off)
sys/uvm/uvm_page.c
1214
pg = RBT_FIND(uvm_objtree, &obj->memt, &p);
sys/uvm/uvm_page.c
1216
KASSERT(pg == NULL || obj->uo_npages != 0);
sys/uvm/uvm_page.c
690
uvm_pagealloc_pg(struct vm_page *pg, struct uvm_object *obj, voff_t off,
sys/uvm/uvm_page.c
695
KASSERT(obj == NULL || anon == NULL);
sys/uvm/uvm_page.c
698
KASSERT(obj == NULL || UVM_OBJ_IS_DUMMY(obj) ||
sys/uvm/uvm_page.c
699
rw_write_held(obj->vmobjlock));
sys/uvm/uvm_page.c
705
pg->uobject = obj;
sys/uvm/uvm_page.c
711
} else if (obj)
sys/uvm/uvm_page.c
804
uvm_pagealloc_multi(struct uvm_object *obj, voff_t off, vsize_t size,
sys/uvm/uvm_page.c
811
KASSERT(UVM_OBJ_IS_BUFCACHE(obj));
sys/uvm/uvm_page.c
825
uvm_pagealloc_pg(pg, obj, off + ptoa(i++), NULL);
sys/uvm/uvm_page.c
837
uvm_pagerealloc_multi(struct uvm_object *obj, voff_t off, vsize_t size,
sys/uvm/uvm_page.c
845
KASSERT(UVM_OBJ_IS_BUFCACHE(obj));
sys/uvm/uvm_page.c
857
tpg = uvm_pagelookup(obj, offset);
sys/uvm/uvm_page.c
867
uvm_pagealloc_pg(pg, obj, offset, NULL);
sys/uvm/uvm_page.c
882
uvm_pagealloc(struct uvm_object *obj, voff_t off, struct vm_anon *anon,
sys/uvm/uvm_page.c
888
KASSERT(obj == NULL || anon == NULL);
sys/uvm/uvm_page.c
891
KASSERT(obj == NULL || UVM_OBJ_IS_DUMMY(obj) ||
sys/uvm/uvm_page.c
892
rw_write_held(obj->vmobjlock));
sys/uvm/uvm_page.c
903
(obj != NULL && UVM_OBJ_IS_KERN_OBJECT(obj)))
sys/uvm/uvm_page.c
912
uvm_pagealloc_pg(pg, obj, off, anon);
usr.bin/dig/lib/isccfg/include/isccfg/cfg.h
119
cfg_map_get(const cfg_obj_t *mapobj, const char* name, const cfg_obj_t **obj);
usr.bin/dig/lib/isccfg/include/isccfg/cfg.h
148
cfg_obj_asstring(const cfg_obj_t *obj);
usr.bin/dig/lib/isccfg/include/isccfg/cfg.h
163
cfg_obj_destroy(cfg_parser_t *pctx, cfg_obj_t **obj);
usr.bin/dig/lib/isccfg/include/isccfg/grammar.h
102
cfg_obj_t *obj;
usr.bin/dig/lib/isccfg/parser.c
1049
fileobj = elt->obj;
usr.bin/dig/lib/isccfg/parser.c
1119
cfg_obj_t *obj;
usr.bin/dig/lib/isccfg/parser.c
1125
obj = malloc(sizeof(cfg_obj_t));
usr.bin/dig/lib/isccfg/parser.c
1126
if (obj == NULL)
usr.bin/dig/lib/isccfg/parser.c
1128
obj->type = type;
usr.bin/dig/lib/isccfg/parser.c
1129
obj->file = current_file(pctx);
usr.bin/dig/lib/isccfg/parser.c
1130
obj->line = pctx->line;
usr.bin/dig/lib/isccfg/parser.c
1131
*ret = obj;
usr.bin/dig/lib/isccfg/parser.c
1139
cfg_obj_t *obj = symval.as_pointer;
usr.bin/dig/lib/isccfg/parser.c
1145
cfg_obj_destroy(pctx, &obj);
usr.bin/dig/lib/isccfg/parser.c
1152
cfg_obj_t *obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
1154
CHECK(cfg_create_obj(pctx, type, &obj));
usr.bin/dig/lib/isccfg/parser.c
1158
obj->value.map.symtab = symtab;
usr.bin/dig/lib/isccfg/parser.c
1159
obj->value.map.id = NULL;
usr.bin/dig/lib/isccfg/parser.c
1161
*ret = obj;
usr.bin/dig/lib/isccfg/parser.c
1165
if (obj != NULL)
usr.bin/dig/lib/isccfg/parser.c
1166
free(obj);
usr.bin/dig/lib/isccfg/parser.c
1171
free_map(cfg_parser_t *pctx, cfg_obj_t *obj) {
usr.bin/dig/lib/isccfg/parser.c
1172
CLEANUP_OBJ(obj->value.map.id);
usr.bin/dig/lib/isccfg/parser.c
1173
isc_symtab_destroy(&obj->value.map.symtab);
usr.bin/dig/lib/isccfg/parser.c
118
free_list(cfg_parser_t *pctx, cfg_obj_t *obj);
usr.bin/dig/lib/isccfg/parser.c
1181
cfg_obj_t *obj;
usr.bin/dig/lib/isccfg/parser.c
1186
obj = *objp;
usr.bin/dig/lib/isccfg/parser.c
1188
obj->type->rep->free(pctx, obj);
usr.bin/dig/lib/isccfg/parser.c
1189
free(obj);
usr.bin/dig/lib/isccfg/parser.c
128
free_string(cfg_parser_t *pctx, cfg_obj_t *obj);
usr.bin/dig/lib/isccfg/parser.c
134
free_map(cfg_parser_t *pctx, cfg_obj_t *obj);
usr.bin/dig/lib/isccfg/parser.c
310
elt->obj = stringobj;
usr.bin/dig/lib/isccfg/parser.c
326
cfg_obj_t *obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
328
result = cfg_parse_obj(pctx, type, &obj);
usr.bin/dig/lib/isccfg/parser.c
346
*ret = obj;
usr.bin/dig/lib/isccfg/parser.c
350
CLEANUP_OBJ(obj);
usr.bin/dig/lib/isccfg/parser.c
402
cfg_obj_t *obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
405
CHECK(cfg_create_obj(pctx, type, &obj));
usr.bin/dig/lib/isccfg/parser.c
407
obj->value.string.length = len;
usr.bin/dig/lib/isccfg/parser.c
408
obj->value.string.base = malloc(len + 1);
usr.bin/dig/lib/isccfg/parser.c
409
if (obj->value.string.base == NULL) {
usr.bin/dig/lib/isccfg/parser.c
410
free(obj);
usr.bin/dig/lib/isccfg/parser.c
413
memmove(obj->value.string.base, contents, len);
usr.bin/dig/lib/isccfg/parser.c
414
obj->value.string.base[len] = '\0';
usr.bin/dig/lib/isccfg/parser.c
416
*ret = obj;
usr.bin/dig/lib/isccfg/parser.c
482
free_string(cfg_parser_t *pctx, cfg_obj_t *obj) {
usr.bin/dig/lib/isccfg/parser.c
484
free(obj->value.string.base);
usr.bin/dig/lib/isccfg/parser.c
488
cfg_obj_asstring(const cfg_obj_t *obj) {
usr.bin/dig/lib/isccfg/parser.c
489
REQUIRE(obj != NULL && obj->type->rep == &cfg_rep_string);
usr.bin/dig/lib/isccfg/parser.c
490
return (obj->value.string.base);
usr.bin/dig/lib/isccfg/parser.c
520
cfg_create_list(cfg_parser_t *pctx, const cfg_type_t *type, cfg_obj_t **obj) {
usr.bin/dig/lib/isccfg/parser.c
525
REQUIRE(obj != NULL && *obj == NULL);
usr.bin/dig/lib/isccfg/parser.c
527
CHECK(cfg_create_obj(pctx, type, obj));
usr.bin/dig/lib/isccfg/parser.c
528
ISC_LIST_INIT((*obj)->value.list);
usr.bin/dig/lib/isccfg/parser.c
541
elt->obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
549
cfg_obj_destroy(pctx, &elt->obj);
usr.bin/dig/lib/isccfg/parser.c
554
free_list(cfg_parser_t *pctx, cfg_obj_t *obj) {
usr.bin/dig/lib/isccfg/parser.c
556
for (elt = ISC_LIST_HEAD(obj->value.list);
usr.bin/dig/lib/isccfg/parser.c
583
elt->obj = value;
usr.bin/dig/lib/isccfg/parser.c
59
#define CLEANUP_OBJ(obj) \
usr.bin/dig/lib/isccfg/parser.c
60
do { if ((obj) != NULL) cfg_obj_destroy(pctx, &(obj)); } while (0)
usr.bin/dig/lib/isccfg/parser.c
615
cfg_obj_t *obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
624
CHECK(create_map(pctx, type, &obj));
usr.bin/dig/lib/isccfg/parser.c
626
obj->value.map.clausesets = clausesets;
usr.bin/dig/lib/isccfg/parser.c
686
result = isc_symtab_lookup(obj->value.map.symtab,
usr.bin/dig/lib/isccfg/parser.c
693
obj->value.map.symtab));
usr.bin/dig/lib/isccfg/parser.c
707
*ret = obj;
usr.bin/dig/lib/isccfg/parser.c
712
CLEANUP_OBJ(obj);
usr.bin/dig/lib/isccfg/parser.c
723
cfg_obj_t *obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
726
CHECK(cfg_parse_obj(pctx, elttype, &obj));
usr.bin/dig/lib/isccfg/parser.c
728
symval.as_pointer = obj;
usr.bin/dig/lib/isccfg/parser.c
735
CLEANUP_OBJ(obj);
usr.bin/dig/lib/isccfg/parser.c
794
cfg_map_get(const cfg_obj_t *mapobj, const char* name, const cfg_obj_t **obj) {
usr.bin/dig/lib/isccfg/parser.c
801
REQUIRE(obj != NULL && *obj == NULL);
usr.bin/dig/lib/isccfg/parser.c
808
*obj = val.as_pointer;
usr.bin/dig/lib/isccfg/parser.c
821
cfg_obj_t *obj = NULL;
usr.bin/dig/lib/isccfg/parser.c
827
CHECK(cfg_create_obj(pctx, &cfg_type_token, &obj));
usr.bin/dig/lib/isccfg/parser.c
837
obj->value.string.base = malloc(r.length + 1);
usr.bin/dig/lib/isccfg/parser.c
838
if (obj->value.string.base == NULL) {
usr.bin/dig/lib/isccfg/parser.c
842
obj->value.string.length = r.length;
usr.bin/dig/lib/isccfg/parser.c
843
memmove(obj->value.string.base, r.base, r.length);
usr.bin/dig/lib/isccfg/parser.c
844
obj->value.string.base[r.length] = '\0';
usr.bin/dig/lib/isccfg/parser.c
845
*ret = obj;
usr.bin/dig/lib/isccfg/parser.c
849
if (obj != NULL)
usr.bin/dig/lib/isccfg/parser.c
850
free(obj);
usr.bin/mandoc/term_ps.c
711
pdf_obj(struct termp *p, size_t obj)
usr.bin/mandoc/term_ps.c
714
assert(obj > 0);
usr.bin/mandoc/term_ps.c
716
if ((obj - 1) >= p->ps->pdfobjsz) {
usr.bin/mandoc/term_ps.c
717
p->ps->pdfobjsz = obj + 128;
usr.bin/mandoc/term_ps.c
722
p->ps->pdfobjs[(int)obj - 1] = p->ps->pdfbytes;
usr.bin/mandoc/term_ps.c
723
ps_printf(p, "%zu 0 obj\n", obj);
usr.bin/openssl/apps.c
893
ASN1_OBJECT *obj;
usr.bin/openssl/apps.c
902
obj = X509_EXTENSION_get_object(ext);
usr.bin/openssl/apps.c
903
idx = X509_get_ext_by_OBJ(x, obj, -1);
usr.bin/openssl/apps.c
914
idx = X509_get_ext_by_OBJ(x, obj, -1);
usr.bin/openssl/ca.c
153
int old_entry_print(BIO *bp, ASN1_OBJECT *obj, ASN1_STRING *str);
usr.bin/openssl/ca.c
1716
ASN1_OBJECT *obj;
usr.bin/openssl/ca.c
1759
obj = X509_NAME_ENTRY_get_object(ne);
usr.bin/openssl/ca.c
1760
if (obj == NULL)
usr.bin/openssl/ca.c
1764
if ((OBJ_obj2nid(obj) == NID_pkcs9_emailAddress) && (!email_dn))
usr.bin/openssl/ca.c
1768
if ((OBJ_obj2nid(obj) == NID_pkcs9_emailAddress) &&
usr.bin/openssl/ca.c
1782
old_entry_print(bio_err, obj, str);
usr.bin/openssl/ca.c
1807
obj = OBJ_nid2obj(j);
usr.bin/openssl/ca.c
1808
if (obj == NULL)
usr.bin/openssl/ca.c
1814
j = X509_NAME_get_index_by_OBJ(name, obj, last);
usr.bin/openssl/ca.c
1851
j = X509_NAME_get_index_by_OBJ(CAname, obj,
usr.bin/openssl/ca.c
2699
old_entry_print(BIO *bp, ASN1_OBJECT *obj, ASN1_STRING *str)
usr.bin/openssl/ca.c
2705
j = i2a_ASN1_OBJECT(bp, obj);
usr.bin/openssl/pkcs12.c
1084
ASN1_OBJECT *obj;
usr.bin/openssl/pkcs12.c
1087
obj = X509_ATTRIBUTE_get0_object(attr);
usr.bin/openssl/pkcs12.c
1091
i2a_ASN1_OBJECT(out, obj);
usr.bin/openssl/s_server.c
2275
X509_OBJECT *obj = NULL;
usr.bin/openssl/s_server.c
2316
if ((obj = X509_OBJECT_new()) == NULL)
usr.bin/openssl/s_server.c
2319
X509_get_issuer_name(x), obj) <= 0) {
usr.bin/openssl/s_server.c
2328
id = OCSP_cert_to_id(NULL, x, X509_OBJECT_get0_X509(obj));
usr.bin/openssl/s_server.c
2329
X509_OBJECT_free(obj);
usr.bin/openssl/s_server.c
2330
obj = NULL;
usr.bin/openssl/s_server.c
2362
X509_OBJECT_free(obj);
usr.bin/ssh/ssh-pkcs11.c
1003
rv = f->C_GetAttributeValue(session, *obj, key_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
1026
rv = f->C_GetAttributeValue(session, *obj, key_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
1084
CK_OBJECT_HANDLE *obj)
usr.bin/ssh/ssh-pkcs11.c
1113
rv = f->C_GetAttributeValue(session, *obj, key_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
1137
rv = f->C_GetAttributeValue(session, *obj, key_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
1194
CK_OBJECT_HANDLE *obj, struct sshkey **keyp, char **labelp)
usr.bin/ssh/ssh-pkcs11.c
1223
rv = f->C_GetAttributeValue(session, *obj, cert_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
1246
rv = f->C_GetAttributeValue(session, *obj, cert_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
1416
CK_OBJECT_HANDLE obj;
usr.bin/ssh/ssh-pkcs11.c
1422
memset(&obj, 0, sizeof(obj));
usr.bin/ssh/ssh-pkcs11.c
1441
rv = f->C_FindObjects(session, &obj, 1, &n);
usr.bin/ssh/ssh-pkcs11.c
1455
rv = f->C_GetAttributeValue(session, obj, key_attr, 1);
usr.bin/ssh/ssh-pkcs11.c
1465
if (pkcs11_fetch_x509_pubkey(p, slotidx, &obj,
usr.bin/ssh/ssh-pkcs11.c
1522
CK_OBJECT_HANDLE obj;
usr.bin/ssh/ssh-pkcs11.c
1527
memset(&obj, 0, sizeof(obj));
usr.bin/ssh/ssh-pkcs11.c
1547
rv = f->C_FindObjects(session, &obj, 1, &n);
usr.bin/ssh/ssh-pkcs11.c
1564
rv = f->C_GetAttributeValue(session, obj, key_attr, 2);
usr.bin/ssh/ssh-pkcs11.c
1575
key = pkcs11_fetch_rsa_pubkey(p, slotidx, &obj);
usr.bin/ssh/ssh-pkcs11.c
1578
key = pkcs11_fetch_ecdsa_pubkey(p, slotidx, &obj);
usr.bin/ssh/ssh-pkcs11.c
1582
key = pkcs11_fetch_ed25519_pubkey(p, slotidx, &obj);
usr.bin/ssh/ssh-pkcs11.c
190
CK_ULONG nattr, CK_OBJECT_HANDLE *obj)
usr.bin/ssh/ssh-pkcs11.c
204
if ((rv = f->C_FindObjects(session, obj, 1, &nfound)) != CKR_OK ||
usr.bin/ssh/ssh-pkcs11.c
2157
CK_OBJECT_HANDLE obj = CK_INVALID_HANDLE;
usr.bin/ssh/ssh-pkcs11.c
2189
if (pkcs11_find(p, slotidx, attrs, nattrs, &obj) == 0 &&
usr.bin/ssh/ssh-pkcs11.c
2190
obj != CK_INVALID_HANDLE) {
usr.bin/ssh/ssh-pkcs11.c
2191
if ((rv = f->C_DestroyObject(session, obj)) != CKR_OK) {
usr.bin/ssh/ssh-pkcs11.c
2205
if (pkcs11_find(p, slotidx, attrs, nattrs, &obj) == 0 &&
usr.bin/ssh/ssh-pkcs11.c
2206
obj != CK_INVALID_HANDLE) {
usr.bin/ssh/ssh-pkcs11.c
2212
rv = f->C_GetAttributeValue(session, obj, attrs, nattrs);
usr.bin/ssh/ssh-pkcs11.c
2221
k = pkcs11_fetch_rsa_pubkey(p, slotidx, &obj);
usr.bin/ssh/ssh-pkcs11.c
2224
k = pkcs11_fetch_ecdsa_pubkey(p, slotidx, &obj);
usr.bin/ssh/ssh-pkcs11.c
2227
k = pkcs11_fetch_ed25519_pubkey(p, slotidx, &obj);
usr.bin/ssh/ssh-pkcs11.c
2234
if ((rv = f->C_DestroyObject(session, obj)) != CKR_OK) {
usr.bin/ssh/ssh-pkcs11.c
284
pkcs11_check_obj_bool_attrib(struct pkcs11_key *k11, CK_OBJECT_HANDLE obj,
usr.bin/ssh/ssh-pkcs11.c
307
rv = f->C_GetAttributeValue(si->session, obj, &attr, 1);
usr.bin/ssh/ssh-pkcs11.c
314
k11->provider->name, k11->slotidx, obj, type, *val);
usr.bin/ssh/ssh-pkcs11.c
323
CK_OBJECT_HANDLE obj;
usr.bin/ssh/ssh-pkcs11.c
364
if (pkcs11_find(k11->provider, k11->slotidx, key_filter, 3, &obj) < 0 &&
usr.bin/ssh/ssh-pkcs11.c
365
pkcs11_find(k11->provider, k11->slotidx, key_filter, 2, &obj) < 0) {
usr.bin/ssh/ssh-pkcs11.c
375
if ((rv = f->C_SignInit(si->session, &mech, obj)) != CKR_OK) {
usr.bin/ssh/ssh-pkcs11.c
380
pkcs11_check_obj_bool_attrib(k11, obj, CKA_ALWAYS_AUTHENTICATE,
usr.bin/ssh/ssh-pkcs11.c
849
CK_OBJECT_HANDLE *obj)
usr.bin/ssh/ssh-pkcs11.c
871
rv = f->C_GetAttributeValue(session, *obj, key_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
894
rv = f->C_GetAttributeValue(session, *obj, key_attr, 3);
usr.bin/ssh/ssh-pkcs11.c
983
CK_OBJECT_HANDLE *obj)
usr.sbin/acme-client/json.c
101
if (n->d.obj == NULL)
usr.sbin/acme-client/json.c
105
&n->d.obj[i].lhs,
usr.sbin/acme-client/json.c
111
&n->d.obj[i].rhs,
usr.sbin/acme-client/json.c
162
free(n->d.obj);
usr.sbin/acme-client/json.c
254
if (n->d.obj[i].lhs->type != JSMN_STRING &&
usr.sbin/acme-client/json.c
255
n->d.obj[i].lhs->type != JSMN_PRIMITIVE)
usr.sbin/acme-client/json.c
257
else if (strcmp(name, n->d.obj[i].lhs->d.str))
usr.sbin/acme-client/json.c
263
if (n->d.obj[i].rhs->type != JSMN_ARRAY)
usr.sbin/acme-client/json.c
265
return n->d.obj[i].rhs;
usr.sbin/acme-client/json.c
281
if (n->d.obj[i].lhs->type != JSMN_STRING &&
usr.sbin/acme-client/json.c
282
n->d.obj[i].lhs->type != JSMN_PRIMITIVE)
usr.sbin/acme-client/json.c
284
else if (strcmp(name, n->d.obj[i].lhs->d.str))
usr.sbin/acme-client/json.c
290
if (n->d.obj[i].rhs->type != JSMN_OBJECT)
usr.sbin/acme-client/json.c
292
return n->d.obj[i].rhs;
usr.sbin/acme-client/json.c
309
if (n->d.obj[i].lhs->type != JSMN_STRING &&
usr.sbin/acme-client/json.c
310
n->d.obj[i].lhs->type != JSMN_PRIMITIVE)
usr.sbin/acme-client/json.c
312
else if (strcmp(name, n->d.obj[i].lhs->d.str))
usr.sbin/acme-client/json.c
318
if (n->d.obj[i].rhs->type != JSMN_STRING &&
usr.sbin/acme-client/json.c
319
n->d.obj[i].rhs->type != JSMN_PRIMITIVE)
usr.sbin/acme-client/json.c
322
cp = strdup(n->d.obj[i].rhs->d.str);
usr.sbin/acme-client/json.c
376
struct jsmnn *array, *obj, *error;
usr.sbin/acme-client/json.c
389
obj = json_getarrayobj(array->d.array[i]);
usr.sbin/acme-client/json.c
390
if (obj == NULL)
usr.sbin/acme-client/json.c
392
type = json_getstr(obj, "type");
usr.sbin/acme-client/json.c
399
p->uri = json_getstr(obj, "url");
usr.sbin/acme-client/json.c
40
struct jsmnp *obj; /* JSMN_OBJECT */
usr.sbin/acme-client/json.c
400
p->token = json_getstr(obj, "token");
usr.sbin/acme-client/json.c
401
p->status = json_parse_response(obj);
usr.sbin/acme-client/json.c
403
error = json_getobj(obj, "error");
usr.sbin/acme-client/json.c
99
n->d.obj = calloc(n->fields,
usr.sbin/bgpd/chash.h
179
const struct _type *obj = ptr; \
usr.sbin/bgpd/chash.h
180
return _hash(obj); \
usr.sbin/config/sem.c
215
has_errobj(struct nvlist *nv, void *obj)
usr.sbin/config/sem.c
219
if (nv->nv_ptr == obj)
usr.sbin/crunchgen/crunchgen.c
660
char *cp, *obj, tempfname[PATH_MAX], cwd[PATH_MAX];
usr.sbin/crunchgen/crunchgen.c
716
obj = cp;
usr.sbin/crunchgen/crunchgen.c
721
add_string(&p->objs, obj);
usr.sbin/ldapd/schema.c
1000
OBJ_NAME(obj), OBJ_NAME(optr->object));
usr.sbin/ldapd/schema.c
1004
if (obj->kind == KIND_ABSTRACT &&
usr.sbin/ldapd/schema.c
1008
OBJ_NAME(obj), OBJ_NAME(optr->object));
usr.sbin/ldapd/schema.c
1018
if (obj != NULL) {
usr.sbin/ldapd/schema.c
1019
if (obj->oid != NULL) {
usr.sbin/ldapd/schema.c
1020
RB_REMOVE(object_tree, &schema->objects, obj);
usr.sbin/ldapd/schema.c
1021
free(obj->oid);
usr.sbin/ldapd/schema.c
1023
free(obj->desc);
usr.sbin/ldapd/schema.c
1024
free(obj);
usr.sbin/ldapd/schema.c
1216
schema_dump_object(struct object *obj, char *buf, size_t size)
usr.sbin/ldapd/schema.c
1219
strlcat(buf, obj->oid, size) >= size)
usr.sbin/ldapd/schema.c
1222
if (schema_dump_names("NAME", obj->names, buf, size) != 0)
usr.sbin/ldapd/schema.c
1225
if (obj->desc != NULL)
usr.sbin/ldapd/schema.c
1227
strlcat(buf, obj->desc, size) >= size ||
usr.sbin/ldapd/schema.c
1231
switch (obj->kind) {
usr.sbin/ldapd/schema.c
1246
if (schema_dump_objlist("SUP", obj->sup, buf, size) != 0)
usr.sbin/ldapd/schema.c
1249
if (obj->obsolete && strlcat(buf, " OBSOLETE", size) >= size)
usr.sbin/ldapd/schema.c
1252
if (schema_dump_attrlist("MUST", obj->must, buf, size) != 0)
usr.sbin/ldapd/schema.c
1255
if (schema_dump_attrlist("MAY", obj->may, buf, size) != 0)
usr.sbin/ldapd/schema.c
227
push_obj(struct obj_list *olist, struct object *obj)
usr.sbin/ldapd/schema.c
244
optr->object = obj;
usr.sbin/ldapd/schema.c
494
schema_link_obj_name(struct schema *schema, const char *name, struct object *obj)
usr.sbin/ldapd/schema.c
504
oidname->on_object = obj;
usr.sbin/ldapd/schema.c
518
schema_link_obj_names(struct schema *schema, struct object *obj)
usr.sbin/ldapd/schema.c
522
SLIST_FOREACH(name, obj->names, next) {
usr.sbin/ldapd/schema.c
523
if (schema_link_obj_name(schema, name->name, obj) != 0)
usr.sbin/ldapd/schema.c
62
static struct obj_list *push_obj(struct obj_list *olist, struct object *obj);
usr.sbin/ldapd/schema.c
627
struct object *obj;
usr.sbin/ldapd/schema.c
633
if ((obj = lookup_object(schema, kw)) == NULL) {
usr.sbin/ldapd/schema.c
638
return push_obj(NULL, obj);
usr.sbin/ldapd/schema.c
656
if ((obj = lookup_object(schema, kw)) == NULL)
usr.sbin/ldapd/schema.c
658
olist = push_obj(olist, obj);
usr.sbin/ldapd/schema.c
897
struct object *obj = NULL, *prev;
usr.sbin/ldapd/schema.c
909
if ((obj = calloc(1, sizeof(*obj))) == NULL) {
usr.sbin/ldapd/schema.c
913
obj->kind = KIND_STRUCTURAL;
usr.sbin/ldapd/schema.c
916
obj->oid = kw;
usr.sbin/ldapd/schema.c
918
obj->oid = lookup_symbolic_oid(schema, kw);
usr.sbin/ldapd/schema.c
919
if (obj->oid == NULL)
usr.sbin/ldapd/schema.c
925
prev = RB_INSERT(object_tree, &schema->objects, obj);
usr.sbin/ldapd/schema.c
927
schema_err(schema, "object class %s already defined", obj->oid);
usr.sbin/ldapd/schema.c
938
obj->names = schema_parse_names(schema);
usr.sbin/ldapd/schema.c
939
if (obj->names == NULL)
usr.sbin/ldapd/schema.c
941
schema_link_obj_names(schema, obj);
usr.sbin/ldapd/schema.c
943
if (schema_lex(schema, &obj->desc) != STRING)
usr.sbin/ldapd/schema.c
946
obj->obsolete = 1;
usr.sbin/ldapd/schema.c
948
obj->sup = schema_parse_objlist(schema);
usr.sbin/ldapd/schema.c
949
if (obj->sup == NULL)
usr.sbin/ldapd/schema.c
952
obj->kind = KIND_ABSTRACT;
usr.sbin/ldapd/schema.c
954
obj->kind = KIND_STRUCTURAL;
usr.sbin/ldapd/schema.c
956
obj->kind = KIND_AUXILIARY;
usr.sbin/ldapd/schema.c
958
obj->must = schema_parse_attrlist(schema);
usr.sbin/ldapd/schema.c
959
if (obj->must == NULL)
usr.sbin/ldapd/schema.c
962
obj->may = schema_parse_attrlist(schema);
usr.sbin/ldapd/schema.c
963
if (obj->may == NULL)
usr.sbin/ldapd/schema.c
986
if (obj->sup != NULL) {
usr.sbin/ldapd/schema.c
987
SLIST_FOREACH(optr, obj->sup, next) {
usr.sbin/ldapd/schema.c
988
if (obj->kind == KIND_STRUCTURAL &&
usr.sbin/ldapd/schema.c
992
OBJ_NAME(obj), OBJ_NAME(optr->object));
usr.sbin/ldapd/schema.c
996
if (obj->kind == KIND_AUXILIARY &&
usr.sbin/ldapd/schema.h
157
int schema_dump_object(struct object *obj,
usr.sbin/ldapd/schema.h
159
int schema_dump_attribute(struct attr_type *obj,
usr.sbin/ldapd/schema.h
22
#define OBJ_NAME(obj) ((obj)->names ? SLIST_FIRST((obj)->names)->name : \
usr.sbin/ldapd/schema.h
23
(obj)->oid)
usr.sbin/ldapd/search.c
550
struct object *obj;
usr.sbin/ldapd/search.c
584
RB_FOREACH(obj, object_tree, &conf->schema->objects) {
usr.sbin/ldapd/search.c
585
if (schema_dump_object(obj, buf, sizeof(buf)) != 0) {
usr.sbin/ldapd/validate.c
211
struct object *obj;
usr.sbin/ldapd/validate.c
218
obj = optr->object;
usr.sbin/ldapd/validate.c
220
if (has_attribute(at, obj->may) ||
usr.sbin/ldapd/validate.c
221
has_attribute(at, obj->must))
usr.sbin/ldapd/validate.c
224
if (validate_allowed_attribute(at, obj->sup) == LDAP_SUCCESS)
usr.sbin/ldapd/validate.c
232
olist_push(struct obj_list *olist, struct object *obj)
usr.sbin/ldapd/validate.c
237
if (optr->object == obj)
usr.sbin/ldapd/validate.c
242
optr->object = obj;
usr.sbin/ldapd/validate.c
247
if (obj->sup != NULL)
usr.sbin/ldapd/validate.c
248
SLIST_FOREACH(sup, obj->sup, next)
usr.sbin/ldapd/validate.c
271
is_super(struct object *sup, struct object *obj)
usr.sbin/ldapd/validate.c
275
if (sup == NULL || obj->sup == NULL)
usr.sbin/ldapd/validate.c
278
SLIST_FOREACH(optr, obj->sup, next)
usr.sbin/ldapd/validate.c
29
validate_required_attributes(struct ber_element *entry, struct object *obj)
usr.sbin/ldapd/validate.c
291
struct object *obj, *structural_obj = NULL;
usr.sbin/ldapd/validate.c
321
if ((obj = lookup_object(conf->schema, s)) == NULL) {
usr.sbin/ldapd/validate.c
327
if (obj->kind == KIND_STRUCTURAL) {
usr.sbin/ldapd/validate.c
329
if (is_super(structural_obj, obj))
usr.sbin/ldapd/validate.c
330
structural_obj = obj;
usr.sbin/ldapd/validate.c
331
else if (!is_super(obj, structural_obj)) {
usr.sbin/ldapd/validate.c
338
structural_obj = obj;
usr.sbin/ldapd/validate.c
34
if (obj->must == NULL)
usr.sbin/ldapd/validate.c
341
olist_push(olist, obj);
usr.sbin/ldapd/validate.c
347
if (strcmp(obj->oid, "1.3.6.1.4.1.1466.101.120.111") == 0)
usr.sbin/ldapd/validate.c
37
SLIST_FOREACH(ap, obj->must, next) {
usr.sbin/nsd/dnstap/dnstap.c
278
dt_tls_writer_destroy(void* obj)
usr.sbin/nsd/dnstap/dnstap.c
280
struct dt_tls_writer* dtw = (struct dt_tls_writer*)obj;
usr.sbin/nsd/dnstap/dnstap.c
287
dt_tls_writer_open(void* obj)
usr.sbin/nsd/dnstap/dnstap.c
294
struct dt_tls_writer* dtw = (struct dt_tls_writer*)obj;
usr.sbin/nsd/dnstap/dnstap.c
423
dt_tls_writer_close(void* obj)
usr.sbin/nsd/dnstap/dnstap.c
425
struct dt_tls_writer* dtw = (struct dt_tls_writer*)obj;
usr.sbin/nsd/dnstap/dnstap.c
443
dt_tls_writer_read(void* obj, void* buf, size_t nbytes)
usr.sbin/nsd/dnstap/dnstap.c
446
struct dt_tls_writer* dtw = (struct dt_tls_writer*)obj;
usr.sbin/nsd/dnstap/dnstap.c
480
dt_tls_writer_write(void* obj, const struct iovec* iov, int iovcnt)
usr.sbin/nsd/dnstap/dnstap.c
482
struct dt_tls_writer* dtw = (struct dt_tls_writer*)obj;
usr.sbin/rpki-client/cert.c
1513
const ASN1_OBJECT *obj;
usr.sbin/rpki-client/cert.c
1531
obj = X509_EXTENSION_get_object((X509_EXTENSION *)ext);
usr.sbin/rpki-client/cert.c
1532
assert(obj != NULL);
usr.sbin/rpki-client/cert.c
1535
switch (nid = OBJ_obj2nid(obj)) {
usr.sbin/rpki-client/cert.c
1604
OBJ_obj2txt(objn, sizeof(objn), obj, 0);
usr.sbin/rpki-client/cms.c
117
obj = X509_ATTRIBUTE_get0_object(attr);
usr.sbin/rpki-client/cms.c
118
if (obj == NULL) {
usr.sbin/rpki-client/cms.c
122
if (OBJ_cmp(obj, cnt_type_oid) == 0) {
usr.sbin/rpki-client/cms.c
128
} else if (OBJ_cmp(obj, msg_dgst_oid) == 0) {
usr.sbin/rpki-client/cms.c
134
} else if (OBJ_cmp(obj, sign_time_oid) == 0) {
usr.sbin/rpki-client/cms.c
143
OBJ_obj2txt(buf, sizeof(buf), obj, 1);
usr.sbin/rpki-client/cms.c
180
const ASN1_OBJECT *obj, *octype;
usr.sbin/rpki-client/cms.c
226
if ((obj = CMS_get0_type(cms)) == NULL) {
usr.sbin/rpki-client/cms.c
230
OBJ_obj2txt(buf, sizeof(buf), obj, 1);
usr.sbin/rpki-client/cms.c
264
X509_ALGOR_get0(&obj, NULL, NULL, pdig);
usr.sbin/rpki-client/cms.c
265
nid = OBJ_obj2nid(obj);
usr.sbin/rpki-client/cms.c
271
X509_ALGOR_get0(&obj, NULL, NULL, psig);
usr.sbin/rpki-client/cms.c
272
nid = OBJ_obj2nid(obj);
usr.sbin/rpki-client/cms.c
286
obj = CMS_get0_eContentType(cms);
usr.sbin/rpki-client/cms.c
287
if (obj == NULL) {
usr.sbin/rpki-client/cms.c
292
if (OBJ_cmp(obj, oid) != 0) {
usr.sbin/rpki-client/cms.c
293
OBJ_obj2txt(buf, sizeof(buf), obj, 1);
usr.sbin/rpki-client/cms.c
318
if (OBJ_cmp(obj, octype) != 0) {
usr.sbin/rpki-client/cms.c
319
OBJ_obj2txt(buf, sizeof(buf), obj, 1);
usr.sbin/rpki-client/cms.c
96
const ASN1_OBJECT *obj;
usr.sbin/rpki-client/filemode.c
347
const ASN1_OBJECT *obj;
usr.sbin/rpki-client/filemode.c
349
if ((obj = CMS_get0_type(cms)) != NULL) {
usr.sbin/rpki-client/filemode.c
350
if (OBJ_cmp(obj, ccr_oid) == 0) {
usr.sbin/rpki-client/filemode.c
361
if ((obj = CMS_get0_eContentType(cms)) == NULL) {
usr.sbin/rpki-client/filemode.c
367
if (OBJ_cmp(obj, aspa_oid) == 0)
usr.sbin/rpki-client/filemode.c
369
else if (OBJ_cmp(obj, mft_oid) == 0)
usr.sbin/rpki-client/filemode.c
371
else if (OBJ_cmp(obj, roa_oid) == 0)
usr.sbin/rpki-client/filemode.c
373
else if (OBJ_cmp(obj, rsc_oid) == 0)
usr.sbin/rpki-client/filemode.c
375
else if (OBJ_cmp(obj, spl_oid) == 0)
usr.sbin/rpki-client/filemode.c
377
else if (OBJ_cmp(obj, tak_oid) == 0)
usr.sbin/rpki-client/rsc.c
202
const ASN1_OBJECT *obj;
usr.sbin/rpki-client/rsc.c
205
X509_ALGOR_get0(&obj, &type, NULL, alg);
usr.sbin/rpki-client/rsc.c
213
if ((nid = OBJ_obj2nid(obj)) != NID_sha256) {
usr.sbin/snmpd/application_internal.c
203
while ((obj = RB_MIN(appl_internal_objects,
usr.sbin/snmpd/application_internal.c
206
&appl_internal_objects_conf, obj);
usr.sbin/snmpd/application_internal.c
207
oid = obj->oid;
usr.sbin/snmpd/application_internal.c
211
free(obj->stringval);
usr.sbin/snmpd/application_internal.c
212
free(obj);
usr.sbin/snmpd/application_internal.c
215
obj);
usr.sbin/snmpd/application_internal.c
256
struct appl_internal_object *obj;
usr.sbin/snmpd/application_internal.c
259
if ((obj = calloc(1, sizeof(*obj))) == NULL)
usr.sbin/snmpd/application_internal.c
261
obj->oid = *oid;
usr.sbin/snmpd/application_internal.c
262
obj->get = get;
usr.sbin/snmpd/application_internal.c
263
obj->getnext = getnext;
usr.sbin/snmpd/application_internal.c
264
obj->stringval = NULL;
usr.sbin/snmpd/application_internal.c
267
&appl_internal_objects, obj) != NULL)
usr.sbin/snmpd/application_internal.c
276
struct appl_internal_object *obj;
usr.sbin/snmpd/application_internal.c
278
if ((obj = calloc(1, sizeof(*obj))) == NULL)
usr.sbin/snmpd/application_internal.c
280
obj->oid = *oid;
usr.sbin/snmpd/application_internal.c
281
obj->get = appl_internal_get_int;
usr.sbin/snmpd/application_internal.c
282
obj->getnext = NULL;
usr.sbin/snmpd/application_internal.c
283
obj->intval = val;
usr.sbin/snmpd/application_internal.c
284
obj->stringval = NULL;
usr.sbin/snmpd/application_internal.c
287
&appl_internal_objects_conf, obj) != NULL) {
usr.sbin/snmpd/application_internal.c
288
free(obj);
usr.sbin/snmpd/application_internal.c
297
struct appl_internal_object *obj;
usr.sbin/snmpd/application_internal.c
299
if ((obj = calloc(1, sizeof(*obj))) == NULL)
usr.sbin/snmpd/application_internal.c
301
obj->oid = *oid;
usr.sbin/snmpd/application_internal.c
302
obj->get = appl_internal_get_string;
usr.sbin/snmpd/application_internal.c
303
obj->getnext = NULL;
usr.sbin/snmpd/application_internal.c
304
obj->stringval = val;
usr.sbin/snmpd/application_internal.c
307
&appl_internal_objects_conf, obj) != NULL) {
usr.sbin/snmpd/application_internal.c
308
free(obj);
usr.sbin/snmpd/application_internal.c
596
struct appl_internal_object *obj;
usr.sbin/snmpd/application_internal.c
598
obj = appl_internal_object_parent(oid);
usr.sbin/snmpd/application_internal.c
599
return ober_add_integer(NULL, obj->intval);
usr.sbin/snmpd/application_internal.c
605
struct appl_internal_object *obj;
usr.sbin/snmpd/application_internal.c
607
obj = appl_internal_object_parent(oid);
usr.sbin/snmpd/application_internal.c
608
return ober_add_string(NULL, obj->stringval);
usr.sbin/snmpd/application_internal.c
98
struct appl_internal_object *obj;
usr.sbin/tcpdump/mib.h
13
struct obj
usr.sbin/tcpdump/print-snmp.c
238
struct obj *child, *next; /* child and next sibling pointers */
usr.sbin/tcpdump/print-snmp.c
258
struct obj *node; /* pointer into object table */